22,406 research outputs found
A Quality and Cost Approach for Comparison of Small-World Networks
We propose an approach based on analysis of cost-quality tradeoffs for
comparison of efficiency of various algorithms for small-world network
construction. A number of both known in the literature and original algorithms
for complex small-world networks construction are shortly reviewed and
compared. The networks constructed on the basis of these algorithms have basic
structure of 1D regular lattice with additional shortcuts providing the
small-world properties. It is shown that networks proposed in this work have
the best cost-quality ratio in the considered class.Comment: 27 pages, 16 figures, 1 tabl
Distributed Robustness Analysis of Interconnected Uncertain Systems Using Chordal Decomposition
Large-scale interconnected uncertain systems commonly have large state and
uncertainty dimensions. Aside from the heavy computational cost of solving
centralized robust stability analysis techniques, privacy requirements in the
network can also introduce further issues. In this paper, we utilize IQC
analysis for analyzing large-scale interconnected uncertain systems and we
evade these issues by describing a decomposition scheme that is based on the
interconnection structure of the system. This scheme is based on the so-called
chordal decomposition and does not add any conservativeness to the analysis
approach. The decomposed problem can be solved using distributed computational
algorithms without the need for a centralized computational unit. We further
discuss the merits of the proposed analysis approach using a numerical
experiment.Comment: 3 figures. Submitted to the 19th IFAC world congres
Effectiveness of segment routing technology in reducing the bandwidth and cloud resources provisioning times in network function virtualization architectures
Network Function Virtualization is a new technology allowing for a elastic cloud and bandwidth resource allocation. The technology requires an orchestrator whose role is the service and resource orchestration. It receives service requests, each one characterized by a Service Function Chain, which is a set of service functions to be executed according to a given order. It implements an algorithm for deciding where both to allocate the cloud and bandwidth resources and to route the SFCs. In a traditional orchestration algorithm, the orchestrator has a detailed knowledge of the cloud and network infrastructures and that can lead to high computational complexity of the SFC Routing and Cloud and Bandwidth resource Allocation (SRCBA) algorithm. In this paper, we propose and evaluate the effectiveness of a scalable orchestration architecture inherited by the one proposed within the European Telecommunications Standards Institute (ETSI) and based on the functional separation of an NFV orchestrator in Resource Orchestrator (RO) and Network Service Orchestrator (NSO). Each cloud domain is equipped with an RO whose task is to provide a simple and abstract representation of the cloud infrastructure. These representations are notified of the NSO that can apply a simplified and less complex SRCBA algorithm. In addition, we show how the segment routing technology can help to simplify the SFC routing by means of an effective addressing of the service functions. The scalable orchestration solution has been investigated and compared to the one of a traditional orchestrator in some network scenarios and varying the number of cloud domains. We have verified that the execution time of the SRCBA algorithm can be drastically reduced without degrading the performance in terms of cloud and bandwidth resource costs
B-LOG: A branch and bound methodology for the parallel execution of logic programs
We propose a computational methodology -"B-LOG"-, which offers the potential for an effective implementation of Logic Programming in a parallel computer. We also propose a weighting scheme to guide the search process through the graph and we apply the concepts of parallel "branch and bound" algorithms in order to perform a "best-first" search using an information theoretic bound. The concept of "session" is used to speed up the search process in a succession of similar queries. Within a session, we strongly modify the bounds in a local database, while bounds kept in a global database are weakly modified to provide a better initial condition for other sessions. We
also propose an implementation scheme based on a database
machine using "semantic paging", and the "B-LOG processor" based on a scoreboard driven controller
Recommended from our members
Chippe : a system for constraint driven behavioral synthesis
This report describes the Chippe system, gives some background previous work and describes several sample design runs of the system. Also presented are the sources of the design tradeoffs used by Chippe, and overview of the internal design model, and experiences using the system
- …