6,474 research outputs found
An algorithm for building an enterprise network topology using widespread data sources
A lot of network management tasks depend on the description of the network topology. However, the lack of standard methods of the network elements detection, coupled with the incompleteness and heterogeneity of the available topology data, complicate the network topology discovery process. In these conditions, formal models and methods of topology discovery are required. Contribution of the paper is an algorithm for automated enterprise network topology discovery based on a previously developed graph model and criteria for building graph elements. The proposed algorithm is capable of dealing with incomplete heterogeneous data about network topology as well as with the presence in the network of uncooperative devices. The paper also evaluates the algorithm and provides the testing results
An Overview of Mobile Ad Hoc Networks for the Existing Protocols and Applications
Mobile Ad Hoc Network (MANET) is a collection of two or more devices or nodes
or terminals with wireless communications and networking capability that
communicate with each other without the aid of any centralized administrator
also the wireless nodes that can dynamically form a network to exchange
information without using any existing fixed network infrastructure. And it's
an autonomous system in which mobile hosts connected by wireless links are free
to be dynamically and some time act as routers at the same time, and we discuss
in this paper the distinct characteristics of traditional wired networks,
including network configuration may change at any time, there is no direction
or limit the movement and so on, and thus needed a new optional path Agreement
(Routing Protocol) to identify nodes for these actions communicate with each
other path, An ideal choice way the agreement should not only be able to find
the right path, and the Ad Hoc Network must be able to adapt to changing
network of this type at any time. and we talk in details in this paper all the
information of Mobile Ad Hoc Network which include the History of ad hoc,
wireless ad hoc, wireless mobile approaches and types of mobile ad Hoc
networks, and then we present more than 13 types of the routing Ad Hoc Networks
protocols have been proposed. In this paper, the more representative of routing
protocols, analysis of individual characteristics and advantages and
disadvantages to collate and compare, and present the all applications or the
Possible Service of Ad Hoc Networks.Comment: 24 Pages, JGraph-Hoc Journa
Towards In-Transit Analytics for Industry 4.0
Industry 4.0, or Digital Manufacturing, is a vision of inter-connected
services to facilitate innovation in the manufacturing sector. A fundamental
requirement of innovation is the ability to be able to visualise manufacturing
data, in order to discover new insight for increased competitive advantage.
This article describes the enabling technologies that facilitate In-Transit
Analytics, which is a necessary precursor for Industrial Internet of Things
(IIoT) visualisation.Comment: 8 pages, 10th IEEE International Conference on Internet of Things
(iThings-2017), Exeter, UK, 201
ANCHOR: logically-centralized security for Software-Defined Networks
While the centralization of SDN brought advantages such as a faster pace of
innovation, it also disrupted some of the natural defenses of traditional
architectures against different threats. The literature on SDN has mostly been
concerned with the functional side, despite some specific works concerning
non-functional properties like 'security' or 'dependability'. Though addressing
the latter in an ad-hoc, piecemeal way, may work, it will most likely lead to
efficiency and effectiveness problems. We claim that the enforcement of
non-functional properties as a pillar of SDN robustness calls for a systemic
approach. As a general concept, we propose ANCHOR, a subsystem architecture
that promotes the logical centralization of non-functional properties. To show
the effectiveness of the concept, we focus on 'security' in this paper: we
identify the current security gaps in SDNs and we populate the architecture
middleware with the appropriate security mechanisms, in a global and consistent
manner. Essential security mechanisms provided by anchor include reliable
entropy and resilient pseudo-random generators, and protocols for secure
registration and association of SDN devices. We claim and justify in the paper
that centralizing such mechanisms is key for their effectiveness, by allowing
us to: define and enforce global policies for those properties; reduce the
complexity of controllers and forwarding devices; ensure higher levels of
robustness for critical services; foster interoperability of the non-functional
property enforcement mechanisms; and promote the security and resilience of the
architecture itself. We discuss design and implementation aspects, and we prove
and evaluate our algorithms and mechanisms, including the formalisation of the
main protocols and the verification of their core security properties using the
Tamarin prover.Comment: 42 pages, 4 figures, 3 tables, 5 algorithms, 139 reference
DEKAS - An evolutionary case-based reasoning system to support protection scheme design
This paper describes a decision support system being developed in conjunction with two UK utility companies to aid the design of electrical power transmission protection systems. A brief overview of the application domain is provided, followed by a description of the work carried out to date concerning the development and deployment of the Design Engineering Knowledge Application System (DEKAS). The paper then discusses the provision of intelligent decision support to the design engineer through the application of case-based reasoning (CBR). The key benefits from this will be outlined in conjunction with a relevant case study
An Algorithm for Network and Data-aware Placement of Multi-Tier Applications in Cloud Data Centers
Today's Cloud applications are dominated by composite applications comprising
multiple computing and data components with strong communication correlations
among them. Although Cloud providers are deploying large number of computing
and storage devices to address the ever increasing demand for computing and
storage resources, network resource demands are emerging as one of the key
areas of performance bottleneck. This paper addresses network-aware placement
of virtual components (computing and data) of multi-tier applications in data
centers and formally defines the placement as an optimization problem. The
simultaneous placement of Virtual Machines and data blocks aims at reducing the
network overhead of the data center network infrastructure. A greedy heuristic
is proposed for the on-demand application components placement that localizes
network traffic in the data center interconnect. Such optimization helps
reducing communication overhead in upper layer network switches that will
eventually reduce the overall traffic volume across the data center. This, in
turn, will help reducing packet transmission delay, increasing network
performance, and minimizing the energy consumption of network components.
Experimental results demonstrate performance superiority of the proposed
algorithm over other approaches where it outperforms the state-of-the-art
network-aware application placement algorithm across all performance metrics by
reducing the average network cost up to 67% and network usage at core switches
up to 84%, as well as increasing the average number of application deployments
up to 18%.Comment: Submitted for publication consideration for the Journal of Network
and Computer Applications (JNCA). Total page: 28. Number of figures: 15
figure
Grid Infrastructure for Domain Decomposition Methods in Computational ElectroMagnetics
The accurate and efficient solution of Maxwell's equation is the problem addressed by the scientific discipline called Computational ElectroMagnetics (CEM). Many macroscopic phenomena in a great number of fields are governed by this set of differential equations: electronic, geophysics, medical and biomedical technologies, virtual EM prototyping, besides the traditional antenna and propagation applications. Therefore, many efforts are focussed on the development of new and more efficient approach to solve Maxwell's equation. The interest in CEM applications is growing on. Several problems, hard to figure out few years ago, can now be easily addressed thanks to the reliability and flexibility of new technologies, together with the increased computational power. This technology evolution opens the possibility to address large and complex tasks. Many of these applications aim to simulate the electromagnetic behavior, for example in terms of input impedance and radiation pattern in antenna problems, or Radar Cross Section for scattering applications. Instead, problems, which solution requires high accuracy, need to implement full wave analysis techniques, e.g., virtual prototyping context, where the objective is to obtain reliable simulations in order to minimize measurement number, and as consequence their cost. Besides, other tasks require the analysis of complete structures (that include an high number of details) by directly simulating a CAD Model. This approach allows to relieve researcher of the burden of removing useless details, while maintaining the original complexity and taking into account all details. Unfortunately, this reduction implies: (a) high computational effort, due to the increased number of degrees of freedom, and (b) worsening of spectral properties of the linear system during complex analysis. The above considerations underline the needs to identify appropriate information technologies that ease solution achievement and fasten required elaborations. The authors analysis and expertise infer that Grid Computing techniques can be very useful to these purposes. Grids appear mainly in high performance computing environments. In this context, hundreds of off-the-shelf nodes are linked together and work in parallel to solve problems, that, previously, could be addressed sequentially or by using supercomputers. Grid Computing is a technique developed to elaborate enormous amounts of data and enables large-scale resource sharing to solve problem by exploiting distributed scenarios. The main advantage of Grid is due to parallel computing, indeed if a problem can be split in smaller tasks, that can be executed independently, its solution calculation fasten up considerably. To exploit this advantage, it is necessary to identify a technique able to split original electromagnetic task into a set of smaller subproblems. The Domain Decomposition (DD) technique, based on the block generation algorithm introduced in Matekovits et al. (2007) and Francavilla et al. (2011), perfectly addresses our requirements (see Section 3.4 for details). In this chapter, a Grid Computing infrastructure is presented. This architecture allows parallel block execution by distributing tasks to nodes that belong to the Grid. The set of nodes is composed by physical machines and virtualized ones. This feature enables great flexibility and increase available computational power. Furthermore, the presence of virtual nodes allows a full and efficient Grid usage, indeed the presented architecture can be used by different users that run different applications
Digital Ecosystems: Ecosystem-Oriented Architectures
We view Digital Ecosystems to be the digital counterparts of biological
ecosystems. Here, we are concerned with the creation of these Digital
Ecosystems, exploiting the self-organising properties of biological ecosystems
to evolve high-level software applications. Therefore, we created the Digital
Ecosystem, a novel optimisation technique inspired by biological ecosystems,
where the optimisation works at two levels: a first optimisation, migration of
agents which are distributed in a decentralised peer-to-peer network, operating
continuously in time; this process feeds a second optimisation based on
evolutionary computing that operates locally on single peers and is aimed at
finding solutions to satisfy locally relevant constraints. The Digital
Ecosystem was then measured experimentally through simulations, with measures
originating from theoretical ecology, evaluating its likeness to biological
ecosystems. This included its responsiveness to requests for applications from
the user base, as a measure of the ecological succession (ecosystem maturity).
Overall, we have advanced the understanding of Digital Ecosystems, creating
Ecosystem-Oriented Architectures where the word ecosystem is more than just a
metaphor.Comment: 39 pages, 26 figures, journa
- …