3,183 research outputs found
The Glasgow raspberry pi cloud: a scale model for cloud computing infrastructures
Data Centers (DC) used to support Cloud services
often consist of tens of thousands of networked machines under a single roof. The significant capital outlay required to replicate such infrastructures constitutes a major obstacle to practical implementation and evaluation of research in this domain. Currently, most research into Cloud computing relies on either limited software simulation, or the use of a testbed environments
with a handful of machines. The recent introduction of the
Raspberry Pi, a low-cost, low-power single-board computer, has made the construction of a miniature Cloud DCs more affordable.
In this paper, we present the Glasgow Raspberry Pi Cloud
(PiCloud), a scale model of a DC composed of clusters of
Raspberry Pi devices. The PiCloud emulates every layer of a
Cloud stack, ranging from resource virtualisation to network
behaviour, providing a full-featured Cloud Computing research and educational environment
Adaptive laser link reconfiguration using constraint propagation
This paper describes Harris AI research performed on the Adaptive Link Reconfiguration (ALR) study for Rome Lab, and focuses on the application of constraint propagation to the problem of link reconfiguration for the proposed space based Strategic Defense System (SDS) Brilliant Pebbles (BP) communications system. According to the concept of operations at the time of the study, laser communications will exist between BP's and to ground entry points. Long-term links typical of RF transmission will not exist. This study addressed an initial implementation of BP's based on the Global Protection Against Limited Strikes (GPALS) SDI mission. The number of satellites and rings studied was representative of this problem. An orbital dynamics program was used to generate line-of-site data for the modeled architecture. This was input into a discrete event simulation implemented in the Harris developed COnstraint Propagation Expert System (COPES) Shell, developed initially on the Rome Lab BM/C3 study. Using a model of the network and several heuristics, the COPES shell was used to develop the Heuristic Adaptive Link Ordering (HALO) Algorithm to rank and order potential laser links according to probability of communication. A reduced set of links based on this ranking would then be used by a routing algorithm to select the next hop. This paper includes an overview of Constraint Propagation as an Artificial Intelligence technique and its embodiment in the COPES shell. It describes the design and implementation of both the simulation of the GPALS BP network and the HALO algorithm in COPES. This is described using a 59 Data Flow Diagram, State Transition Diagrams, and Structured English PDL. It describes a laser communications model and the heuristics involved in rank-ordering the potential communication links. The generation of simulation data is described along with its interface via COPES to the Harris developed View Net graphical tool for visual analysis of communications networks. Conclusions are presented, including a graphical analysis of results depicting the ordered set of links versus the set of all possible links based on the computed Bit Error Rate (BER). Finally, future research is discussed which includes enhancements to the HALO algorithm, network simulation, and the addition of an intelligent routing algorithm for BP
Reference Nodes Selection for Anchor-Free Localization in Wireless Sensor Networks
Dizertační práce se zabývá návrhem nového bezkotevního lokalizačního algoritmu sloužícího pro výpočet pozice uzlů v bezdrátových senzorových sítích. Provedené studie ukázaly, že dosavadní bezkotevní lokalizační algoritmy, pracující v paralelním režimu, dosahují malých lokalizačních chyb. Jejich nevýhodou ovšem je, že při sestavení množiny referenčních uzlu spotřebovávají daleko větší množství energie než algoritmy pracující v inkrementálním režimu. Paralelní lokalizační algoritmy využívají pro určení pozice referenční uzly nacházející se na protilehlých hranách bezdrátové sítě. Nový lokalizační algoritmus označený jako BRL (Boundary Recognition aided Localization) je založen na myšlence decentralizovaně detekovat uzly ležící na hranici síti a pouze z této množiny vybrat potřebný počet referenčních uzlu. Pomocí navrženého přístupu lze znažně snížit množství energie spotřebované v průběhu procesu výběru referenčních uzlů v senzorovém poli. Dalším přínosem ke snížení energetických nároku a zároveň zachování nízké lokalizační chyby je využití procesu multilaterace se třemi, eventuálně čtyřmi referenčními body. V rámci práce byly provedeny simulace několika dílčích algoritmu a jejich funkčnost byla ověřena experimentálně v reálné senzorové síti. Navržený algoritmus BRL byl porovnán z hlediska lokalizační chyby a počtu zpracovaných paketů s několika známými lokalizačními algoritmy. Výsledky simulací dokázaly, že navržený algoritmus představuje efektivní řešení pro přesnou a zároveň nízkoenergetickou lokalizaci uzlů v bezdrátových senzorových sítích.The doctoral thesis is focused on a design of a novel anchor free localization algorithm for wireless sensor networks. As introduction, the incremental and concurrent anchor free localization algorithms are presented and their performance is compared. It was found that contemporary anchor free localization algorithms working in the concurrent manner achieve a low localization error, but dissipate signicant energy reserves. A new Boundary Recognition Aided Localization algorithm presented in this thesis is based on an idea to recognize the nodes placed on the boundary of network and thus reduce the number of transmission realized during the reference nodes selection phase of the algorithm. For the position estimation, the algorithm employs the multilateration technique that work eectively with the low number of the reference nodes. Proposed algorithms are tested through the simulations and validated by the real experiment with the wireless sensor network. The novel Boundary Recognition Aided Localization algorithm is compared with the known algorithms in terms of localization error and the communication cost. The results show that the novel algorithm presents powerful solution for the anchor free localization.
A deep representation for depth images from synthetic data
Convolutional Neural Networks (CNNs) trained on large scale RGB databases
have become the secret sauce in the majority of recent approaches for object
categorization from RGB-D data. Thanks to colorization techniques, these
methods exploit the filters learned from 2D images to extract meaningful
representations in 2.5D. Still, the perceptual signature of these two kind of
images is very different, with the first usually strongly characterized by
textures, and the second mostly by silhouettes of objects. Ideally, one would
like to have two CNNs, one for RGB and one for depth, each trained on a
suitable data collection, able to capture the perceptual properties of each
channel for the task at hand. This has not been possible so far, due to the
lack of a suitable depth database. This paper addresses this issue, proposing
to opt for synthetically generated images rather than collecting by hand a 2.5D
large scale database. While being clearly a proxy for real data, synthetic
images allow to trade quality for quantity, making it possible to generate a
virtually infinite amount of data. We show that the filters learned from such
data collection, using the very same architecture typically used on visual
data, learns very different filters, resulting in depth features (a) able to
better characterize the different facets of depth images, and (b) complementary
with respect to those derived from CNNs pre-trained on 2D datasets. Experiments
on two publicly available databases show the power of our approach
Emulating opportunistic networks with KauNet Triggers
In opportunistic networks the availability of an end-to-end path is no longer required. Instead opportunistic networks may take advantage of temporary connectivity opportunities.
Opportunistic networks present a demanding environment for network emulation as the traditional emulation setup, where application/transport endpoints only send and receive packets from the network following a black box approach,
is no longer applicable. Opportunistic networking protocols
and applications additionally need to react to the dynamics of the underlying network beyond what is conveyed through the exchange of packets.
In order to support IP-level emulation evaluations of applications and protocols that react to lower layer events, we have proposed the use of emulation triggers. Emulation triggers can emulate arbitrary cross-layer feedback and can be synchronized with other emulation effects. After introducing the design and implementation of
triggers in the KauNet emulator, we describe the integration of triggers with the DTN2 reference implementation and illustrate how the functionality can be used to emulate a classical DTN data-mule scenario
- …