434 research outputs found

    MFACE: A Multicast Backbone-Assisted Face Traversal Algorithm for Arbitrary Planar Ad Hoc and Sensor Network Topologies

    Get PDF
    Face is a well-known localized routing protocol for ad hoc and sensor networks which guarantees delivery of the message as long as a path exists between the source and the destination. This is achieved by employing a left/right hand rule to route the message along the faces of a planar topology. Although face was developed for the unicast case, it has recently been used in combination with multicasting protocols, where there are multiple destinations. Some of the proposed solutions handle each destination separately and lead thus to increased energy consumption. Extensions of face recovery to the multicast case described so far are either limited to certain planar graphs or do not provide delivery guarantees. A recently described scheme employs multicast face recovery based on a so called multicast backbone. A multicast backbone is a Euclidean spanning tree which contains at least the source and the destination nodes. The idea of backbone assisted routing it to follow the edges of the backbone in order to deliver a multicast message to all spanned destination nodes. The existing backbone face routing scheme is however limited to a certain planar graph type and a certain backbone construction. One of the key aspects of the multicast face algorithm MFACE we propose in this work is that it may be applied on top of any planar topology. Moreover, our solution may be used as a generic framework since it is able to work with any arbitrary multicast backbone. In MFACE, any edge of the backbone originated at the source node will generate a new copy of the message which will be routed toward the set of destination nodes spanned by the corresponding edge. Whenever the message arrives at a face edge intersected by a backbone edge different from the initial edge, the message is split into two copies, both handling a disjoint subset of the multicast destinations which are defined by splitting the multicast backbone at that intersection point

    Fault Tolerant Mobile Sensor Node Traversal Schemes Based on Hexagonal Coverage

    Get PDF
    Mobile sensor nodes (MSNs) are equipped with locomotive can move around after having been deployed. They are equipped with limited energy. A large portion of energy is drained during the traversal. In order to extend the life time of a MSN, the traveling distance must be minimized. Region of interest (ROI) is covered with multiple MSNs using coverage based pattern movement. When a group of MSNs are deployed to cover a given ROI, all the deployed MSNs should travel an approximately equal distance. Otherwise, the MSN which travels longer distance depletes more energy compared to the MSN which travels a shorter distance. In this work we show that, ROI partition plays great role in hole free coverage and makes the MSNs have optimized movement cost with fault tolerant support

    Square grid Path Planning for Mobile Anchor-Based Localization in Wireless Sensor Networks

    Get PDF
    Localization is to provide all sensor nodes with their geographical positions. A mobile anchor-based localization in WSNs uses a mobile anchor equipped with GPS, which travels along a predetermined path. At each specified beacon point, it broadcasts its current known position to help other sensor nodes with unknown locations estimate their positions. In this paper, we analyze the determination of beacon points based on a square grid. We propose an improved path planning model named Union-curve. Our proposed model incorporates all beacon points of five previously developed paths, namely, SCAN, HILBERT, S-type, Z-curve, and ÎŁ\Sigma-Scan on the commonly used square grid decomposition of area. Unknown sensor nodes estimate their positions using two techniques, APT and WCWCL-RSSI. Simulation results show that the proposed model has higher accuracy, with a big difference in error rate compared to the other models. In addition, this model guarantees maximum coverage with less path resolution value

    Component-based synthesis of motion planning algorithms

    Get PDF
    Combinatory Logic Synthesis generates data or runnable programs according to formal type specifications. Synthesis results are composed based on a user-specified repository of components, which brings several advantages for representing spaces of high variability. This work suggests strategies to manage the resulting variations by proposing a domain-specific brute-force search and a machine learning-based optimization procedure. The brute-force search involves the iterative generation and evaluation of machining strategies. In contrast, machine learning optimization uses statistical models to enable the exploration of the design space. The approaches involve synthesizing programs and meta-programs that manipulate, run, and evaluate programs. The methodologies are applied to the domain of motion planning algorithms, and they include the configuration of programs belonging to different algorithmic families. The study of the domain led to the identification of variability points and possible variations. Proof-of-concept repositories represent these variability points and incorporate them into their semantic structure. The selected algorithmic families involve specific computation steps or data structures, and corresponding software components represent possible variations. Experimental results demonstrate that CLS enables synthesis-driven domain-specific optimization procedures to solve complex problems by exploring spaces of high variability.Combinatory Logic Synthesis (CLS) generiert Daten oder lauffähige Programme anhand von formalen Typspezifikationen. Die Ergebnisse der Synthese werden auf Basis eines benutzerdefinierten Repositories von Komponenten zusammengestellt, was diverse Vorteile für die Beschreibung von Räumen mit hoher Variabilität mit sich bringt. Diese Arbeit stellt Strategien für den Umgang mit den resultierenden Variationen vor, indem eine domänen-spezifische Brute-Force Suche und ein maschinelles Lernverfahren für die Untersuchung eines Optimierungsproblems aufgezeigt werden. Die Brute-Force Suche besteht aus der iterativen Generierung und Evaluation von Frässtrategien. Im Gegensatz dazu nutzt der Optimierungsansatz statistische Modelle zur Erkundung des Entwurfsraums. Beide Ansätze synthetisieren Programme und Metaprogramme, welche Programme bearbeiten, ausführen und evaluieren. Diese Methoden werden auf die Domäne der Bewegungsplanungsalgorithmen angewendet und sie beinhalten die Konfiguration von Programmen, welche zu unterschiedlichen algorithmischen Familien gehören. Die Untersuchung der Domäne führte zur Identifizierung der Variabilitätspunkte und der möglichen Variationen. Entsprechende Proof of Concept Implementierungen in Form von Repositories repräsentieren jene Variabilitätspunkte und beziehen diese in ihre semantische Struktur ein. Die gewählten algorithmischen Familien sehen bestimmte Berechnungsschritte oder Datenstrukturen vor, und entsprechende Software Komponenten stellen mögliche Variationen dar. Versuchsergebnisse belegen, dass CLS synthese-getriebene domänenspezifische Optimierungsverfahren ermöglicht, welche komplexe Probleme durch die Exploration von Räumen hoher Variabilität lösen

    3D exemplar-based image inpainting in electron microscopy

    Get PDF
    In electron microscopy (EM) a common problem is the non-availability of data, which causes artefacts in reconstructions. In this thesis the goal is to generate artificial data where missing in EM by using exemplar-based inpainting (EBI). We implement an accelerated 3D version tailored to applications in EM, which reduces reconstruction times from days to minutes. We develop intelligent sampling strategies to find optimal data as input for reconstruction methods. Further, we investigate approaches to reduce electron dose and acquisition time. Sparse sampling followed by inpainting is the most promising approach. As common evaluation measures may lead to misinterpretation of results in EM and falsify a subsequent analysis, we propose to use application driven metrics and demonstrate this in a segmentation task. A further application of our technique is the artificial generation of projections in tiltbased EM. EBI is used to generate missing projections, such that the full angular range is covered. Subsequent reconstructions are significantly enhanced in terms of resolution, which facilitates further analysis of samples. In conclusion, EBI proves promising when used as an additional data generation step to tackle the non-availability of data in EM, which is evaluated in selected applications. Enhancing adaptive sampling methods and refining EBI, especially considering the mutual influence, promotes higher throughput in EM using less electron dose while not lessening quality.Ein häufig vorkommendes Problem in der Elektronenmikroskopie (EM) ist die Nichtverfügbarkeit von Daten, was zu Artefakten in Rekonstruktionen führt. In dieser Arbeit ist es das Ziel fehlende Daten in der EM künstlich zu erzeugen, was durch Exemplar-basiertes Inpainting (EBI) realisiert wird. Wir implementieren eine auf EM zugeschnittene beschleunigte 3D Version, welche es ermöglicht, Rekonstruktionszeiten von Tagen auf Minuten zu reduzieren. Wir entwickeln intelligente Abtaststrategien, um optimale Datenpunkte für die Rekonstruktion zu erhalten. Ansätze zur Reduzierung von Elektronendosis und Aufnahmezeit werden untersucht. Unterabtastung gefolgt von Inpainting führt zu den besten Resultaten. Evaluationsmaße zur Beurteilung der Rekonstruktionsqualität helfen in der EM oft nicht und können zu falschen Schlüssen führen, weswegen anwendungsbasierte Metriken die bessere Wahl darstellen. Dies demonstrieren wir anhand eines Beispiels. Die künstliche Erzeugung von Projektionen in der neigungsbasierten Elektronentomographie ist eine weitere Anwendung. EBI wird verwendet um fehlende Projektionen zu generieren. Daraus resultierende Rekonstruktionen weisen eine deutlich erhöhte Auflösung auf. EBI ist ein vielversprechender Ansatz, um nicht verfügbare Daten in der EM zu generieren. Dies wird auf Basis verschiedener Anwendungen gezeigt und evaluiert. Adaptive Aufnahmestrategien und EBI können also zu einem höheren Durchsatz in der EM führen, ohne die Bildqualität merklich zu verschlechtern

    Towards Physarum Binary Adders

    Get PDF
    Plasmodium of \emph{Physarum polycephalum} is a single cell visible by unaided eye. The plasmodium's foraging behaviour is interpreted in terms of computation. Input data is a configuration of nutrients, result of computation is a network of plasmodium's cytoplasmic tubes spanning sources of nutrients. Tsuda et al (2004) experimentally demonstrated that basic logical gates can be implemented in foraging behaviour of the plasmodium. We simplify the original designs of the gates and show --- in computer models --- that the plasmodium is capable for computation of two-input two-output gate → \to and three-input two-output → \to . We assemble the gates in a binary one-bit adder and demonstrate validity of the design using computer simulation.Comment: Biosystems (2010), in press. Please download final version of the paper from the Publishers's sit

    A Voxel-Based Approach for Imaging Voids in Three-Dimensional Point Clouds

    Get PDF
    Geographically accurate scene models have enormous potential beyond that of just simple visualizations in regard to automated scene generation. In recent years, thanks to ever increasing computational efficiencies, there has been significant growth in both the computer vision and photogrammetry communities pertaining to automatic scene reconstruction from multiple-view imagery. The result of these algorithms is a three-dimensional (3D) point cloud which can be used to derive a final model using surface reconstruction techniques. However, the fidelity of these point clouds has not been well studied, and voids often exist within the point cloud. Voids exist in texturally difficult areas, as well as areas where multiple views were not obtained during collection, constant occlusion existed due to collection angles or overlapping scene geometry, or in regions that failed to triangulate accurately. It may be possible to fill in small voids in the scene using surface reconstruction or hole-filling techniques, but this is not the case with larger more complex voids, and attempting to reconstruct them using only the knowledge of the incomplete point cloud is neither accurate nor aesthetically pleasing. A method is presented for identifying voids in point clouds by using a voxel-based approach to partition the 3D space. By using collection geometry and information derived from the point cloud, it is possible to detect unsampled voxels such that voids can be identified. This analysis takes into account the location of the camera and the 3D points themselves to capitalize on the idea of free space, such that voxels that lie on the ray between the camera and point are devoid of obstruction, as a clear line of sight is a necessary requirement for reconstruction. Using this approach, voxels are classified into three categories: occupied (contains points from the point cloud), free (rays from the camera to the point passed through the voxel), and unsampled (does not contain points and no rays passed through the area). Voids in the voxel space are manifested as unsampled voxels. A similar line-of-sight analysis can then be used to pinpoint locations at aircraft altitude at which the voids in the point clouds could theoretically be imaged. This work is based on the assumption that inclusion of more images of the void areas in the 3D reconstruction process will reduce the number of voids in the point cloud that were a result of lack of coverage. Voids resulting from texturally difficult areas will not benefit from more imagery in the reconstruction process, and thus are identified and removed prior to the determination of future potential imaging locations
    • …
    corecore