3,721 research outputs found
On the Topology Maintenance of Dynamic P2P Overlays through Self-Healing Local Interactions
This paper deals with the use of self-organizing protocols to improve the
reliability of dynamic Peer-to-Peer (P2P) overlay networks. We present two
approaches, that employ local knowledge of the 2nd neighborhood of nodes. The
first scheme is a simple protocol requiring interactions among nodes and their
direct neighbors. The second scheme extends this approach by resorting to the
Edge Clustering Coefficient (ECC), a local measure that allows to identify
those edges that connect different clusters in an overlay. A simulation
assessment is presented, which evaluates these protocols over uniform networks,
clustered networks and scale-free networks. Different failure modes are
considered. Results demonstrate the viability of the proposal.Comment: A revised version of the paper appears in Proc. of the IFIP
Networking 2014 Conference, IEEE, Trondheim, (Norway), June 201
Self-Healing Protocols for Connectivity Maintenance in Unstructured Overlays
In this paper, we discuss on the use of self-organizing protocols to improve
the reliability of dynamic Peer-to-Peer (P2P) overlay networks. Two similar
approaches are studied, which are based on local knowledge of the nodes' 2nd
neighborhood. The first scheme is a simple protocol requiring interactions
among nodes and their direct neighbors. The second scheme adds a check on the
Edge Clustering Coefficient (ECC), a local measure that allows determining
edges connecting different clusters in the network. The performed simulation
assessment evaluates these protocols over uniform networks, clustered networks
and scale-free networks. Different failure modes are considered. Results
demonstrate the effectiveness of the proposal.Comment: The paper has been accepted to the journal Peer-to-Peer Networking
and Applications. The final publication is available at Springer via
http://dx.doi.org/10.1007/s12083-015-0384-
A Peer-to-Peer Middleware Framework for Resilient Persistent Programming
The persistent programming systems of the 1980s offered a programming model
that integrated computation and long-term storage. In these systems, reliable
applications could be engineered without requiring the programmer to write
translation code to manage the transfer of data to and from non-volatile
storage. More importantly, it simplified the programmer's conceptual model of
an application, and avoided the many coherency problems that result from
multiple cached copies of the same information. Although technically
innovative, persistent languages were not widely adopted, perhaps due in part
to their closed-world model. Each persistent store was located on a single
host, and there were no flexible mechanisms for communication or transfer of
data between separate stores. Here we re-open the work on persistence and
combine it with modern peer-to-peer techniques in order to provide support for
orthogonal persistence in resilient and potentially long-running distributed
applications. Our vision is of an infrastructure within which an application
can be developed and distributed with minimal modification, whereupon the
application becomes resilient to certain failure modes. If a node, or the
connection to it, fails during execution of the application, the objects are
re-instantiated from distributed replicas, without their reference holders
being aware of the failure. Furthermore, we believe that this can be achieved
within a spectrum of application programmer intervention, ranging from minimal
to totally prescriptive, as desired. The same mechanisms encompass an
orthogonally persistent programming model. We outline our approach to
implementing this vision, and describe current progress.Comment: Submitted to EuroSys 200
Management and Service-aware Networking Architectures (MANA) for Future Internet Position Paper: System Functions, Capabilities and Requirements
Future Internet (FI) research and development threads have recently been gaining momentum all over the world and as such the international race to create a new generation Internet is in full swing: GENI, Asia Future Internet, Future Internet Forum Korea, European Union Future Internet Assembly (FIA). This is a position paper identifying the research orientation with a time horizon of 10 years, together with the key challenges for the capabilities in the Management and Service-aware Networking Architectures (MANA) part of the Future Internet (FI) allowing for parallel and federated Internet(s)
Recommended from our members
Multimedia delivery in the future internet
The term âNetworked Mediaâ implies that all kinds of media including text, image, 3D graphics, audio
and video are produced, distributed, shared, managed and consumed on-line through various networks,
like the Internet, Fiber, WiFi, WiMAX, GPRS, 3G and so on, in a convergent manner [1]. This white
paper is the contribution of the Media Delivery Platform (MDP) cluster and aims to cover the Networked
challenges of the Networked Media in the transition to the Future of the Internet.
Internet has evolved and changed the way we work and live. End users of the Internet have been confronted
with a bewildering range of media, services and applications and of technological innovations concerning
media formats, wireless networks, terminal types and capabilities. And there is little evidence that the pace
of this innovation is slowing. Today, over one billion of users access the Internet on regular basis, more
than 100 million users have downloaded at least one (multi)media file and over 47 millions of them do so
regularly, searching in more than 160 Exabytes1 of content. In the near future these numbers are expected
to exponentially rise. It is expected that the Internet content will be increased by at least a factor of 6, rising
to more than 990 Exabytes before 2012, fuelled mainly by the users themselves. Moreover, it is envisaged
that in a near- to mid-term future, the Internet will provide the means to share and distribute (new)
multimedia content and services with superior quality and striking flexibility, in a trusted and personalized
way, improving citizensâ quality of life, working conditions, edutainment and safety.
In this evolving environment, new transport protocols, new multimedia encoding schemes, cross-layer inthe
network adaptation, machine-to-machine communication (including RFIDs), rich 3D content as well as
community networks and the use of peer-to-peer (P2P) overlays are expected to generate new models of
interaction and cooperation, and be able to support enhanced perceived quality-of-experience (PQoE) and
innovative applications âon the moveâ, like virtual collaboration environments, personalised services/
media, virtual sport groups, on-line gaming, edutainment. In this context, the interaction with content
combined with interactive/multimedia search capabilities across distributed repositories, opportunistic P2P
networks and the dynamic adaptation to the characteristics of diverse mobile terminals are expected to
contribute towards such a vision.
Based on work that has taken place in a number of EC co-funded projects, in Framework Program 6 (FP6)
and Framework Program 7 (FP7), a group of experts and technology visionaries have voluntarily
contributed in this white paper aiming to describe the status, the state-of-the art, the challenges and the way
ahead in the area of Content Aware media delivery platforms
A peer-to-peer infrastructure for resilient web services
This work is funded by GR/M78403 âSupporting Internet Computation in Arbitrary Geographical Locationsâ and GR/R51872 âReflective Application Framework for Distributed Architecturesâ, and by Nuffield Grant URB/01597/G âPeer-to-Peer Infrastructure for Autonomic Storage ArchitecturesâThis paper describes an infrastructure for the deployment and use of Web Services that are resilient to the failure of the nodes that host those services. The infrastructure presents a single interface that provides mechanisms for users to publish services and to find hosted services. The infrastructure supports the autonomic deployment of services and the brokerage of hosts on which services may be deployed. Once deployed, services are autonomically managed in a number of aspects including load balancing, availability, failure detection and recovery, and lifetime management. Services are published and deployed with associated metadata describing the service type. This same metadata may be used subsequently by interested parties to discover services. The infrastructure uses peer-to-peer (P2P) overlay technologies to abstract over the underlying network to deploy and locate instances of those services. It takes advantage of the P2P network to replicate directory services used to locate service instances (for using a service), Service Hosts (for deployment of services) and Autonomic Managers which manage the deployed services. The P2P overlay network is itself constructed using novel Web Services-based middleware and a variation of the Chord P2P protocol, which is self-managing.Postprin
Innovation as a Nonlinear Process, the Scientometric Perspective, and the Specification of an "Innovation Opportunities Explorer"
The process of innovation follows non-linear patterns across the domains of
science, technology, and the economy. Novel bibliometric mapping techniques can
be used to investigate and represent distinctive, but complementary
perspectives on the innovation process (e.g., "demand" and "supply") as well as
the interactions among these perspectives. The perspectives can be represented
as "continents" of data related to varying extents over time. For example, the
different branches of Medical Subject Headings (MeSH) in the Medline database
provide sources of such perspectives (e.g., "Diseases" versus "Drugs and
Chemicals"). The multiple-perspective approach enables us to reconstruct facets
of the dynamics of innovation, in terms of selection mechanisms shaping
localizable trajectories and/or resulting in more globalized regimes. By
expanding the data with patents and scholarly publications, we demonstrate the
use of this multi-perspective approach in the case of RNA Interference (RNAi).
The possibility to develop an "Innovation Opportunities Explorer" is specified.Comment: Technology Analysis and Strategic Management (forthcoming in 2013
MOSAIC: Unified Platform for Dynamic Overlay Selection and Composition
MOSAIC constructs new overlay networks with desired characteristics by composing existing overlays with subsets of those attributes. Thus, MOSAIC overcomes the problem of multiple network infrastructures that are partial solutions, while preserving deployability. Composition of control and/or data planes is possible in the system. MOSAIC overlays are specified in Mozlog, a declarative language that specifies overlay properties without binding them to a particular implementation or underlying network.
This paper focuses on the runtime aspects of MOSAIC: how it enables interoperability between different overlay networks and how it implements switching between different overlay compositions, permitting dynamic compositions with both existing overlay networks and legacy applications. The system is validated experimentally using declarative overlay compositions concisely specified in Mozlog: an indirection overlay that supports mobility (i3), a resilient overlay (RON), and scalable lookups (Chord), all of which are combined to provide new functionality. MOSAIC provides the benefits of runtime composition to simultaneously deliver application-aware mobility, NAT traversal and reliability with low performance overhead, demonstrated by measurements on both a local cluster and PlanetLab
Symbolic representation of scenarios in Bologna airport on virtual reality concept
This paper is a part of a big Project named Retina Project, which is focused in reduce the workload of an ATCO. It uses the last technological advances as Virtual Reality concept. The work has consisted in studying the different awareness situations that happens daily in Bologna Airport. It has been analysed one scenario with good visibility where the sun predominates and two other scenarios with poor visibility where the rain and the fog dominate. Due to the study of visibility in the three scenarios computed, the conclusion obtained is that the overlay must be shown with a constant dimension regardless the position of the aircraft to be readable by the ATC and also, the frame and the flight strip should be coloured in a showy colour (like red) for a better control by the ATCO
- âŠ