806 research outputs found

    Monitoring Cluster on Online Compiler with Ganglia

    Get PDF
    Ganglia is an open source monitoring system for high performance computing (HPC) that collect both a whole cluster and every nodes status and report to the user. We use Ganglia to monitor our spasi.informatika.lipi.go.id (SPASI), a customized-fedora10-based cluster, for our cluster online compiler, CLAW (cluster access through web). Our experience on using Ganglia shows that Ganglia has a capability to view our cluster status and allow us to track them

    The Changing Patterns of Internet Usage

    Get PDF
    The Internet unquestionably represents one of the most important technological developments in recent history. It has revolutionized the way people communicate with one another and obtain information and created an unimaginable variety of commercial and leisure activities. Interestingly, many members of the engineering community often observe that the current network is ill-suited to handle the demands that end users are placing on it. Indeed, engineering researchers often describe the network as ossified and impervious to significant architectural change. As a result, both the U.S. and the European Commission are sponsoring “clean slate” projects to study how the Internet might be designed differently if it were designed from scratch today. This Essay explores emerging trends that are transforming the way end users are using the Internet and examine their implications both for network architecture and public policy. These trends include Internet protocol video, wireless broadband, cloud computing programmable networking, and pervasive computing and sensor networks. It discusses how these changes in the way people are using the network may require the network to evolve in new directions

    The Changing Patterns of Internet Usage

    Get PDF
    The Internet unquestionably represents one of the most important technological developments in recent history. It has revolutionized the way people communicate with one another and obtain information and created an unimaginable variety of commercial and leisure activities. Interestingly, many members of the engineering community often observe that the current network is ill-suited to handle the demands that end users are placing on it. Indeed, engineering researchers often describe the network as ossified and impervious to significant architectural change. As a result, both the U.S. and the European Commission are sponsoring “clean slate” projects to study how the Internet might be designed differently if it were designed from scratch today. This Essay explores emerging trends that are transforming the way end users are using the Internet and examine their implications both for network architecture and public policy. These trends include Internet protocol video, wireless broadband, cloud computing programmable networking, and pervasive computing and sensor networks. It discusses how these changes in the way people are using the network may require the network to evolve in new directions

    Multicasting in Network Function Virtualization (NFV) Environment

    Get PDF
    Network Function Virtualization is a growing concept in the research field because of its ability to decouple network functions, like network address translation (NAT), domain name service (DNS), firewall, intrusion detection (IDS) etc., from proprietary hardware equipment. They can now run in software making the network more flexible and agile. This also reduces hardware and maintenance costs of the network. Nowadays many applications use multicasting as it saves a huge amount of communication bandwidth. But many packets need intermediary processing before reaching their destinations. For this processing, Virtual Network functions (VNFs) are implemented in the network where processing of packets takes place. Because of this the path through which the packets traverse changes, and delay increases. This project considers different number and placements of VNFs in four real-world topologies namely NSFNET, Cost239, Arpanet and Random12, and observes the delay for every case. As the VNFs are duplicated on different nodes in the network, the cost of deployment and maintenance of VNFs is increased, but the delay decreases up to a certain number of VNFs. After this, the delay becomes constant. This project presents this trade-off between cost and delay

    The Changing Patterns of Internet Usage

    Get PDF
    Symposium: Essays from Time Warner Cable\u27s Research Program on Digital Communications

    Resource sharing and hybrid libraries: the MALIBU Project

    Get PDF
    The British electronic libraries programme has funded four projects to develop thinking on so-called hybrid libraries and one of these is MALIBU (Modernising Academic Libraries in British Universities), based at King's College London. Working with the libraries of Oxford and Southampton Universities it will take a very rich set of humanities resources from archives and incunabula to digital products and networked resources and create a seamless single access point to all the available resources. It will also explore how strategies can be developed to make resources available locally as operational services rather than unreliably over the Internet. It does not set out to create its own new tools but rather to find ways of integrating tools already in existence or being developed

    Continuum: an architecture for user evolvable collaborative virtual environments

    Get PDF
    Continuum is a software platform for collaborative virtual environments. Continuum\u27s architecture supplies a world model and defines how to combine object state, behavior code, and resource data into this single shared structure. The system frees distributed users from the constraints of monolithic centralized virtual world architectures and instead allows individual users to extend and evolve the virtual world by creating and controlling their own individual pieces of the larger world model. The architecture provides support for data distribution, code management, resource management, and rapid deployment through standardized viewers. This work not only provides this architecture, but it includes a proven implementation and the associated development tools to allow for creation of these worlds

    A Policy-Based Resource Brokering Environment for Computational Grids

    Get PDF
    With the advances in networking infrastructure in general, and the Internet in particular, we can build grid environments that allow users to utilize a diverse set of distributed and heterogeneous resources. Since the focus of such environments is the efficient usage of the underlying resources, a critical component is the resource brokering environment that mediates the discovery, access and usage of these resources. With the consumer\u27s constraints, provider\u27s rules, distributed heterogeneous resources and the large number of scheduling choices, the resource brokering environment needs to decide where to place the user\u27s jobs and when to start their execution in a way that yields the best performance for the user and the best utilization for the resource provider. As brokering and scheduling are very complicated tasks, most current resource brokering environments are either specific to a particular grid environment or have limited features. This makes them unsuitable for large applications with heterogeneous requirements. In addition, most of these resource brokering environments lack flexibility. Policies at the resource-, application-, and system-levels cannot be specified and enforced to provide commitment to the guaranteed level of allocation that can help in attracting grid users and contribute to establishing credibility for existing grid environments. In this thesis, we propose and prototype a flexible and extensible Policy-based Resource Brokering Environment (PROBE) that can be utilized by various grid systems. In designing PROBE, we follow a policy-based approach that provides PROBE with the intelligence to not only match the user\u27s request with the right set of resources, but also to assure the guaranteed level of the allocation. PROBE looks at the task allocation as a Service Level Agreement (SLA) that needs to be enforced between the resource provider and the resource consumer. The policy-based framework is useful in a typical grid environment where resources, most of the time, are not dedicated. In implementing PROBE, we have utilized a layered architecture and façade design patterns. These along with the well-defined API, make the framework independent of any architecture and allow for the incorporation of different types of scheduling algorithms, applications and platform adaptors as the underlying environment requires. We have utilized XML as a base for all the specification needs. This provides a flexible mechanism to specify the heterogeneous resources and user\u27s requests along with their allocation constraints. We have developed XML-based specifications by which high-level internal structures of resources, jobs and policies can be specified. This provides interoperability in which a grid system can utilize PROBE to discover and use resources controlled by other grid systems. We have implemented a prototype of PROBE to demonstrate its feasibility. We also describe a test bed environment and the evaluation experiments that we have conducted to demonstrate the usefulness and effectiveness of our approach
    corecore