229 research outputs found

    Clustering Techniques : A solution for e-business

    Get PDF
    The purpose of this thesis was to provide the best clustering solution for the Archipelago web site project which would have been part of the Central Baltic Intereg IV programme 2007-2013. The entire program is a merger between the central Baltic regions of Finland, including the Åland Islands, Sweden and Estonia. A literature review of articles and research on various clustering techniques for the different sections of the project led to the findings of this document. Clustering was needed for web servers and the underlying database implementation. Additionally, the operating system used for all servers in both sections was required to present the best clustering solution. Implementing OSI layer 7 clustering for the web server cluster, MySQL database clustering and using Linux operating system would have provided the best solution for the Archipelago website. This implementation would have provided unlimited scalability, availability and high performance for the web site. Also, it is the most cost effective solution because it would utilize the commodity hardware

    Experience and Lessons learnt from running High Availability Databases on Network Attached Storage

    Get PDF
    The Database and Engineering Services Group of CERN's Information Technology Department supplies the Oracle Central Database services used in many activities at CERN. In order to provide High Availability and ease management for those services, a NAS (Network Attached Storage) based infrastructure has been setup. It runs several instances of the Oracle RAC (Real Application Cluster) using NFS (Network File System) as shared disk space for RAC purposes and Data hosting. It is composed of two private LANs (Local Area Network), one to provide access to the NAS filers and a second to implement the Oracle RAC private interconnect, both using Network Bonding. NAS filers are configured in partnership to prevent having single points of failure and to provide automatic NAS filer fail-over

    Databases in High Energy Physics: a critial review

    Get PDF
    The year 2000 is marked by a plethora of significant milestones in the history of High Energy Physics. Not only the true numerical end to the second millennium, this watershed year saw the final run of CERN's Large Electron-Positron collider (LEP) - the world-class machine that had been the focus of the lives of many of us for such a long time. It is also closely related to the subject of this chapter in the following respects: - Classified as a nuclear installation, information on the LEP machine must be retained indefinitely. This represents a challenge to the database community that is almost beyond discussion - archiving of data for a relatively small number of years is indeed feasible, but retaining it for centuries, millennia or more is a very different issue; - There are strong scientific arguments as to why the data from the LEP machine should be retained for a short period. However, the complexity of the data itself, the associated metadata and the programs that manipulate it make even this a huge challenge; - The story of databases in HEP is closely linked to that of LEP itself: what were the basic requirements that were identified in the early years of LEP preparation? How well have these been satisfied? What are the remaining issues and key messages? - Finally, the year 2000 also marked the entry of Grid architectures into the central stage of HEP computing. How has the Grid affected the requirements on databases or the manner in which they are deployed? Furthermore, as the LEP tunnel and even parts of the detectors that it housed are readied for re-use for the Large Hadron Collider (LHC), how have our requirements on databases evolved at this new scale of computing? A number of the key players in the field of databases - as can be seen from the author list of the various publications - have since retired from the field or else this world. Given the fallibility of human memory, the need for a record of the use of databases for physics data processing is clearly needed before memories fade completely and the story is lost forever. It is necessarily somewhat CERN-centric, although effort has been made to cover important developments and events elsewhere. Frequent reference is made to the Computing in High Energy Physics (CHEP) conference series - the most accessible and consistent record of this field

    April-June 2008

    Get PDF

    Database System Architecture for Fault tolerance and Disaster Recovery

    Get PDF
    Application systems being used today rely heavily on the availability of the database system. Disruption of database system can be damaging and catastrophic to the organization that depends on the availability of the database system for its business and service operations. To ensure business continuity under foreseeable and unforeseeable man-made or natural disasters, the database system has to be designed and built with fault tolerance and disaster recovery capabilities. This project explored existing technologies and solutions to design, build, and implement database system architecture for fault tolerance and disaster recovery using Oracle database software products. The project goal was to implement database system architecture for migrating multiple web applications and databases onto a consolidated system architecture providing high availability database application systems

    The End of a Myth: Distributed Transactions Can Scale

    Full text link
    The common wisdom is that distributed transactions do not scale. But what if distributed transactions could be made scalable using the next generation of networks and a redesign of distributed databases? There would be no need for developers anymore to worry about co-partitioning schemes to achieve decent performance. Application development would become easier as data placement would no longer determine how scalable an application is. Hardware provisioning would be simplified as the system administrator can expect a linear scale-out when adding more machines rather than some complex sub-linear function, which is highly application specific. In this paper, we present the design of our novel scalable database system NAM-DB and show that distributed transactions with the very common Snapshot Isolation guarantee can indeed scale using the next generation of RDMA-enabled network technology without any inherent bottlenecks. Our experiments with the TPC-C benchmark show that our system scales linearly to over 6.5 million new-order (14.5 million total) distributed transactions per second on 56 machines.Comment: 12 page

    The LHC Logging Service: Capturing, storing and using time-series data for the world's largest scientific instrument

    Get PDF
    CERN, the European Laboratory for Particle Physics, is well underway in building the most powerful particle accelerator called LHC (Large Hadron Collider), which will probe deeper into matter than ever before. This circular 27-km long superconducting installation is extremely complex, and its functioning has to be closely monitored. The LHC Logging service is aimed to satisfy the requirement of capturing and storing of any relevant piece of information to track its variation over time. Web-deployed tools have been developed to visualize, correlate and export the data into dedicated off-line analysis tools. The quality of the data, the manageability of the service and the overall system performance are key factors for the service. Oracle technology has been used extensively to support this mission-critical service, which has proven already to be useful during the commissioning phase of individual subsystems of the LHC. The architecture, design and implementation of the LHC Logging service, based on Oracle Database, Application Servers and the Enterprise Manager, are described in this paper

    Palvelulähtöisen arkkitehtuurin sovellusalustan käyttö pilvilaskentainfrastruktuurissa

    Get PDF
    Cloud computing is a growing Information Technology (IT) trend. It is said that cloud computing can change the essence IT business. Cloud computing is, like many other techniques, a convergence of old ideologies and techniques into something new. It combines features from cluster and grid computing with the help of virtualization. Cloud computing offers its users a seemingly infinite pooled computing resource over the network. Users can start, stop, and scale its power at will. The goal of this thesis is to give a high level picture of cloud computing and implement a working prototype of a Service Oriented Architecture (SOA) platform on Amazon's Elastic Compute Cloud (EC2). The theoretical part studies what is cloud computing. Its supporting techniques and defining features are looked into as well as what benefits and drawbacks there are. Then a SOA Platform is implemented in EC2 in two separate configurations. The former implementation is an automatically scaling SOA cluster configuration. The cluster consists of Oracle SOA stack and uses the cloud computing. The cluster is used as platform to enable building of other enterprise systems using SOA technologies. The latter is a SOA training environment, which is used for training people to use Oracle SOA stack. Cloud computing enables trainees to take part in training sessions even with slow computers, because the necessary resources are accumulated from the cloud. Both implementations successfully demonstrate some benefits and drawbacks of cloud computing. For example, the lack of upfront costs made this kind of project possible. Meanwhile, need for absent special requirements prevented the use of Real Application Cluster (RAC) database. Cloud computing is still evolving to smooth out some of its rough edges. /Kir1

    The Research of Massive Data Processing Capability of DML Operation Based on Oracle RAC

    Get PDF
    随着IT应用的广泛与深入发展,传统的计算架构难于适应企业急速发展带来的需求。企业需要他们的系统以最低的成本、最高的效率、最出色的可用性为企业提供服务。而事实上,他们的信息系统却往往响应缓慢,为了解决这个问题,各企业不得不构建昂贵的基础架构,而这些架构的设计能力和冗余度却远远超出应付偶然事件的需要。同时,随着TB时代的来临,海量数据大大降低了数据库日常操作和批量处理的性能与效率。使用传统的数据库DML处理方法,即使在硬件支撑能力充足的情况下,数据库处理TB级数据的效率仍然十分低下。Oracle10g是第一个为企业网格计算而设计的数据库,它带来了许多新技术,并大幅度改良了数据库性能。IT软硬件技术...With the extensive and further development of IT application, the traditional structure of the calculation can’t meet the needs of enterprises with rapid development. Enterprises need their systems with the lowest cost, the most efficient and outstanding availability to serve enterprises. Indeed, their information systems usually respond slowly. To solve these problems, all enterprises had to buil...学位:工学硕士院系专业:信息科学与技术学院计算机科学系_计算机应用技术学号:20042803
    corecore