913 research outputs found

    A scalable application server on Beowulf clusters : a thesis presented in partial fulfilment of the requirement for the degree of Master of Information Science at Albany, Auckland, Massey University, New Zealand

    Get PDF
    Application performance and scalability of a large distributed multi-tiered application is a core requirement for most of today's critical business applications. I have investigated the scalability of a J2EE application server using the standard ECperf benchmark application in the Massey Beowulf Clusters namely the Sisters and the Helix. My testing environment consists of Open Source software: The integrated JBoss-Tomcat as the application server and the web server, along with PostgreSQL as the database. My testing programs were run on the clustered application server, which provide replication of the Enterprise Java Bean (EJB) objects. I have completed various centralized and distributed tests using the JBoss Cluster. I concluded that clustering of the application server and web server will effectively increase the performance of the application running on them given sufficient system resources. The application performance will scale to a point where a bottleneck has occurred in the testing system, the bottleneck could be any resources included in the testing environment: the hardware, software, network and the application that is running. Performance tuning for a large-scale J2EE application is a complicated issue, which is related to the resources available. However, by carefully identifying the performance bottleneck in the system with hardware, software, network, operating system and application configuration. I can improve the performance of the J2EE applications running in a Beowulf Cluster. The software bottleneck can be solved by changing the default settings, on the other hand, hardware bottlenecks are harder unless more investment are made to purchase higher speed and capacity hardware

    Software engineering and middleware: a roadmap (Invited talk)

    Get PDF
    The construction of a large class of distributed systems can be simplified by leveraging middleware, which is layered between network operating systems and application components. Middleware resolves heterogeneity and facilitates communication and coordination of distributed components. Existing middleware products enable software engineers to build systems that are distributed across a local-area network. State-of-the-art middleware research aims to push this boundary towards Internet-scale distribution, adaptive and reconfigurable middleware and middleware for dependable and wireless systems. The challenge for software engineering research is to devise notations, techniques, methods and tools for distributed system construction that systematically build and exploit the capabilities that middleware deliver

    DISTRIBUTED BROKERAGE OFFICES THROUGH INFORMATION TECHNOLOGY

    Get PDF
    This paper describes some novel ways in which Edward D. Jones and Co., a successful brokerage firm with 1650 offices nationwide, uses information technology to pursue a unique market niche: single-broker offices in communities too small to support a traditional, typically much larger, brokerage branch office. The paper focuses on the use of mainframes with "dumb" CRT terminals, rather than workstations or personal computers, to coordinate distributed operational work on a day-to-day basis.Information Systems Working Papers Serie

    Methodologies for CIM systems integration in small batch manufacturing

    Get PDF
    This thesis is concerned with identifying the problems and constraints faced by small batch manufacturing companies during the implementation of Computer Integrated Manufacturing (CIM). The main aim of this work is to recommend generic solutions to these problems with particular regard to those constraints arising because of the need for ClM systems integration involving both new and existing systems and procedures. The work has involved the application of modern computer technologies, including suitable hardware and software tools, in an industrial environment. Since the research has been undertaken with particular emphasis on the industrial implementor's viewpoint, it is supported by the results of a two phased implementation of computer based control systems within the machine shop of a manufacturing company. This involved the specific implementation of a Distributed Numerical Control system on a single machine in a group technology cell of machines followed by the evolution of this system into Cell and Machine Management Systems to provide a comprehensive decision support and information distribution facility for the foremen and uperators within the cell. The work also required the integration of these systems with existing Factory level manufacturing control and CADCAM functions. Alternative approaches have been investigated which may have been applicable under differing conditions and the implications that this specific work has for CIM systems integration in small batch manufacturing companies evaluated with regard not only to the users within an industrial company but also the systems suppliers external to the company. The work has resulted in certain generic contributions to knowledge by complementing ClM systems integration research with regard to problems encountered; cost implications; the use of appropriate methodologies including the role of emerging international standard methods, tools and technologies and also the importance of 'human integration' when implementing CIM systems in a real industrial situation

    High Availability and Scalability of Mainframe Environments using System z and z/OS as example

    Get PDF
    Mainframe computers are the backbone of industrial and commercial computing, hosting the most relevant and critical data of businesses. One of the most important mainframe environments is IBM System z with the operating system z/OS. This book introduces mainframe technology of System z and z/OS with respect to high availability and scalability. It highlights their presence on different levels within the hardware and software stack to satisfy the needs for large IT organizations

    Database recovery

    Get PDF
    Recovery techniques are an important aspect of database systems. They are essential to ensure that data integrity is maintained after any type of failure occurs. The recovery mechanism must be designed so that the availability and performance of the system are not unacceptably impacted by the recovery algorithms running during normal execution. On the other hand, enough information must be stored so that the database can be restored or transactions backed out in a reasonable amount of time. Concepts, techniques, and problems associated with database recovery will be presented in this thesis. The recovery issues for both centralized and distributed systems will be discussed, along with the tradeoffs of different recovery tools. The database recovery schemes in IMS/VS, DB2 and SDD-1 will be described to show approaches in existing systems

    Integrating legacy mainframe systems: architectural issues and solutions

    Get PDF
    For more than 30 years, mainframe computers have been the backbone of computing systems throughout the world. Even today it is estimated that some 80% of the worlds' data is held on such machines. However, new business requirements and pressure from evolving technologies, such as the Internet is pushing these existing systems to their limits and they are reaching breaking point. The Banking and Financial Sectors in particular have been relying on mainframes for the longest time to do their business and as a result it is they that feel these pressures the most. In recent years there have been various solutions for enabling a re-engineering of these legacy systems. It quickly became clear that to completely rewrite them was not possible so various integration strategies emerged. Out of these new integration strategies, the CORBA standard by the Object Management Group emerged as the strongest, providing a standards based solution that enabled the mainframe applications become a peer in a distributed computing environment. However, the requirements did not stop there. The mainframe systems were reliable, secure, scalable and fast, so any integration strategy had to ensure that the new distributed systems did not lose any of these benefits. Various patterns or general solutions to the problem of meeting these requirements have arisen and this research looks at applying some of these patterns to mainframe based CORBA applications. The purpose of this research is to examine some of the issues involved with making mainframebased legacy applications inter-operate with newer Object Oriented Technologies
    • …
    corecore