22,348 research outputs found

    Distributed data mining in grid computing environments

    Get PDF
    The official published version of this article can be found at the link below.The computing-intensive data mining for inherently Internet-wide distributed data, referred to as Distributed Data Mining (DDM), calls for the support of a powerful Grid with an effective scheduling framework. DDM often shares the computing paradigm of local processing and global synthesizing. It involves every phase of Data Mining (DM) processes, which makes the workflow of DDM very complex and can be modelled only by a Directed Acyclic Graph (DAG) with multiple data entries. Motivated by the need for a practical solution of the Grid scheduling problem for the DDM workflow, this paper proposes a novel two-phase scheduling framework, including External Scheduling and Internal Scheduling, on a two-level Grid architecture (InterGrid, IntraGrid). Currently a DM IntraGrid, named DMGCE (Data Mining Grid Computing Environment), has been developed with a dynamic scheduling framework for competitive DAGs in a heterogeneous computing environment. This system is implemented in an established Multi-Agent System (MAS) environment, in which the reuse of existing DM algorithms is achieved by encapsulating them into agents. Practical classification problems from oil well logging analysis are used to measure the system performance. The detailed experiment procedure and result analysis are also discussed in this paper

    Development of Grid e-Infrastructure in South-Eastern Europe

    Full text link
    Over the period of 6 years and three phases, the SEE-GRID programme has established a strong regional human network in the area of distributed scientific computing and has set up a powerful regional Grid infrastructure. It attracted a number of user communities and applications from diverse fields from countries throughout the South-Eastern Europe. From the infrastructure point view, the first project phase has established a pilot Grid infrastructure with more than 20 resource centers in 11 countries. During the subsequent two phases of the project, the infrastructure has grown to currently 55 resource centers with more than 6600 CPUs and 750 TBs of disk storage, distributed in 16 participating countries. Inclusion of new resource centers to the existing infrastructure, as well as a support to new user communities, has demanded setup of regionally distributed core services, development of new monitoring and operational tools, and close collaboration of all partner institution in managing such a complex infrastructure. In this paper we give an overview of the development and current status of SEE-GRID regional infrastructure and describe its transition to the NGI-based Grid model in EGI, with the strong SEE regional collaboration.Comment: 22 pages, 12 figures, 4 table

    Does "thin client" mean "energy efficient"?

    Get PDF
    The thick client –a personal computer with integral disk storage and local processing capability, which also has access to data and other resources via a network connection – is accepted as the model for providing computing resource in most office environments. The Further and Higher Education sector is no exception to that, and therefore most academic and administrative offices are equipped with desktop computers of this form to support users in their day to day tasks. This system structure has a number of advantages: there is a reduced reliance on network resources; users access a system appropriate to their needs, and may customise “their” system to meet their own personal requirements and working patterns. However it also has disadvantages: some are outside the scope of this project, but of most relevance to the green IT agenda is the fact that relatively complex and expensive (in first cost and in running cost) desktop systems and servers are underutilised – especially in respect of processing power. While some savings are achieved through use of “sleep” modes and similar power reducing mechanisms, in most configurations only a small portion of the overall total available processor resource is utilised. This realisation has led to the promotion of an alternative paradigm, the thin client. In a thin client system, the desktop is shorn of most of its local processing and data storage capability, and essentially acts as a terminal to the server, which now takes on responsibility for data storage and processing. The energy benefit is derived through resource sharing: the processor of the server does the work, and because that processor is shared by all users, a number of users are supported by a single system. Therefore – according to proponents of thin client – the total energy required to support a user group is reduced, since a shared physical resource is used more efficiently. These claims are widely reported: indeed there are a number of estimation tools which show these savings can be achieved; however there appears to be little or no actual measured data to confirm this. The community does not appear to have access to measured data comparing thin and thick client systems in operation in the same situation, allowing direct comparisons to be drawn. This is the main goal of this project. One specific question relates to the overall power use, while it would seem to be obvious that the thin client would require less electricity, what of the server? Two other variations are also considered: it is not uncommon for thin client deployments to continue to use their existing PCs as thin client workstations, with or without modification. Also, attempts by PC makers to reduce the power requirements of their products have given rise to a further variation: the incorporation of low power features in otherwise standard PC technology, working as thick clients. This project was devised to conduct actual measurements in use in a typical university environment. We identified a test area: a mixed administrative and academic office location which supported a range of users, and we made a direct replacement of the current thick client systems with thin client equivalents; in addition, we exchanged a number of PCs operating in thin and thick client mode with devices specifically branded as “low power” PCs and measured their power requirements in both thin and thick modes. We measured the energy consumption at each desktop for the duration of our experiments, and also measured the energy draw of the server designated to supporting the thin client setup, giving us the opportunity to determine the power per user of each technology. Our results show a significant difference in power use between the various candidate technologies, and that a configuration of low power PC in thick client mode returned the lowest power use during our study. We were also aware of other factors surrounding a change such as this: we have addressed the technical issues of implementation and management, and the non-technical or human factors of acceptance and use: all are reported within this document. Finally, our project is necessarily limited to a set of experiments carried out in a particular situation, therefore we use estimation methods to draw wider conclusions and make general observations which should allow others to select appropriate thick or thin client solutions in their situation

    The measurement and interpretation methodology of resistivity logs affected by the Groningen effect – a Polish case study

    Get PDF
    True formation resistivity Rt measurement is one of the fundamental logs in the calculation of hydrocarbon resources. That is why it is very important to have the most reliable resistivity data possible. In this paper, the various outcomes obtained by Polish well log analysts and engineers for the proper determination of hydrocarbon saturation in the Main Dolomite deposits in the Polish Lowland are presented. The long history of efforts directed to make proper exploitation decisions in wells where the Groningen effect has been observed is illustrated, starting with the standard measurement and interpretational approach, through the modified construction of a reference electrode in a Laterolog device and ending with an examination of HRLA (High-Resolution Laterolog Array) or Array Compensated Resistivity Tool) ACRt results. The processing of resistivity logs with the special Poprawki software is included

    TechNews digests: Jan - Mar 2010

    Get PDF
    TechNews is a technology, news and analysis service aimed at anyone in the education sector keen to stay informed about technology developments, trends and issues. TechNews focuses on emerging technologies and other technology news. TechNews service : digests september 2004 till May 2010 Analysis pieces and News combined publish every 2 to 3 month

    Assessing and augmenting SCADA cyber security: a survey of techniques

    Get PDF
    SCADA systems monitor and control critical infrastructures of national importance such as power generation and distribution, water supply, transportation networks, and manufacturing facilities. The pervasiveness, miniaturisations and declining costs of internet connectivity have transformed these systems from strictly isolated to highly interconnected networks. The connectivity provides immense benefits such as reliability, scalability and remote connectivity, but at the same time exposes an otherwise isolated and secure system, to global cyber security threats. This inevitable transformation to highly connected systems thus necessitates effective security safeguards to be in place as any compromise or downtime of SCADA systems can have severe economic, safety and security ramifications. One way to ensure vital asset protection is to adopt a viewpoint similar to an attacker to determine weaknesses and loopholes in defences. Such mind sets help to identify and fix potential breaches before their exploitation. This paper surveys tools and techniques to uncover SCADA system vulnerabilities. A comprehensive review of the selected approaches is provided along with their applicability
    corecore