2,149 research outputs found
Against the ânetworked information economyâ: rethinking decentralization, community, and free software development
No description supplie
Distributed Computing in a Pandemic: A Review of Technologies Available for Tackling COVID-19
The current COVID-19 global pandemic caused by the SARS-CoV-2 betacoronavirus
has resulted in over a million deaths and is having a grave socio-economic
impact, hence there is an urgency to find solutions to key research challenges.
Much of this COVID-19 research depends on distributed computing. In this
article, I review distributed architectures -- various types of clusters, grids
and clouds -- that can be leveraged to perform these tasks at scale, at
high-throughput, with a high degree of parallelism, and which can also be used
to work collaboratively. High-performance computing (HPC) clusters will be used
to carry out much of this work. Several bigdata processing tasks used in
reducing the spread of SARS-CoV-2 require high-throughput approaches, and a
variety of tools, which Hadoop and Spark offer, even using commodity hardware.
Extremely large-scale COVID-19 research has also utilised some of the world's
fastest supercomputers, such as IBM's SUMMIT -- for ensemble docking
high-throughput screening against SARS-CoV-2 targets for drug-repurposing, and
high-throughput gene analysis -- and Sentinel, an XPE-Cray based system used to
explore natural products. Grid computing has facilitated the formation of the
world's first Exascale grid computer. This has accelerated COVID-19 research in
molecular dynamics simulations of SARS-CoV-2 spike protein interactions through
massively-parallel computation and was performed with over 1 million volunteer
computing devices using the Folding@home platform. Grids and clouds both can
also be used for international collaboration by enabling access to important
datasets and providing services that allow researchers to focus on research
rather than on time-consuming data-management tasks.Comment: 21 pages (15 excl. refs), 2 figures, 3 table
A gentle transition from Java programming to Web Services using XML-RPC
Exposing students to leading edge vocational areas of relevance such as Web Services can be difficult. We show a lightweight approach by embedding a key component of Web Services within a Level 3 BSc module in Distributed Computing. We present a ready to use collection of lecture slides and student activities based on XML-RPC. In
addition we show that this material addresses the central topics in the context of web services as identified by Draganova (2003)
Ontology-based solutions for interoperability among product lifecycle management systems: A systematic literature review
During recent years, globalization has had an impact on the competitive capacity of industries, forcing them to integrate their productive processes with other, geographically distributed, facilities. This requires the information systems that support such processes to interoperate. Significant attention has been paid to the development of ontology-based solutions, which are meant to tackle issues from inconsistency to semantic interoperability and knowledge reusability. This paper looks into how the available technology, models and ontology-based solutions might interact within the manufacturing industry environment to achieve semantic interoperability among industrial information systems. Through a systematic literature review, this paper has aimed to identify the most relevant elements to consider in the development of an ontology-based solution and how these solutions are being deployed in industry. The research analyzed 54 studies in alignment with the specific requirements of our research questions. The most relevant results show that ontology-based solutions can be set up using OWL as the ontology language, ProtĂ©gĂ© as the ontology modeling tool, Jena as the application programming interface to interact with the built ontology, and different standards from the International Organization for Standardization Technical Committee 184, Subcommittee 4 or 5, to get the foundational concepts, axioms, and relationships to develop the knowledge base. We believe that the findings of this study make an important contribution to practitioners and researchers as they provide useful information about different projects and choices involved in undertaking projects in the field of industrial ontology application.Fil: Fraga, Alvaro Luis. Consejo Nacional de Investigaciones CientĂficas y TĂ©cnicas. Centro CientĂfico TecnolĂłgico Conicet - Santa Fe. Instituto de Desarrollo y Diseño. Universidad TecnolĂłgica Nacional. Facultad Regional Santa Fe. Instituto de Desarrollo y Diseño; ArgentinaFil: Vegetti, Maria Marcela. Consejo Nacional de Investigaciones CientĂficas y TĂ©cnicas. Centro CientĂfico TecnolĂłgico Conicet - Santa Fe. Instituto de Desarrollo y Diseño. Universidad TecnolĂłgica Nacional. Facultad Regional Santa Fe. Instituto de Desarrollo y Diseño; ArgentinaFil: Leone, Horacio Pascual. Consejo Nacional de Investigaciones CientĂficas y TĂ©cnicas. Centro CientĂfico TecnolĂłgico Conicet - Santa Fe. Instituto de Desarrollo y Diseño. Universidad TecnolĂłgica Nacional. Facultad Regional Santa Fe. Instituto de Desarrollo y Diseño; Argentin
Recommended from our members
Kafka, Samza and the Unix Philosophy of Distributed Data
Apache Kafka is a scalable message broker, and Apache Samza is a stream processing framework built upon Kafka. They are widely used as infrastructure for implementing personalized online services and real-time predictive analytics. Besides providing high throughput and low latency, Kafka and Samza are designed with operational robustness and long-term maintenance of applications in mind. In this paper we explain the reasoning behind the design of Kafka and Samza, which allow complex applications to be built by composing a small number of simple primitives â replicated logs and stream operators. We draw parallels between the design of Kafka and Samza, batch processing pipelines, database architecture, and the design philosophy of Unix
Software solutions for web information systems in digital humanities: review, analysis and comparative study
Research in the humanities increasingly depends on how information is structured and managed and how, on the basis of that information, new knowledge is produced. Additionally, participatory approaches, which often rely on web information systems as their supportive infrastructure, have made an impact on the most recent historiographical trends, in particular in the methodological framework of digital humanities. The aim of this paper was to produce, from an operational and implementation perspective, a review of software solutions frequently used to develop web information systems for research projects in humanities and cultural heritage, in order to provide an understanding of the various possibilities available and their positives and limitations, also based on different usersâ requirements. An individual and comparative analysis of sixteen different application frameworks commonly used in these fields, either generic or developed for a specific research domain, has been carried out, considering their main functionalities, strengths, and weaknesses. The achieved results facilitate critical and reasoned decision-making among several available options, guiding the makers of those systems, both researcher(s) and developers(s), and providing them also with a common ground of terms and use cases to facilitate their necessary dialogue
ACHIEVING AUTONOMIC SERVICE ORIENTED ARCHITECTURE USING CASE BASED REASONING
Service-Oriented Architecture (SOA) enables composition of large and complex
computational units out of the available atomic services. However, implementation of
SOA, for its dynamic nature, could bring about challenges in terms of service
discovery, service interaction, service composition, robustness, etc. In the near future,
SOA will often need to dynamically re-configuring and re-organizing its topologies of
interactions between the web services because of some unpredictable events, such as
crashes or network problems, which will cause service unavailability. Complexity and
dynamism of the current and future global network system require service architecture
that is capable of autonomously changing its structure and functionality to meet
dynamic changes in the requirements and environment with little human intervention.
This then needs to motivate the research described throughout this thesis.
In this thesis, the idea of introducing autonomy and adapting case-based reasoning
into SOA in order to extend the intelligence and capability of SOA is contributed and
elaborated. It is conducted by proposing architecture of an autonomic SOA
framework based on case-based reasoning and the architectural considerations of
autonomic computing paradigm. It is then followed by developing and analyzing
formal models of the proposed architecture using Petri Net. The framework is also
tested and analyzed through case studies, simulation, and prototype development. The
case studies show feasibility to employing case-based reasoning and autonomic
computing into SOA domain and the simulation results show believability that it
would increase the intelligence, capability, usability and robustness of SOA. It was
shown that SOA can be improved to cope with dynamic environment and services
unavailability by incorporating case-based reasoning and autonomic computing
paradigm to monitor and analyze events and service requests, then to plan and execute
the appropriate actions using the knowledge stored in knowledge database
- âŠ