2,024 research outputs found
CMS dashboard task monitoring: A user-centric monitoring view
We are now in a phase change of the CMS experiment where people are turning more intensely to physics analysis and away from construction. This brings a lot of challenging issues with respect to monitoring of the user analysis. The physicists must be able to monitor the execution status, application and grid-level messages of their tasks that may run at any site within the CMS Virtual Organisation. The CMS Dashboard Task Monitoring project provides this information towards individual analysis users by collecting and exposing a user-centric set of information regarding submitted tasks including reason of failure, distribution by site and
over time, consumed time and efficiency. The development was user-driven with physicists invited to test the prototype in order to assemble further requirements and identify weaknesses with the application
Adaptability as a Strategy of Self-Identity in the Modern Information Space
The values and meanings for education of the new century information technologies dictate new teaching strategies, development and self-development. Innovative co-adaptation in a rapidly changing society is an active strategy for the development of the individual, due to cooperation, communication and creative activity.Ценности и смыслы, цели образования нового века информационных технологий диктуют новые стратегии обучения, развития и саморазвития личности. Инновационная ко-адаптация в условиях стремительно меняющегося социума – это активная стратегия развития личности, обусловленная сотрудничеством, коммуникативной и созидательной деятельностью
Producing phrasal prominence in German
This study examines the relative change in a number of acoustic parameters usually associated with the production of prominences. The production of six German sentences under different question answer conditions provide de-accented and accented versions of the same words in broad and narrow focus. Normalised energy, F0, duration and spectral measures were found to form a stable hierarchy in their exponency of the three degrees of accentuation
Data growth and its impact on the SCOP database: new developments
The Structural Classification of Proteins (SCOP) database is a comprehensive ordering of all proteins of known structure, according to their evolutionary and structural relationships. The SCOP hierarchy comprises the following levels: Species, Protein, Family, Superfamily, Fold and Class. While keeping the original classification scheme intact, we have changed the production of SCOP in order to cope with a rapid growth of new structural data and to facilitate the discovery of new protein relationships. We describe ongoing developments and new features implemented in SCOP. A new update protocol supports batch classification of new protein structures by their detected relationships at Family and Superfamily levels in contrast to our previous sequential handling of new structural data by release date. We introduce pre-SCOP, a preview of the SCOP developmental version that enables earlier access to the information on new relationships. We also discuss the impact of worldwide Structural Genomics initiatives, which are producing new protein structures at an increasing rate, on the rates of discovery and growth of protein families and superfamilies. SCOP can be accessed at http://scop.mrc-lmb.cam.ac.uk/scop
An efficient strategy for the collection and storage of large volumes of data for computation
In recent years, there has been an increasing amount of data being produced and stored, which is known as Big Data. The social networks, internet of things, scientific experiments and commercial services play a significant role in generating a vast amount of data. Three main factors are important in Big Data; Volume, Velocity and Variety. One needs to consider all three factors when designing a platform to support Big Data. The Large Hadron Collider (LHC) particle accelerator at CERN consists of a number of data-intensive experiments, which are estimated to produce a volume of about 30 PB of data, annually. The velocity of these data that are propagated will be extremely fast. Traditional methods of collecting, storing and analysing data have become insufficient in managing the rapidly growing volume of data. Therefore, it is essential to have an efficient strategy to capture these data as they are produced. In this paper, a number of models are explored to understand what should be the best approach for collecting and storing Big Data for analytics. An evaluation of the performance of full execution cycles of these approaches on the monitoring of the Worldwide LHC Computing Grid (WLCG) infrastructure for collecting, storing and analysing data is presented. Moreover, the models discussed are applied to a community driven software solution, Apache Flume, to show how they can be integrated, seamlessly
Search for a signal on intermediate baryon systems formation in hadron-nuclear and nuclear-nuclear interactions at high energies
We have analyzed the behavior of different characteristics of hadron-nuclear
and nuclear-nuclear interactions as a function of centrality to get a signal on
the formation of intermediate baryon systems. We observed that the data
demonstrate the regime change and saturation. The angular distributions of slow
particles exhibit some structure in the above mentioned reactions at low
energy. We believe that the structure could be connected with the formation and
decay of the percolation cluster. With increasing the mass of colliding nuclei,
the structure starts to become weak and almost disappears ultimately. This
shows that the number of secondary internuclear interactions increases with
increasing the mass of the colliding nuclei. The latter could be a reason of
the disintegration of any intermediate formations as well as clusters, which
decrease their influence on the angular distribution of the emitted particles.Comment: 2 pages and one figur
Use of grid tools to support CMS distributed analysis
In order to prepare the Physics Technical Design Report, due by end of 2005, the CMS experiment needs to simulate, reconstruct and analyse about 100 million events, corresponding to more than 200 TB of data. The data will be distributed to several Computing Centres. In order to provide access to the whole data sample to all the world-wide dispersed physicists, CMS is developing a layer of software that uses the Grid tools provided by the LCG project to gain access to data and resources and that aims to provide a user friendly interface to the physicists submitting the analysis jobs. To achieve these aims CMS will use Grid tools from both the LCG-2 release and those being developed in the framework of the ARDA project. This work describes the current status and the future developments of the CMS analysis system
- …