10,059 research outputs found

    Long-term monitoring of a winter bat assemblage revealed large fluctuations and trends in species abundance

    Get PDF
    Monitoring studies in Strzaliny, one of the greatest hibernacula in Poland, comprised 31 annual bat censuses (1989ā€“2019). The abundance peaked in 2002 for Myotis myotis, 2009 for Myotis nattereri and 2008 for the whole assemblage. Comparison of the maximum abundance in the monitoring period with that from 1980 to 1982 showed an almost fourfold increase for the whole assemblage, tenfold increase for M. nattereri and fourfold increase for M. myotis. In 1989ā€“2019, the numbers of M. myotis, M. nattereri, Myotis daubentonii and Plecotus auritus were fluctuating, but most of the recorded changes could not be explained by methodological problems or a direct human impact. Therefore, the cumulative results largely reflected the real trends in the species abundance. A long-term upward trend in the whole bat assemblage was recognisable, but with a stable or slightly decreasing phase in the last decade. An upward trend in M. nattereri was even stronger and has only slightly flattened recently. In M. myotis, the trend was clearly upwards up to the early 2000s, but weakly downwards in the following years. In M. daubentonii and P. auritus, no significant trend was determined. In strongly fluctuating M. daubentonii, the numbers were mostly moderate or high, and even increasing, up to 2008 and only moderate or low in the following years. In P. auritus, an increase occurred in the 1980s and early 1990s, and then, after the stochastic human-induced drop in 1994, its abundance remained relatively stable. The population trends in Strzaliny largely reflected the general trends assessed for a large part of Europe. This suggests that the general population trends may be recognisable even in one large winter assemblage if it is reliably and consistently monitored through a long period. In this context, the hibernaculum in Strzaliny appeared to be a model object for such studies

    The Parma Polyhedra Library: Toward a Complete Set of Numerical Abstractions for the Analysis and Verification of Hardware and Software Systems

    Get PDF
    Since its inception as a student project in 2001, initially just for the handling (as the name implies) of convex polyhedra, the Parma Polyhedra Library has been continuously improved and extended by joining scrupulous research on the theoretical foundations of (possibly non-convex) numerical abstractions to a total adherence to the best available practices in software development. Even though it is still not fully mature and functionally complete, the Parma Polyhedra Library already offers a combination of functionality, reliability, usability and performance that is not matched by similar, freely available libraries. In this paper, we present the main features of the current version of the library, emphasizing those that distinguish it from other similar libraries and those that are important for applications in the field of analysis and verification of hardware and software systems.Comment: 38 pages, 2 figures, 3 listings, 3 table

    Instrumenting self-modifying code

    Full text link
    Adding small code snippets at key points to existing code fragments is called instrumentation. It is an established technique to debug certain otherwise hard to solve faults, such as memory management issues and data races. Dynamic instrumentation can already be used to analyse code which is loaded or even generated at run time.With the advent of environments such as the Java Virtual Machine with optimizing Just-In-Time compilers, a new obstacle arises: self-modifying code. In order to instrument this kind of code correctly, one must be able to detect modifications and adapt the instrumentation code accordingly, preferably without incurring a high penalty speedwise. In this paper we propose an innovative technique that uses the hardware page protection mechanism of modern processors to detect such modifications. We also show how an instrumentor can adapt the instrumented version depending on the kind of modificiations as well as an experimental evaluation of said techniques.Comment: In M. Ronsse, K. De Bosschere (eds), proceedings of the Fifth International Workshop on Automated Debugging (AADEBUG 2003), September 2003, Ghent. cs.SE/030902

    Semantically Resolving Type Mismatches in Scientific Workflows

    No full text
    Scientists are increasingly utilizing Grids to manage large data sets and execute scientific experiments on distributed resources. Scientific workflows are used as means for modeling and enacting scientific experiments. Windows Workflow Foundation (WF) is a major component of Microsoftā€™s .NET technology which offers lightweight support for long-running workflows. It provides a comfortable graphical and programmatic environment for the development of extended BPEL-style workflows. WFā€™s visual features ease the syntactic composition of Web services into scientific workflows but do nothing to assure that information passed between services has consistent semantic types or representations or that deviant flows, errors and compensations are handled meaningfully. In this paper we introduce SAWSDL-compliant annotations for WF and use them with a semantic reasoner to guarantee semantic type correctness in scientific workflows. Examples from bioinformatics are presented

    Four Lessons in Versatility or How Query Languages Adapt to the Web

    Get PDF
    Exposing not only human-centered information, but machine-processable data on the Web is one of the commonalities of recent Web trends. It has enabled a new kind of applications and businesses where the data is used in ways not foreseen by the data providers. Yet this exposition has fractured the Web into islands of data, each in different Web formats: Some providers choose XML, others RDF, again others JSON or OWL, for their data, even in similar domains. This fracturing stifles innovation as application builders have to cope not only with one Web stack (e.g., XML technology) but with several ones, each of considerable complexity. With Xcerpt we have developed a rule- and pattern based query language that aims to give shield application builders from much of this complexity: In a single query language XML and RDF data can be accessed, processed, combined, and re-published. Though the need for combined access to XML and RDF data has been recognized in previous work (including the W3Cā€™s GRDDL), our approach differs in four main aspects: (1) We provide a single language (rather than two separate or embedded languages), thus minimizing the conceptual overhead of dealing with disparate data formats. (2) Both the declarative (logic-based) and the operational semantics are unified in that they apply for querying XML and RDF in the same way. (3) We show that the resulting query language can be implemented reusing traditional database technology, if desirable. Nevertheless, we also give a unified evaluation approach based on interval labelings of graphs that is at least as fast as existing approaches for tree-shaped XML data, yet provides linear time and space querying also for many RDF graphs. We believe that Web query languages are the right tool for declarative data access in Web applications and that Xcerpt is a significant step towards a more convenient, yet highly efficient data access in a ā€œWeb of Dataā€
    • ā€¦
    corecore