10,059 research outputs found
Long-term monitoring of a winter bat assemblage revealed large fluctuations and trends in species abundance
Monitoring studies in Strzaliny, one of the greatest hibernacula in Poland, comprised 31 annual bat censuses (1989ā2019). The abundance peaked in 2002 for Myotis myotis, 2009 for Myotis nattereri and 2008 for the whole assemblage. Comparison of the maximum abundance in the monitoring period with that from 1980 to 1982 showed an almost fourfold increase for the whole assemblage, tenfold increase for M. nattereri and fourfold increase for M. myotis.
In 1989ā2019, the numbers of M. myotis, M. nattereri, Myotis daubentonii and Plecotus auritus were fluctuating, but most of the recorded changes could not be explained by methodological problems or a direct human impact. Therefore, the cumulative results largely reflected the real trends in the species abundance. A long-term upward trend in the whole bat assemblage was recognisable, but with a stable or slightly decreasing phase in the last decade. An upward trend in M. nattereri was even stronger and has only slightly flattened recently. In M. myotis, the trend was clearly upwards up to the early 2000s, but weakly downwards in the following years. In M. daubentonii and P. auritus, no significant trend was determined. In strongly fluctuating M. daubentonii, the numbers were mostly moderate or high, and even increasing, up to 2008 and only moderate or low in the following years. In P. auritus, an increase occurred in the 1980s and early 1990s, and then, after the stochastic human-induced drop in 1994, its abundance remained relatively stable.
The population trends in Strzaliny largely reflected the general trends assessed for a large part of Europe. This suggests that the general population trends may be recognisable even in one large winter assemblage if it is reliably and consistently monitored through a long period. In this context, the hibernaculum in Strzaliny appeared to be a model object for such studies
The Parma Polyhedra Library: Toward a Complete Set of Numerical Abstractions for the Analysis and Verification of Hardware and Software Systems
Since its inception as a student project in 2001, initially just for the
handling (as the name implies) of convex polyhedra, the Parma Polyhedra Library
has been continuously improved and extended by joining scrupulous research on
the theoretical foundations of (possibly non-convex) numerical abstractions to
a total adherence to the best available practices in software development. Even
though it is still not fully mature and functionally complete, the Parma
Polyhedra Library already offers a combination of functionality, reliability,
usability and performance that is not matched by similar, freely available
libraries. In this paper, we present the main features of the current version
of the library, emphasizing those that distinguish it from other similar
libraries and those that are important for applications in the field of
analysis and verification of hardware and software systems.Comment: 38 pages, 2 figures, 3 listings, 3 table
Instrumenting self-modifying code
Adding small code snippets at key points to existing code fragments is called
instrumentation. It is an established technique to debug certain otherwise hard
to solve faults, such as memory management issues and data races. Dynamic
instrumentation can already be used to analyse code which is loaded or even
generated at run time.With the advent of environments such as the Java Virtual
Machine with optimizing Just-In-Time compilers, a new obstacle arises:
self-modifying code. In order to instrument this kind of code correctly, one
must be able to detect modifications and adapt the instrumentation code
accordingly, preferably without incurring a high penalty speedwise. In this
paper we propose an innovative technique that uses the hardware page protection
mechanism of modern processors to detect such modifications. We also show how
an instrumentor can adapt the instrumented version depending on the kind of
modificiations as well as an experimental evaluation of said techniques.Comment: In M. Ronsse, K. De Bosschere (eds), proceedings of the Fifth
International Workshop on Automated Debugging (AADEBUG 2003), September 2003,
Ghent. cs.SE/030902
Semantically Resolving Type Mismatches in Scientific Workflows
Scientists are increasingly utilizing Grids to manage large data sets and execute scientific experiments on distributed resources. Scientific workflows are used as means for modeling and enacting scientific experiments. Windows Workflow Foundation (WF) is a major component of Microsoftās .NET technology which offers lightweight support for long-running workflows. It provides a comfortable graphical and programmatic environment for the development of extended BPEL-style workflows. WFās visual features ease the syntactic composition of Web services into scientific workflows but do nothing to assure that information passed between services has consistent semantic types or representations or that deviant flows, errors and compensations are handled meaningfully. In this paper we introduce SAWSDL-compliant annotations for WF and use them with a semantic reasoner to guarantee semantic type correctness in scientific workflows. Examples from bioinformatics are presented
Four Lessons in Versatility or How Query Languages Adapt to the Web
Exposing not only human-centered information, but machine-processable data on the Web is one of the commonalities of recent Web trends. It has enabled a new kind of applications and businesses where the data is used in ways not foreseen by the data providers. Yet this exposition has fractured the Web into islands of data, each in different Web formats: Some providers choose XML, others RDF, again others JSON or OWL, for their data, even in similar domains. This fracturing stifles innovation as application builders have to cope not only with one Web stack (e.g., XML technology) but with several ones, each of considerable complexity. With Xcerpt we have developed a rule- and pattern based query language that aims to give shield application builders from much of this complexity: In a single query language XML and RDF data can be accessed, processed, combined, and re-published. Though the need for combined access to XML and RDF data has been recognized in previous work (including the W3Cās GRDDL), our approach differs in four main aspects: (1) We provide a single language (rather than two separate or embedded languages), thus minimizing the conceptual overhead of dealing with disparate data formats. (2) Both the declarative (logic-based) and the operational semantics are unified in that they apply for querying XML and RDF in the same way. (3) We show that the resulting query language can be implemented reusing traditional database technology, if desirable. Nevertheless, we also give a unified evaluation approach based on interval labelings of graphs that is at least as fast as existing approaches for tree-shaped XML data, yet provides linear time and space querying also for many RDF graphs. We believe that Web query languages are the right tool for declarative data access in Web applications and that Xcerpt is a significant step towards a more convenient, yet highly efficient data access in a āWeb of Dataā
- ā¦