18,682 research outputs found
Ontology of core data mining entities
In this article, we present OntoDM-core, an ontology of core data mining
entities. OntoDM-core defines themost essential datamining entities in a three-layered
ontological structure comprising of a specification, an implementation and an application
layer. It provides a representational framework for the description of mining
structured data, and in addition provides taxonomies of datasets, data mining tasks,
generalizations, data mining algorithms and constraints, based on the type of data.
OntoDM-core is designed to support a wide range of applications/use cases, such as
semantic annotation of data mining algorithms, datasets and results; annotation of
QSAR studies in the context of drug discovery investigations; and disambiguation of
terms in text mining. The ontology has been thoroughly assessed following the practices
in ontology engineering, is fully interoperable with many domain resources and
is easy to extend
Seasonal variation of arsenic concentrations in tubewells in west Bengal, India.
This study was conducted to monitor the changes in arsenic concentration during different seasons in a one-year period during 2002-2003 in selected tubewells in an arsenic-affected area in the district of South 24 Parganas in West Bengal, India, and to map the location of the wells. Seasonal variations in concentrations of arsenic in water were measured from 74 selected tubewells, ranging in depth from 40 to 500 feet. Water samples were collected from these wells during winter, summer, monsoon, and the following winter in 2002-2003. A global positioning system was used for locating the tubewells, and a geographic information system was used for mapping. There was evidence of seasonal variation in concentrations of arsenic in water (p=0.02) with the minimum average concentration occurring in the summer season (694 microg/L) and the maximum in the monsoon season (906 microg/L). From the winter of 2002 to the winter of 2003, arsenic concentrations increased, irrespective of the depth of the tubewells, from an average of 464 microg/L to 820 microg/L (p<0.001). This extent of variation in arsenic concentration, if confirmed, has important implications for both epidemiological research and mitigation programmes
Examples of minimal-memory, non-catastrophic quantum convolutional encoders
One of the most important open questions in the theory of quantum
convolutional coding is to determine a minimal-memory, non-catastrophic,
polynomial-depth convolutional encoder for an arbitrary quantum convolutional
code. Here, we present a technique that finds quantum convolutional encoders
with such desirable properties for several example quantum convolutional codes
(an exposition of our technique in full generality will appear elsewhere). We
first show how to encode the well-studied Forney-Grassl-Guha (FGG) code with an
encoder that exploits just one memory qubit (the former Grassl-Roetteler
encoder requires 15 memory qubits). We then show how our technique can find an
online decoder corresponding to this encoder, and we also detail the operation
of our technique on a different example of a quantum convolutional code.
Finally, the reduction in memory for the FGG encoder makes it feasible to
simulate the performance of a quantum turbo code employing it, and we present
the results of such simulations.Comment: 5 pages, 2 figures, Accepted for the International Symposium on
Information Theory 2011 (ISIT 2011), St. Petersburg, Russia; v2 has minor
change
ADsafety: Type-Based Verification of JavaScript Sandboxing
Web sites routinely incorporate JavaScript programs from several sources into
a single page. These sources must be protected from one another, which requires
robust sandboxing. The many entry-points of sandboxes and the subtleties of
JavaScript demand robust verification of the actual sandbox source. We use a
novel type system for JavaScript to encode and verify sandboxing properties.
The resulting verifier is lightweight and efficient, and operates on actual
source. We demonstrate the effectiveness of our technique by applying it to
ADsafe, which revealed several bugs and other weaknesses.Comment: in Proceedings of the USENIX Security Symposium (2011
A Fast Compiler for NetKAT
High-level programming languages play a key role in a growing number of
networking platforms, streamlining application development and enabling precise
formal reasoning about network behavior. Unfortunately, current compilers only
handle "local" programs that specify behavior in terms of hop-by-hop forwarding
behavior, or modest extensions such as simple paths. To encode richer "global"
behaviors, programmers must add extra state -- something that is tricky to get
right and makes programs harder to write and maintain. Making matters worse,
existing compilers can take tens of minutes to generate the forwarding state
for the network, even on relatively small inputs. This forces programmers to
waste time working around performance issues or even revert to using
hardware-level APIs.
This paper presents a new compiler for the NetKAT language that handles rich
features including regular paths and virtual networks, and yet is several
orders of magnitude faster than previous compilers. The compiler uses symbolic
automata to calculate the extra state needed to implement "global" programs,
and an intermediate representation based on binary decision diagrams to
dramatically improve performance. We describe the design and implementation of
three essential compiler stages: from virtual programs (which specify behavior
in terms of virtual topologies) to global programs (which specify network-wide
behavior in terms of physical topologies), from global programs to local
programs (which specify behavior in terms of single-switch behavior), and from
local programs to hardware-level forwarding tables. We present results from
experiments on real-world benchmarks that quantify performance in terms of
compilation time and forwarding table size
Ontologies on the semantic web
As an informational technology, the World Wide Web has enjoyed spectacular success. In just ten years it has transformed the way information is produced, stored, and shared in arenas as diverse as shopping, family photo albums, and high-level academic research. The “Semantic Web” was touted by its developers as equally revolutionary but has not yet achieved anything like the Web’s exponential uptake. This 17 000 word survey article explores why this might be so, from a perspective that bridges both philosophy and IT
A Unifying Approach to Decide Relations for Timed Automata and their Game Characterization
In this paper we present a unifying approach for deciding various
bisimulations, simulation equivalences and preorders between two timed automata
states. We propose a zone based method for deciding these relations in which we
eliminate an explicit product construction of the region graphs or the zone
graphs as in the classical methods. Our method is also generic and can be used
to decide several timed relations. We also present a game characterization for
these timed relations and show that the game hierarchy reflects the hierarchy
of the timed relations. One can obtain an infinite game hierarchy and thus the
game characterization further indicates the possibility of defining new timed
relations which have not been studied yet. The game characterization also helps
us to come up with a formula which encodes the separation between two states
that are not timed bisimilar. Such distinguishing formulae can also be
generated for many relations other than timed bisimilarity.Comment: In Proceedings EXPRESS/SOS 2013, arXiv:1307.690
- …