1,539 research outputs found
Temporal Data Modeling and Reasoning for Information Systems
Temporal knowledge representation and reasoning is a major research field in Artificial
Intelligence, in Database Systems, and in Web and Semantic Web research. The ability to
model and process time and calendar data is essential for many applications like appointment
scheduling, planning, Web services, temporal and active database systems, adaptive
Web applications, and mobile computing applications. This article aims at three complementary
goals. First, to provide with a general background in temporal data modeling
and reasoning approaches. Second, to serve as an orientation guide for further specific
reading. Third, to point to new application fields and research perspectives on temporal
knowledge representation and reasoning in the Web and Semantic Web
Modeling Multiple Granularities of Spatial Objects
People conceptualize objects in an information space over different levels of details or granularities and shift among these granularities as necessary for the task at hand. Shifting among granularities is fundamental for understanding and reasoning about an information space. In general, shifting to a coarser granularity can improve one\u27s understanding of a complex information space, whereas shifting to a more detailed granularity reveals information that is otherwise unknown. To arrive at a coarser granularity. objects must be generalized. There are multiple ways to perform generalization. Several generalization methods have been adopted from the abstraction processes that are intuitively carried out by people. Although, people seem to be able to carry out abstractions and generalize objects with ease. formalizing these generalization and shifts between them in an information system, such as geographic information system, still offers many challenges. A set of rules capturing multiple granularities of objects and the use of these granularities for enhanced reasoning and browsing is yet to be well researched. This thesis pursues an approach for arriving at multiple granularities of spatial objects based on the concept of coarsening. Coarsening refers to the process of transforming a representation of objects into a less detailed representation. The focus of this thesis is to develop a set of coarsening operators that are based on the objects\u27 attributes, attribute values and relations with other objects, such as containment, connectivity, and nearness. for arriving at coarser or amalgamated objects. As a result. a set of four coarsening operators—group, group, compose, coexist, and filter are defined. A framework, called a granularity graph. is presented for modeling the application of coarsening operators iteratively to form amalgamated objects. A granularity graph can be used to browse through objects at different granularities, to retrieve objects that are at different granularities, and to examine how the granularities are related to each other. There can occur long sequences of operators between objects in the graph, which need to be simplified. Compositions of coarsening operators are derived to collapse or simplify the chain of operators. The semantics associated with objects amalgamations enable to determine correct results of the compositions of coarsening operators. The composition of operators enables to determine all the possible ways for arriving at a coarser granularity of objects from a set of objects. Capturing these different ways facilitates enhanced reasoning of how objects at multiple granularities are related to each other
Recommended from our members
Exploring Temporal Granularities with Visualization
Time can be expressed and aggregated into concepts called granularities. Granularities are defined in a structure with their rules of conversion that may take the form of trees or graphs, thus it’s possible to design tools that dynamically explore different granularities that might reveal patterns hidden in other levels. We described an intial investigation of the use of interactive visualization techniques for that purpose and define future work to be done
Using treemaps for variable selection in spatio-temporal visualisation
We demonstrate and reflect upon the use of enhanced treemaps that incorporate spatial and temporal ordering for exploring a large multivariate spatio-temporal data set. The resulting data-dense views summarise and simultaneously present hundreds of space-, time-, and variable-constrained subsets of a large multivariate data set in a structure that facilitates their meaningful comparison and supports visual analysis. Interactive techniques allow localised patterns to be explored and subsets of interest selected and compared with the spatial aggregate. Spatial variation is considered through interactive raster maps and high-resolution local road maps. The techniques are developed in the context of 42.2 million records of vehicular activity in a 98 km(2) area of central London and informally evaluated through a design used in the exploratory visualisation of this data set. The main advantages of our technique are the means to simultaneously display hundreds of summaries of the data and to interactively browse hundreds of variable combinations with ordering and symbolism that are consistent and appropriate for space- and time- based variables. These capabilities are difficult to achieve in the case of spatio-temporal data with categorical attributes using existing geovisualisation methods. We acknowledge limitations in the treemap representation but enhance the cognitive plausibility of this popular layout through our two-dimensional ordering algorithm and interactions. Patterns that are expected (e.g. more traffic in central London), interesting (e.g. the spatial and temporal distribution of particular vehicle types) and anomalous (e.g. low speeds on particular road sections) are detected at various scales and locations using the approach. In many cases, anomalies identify biases that may have implications for future use of the data set for analyses and applications. Ordered treemaps appear to have potential as interactive interfaces for variable selection in spatio-temporal visualisation. Information Visualization (2008) 7, 210-224. doi: 10.1057/palgrave.ivs.950018
Ontology-Based Consistent Specification of Sensor Data Acquisition Plans in Cross-Domain IoT Platforms
Nowadays there is an high number of IoT applications that seldom can interact with each other because developed within different Vertical IoT Platforms that adopt different standards. Several efforts are devoted to the construction of cross-layered frameworks that facilitate the interoperability among cross-domain IoT platforms for the development of horizontal applications. Even if their realization poses different challenges across all layers of the network stack, in this paper we focus on the interoperability issues that arise at the data management layer. Specifically, starting from a flexible multi-granular Spatio-Temporal-Thematic data model according to which events generated by different kinds of sensors can be represented, we propose a Semantic Virtualization approach according to which the sensors belonging to different IoT platforms and the schema of the produced event streams are described in a Domain Ontology, obtained through the extension of the well-known Semantic Sensor Network ontology. Then, these sensors can be exploited for the creation of Data Acquisition Plans by means of which the streams of events can be filtered, merged, and aggregated in a meaningful way. A notion of consistency is introduced to bind the output streams of the services contained in the Data Acquisition Plan with the Domain Ontology in order to provide a semantic description of its final output. When these plans meet the consistency constraints, it means that the data they handle are well described at the Ontological level and thus the data acquisition process over passed the interoperability barriers occurring in the original sources. The facilities of the StreamLoader prototype are finally presented for supporting the user in the Semantic Virtualization process and for the construction of meaningful Data Acquisition Plans
SODA: Generating SQL for Business Users
The purpose of data warehouses is to enable business analysts to make better
decisions. Over the years the technology has matured and data warehouses have
become extremely successful. As a consequence, more and more data has been
added to the data warehouses and their schemas have become increasingly
complex. These systems still work great in order to generate pre-canned
reports. However, with their current complexity, they tend to be a poor match
for non tech-savvy business analysts who need answers to ad-hoc queries that
were not anticipated. This paper describes the design, implementation, and
experience of the SODA system (Search over DAta Warehouse). SODA bridges the
gap between the business needs of analysts and the technical complexity of
current data warehouses. SODA enables a Google-like search experience for data
warehouses by taking keyword queries of business users and automatically
generating executable SQL. The key idea is to use a graph pattern matching
algorithm that uses the metadata model of the data warehouse. Our results with
real data from a global player in the financial services industry show that
SODA produces queries with high precision and recall, and makes it much easier
for business users to interactively explore highly-complex data warehouses.Comment: VLDB201
On-line analytical processing
On-line analytical processing (OLAP) describes an approach to decision support, which aims to extract knowledge from a data warehouse, or more specifically, from data marts. Its main idea is providing navigation through data to non-expert users, so that they are able to interactively generate ad hoc queries without the intervention of IT professionals. This name was introduced in contrast to on-line transactional processing (OLTP), so that it reflected the different requirements and characteristics between these classes of uses. The concept falls in the area of business intelligence.Peer ReviewedPostprint (author's final draft
- …