13,301 research outputs found
Conceptual design and implementation of the fuzzy semantic model
FSM is one of few database models that support
fuzziness, uncertainty and impreciseness of real-world at the class
deïŹnition level. FSM authorizes an entity to be partially member
of its class according to a given degree of membership that reïŹects
the level to which the entity veriïŹes the extent properties of this
class. This paper deals with the conceptual design of FSM and
adresses some implementation issues.ou
Some notes on an extended query language for FSM
FSM is a database model that has been recently proposed by the authors. FSM uses basic concepts of
classification, generalization, aggregation and association that are commonly used in semantic modelling and
supports the fuzziness of real-world at attribute, entity, class and relations intra and inter-classes levels. Hence, it
provides tools to formalize and conceptualize real-world within a manner adapted to human perception of and
reasoning about this real-word. In this paper we briefly review basic concepts of FSM and provide some notes on an
extended query language adapted to it.ou
Data modeling dealing with uncertainty in fuzzy logic
This paper shows models of data description that incorporate uncertainty like models of data extension EER, IFO among others. These database modeling tools are compared with the pattern FuzzyEER proposed by us, which is an extension of the EER model in order to manage uncertainty with fuzzy logic in fuzzy databases. Finally, a table shows the components of EER tool with the representation of all the revised models.The past and the future of information systems: 1976-2006 and beyondRed de Universidades con Carreras en InformĂĄtica (RedUNCI
Historical collaborative geocoding
The latest developments in digital have provided large data sets that can
increasingly easily be accessed and used. These data sets often contain
indirect localisation information, such as historical addresses. Historical
geocoding is the process of transforming the indirect localisation information
to direct localisation that can be placed on a map, which enables spatial
analysis and cross-referencing. Many efficient geocoders exist for current
addresses, but they do not deal with the temporal aspect and are based on a
strict hierarchy (..., city, street, house number) that is hard or impossible
to use with historical data. Indeed historical data are full of uncertainties
(temporal aspect, semantic aspect, spatial precision, confidence in historical
source, ...) that can not be resolved, as there is no way to go back in time to
check. We propose an open source, open data, extensible solution for geocoding
that is based on the building of gazetteers composed of geohistorical objects
extracted from historical topographical maps. Once the gazetteers are
available, geocoding an historical address is a matter of finding the
geohistorical object in the gazetteers that is the best match to the historical
address. The matching criteriae are customisable and include several dimensions
(fuzzy semantic, fuzzy temporal, scale, spatial precision ...). As the goal is
to facilitate historical work, we also propose web-based user interfaces that
help geocode (one address or batch mode) and display over current or historical
topographical maps, so that they can be checked and collaboratively edited. The
system is tested on Paris city for the 19-20th centuries, shows high returns
rate and is fast enough to be used interactively.Comment: WORKING PAPE
An overview of decision table literature 1982-1995.
This report gives an overview of the literature on decision tables over the past 15 years. As much as possible, for each reference, an author supplied abstract, a number of keywords and a classification are provided. In some cases own comments are added. The purpose of these comments is to show where, how and why decision tables are used. The literature is classified according to application area, theoretical versus practical character, year of publication, country or origin (not necessarily country of publication) and the language of the document. After a description of the scope of the interview, classification results and the classification by topic are presented. The main body of the paper is the ordered list of publications with abstract, classification and comments.
Context Aware Computing for The Internet of Things: A Survey
As we are moving towards the Internet of Things (IoT), the number of sensors
deployed around the world is growing at a rapid pace. Market research has shown
a significant growth of sensor deployments over the past decade and has
predicted a significant increment of the growth rate in the future. These
sensors continuously generate enormous amounts of data. However, in order to
add value to raw sensor data we need to understand it. Collection, modelling,
reasoning, and distribution of context in relation to sensor data plays
critical role in this challenge. Context-aware computing has proven to be
successful in understanding sensor data. In this paper, we survey context
awareness from an IoT perspective. We present the necessary background by
introducing the IoT paradigm and context-aware fundamentals at the beginning.
Then we provide an in-depth analysis of context life cycle. We evaluate a
subset of projects (50) which represent the majority of research and commercial
solutions proposed in the field of context-aware computing conducted over the
last decade (2001-2011) based on our own taxonomy. Finally, based on our
evaluation, we highlight the lessons to be learnt from the past and some
possible directions for future research. The survey addresses a broad range of
techniques, methods, models, functionalities, systems, applications, and
middleware solutions related to context awareness and IoT. Our goal is not only
to analyse, compare and consolidate past research work but also to appreciate
their findings and discuss their applicability towards the IoT.Comment: IEEE Communications Surveys & Tutorials Journal, 201
- âŠ