1,125 research outputs found
Km4City Ontology Building vs Data Harvesting and Cleaning for Smart-city Services
Presently, a very large number of public and private data sets are available
from local governments. In most cases, they are not semantically interoperable
and a huge human effort would be needed to create integrated ontologies and
knowledge base for smart city. Smart City ontology is not yet standardized, and
a lot of research work is needed to identify models that can easily support the
data reconciliation, the management of the complexity, to allow the data
reasoning. In this paper, a system for data ingestion and reconciliation of
smart cities related aspects as road graph, services available on the roads,
traffic sensors etc., is proposed. The system allows managing a big data volume
of data coming from a variety of sources considering both static and dynamic
data. These data are mapped to a smart-city ontology, called KM4City (Knowledge
Model for City), and stored into an RDF-Store where they are available for
applications via SPARQL queries to provide new services to the users via
specific applications of public administration and enterprises. The paper
presents the process adopted to produce the ontology and the big data
architecture for the knowledge base feeding on the basis of open and private
data, and the mechanisms adopted for the data verification, reconciliation and
validation. Some examples about the possible usage of the coherent big data
knowledge base produced are also offered and are accessible from the RDF-Store
and related services. The article also presented the work performed about
reconciliation algorithms and their comparative assessment and selection
Semantic Technologies for Manuscript Descriptions — Concepts and Visions
The contribution at hand relates recent developments in the area of the World Wide
Web to codicological research. In the last number of years, an informational extension
of the internet has been discussed and extensively researched: the Semantic Web. It
has already been applied in many areas, including digital information processing of
cultural heritage data. The Semantic Web facilitates the organisation and linking of
data across websites, according to a given semantic structure. Software can then process
this structural and semantic information to extract further knowledge. In the area
of codicological research, many institutions are making efforts to improve the online
availability of handwritten codices. If these resources could also employ Semantic
Web techniques, considerable research potential could be unleashed. However, data
acquisition from less structured data sources will be problematic. In particular, data
stemming from unstructured sources needs to be made accessible to SemanticWeb tools
through information extraction techniques. In the area of museum research, the CIDOC
Conceptual Reference Model (CRM) has been widely examined and is being adopted
successfully. The CRM translates well to Semantic Web research, and its concentration
on contextualization of objects could support approaches in codicological research.
Further concepts for the creation and management of bibliographic coherences and
structured vocabularies related to the CRM will be considered in this chapter. Finally, a
user scenario showing all processing steps in their context will be elaborated on
SWI-Prolog and the Web
Where Prolog is commonly seen as a component in a Web application that is
either embedded or communicates using a proprietary protocol, we propose an
architecture where Prolog communicates to other components in a Web application
using the standard HTTP protocol. By avoiding embedding in external Web servers
development and deployment become much easier. To support this architecture, in
addition to the transfer protocol, we must also support parsing, representing
and generating the key Web document types such as HTML, XML and RDF.
This paper motivates the design decisions in the libraries and extensions to
Prolog for handling Web documents and protocols. The design has been guided by
the requirement to handle large documents efficiently. The described libraries
support a wide range of Web applications ranging from HTML and XML documents to
Semantic Web RDF processing.
To appear in Theory and Practice of Logic Programming (TPLP)Comment: 31 pages, 24 figures and 2 tables. To appear in Theory and Practice
of Logic Programming (TPLP
Using Protege for automatic ontology instantiation
This paper gives an overview on the use of Protégé in the Artequakt system, which integrated Protégé with a set of natural language tools to automatically extract knowledge about artists from web documents and instantiate a given ontology. Protégé was also linked to structured templates that generate documents from the knowledge fragments it maintains
- …