744,631 research outputs found
A framework for deriving semantic web services
Web service-based development represents an emerging approach for the development of distributed information systems. Web services have been mainly applied by software practitioners as a means to modularize system functionality that can be offered across a network (e.g., intranet and/or the Internet). Although web services have been
predominantly developed as a technical solution for integrating software systems, there is a more business-oriented aspect that developers and enterprises need to deal with in order to benefit from the full potential of web services in an electronic market. This âignoredâ aspect is the representation of the semantics underlying the services themselves as well as the âthingsâ that the services manage. Currently languages like the Web Services Description Language (WSDL) provide the syntactic means to describe web services, but
lack in providing a semantic underpinning. In order to harvest all the benefits of web services technology, a framework has been developed for deriving business semantics from syntactic descriptions of web services. The benefits of such a framework are two-fold. Firstly, the framework provides a way to gradually construct domain ontologies from previously defined technical services. Secondly, the framework enables the
migration of syntactically defined web services toward semantic web services. The study follows a design research approach which (1) identifies the problem area and its relevance from an industrial case study and previous research, (2) develops the
framework as a design artifact and (3) evaluates the application of the framework through a relevant scenario
Modular Web Queries â From Rules to Stores
Even with all the progress in Semantic technology, accessing Web
data remains a challenging issue with new Web query languages and approaches
appearing regularly. Yet most of these languages, including W3C approaches
such as XQuery and SPARQL, do little to cope with the explosion of the data
size and schemata diversity and richness on the Web. In this paper we propose
a straightforward step toward the improvement of this situation that is simple to
realize and yet effective: Advanced module systems that make partitioning of (a)
the evaluation and (b) the conceptual design of complex Web queries possible.
They provide the query programmer with a powerful, but easy to use high-level
abstraction for packaging, encapsulating, and reusing conceptually related parts
(in our case, rules) of a Web query. The proposed module system combines ease
of use thanks to a simple core concept, the partitioning of rules and their consequences
in flexible âstoresâ, with ease of deployment thanks to a reduction
semantics. We focus on extending the rule-based Semantic Web query language
Xcerpt with such a module system though the same approach can be applied to
other (rule-based) languages as well
Embarrassingly Shallow Autoencoders for Sparse Data
Combining simple elements from the literature, we define a linear model that
is geared toward sparse data, in particular implicit feedback data for
recommender systems. We show that its training objective has a closed-form
solution, and discuss the resulting conceptual insights. Surprisingly, this
simple model achieves better ranking accuracy than various state-of-the-art
collaborative-filtering approaches, including deep non-linear models, on most
of the publicly available data-sets used in our experiments.Comment: In the proceedings of the Web Conference (WWW) 2019 (7 pages
The web of Calabi-Yau hypersurfaces in toric varieties
Recent results on duality between string theories and connectedness of their
moduli spaces seem to go a long way toward establishing the uniqueness of an
underlying theory. For the large class of Calabi-Yau 3-folds that can be
embedded as hypersurfaces in toric varieties the proof of mathematical
connectedness via singular limits is greatly simplified by using polytopes that
are maximal with respect to certain single or multiple weight systems. We
identify the multiple weight systems occurring in this approach. We show that
all of the corresponding Calabi-Yau manifolds are connected among themselves
and to the web of CICY's. This almost completes the proof of connectedness for
toric Calabi-Yau hypersurfaces.Comment: TeX, epsf.tex; 24 page
A Benchmark for Image Retrieval using Distributed Systems over the Internet: BIRDS-I
The performance of CBIR algorithms is usually measured on an isolated
workstation. In a real-world environment the algorithms would only constitute a
minor component among the many interacting components. The Internet
dramati-cally changes many of the usual assumptions about measuring CBIR
performance. Any CBIR benchmark should be designed from a networked systems
standpoint. These benchmarks typically introduce communication overhead because
the real systems they model are distributed applications. We present our
implementation of a client/server benchmark called BIRDS-I to measure image
retrieval performance over the Internet. It has been designed with the trend
toward the use of small personalized wireless systems in mind. Web-based CBIR
implies the use of heteroge-neous image sets, imposing certain constraints on
how the images are organized and the type of performance metrics applicable.
BIRDS-I only requires controlled human intervention for the compilation of the
image collection and none for the generation of ground truth in the measurement
of retrieval accuracy. Benchmark image collections need to be evolved
incrementally toward the storage of millions of images and that scaleup can
only be achieved through the use of computer-aided compilation. Finally, our
scoring metric introduces a tightly optimized image-ranking window.Comment: 24 pages, To appear in the Proc. SPIE Internet Imaging Conference
200
Next Generation Teaching and Learning ??? Technologies and Trends
The landscape of teaching and learning has been radically shifted
in the last 15 years by the advent of web technologies, which
enabled the emergence of Learning Management Systems (LMS).
These systems changed the educational paradigm by extending the
classroom borders, capturing and persisting course content and
giving instructors more flexibility and access to students and other
resources. However, they also constrained and limited the
evolution of teaching and learning by imposing a traditional,
instructional framework. With the advent of Web 2.0
technologies, participation and collaboration have become
predominant experiences on the Web. The teaching and learning
community, as a whole, has been late to capitalize on these
technologies in the classroom. Part of this trend is due to
constraints in the technology (LMS), and part is due to the fact
that participatory media tools require an additional shift in
educational paradigms, from instructional, on-the-pulpit type of
teaching, to a student-centered, adaptive environment where
students can contribute to the course material and learn from one
another. This panel will discuss the next generation of teaching
and learning, involving more lightweight, modular systems to
empower instructors to be flexible, explore new student-centered
paradigms, and plug and play tools as needed. We will also
discuss how the iSchools are and should be increasingly involved
in studying these new forms, formulating best practices and
supporting the needs of teachers as they move toward more
collaborative learning environments
Recommended from our members
Computerization of workflows, guidelines and care pathways: a review of implementation challenges for process-oriented health information systems
There is a need to integrate the various theoretical frameworks and formalisms for modeling clinical guidelines, workflows, and pathways, in order to move beyond providing support for individual clinical decisions and toward the provision of process-oriented, patient-centered, health information systems (HIS). In this review, we analyze the challenges in developing process-oriented HIS that formally model guidelines, workflows, and care pathways. A qualitative meta-synthesis was performed on studies published in English between 1995 and 2010 that addressed the modeling process and reported the exposition of a new methodology, model, system implementation, or system architecture. Thematic analysis, principal component analysis (PCA) and data visualisation techniques were used to identify and cluster the underlying implementation âchallengeâ themes. One hundred and eight relevant studies were selected for review. Twenty-five underlying âchallengeâ themes were identified. These were clustered into 10 distinct groups, from which a conceptual model of the implementation process was developed. We found that the development of systems supporting individual clinical decisions is evolving toward the implementation of adaptable care pathways on the semantic web, incorporating formal, clinical, and organizational ontologies, and the use of workflow management systems. These architectures now need to be implemented and evaluated on a wider scale within clinical settings
Basis Token Consistency: A Practical Mechanism for Strong Web Cache Consistency
With web caching and cache-related services like CDNs and edge services playing an increasingly significant role in the modern internet, the problem of the weak consistency and coherence provisions in current web protocols is becoming increasingly significant and drawing the attention of the standards community [LCD01]. Toward this end, we present definitions of consistency and coherence for web-like environments, that is, distributed client-server information systems where the semantics of interactions with resource are more general than the read/write operations found in memory hierarchies and distributed file systems. We then present a brief review of proposed mechanisms which strengthen the consistency of caches in the web, focusing upon their conceptual contributions and their weaknesses in real-world practice. These insights motivate a new mechanism, which we call "Basis Token Consistency" or BTC; when implemented at the server, this mechanism allows any client (independent of the presence and conformity of any intermediaries) to maintain a self-consistent view of the server's state. This is accomplished by annotating responses with additional per-resource application information which allows client caches to recognize the obsolescence of currently cached entities and identify responses from other caches which are already stale in light of what has already been seen. The mechanism requires no deviation from the existing client-server communication model, and does not require servers to maintain any additional per-client state. We discuss how our mechanism could be integrated into a fragment-assembling Content Management System (CMS), and present a simulation-driven performance comparison between the BTC algorithm and the use of the Time-To-Live (TTL) heuristic.National Science Foundation (ANI-9986397, ANI-0095988
- âŠ