2,883 research outputs found
Layers in the Fabric of Mind: A Critical Review of Cognitive Ontogeny
The essay is critically examines the conceptual problems with the influential modularity model of mind. We shall see that one of the essential characters of modules, namely informational encapsulation, is not only inessential, it ties a knot at a crucial place blocking the solution to the problem of understanding the formation of concepts from percepts (nodes of procedural knowledge). Subsequently I propose that concept formation takes place by modulation of modules leading to cross-representations, which were otherwise prevented by encapsulation. It must be noted that the argument is not against modular architecture, but a variety of an architecture that prevents interaction among modules. This is followed by a brief argument demonstrating that module without modularization, i.e. without developmental history, is impossible. Finally the emerging picture of cognitive development is drawn in the form of the layers in the fabric of mind, with a brief statement of the possible implications
The Knowledge Level in Cognitive Architectures: Current Limitations and Possible Developments
In this paper we identify and characterize an analysis of two problematic aspects affecting the representational level of cognitive architectures (CAs), namely: the limited size and the homogeneous typology of the encoded and processed knowledge.
We argue that such aspects may constitute not only a technological problem that, in our opinion, should be addressed in order to build articial agents able to exhibit intelligent behaviours in general scenarios, but also an epistemological one, since they limit the plausibility of the comparison of the CAs' knowledge representation and processing mechanisms with those executed by humans in their everyday activities. In the final part of the paper further directions of research will be explored, trying to address current limitations and
future challenges
Preserving the knowledge of long clinical texts using aggregated ensembles of large language models
Clinical texts, such as admission notes, discharge summaries, and progress
notes, contain rich and valuable information that can be used for various
clinical outcome prediction tasks. However, applying large language models,
such as BERT-based models, to clinical texts poses two major challenges: the
limitation of input length and the diversity of data sources. This paper
proposes a novel method to preserve the knowledge of long clinical texts using
aggregated ensembles of large language models. Unlike previous studies which
use model ensembling or text aggregation methods separately, we combine
ensemble learning with text aggregation and train multiple large language
models on two clinical outcome tasks: mortality prediction and length of stay
prediction. We show that our method can achieve better results than baselines,
ensembling, and aggregation individually, and can improve the performance of
large language models while handling long inputs and diverse datasets. We
conduct extensive experiments on the admission notes from the MIMIC-III
clinical database by combining multiple unstructured and high-dimensional
datasets, demonstrating our method's effectiveness and superiority over
existing approaches. We also provide a comprehensive analysis and discussion of
our results, highlighting our method's applications and limitations for future
research in the domain of clinical healthcare. The results and analysis of this
study is supportive of our method assisting in clinical healthcare systems by
enabling clinical decision-making with robust performance overcoming the
challenges of long text inputs and varied datasets.Comment: 17 pages, 4 figures, 4 tables, 9 equations and 1 algorith
Cooperating systems: Layered MAS
Distributed intelligent systems can be distinguished by the models that they use. The model developed focuses on layered multiagent system conceived of as a bureaucracy in which a distributed data base serves as a central means of communication. The various generic bureaus of such a system is described and a basic vocabulary for such systems is presented. In presenting the bureaus and vocabularies, special attention is given to the sorts of reasonings that are appropriate. A bureaucratic model has a hierarchy of master system and work group that organizes E agents and B agents. The master system provides the administrative services and support facilities for the work groups
Hypermedia Learning Objects System - On the Way to a Semantic Educational Web
While eLearning systems become more and more popular in daily education,
available applications lack opportunities to structure, annotate and manage
their contents in a high-level fashion. General efforts to improve these
deficits are taken by initiatives to define rich meta data sets and a
semanticWeb layer. In the present paper we introduce Hylos, an online learning
system. Hylos is based on a cellular eLearning Object (ELO) information model
encapsulating meta data conforming to the LOM standard. Content management is
provisioned on this semantic meta data level and allows for variable,
dynamically adaptable access structures. Context aware multifunctional links
permit a systematic navigation depending on the learners and didactic needs,
thereby exploring the capabilities of the semantic web. Hylos is built upon the
more general Multimedia Information Repository (MIR) and the MIR adaptive
context linking environment (MIRaCLE), its linking extension. MIR is an open
system supporting the standards XML, Corba and JNDI. Hylos benefits from
manageable information structures, sophisticated access logic and high-level
authoring tools like the ELO editor responsible for the semi-manual creation of
meta data and WYSIWYG like content editing.Comment: 11 pages, 7 figure
Tracking time evolving data streams for short-term traffic forecasting
YesData streams have arisen as a relevant topic during the last few years as an efficient method for extracting knowledge from big data. In the robust layered ensemble model (RLEM) proposed in this paper for short-term traffic flow forecasting, incoming traffic flow data of all connected road links are organized in chunks corresponding to an optimal time lag. The RLEM model is composed of two layers. In the first layer, we cluster the chunks by using the Graded Possibilistic c-Means method. The second layer is made up by an ensemble of forecasters, each of them trained for short-term traffic flow forecasting on the chunks belonging to a specific cluster. In the operational phase, as a new chunk of traffic flow data presented as input to the RLEM, its memberships to all clusters are evaluated, and if it is not recognized as an outlier, the outputs of all forecasters are combined in an ensemble, obtaining in this a way a forecasting of traffic flow for a short-term time horizon. The proposed RLEM model is evaluated on a synthetic data set, on a traffic flow data simulator and on two real-world traffic flow data sets. The model gives an accurate forecasting of the traffic flow rates with outlier detection and shows a good adaptation to non-stationary traffic regimes. Given its characteristics of outlier detection, accuracy, and robustness, RLEM can be fruitfully integrated in traffic flow management systems
A User-Centered Concept Mining System for Query and Document Understanding at Tencent
Concepts embody the knowledge of the world and facilitate the cognitive
processes of human beings. Mining concepts from web documents and constructing
the corresponding taxonomy are core research problems in text understanding and
support many downstream tasks such as query analysis, knowledge base
construction, recommendation, and search. However, we argue that most prior
studies extract formal and overly general concepts from Wikipedia or static web
pages, which are not representing the user perspective. In this paper, we
describe our experience of implementing and deploying ConcepT in Tencent QQ
Browser. It discovers user-centered concepts at the right granularity
conforming to user interests, by mining a large amount of user queries and
interactive search click logs. The extracted concepts have the proper
granularity, are consistent with user language styles and are dynamically
updated. We further present our techniques to tag documents with user-centered
concepts and to construct a topic-concept-instance taxonomy, which has helped
to improve search as well as news feeds recommendation in Tencent QQ Browser.
We performed extensive offline evaluation to demonstrate that our approach
could extract concepts of higher quality compared to several other existing
methods. Our system has been deployed in Tencent QQ Browser. Results from
online A/B testing involving a large number of real users suggest that the
Impression Efficiency of feeds users increased by 6.01% after incorporating the
user-centered concepts into the recommendation framework of Tencent QQ Browser.Comment: Accepted by KDD 201
Accelerating COVID-19 research with graph mining and transformer-based learning
In 2020, the White House released the, "Call to Action to the Tech Community
on New Machine Readable COVID-19 Dataset," wherein artificial intelligence
experts are asked to collect data and develop text mining techniques that can
help the science community answer high-priority scientific questions related to
COVID-19. The Allen Institute for AI and collaborators announced the
availability of a rapidly growing open dataset of publications, the COVID-19
Open Research Dataset (CORD-19). As the pace of research accelerates,
biomedical scientists struggle to stay current. To expedite their
investigations, scientists leverage hypothesis generation systems, which can
automatically inspect published papers to discover novel implicit connections.
We present an automated general purpose hypothesis generation systems AGATHA-C
and AGATHA-GP for COVID-19 research. The systems are based on graph-mining and
the transformer model. The systems are massively validated using retrospective
information rediscovery and proactive analysis involving human-in-the-loop
expert analysis. Both systems achieve high-quality predictions across domains
(in some domains up to 0.97% ROC AUC) in fast computational time and are
released to the broad scientific community to accelerate biomedical research.
In addition, by performing the domain expert curated study, we show that the
systems are able to discover on-going research findings such as the
relationship between COVID-19 and oxytocin hormone
It's LeVAsa not LevioSA! Latent Encodings for Valence-Arousal Structure Alignment
In recent years, great strides have been made in the field of affective
computing. Several models have been developed to represent and quantify
emotions. Two popular ones include (i) categorical models which represent
emotions as discrete labels, and (ii) dimensional models which represent
emotions in a Valence-Arousal (VA) circumplex domain. However, there is no
standard for annotation mapping between the two labelling methods. We build a
novel algorithm for mapping categorical and dimensional model labels using
annotation transfer across affective facial image datasets. Further, we utilize
the transferred annotations to learn rich and interpretable data
representations using a variational autoencoder (VAE). We present "LeVAsa", a
VAE model that learns implicit structure by aligning the latent space with the
VA space. We evaluate the efficacy of LeVAsa by comparing performance with the
Vanilla VAE using quantitative and qualitative analysis on two benchmark
affective image datasets. Our results reveal that LeVAsa achieves high
latent-circumplex alignment which leads to improved downstream categorical
emotion prediction. The work also demonstrates the trade-off between degree of
alignment and quality of reconstructions.Comment: 5 pages, 4 figures and 3 table
- …