690 research outputs found
Learning Language from a Large (Unannotated) Corpus
A novel approach to the fully automated, unsupervised extraction of
dependency grammars and associated syntax-to-semantic-relationship mappings
from large text corpora is described. The suggested approach builds on the
authors' prior work with the Link Grammar, RelEx and OpenCog systems, as well
as on a number of prior papers and approaches from the statistical language
learning literature. If successful, this approach would enable the mining of
all the information needed to power a natural language comprehension and
generation system, directly from a large, unannotated corpus.Comment: 29 pages, 5 figures, research proposa
Context-Aware Personalized Point-of-Interest Recommendation System
The increasing volume of information has created overwhelming challenges to extract the relevant items manually. Fortunately, the online systems, such as e-commerce (e.g., Amazon), location-based social networks (LBSNs) (e.g., Facebook) among many others have the ability to track end users\u27 browsing and consumption experiences. Such explicit experiences (e.g., ratings) and many implicit contexts (e.g., social, spatial, temporal, and categorical) are useful in preference elicitation and recommendation. As an emerging branch of information filtering, the recommendation systems are already popular in many domains, such as movies (e.g., YouTube), music (e.g., Pandora), and Point-of-Interest (POI) (e.g., Yelp).
The POI domain has many contextual challenges (e.g., spatial (preferences to a near place), social (e.g., friend\u27s influence), temporal (e.g., popularity at certain time), categorical (similar preferences to places with same category), locality of POI, etc.) that can be crucial for an efficient recommendation. The user reviews shared across different social networks provide granularity in users\u27 consumption experience. From the data mining and machine learning perspective, following three research directions are identified and considered relevant to an efficient context-aware POI recommendation, (1) incorporation of major contexts into a single model and a detailed analysis of the impact of those contexts, (2) exploitation of user activity and location influence to model hierarchical preferences, and (3) exploitation of user reviews to formulate the aspect opinion relation and to generate explanation for recommendation.
This dissertation presents different machine learning and data mining-based solutions to address the above-mentioned research problems, including, (1) recommendation models inspired from contextualized ranking and matrix factorization that incorporate the major contexts and help in analysis of their importance, (2) hierarchical and matrix-factorization models that formulate users\u27 activity and POI influences on different localities that model hierarchical preferences and generate individual and sequence recommendations, and (3) graphical models inspired from natural language processing and neural networks to generate recommendations augmented with aspect-based explanations
Recommended from our members
Efficient Latent Semantic Extraction from Cross Domain Data with Declarative Language
With large amounts of data continuously generated by intelligence devices, efficient analysis of huge data collections to unearth valuable insights has become one of the most elusive challenges for both academia and industry. The key elements to establishing a scalable analyzing framework should involve (1) an intuitive interface to describe the desired outcome, (2) a well-crafted model that integrates all available information sources to derive the optimal outcome and (3) an efficient algorithm that performs the data integration and extraction within a reasonable amount of time. In this dissertation, we address these challenges by proposing (1) a cross-language interface for a succinct expression of recursive queries, (2) a domain specific neural network model that can incorporate information of multiple modalities, and (3) a sample efficient training method that can be used even for extremely-large output-class classifiers. Our contributions in this thesis are thus threefold: First, for the ubiquitous recursive queries in advanced data analytics, on top of BigDatalog and Apache Spark, we design a succinct and expressive analytics tool encapsulating the functionality and classical algorithms of Datalog, a quintessential logic programming language. We provide the Logical Library (LLib), a Spark MLlib-like high-level API supporting a wide range of recursive algorithms and the Logical DataFrame (LFrame), an extension to Spark DataFrame supporting both relational and logical operations. The LLib and LFrame enable smooth collaborations between logical applications and other Spark libraries and cross-language logical programming in Scala, Java, or Python. Second, we utilize variants of recurrent neural network (RNN) to incorporate some enlightening sequential information overlooked by the conventional works in two different domains including Spoken Language Understanding (SLU) and Internet Embedding (IE). In SLU, we address the problem caused by solely relying on the first best interpretation (hypothesis) of an audio command through a series of new architectures comprising bidirectional LSTM and pooling layers to jointly utilize the other hypotheses' texts or embedding vectors, which are neglected but with valuable information missed by the first best hypothesis. In IE, we propose the DIP, an extension of RNN, to build up the internet coordinate system with the IP address sequences, which are also unnoticed in conventional distance-based internet embedding algorithms but encode structural information of the network. Both DIP and the integration of all hypotheses bring significant performance improvements for the corresponding downstream tasks. Finally, we investigate the training algorithm for multi-class classifiers with a large output-class size, which is common in deep neural networks and typically implemented as a softmax final layer with one output neuron per each class. To avoid expensive computing the intractable normalizing constant of softmax for each training data point, we analyze the well-known negative sampling and improve it to the amplified negative sampling algorithm, which gains much higher performance with lower training cost
The Processing of Emotional Sentences by Young and Older Adults: A Visual World Eye-movement Study
Carminati MN, Knoeferle P. The Processing of Emotional Sentences by Young and Older Adults: A Visual World Eye-movement Study. Presented at the Architectures and Mechanisms of Language and Processing (AMLaP), Riva del Garda, Italy
- …