1,596 research outputs found

    Resolving Semantic Ambiguities in Sentences: Cognitive Processes and Brain Mechanisms

    Get PDF
    fMRI studies of how the brain processes sentences containing semantically ambiguous words have consistently implicated (i) the left inferior frontal gyrus (LIFG) and (ii) posterior regions of the left temporal lobe in processing high-ambiguity sentences. This article reviews recent findings on this topic and relates them to (i) psycholinguistic theories about the underlying cognitive processes and (ii) general neuro-cognitive accounts of the relevant brain regions. We suggest that the LIFG plays a general role in the cognitive control process that are necessary to select contextually relevant meanings and to reinterpret sentences that were initially misunderstood, but it is currently unclear whether these control processes should best be characterised in terms of specific processes such as conflict resolution and controlled retrieval that are required for high-ambiguity sentences, or whether its function is better characterised in terms of a more general set of ‘unification’ processes. In contrast to the relatively rapid progress that has been made in understanding the function of the LIFG, we suggest that the contribution of the posterior temporal lobe is less well understood and future work is needed to clarify its role in sentence comprehension

    NLSC: Unrestricted Natural Language-based Service Composition through Sentence Embeddings

    Full text link
    Current approaches for service composition (assemblies of atomic services) require developers to use: (a) domain-specific semantics to formalize services that restrict the vocabulary for their descriptions, and (b) translation mechanisms for service retrieval to convert unstructured user requests to strongly-typed semantic representations. In our work, we argue that effort to developing service descriptions, request translations, and matching mechanisms could be reduced using unrestricted natural language; allowing both: (1) end-users to intuitively express their needs using natural language, and (2) service developers to develop services without relying on syntactic/semantic description languages. Although there are some natural language-based service composition approaches, they restrict service retrieval to syntactic/semantic matching. With recent developments in Machine learning and Natural Language Processing, we motivate the use of Sentence Embeddings by leveraging richer semantic representations of sentences for service description, matching and retrieval. Experimental results show that service composition development effort may be reduced by more than 44\% while keeping a high precision/recall when matching high-level user requests with low-level service method invocations.Comment: This paper will appear on SCC'19 (IEEE International Conference on Services Computing) on July 1

    Applying Wikipedia to Interactive Information Retrieval

    Get PDF
    There are many opportunities to improve the interactivity of information retrieval systems beyond the ubiquitous search box. One idea is to use knowledge bases—e.g. controlled vocabularies, classification schemes, thesauri and ontologies—to organize, describe and navigate the information space. These resources are popular in libraries and specialist collections, but have proven too expensive and narrow to be applied to everyday webscale search. Wikipedia has the potential to bring structured knowledge into more widespread use. This online, collaboratively generated encyclopaedia is one of the largest and most consulted reference works in existence. It is broader, deeper and more agile than the knowledge bases put forward to assist retrieval in the past. Rendering this resource machine-readable is a challenging task that has captured the interest of many researchers. Many see it as a key step required to break the knowledge acquisition bottleneck that crippled previous efforts. This thesis claims that the roadblock can be sidestepped: Wikipedia can be applied effectively to open-domain information retrieval with minimal natural language processing or information extraction. The key is to focus on gathering and applying human-readable rather than machine-readable knowledge. To demonstrate this claim, the thesis tackles three separate problems: extracting knowledge from Wikipedia; connecting it to textual documents; and applying it to the retrieval process. First, we demonstrate that a large thesaurus-like structure can be obtained directly from Wikipedia, and that accurate measures of semantic relatedness can be efficiently mined from it. Second, we show that Wikipedia provides the necessary features and training data for existing data mining techniques to accurately detect and disambiguate topics when they are mentioned in plain text. Third, we provide two systems and user studies that demonstrate the utility of the Wikipedia-derived knowledge base for interactive information retrieval

    A Survey on Semantic Processing Techniques

    Full text link
    Semantic processing is a fundamental research domain in computational linguistics. In the era of powerful pre-trained language models and large language models, the advancement of research in this domain appears to be decelerating. However, the study of semantics is multi-dimensional in linguistics. The research depth and breadth of computational semantic processing can be largely improved with new technologies. In this survey, we analyzed five semantic processing tasks, e.g., word sense disambiguation, anaphora resolution, named entity recognition, concept extraction, and subjectivity detection. We study relevant theoretical research in these fields, advanced methods, and downstream applications. We connect the surveyed tasks with downstream applications because this may inspire future scholars to fuse these low-level semantic processing tasks with high-level natural language processing tasks. The review of theoretical research may also inspire new tasks and technologies in the semantic processing domain. Finally, we compare the different semantic processing techniques and summarize their technical trends, application trends, and future directions.Comment: Published at Information Fusion, Volume 101, 2024, 101988, ISSN 1566-2535. The equal contribution mark is missed in the published version due to the publication policies. Please contact Prof. Erik Cambria for detail

    Representation and processing of semantic ambiguity

    Get PDF
    One of the established findings in the psycholinguistic literature is that semantic ambiguity (e.g., “dog/tree bark”) slows word comprehension in neutral/ minimal context, though it is not entirely clear why this happens. Under the “semantic competition” account, this ambiguity disadvantage effect is due to competition between multiple semantic representations in the race for activation. Under the alternative “decision-making” account, it is due to decision-making difficulties in response selection. This thesis tests the two accounts by investigating in detail the ambiguity disadvantage in semantic relatedness decisions. Chapters 2-4 concentrate on homonyms, words with multiple unrelated meanings. The findings show that the ambiguity disadvantage effect arises only when the different meanings of homonyms are of comparable frequency (e.g., “football/electric fan”), and are therefore initially activated in parallel. Critically, homonymy has this effect during semantic activation of the ambiguous word, not during response selection. This finding, in particular, refutes any idea that the ambiguity disadvantage is due to decision making in response selection. Chapters 5 and 6 concentrate on polysemes, words with multiple related senses. The findings show that the ambiguity disadvantage effect arises for polysemes with irregular sense extension (e.g., “restaurant/website menu”), but not for polysemes with regular (e.g., “fluffy/marinated rabbit”) or figurative sense extension (e.g., “wooden/authoritative chair”). The latter two escape competition because they have only one semantic representation for the dominant sense, with rules of sense extension to derive the alternative sense on-line. Taken together, this thesis establishes that the ambiguity disadvantage is due to semantic competition but is restricted to some forms of ambiguity only. This is because ambiguous words differ in how their meanings are represented and processed, as delineated in this work

    NLP Driven Models for Automatically Generating Survey Articles for Scientific Topics.

    Full text link
    This thesis presents new methods that use natural language processing (NLP) driven models for summarizing research in scientific fields. Given a topic query in the form of a text string, we present methods for finding research articles relevant to the topic as well as summarization algorithms that use lexical and discourse information present in the text of these articles to generate coherent and readable extractive summaries of past research on the topic. In addition to summarizing prior research, good survey articles should also forecast future trends. With this motivation, we present work on forecasting future impact of scientific publications using NLP driven features.PhDComputer Science and EngineeringUniversity of Michigan, Horace H. Rackham School of Graduate Studieshttp://deepblue.lib.umich.edu/bitstream/2027.42/113407/1/rahuljha_1.pd

    The neurocognition of syntactic processing

    Get PDF

    Unsupervised quantification of entity consistency between photos and text in real-world news

    Get PDF
    Das World Wide Web und die sozialen Medien ĂŒbernehmen im heutigen Informationszeitalter eine wichtige Rolle fĂŒr die Vermittlung von Nachrichten und Informationen. In der Regel werden verschiedene ModalitĂ€ten im Sinne der Informationskodierung wie beispielsweise Fotos und Text verwendet, um Nachrichten effektiver zu vermitteln oder Aufmerksamkeit zu erregen. Kommunikations- und Sprachwissenschaftler erforschen das komplexe Zusammenspiel zwischen ModalitĂ€ten seit Jahrzehnten und haben unter Anderem untersucht, wie durch die Kombination der ModalitĂ€ten zusĂ€tzliche Informationen oder eine neue Bedeutungsebene entstehen können. Die Anzahl gemeinsamer Konzepte oder EntitĂ€ten (beispielsweise Personen, Orte und Ereignisse) zwischen Fotos und Text stellen einen wichtigen Aspekt fĂŒr die Bewertung der Gesamtaussage und Bedeutung eines multimodalen Artikels dar. Automatisierte AnsĂ€tze zur Quantifizierung von Bild-Text-Beziehungen können fĂŒr zahlreiche Anwendungen eingesetzt werden. Sie ermöglichen beispielsweise eine effiziente Exploration von Nachrichten, erleichtern die semantische Suche von Multimedia-Inhalten in (Web)-Archiven oder unterstĂŒtzen menschliche Analysten bei der Evaluierung der GlaubwĂŒrdigkeit von Nachrichten. Allerdings gibt es bislang nur wenige AnsĂ€tze, die sich mit der Quantifizierung von Beziehungen zwischen Fotos und Text beschĂ€ftigen. Diese AnsĂ€tze berĂŒcksichtigen jedoch nicht explizit die intermodalen Beziehungen von EntitĂ€ten, welche eine wichtige Rolle in Nachrichten darstellen, oder basieren auf ĂŒberwachten multimodalen Deep-Learning-Techniken. Diese ĂŒberwachten Lernverfahren können ausschließlich die intermodalen Beziehungen von EntitĂ€ten detektieren, die in annotierten Trainingsdaten enthalten sind. Um diese ForschungslĂŒcke zu schließen, wird in dieser Arbeit ein unĂŒberwachter Ansatz zur Quantifizierung der intermodalen Konsistenz von EntitĂ€ten zwischen Fotos und Text in realen multimodalen Nachrichtenartikeln vorgestellt. Im ersten Teil dieser Arbeit werden neuartige Verfahren auf Basis von Deep Learning zur Extrahierung von Informationen aus Fotos vorgestellt, um Ereignisse (Events), Orte, Zeitangaben und Personen automatisch zu erkennen. Diese Verfahren bilden eine wichtige Voraussetzung, um die Beziehungen von EntitĂ€ten zwischen Bild und Text zu bewerten. ZunĂ€chst wird ein Ansatz zur Ereignisklassifizierung prĂ€sentiert, der neuartige Optimierungsfunktionen und Gewichtungsschemata nutzt um Ontologie-Informationen aus einer Wissensdatenbank in ein Deep-Learning-Verfahren zu integrieren. Das Training erfolgt anhand eines neu vorgestellten Datensatzes, der 570.540 Fotos und eine Ontologie mit 148 Ereignistypen enthĂ€lt. Der Ansatz ĂŒbertrifft die Ergebnisse von Referenzsystemen die keine strukturierten Ontologie-Informationen verwenden. Weiterhin wird ein DeepLearning-Ansatz zur SchĂ€tzung des Aufnahmeortes von Fotos vorgeschlagen, der Kontextinformationen ĂŒber die Umgebung (Innen-, Stadt-, oder Naturaufnahme) und von Erdpartitionen unterschiedlicher GranularitĂ€t verwendet. Die vorgeschlagene Lösung ĂŒbertrifft die bisher besten Ergebnisse von aktuellen Forschungsarbeiten, obwohl diese deutlich mehr Fotos zum Training verwenden. DarĂŒber hinaus stellen wir den ersten Datensatz zur SchĂ€tzung des Aufnahmejahres von Fotos vor, der mehr als eine Million Bilder aus den Jahren 1930 bis 1999 umfasst. Dieser Datensatz wird fĂŒr das Training von zwei Deep-Learning-AnsĂ€tzen zur SchĂ€tzung des Aufnahmejahres verwendet, welche die Aufgabe als Klassifizierungs- und Regressionsproblem behandeln. Beide AnsĂ€tze erzielen sehr gute Ergebnisse und ĂŒbertreffen Annotationen von menschlichen Probanden. Schließlich wird ein neuartiger Ansatz zur Identifizierung von Personen des öffentlichen Lebens und ihres gemeinsamen Auftretens in Nachrichtenfotos aus der digitalen Bibliothek Internet Archiv prĂ€sentiert. Der Ansatz ermöglicht es unstrukturierte Webdaten aus dem Internet Archiv mit Metadaten, beispielsweise zur semantischen Suche, zu erweitern. Experimentelle Ergebnisse haben die EffektivitĂ€t des zugrundeliegenden Deep-Learning-Ansatzes zur Personenerkennung bestĂ€tigt. Im zweiten Teil dieser Arbeit wird ein unĂŒberwachtes System zur Quantifizierung von BildText-Beziehungen in realen Nachrichten vorgestellt. Im Gegensatz zu bisherigen Verfahren liefert es automatisch neuartige Maße der intermodalen Konsistenz fĂŒr verschiedene EntitĂ€tstypen (Personen, Orte und Ereignisse) sowie den Gesamtkontext. Das System ist nicht auf vordefinierte DatensĂ€tze angewiesen, und kann daher mit der Vielzahl und DiversitĂ€t von EntitĂ€ten und Themen in Nachrichten umgehen. Zur Extrahierung von EntitĂ€ten aus dem Text werden geeignete Methoden der natĂŒrlichen Sprachverarbeitung eingesetzt. Examplarbilder fĂŒr diese EntitĂ€ten werden automatisch aus dem Internet beschafft. Die vorgeschlagenen Methoden zur Informationsextraktion aus Fotos werden auf die Nachrichten- und heruntergeladenen Exemplarbilder angewendet, um die intermodale Konsistenz von EntitĂ€ten zu quantifizieren. Es werden zwei Aufgaben untersucht um die QualitĂ€t des vorgeschlagenen Ansatzes in realen Anwendungen zu bewerten. Experimentelle Ergebnisse fĂŒr die Dokumentverifikation und die Beschaffung von Nachrichten mit geringer (potenzielle Fehlinformation) oder hoher multimodalen Konsistenz zeigen den Nutzen und das Potenzial des Ansatzes zur UnterstĂŒtzung menschlicher Analysten bei der Untersuchung von Nachrichten.In today’s information age, the World Wide Web and social media are important sources for news and information. Different modalities (in the sense of information encoding) such as photos and text are typically used to communicate news more effectively or to attract attention. Communication scientists, linguists, and semioticians have studied the complex interplay between modalities for decades and investigated, e.g., how their combination can carry additional information or add a new level of meaning. The number of shared concepts or entities (e.g., persons, locations, and events) between photos and text is an important aspect to evaluate the overall message and meaning of an article. Computational models for the quantification of image-text relations can enable many applications. For example, they allow for more efficient exploration of news, facilitate semantic search and multimedia retrieval in large (web) archives, or assist human assessors in evaluating news for credibility. To date, only a few approaches have been suggested that quantify relations between photos and text. However, they either do not explicitly consider the cross-modal relations of entities – which are important in the news – or rely on supervised deep learning approaches that can only detect the cross-modal presence of entities covered in the labeled training data. To address this research gap, this thesis proposes an unsupervised approach that can quantify entity consistency between photos and text in multimodal real-world news articles. The first part of this thesis presents novel approaches based on deep learning for information extraction from photos to recognize events, locations, dates, and persons. These approaches are an important prerequisite to measure the cross-modal presence of entities in text and photos. First, an ontology-driven event classification approach that leverages new loss functions and weighting schemes is presented. It is trained on a novel dataset of 570,540 photos and an ontology with 148 event types. The proposed system outperforms approaches that do not use structured ontology information. Second, a novel deep learning approach for geolocation estimation is proposed that uses additional contextual information on the environmental setting (indoor, urban, natural) and from earth partitions of different granularity. The proposed solution outperforms state-of-the-art approaches, which are trained with significantly more photos. Third, we introduce the first large-scale dataset for date estimation with more than one million photos taken between 1930 and 1999, along with two deep learning approaches that treat date estimation as a classification and regression problem. Both approaches achieve very good results that are superior to human annotations. Finally, a novel approach is presented that identifies public persons and their co-occurrences in news photos extracted from the Internet Archive, which collects time-versioned snapshots of web pages that are rarely enriched with metadata relevant to multimedia retrieval. Experimental results confirm the effectiveness of the deep learning approach for person identification. The second part of this thesis introduces an unsupervised approach capable of quantifying image-text relations in real-world news. Unlike related work, the proposed solution automatically provides novel measures of cross-modal consistency for different entity types (persons, locations, and events) as well as the overall context. The approach does not rely on any predefined datasets to cope with the large amount and diversity of entities and topics covered in the news. State-of-the-art tools for natural language processing are applied to extract named entities from the text. Example photos for these entities are automatically crawled from the Web. The proposed methods for information extraction from photos are applied to both news images and example photos to quantify the cross-modal consistency of entities. Two tasks are introduced to assess the quality of the proposed approach in real-world applications. Experimental results for document verification and retrieval of news with either low (potential misinformation) or high cross-modal similarities demonstrate the feasibility of the approach and its potential to support human assessors to study news

    Deep Learning Methods for Register Classification

    Get PDF
    For this project the data used is the one collected by, Biber and Egbert (2018) related to various language articles from the internet. I am using BERT model (Bidirectional Encoder Representations from Transformers), which is a deep neural network and FastText, which is a shallow neural network, as a baseline to perform text classiïŹcation. Also, I am using Deep Learning models like XLNet to see if classiïŹcation accuracy is improved. Also, it has been described by Biber and Egbert (2018) what is register. We can think of register as genre. According to Biber (1988), register is varieties deïŹned in terms of general situational parameters. Hence, it can be inferred that there is a close relation between the language and the context of the situation in which it is being used. This work attempts register classiïŹcation using deep learning methods that use attention mechanism. Working with the models, dealing with the imbalanced datasets in real life problems, tuning the hyperparameters for training the models was accomplished throughout the work. Also, proper evaluation metrics for various kind of data was determined. The background study shows that how cumbersome the use classical Machine Learning approach used to be. Deep Learning, on the other hand, can accomplish the task with ease. The metric to be selected for the classiïŹcation task for different types of datasets (balanced vs imbalanced), dealing with overïŹtting was also accomplished

    A test of the role of the medial temporal lobe in single-word decoding.

    Get PDF
    The degree to which the MTL system contributes to effective language skills is not well delineated. We sought to determine if the MTL plays a role in single-word decoding in healthy, normal skilled readers. The experiment follows from the implications of the dual-process model of single-word decoding, which provides distinct predictions about the nature of MTL involvement. The paradigm utilized word (regular and irregularly spelled words) and pseudoword (phonetically regular) stimuli that differed in their demand for non-lexical as opposed lexical decoding. The data clearly showed that the MTL system was not involved in single word decoding in skilled, native English readers. Neither the hippocampus nor the MTL system as a whole showed significant activation during lexical or non-lexical based decoding. The results provide evidence that lexical and non-lexical decoding are implemented by distinct but overlapping neuroanatomical networks. Non-lexical decoding appeared most uniquely associated with cuneus and fusiform gyrus activation biased toward the left hemisphere. In contrast, lexical decoding appeared associated with right middle frontal and supramarginal, and bilateral cerebellar activation. Both these decoding operations appeared in the context of a shared widespread network of activations including bilateral occipital cortex and superior frontal regions. These activations suggest that the absence of MTL involvement in either lexical or non-lexical decoding appears likely a function of the skilled reading ability of our sample such that whole-word recognition and retrieval processes do not utilize the declarative memory system, in the case of lexical decoding, and require only minimal analysis and recombination of the phonetic elements of a word, in the case of non-lexical decoding
    • 

    corecore