21,100 research outputs found
Recommended from our members
Abbreviated text input using language modeling.
We address the problem of improving the efficiency of natural language text input under degraded conditions (for instance, on mobile computing devices or by disabled users), by taking advantage of the informational redundancy in natural language. Previous approaches to this problem have been based on the idea of prediction of the text, but these require the user to take overt action to verify or select the system’s predictions. We propose taking advantage of the duality between prediction and compression. We allow the
user to enter text in compressed form, in particular, using a simple stipulated abbreviation method that reduces characters by 26.4%, yet is simple enough that it can be learned
easily and generated relatively fluently. We decode the abbreviated text using a statistical generative model of abbreviation, with a residual word error rate of 3.3%. The chief
component of this model is an n-gram language model. Because the system’s operation is
completely independent from the user’s, the overhead from cognitive task switching and
attending to the system’s actions online is eliminated, opening up the possibility that
the compression-based method can achieve text input efficiency improvements where the
prediction-based methods have not. We report the results of a user study evaluating this
method.Engineering and Applied Science
Intelligent Techniques to Accelerate Everyday Text Communication
People with some form of speech- or motor-impairments usually use a high-tech augmentative and alternative communication (AAC) device to communicate with other people in writing or in face-to-face conversations. Their text entry rate on these devices is slow due to their motor abilities. Making good letter or word predictions can help accelerate the communication of such users. In this dissertation, we investigated several approaches to accelerate input for AAC users. First, considering that an AAC user is participating in a face-to-face conversation, we investigated whether performing speech recognition on the speaking-side can improve next word predictions. We compared the accuracy of three plausible microphone deployment options and the accuracy of two commercial speech recognition engines. We found that despite recognition word error rates of 7-16%, our ensemble of n-gram and recurrent neural network language models made predictions nearly as good as when they used the reference transcripts. In a user study with 160 participants, we also found that increasing number of prediction slots in a keyboard interface does not necessarily correlate to improved performance. Second, typing every character in a text message may require an AAC user more time or effort than strictly necessary. Skipping spaces or other characters may be able to speed input and reduce an AAC user\u27s physical input effort. We designed a recognizer optimized for expanding noisy abbreviated input where users often omitted spaces and mid-word vowels. We showed using neural language models for selecting conversational-style training text and for rescoring the recognizer\u27s n-best sentences improved accuracy. We found accurate abbreviated input was possible even if a third of characters was omitted. In a study where users had to dwell for a second on each key, we found sentence abbreviated input was competitive with a conventional keyboard with word predictions. Finally, AAC keyboards rely on language modeling to auto-correct noisy typing and to offer word predictions. While today language models can be trained on huge amounts of text, pre-trained models may fail to capture the unique writing style and vocabulary of individual users. We demonstrated improved performance compared to a unigram cache by adapting to a user\u27s text via language models based on prediction by partial match (PPM) and recurrent neural networks. Our best model ensemble increased keystroke savings by 9.6%
Metamodeling in EIA/CDIF - Meta-Metamodel and Metamodels
This article introduces the EIA/CDIF set of standards for the modeling of information systems and its exchange among computer-aided software tools of different vendors. It lays out the meta-metamodel and the standardized metamodels which are fully depicted in a hierarchical layout and annotated with the unique identifiers of all the standardized modeling concepts. The article also stresses the fact that EIA/CDIF has been used as the baseline in the creation of an international standard, the ISO/CDIF set of models, an ongoing project
An Agent-based Simulation of the Effectiveness of Creative Leadership\ud
This paper investigates the effectiveness of creative versus\ud
uncreative leadership using EVOC, an agent-based model of\ud
cultural evolution. Each iteration, each agent in the artificial society invents a new action, or imitates a neighbor’s action. Only the leader’s actions can be imitated by all other agents, referred to as followers. Two measures of creativity were used: (1) invention-to-imitation ratio, iLeader, which measures how often an agent invents, and (2) rate of conceptual change, cLeader, which measures how creative an invention is. High iLeader increased mean fitness of ideas, but only when creativity of followers was low. High iLeader was associated with greater diversity of ideas in the early stage of idea generation only. High Leader increased mean fitness of ideas in the early stage of idea generation; in the later stage it decreased idea fitness. Reasons for these findings and tentative implications for creative leadership in human society are discussed
Doc2EDAG: An End-to-End Document-level Framework for Chinese Financial Event Extraction
Most existing event extraction (EE) methods merely extract event arguments
within the sentence scope. However, such sentence-level EE methods struggle to
handle soaring amounts of documents from emerging applications, such as
finance, legislation, health, etc., where event arguments always scatter across
different sentences, and even multiple such event mentions frequently co-exist
in the same document. To address these challenges, we propose a novel
end-to-end model, Doc2EDAG, which can generate an entity-based directed acyclic
graph to fulfill the document-level EE (DEE) effectively. Moreover, we
reformalize a DEE task with the no-trigger-words design to ease the
document-level event labeling. To demonstrate the effectiveness of Doc2EDAG, we
build a large-scale real-world dataset consisting of Chinese financial
announcements with the challenges mentioned above. Extensive experiments with
comprehensive analyses illustrate the superiority of Doc2EDAG over
state-of-the-art methods. Data and codes can be found at
https://github.com/dolphin-zs/Doc2EDAG.Comment: Accepted by EMNLP 201
- …