2,089 research outputs found
Extending Machine Language Models toward Human-Level Language Understanding
Language is central to human intelligence. We review recent break- throughs in machine language processing and consider what re- mains to be achieved. Recent approaches rely on domain general principles of learning and representation captured in artificial neu- ral networks. Most current models, however, focus too closely on language itself. In humans, language is part of a larger system for acquiring, representing, and communicating about objects and sit- uations in the physical and social world, and future machine lan- guage models should emulate such a system. We describe exist- ing machine models linking language to concrete situations, and point toward extensions to address more abstract cases. Human language processing exploits complementary learning systems, in- cluding a deep neural network-like learning system that learns grad- ually as machine systems do, as well as a fast-learning system that supports learning new information quickly. Adding such a system to machine language models will be an important further step toward truly human-like language understanding
Neurobehavioral Correlates of Surprisal in Language Comprehension : A Neurocomputational Model
Expectation-based theories of language comprehension, in particular Surprisal Theory,
go a long way in accounting for the behavioral correlates of word-by-word processing
difficulty, such as reading times. An open question, however, is in which component(s)
of the Event-Related brain Potential (ERP) signal Surprisal is reflected, and how these
electrophysiological correlates relate to behavioral processing indices. Here, we address
this question by instantiating an explicit neurocomputational model of incremental,
word-by-word language comprehension that produces estimates of the N400 and
the P600âthe two most salient ERP components for language processingâas well
as estimates of âcomprehension-centricâ Surprisal for each word in a sentence. We
derive model predictions for a recent experimental design that directly investigates
âworld-knowledgeâ-induced Surprisal. By relating these predictions to both empirical
electrophysiological and behavioral results, we establish a close link between Surprisal,
as indexed by reading times, and the P600 component of the ERP signal. The resultant
model thus offers an integrated neurobehavioral account of processing difficulty in
language comprehension
A compositional neural architecture for language
Hierarchical structure and compositionality imbue human language with unparalleled expressive power and set it apart from other perceptionâaction systems. However, neither formal nor neurobiological models account for how these defining computational properties might arise in a physiological system. I attempt to reconcile hierarchy and compositionality with principles from cell assembly computation in neuroscience; the result is an emerging theory of how the brain could convert distributed perceptual representations into hierarchical structures across multiple timescales while representing interpretable incremental stages of (de) compositional meaning. The model's architectureâa multidimensional coordinate system based on neurophysiological models of sensory processingâproposes that a manifold of neural trajectories encodes sensory, motor, and abstract linguistic states. Gain modulation, including inhibition, tunes the path in the manifold in accordance with behavior and is how latent structure is inferred. As a consequence, predictive information about upcoming sensory input during production and comprehension is available without a separate operation. The proposed processing mechanism is synthesized from current models of neural entrainment to speech, concepts from systems neuroscience and category theory, and a symbolic-connectionist computational model that uses time and rhythm to structure information. I build on evidence from cognitive neuroscience and computational modeling that suggests a formal and mechanistic alignment between structure building and neural oscillations and moves toward unifying basic insights from linguistics and psycholinguistics with the currency of neural computation
- âŠ