41 research outputs found
Color naming reflects both perceptual structure and communicative need
Gibson et al. (2017) argued that color naming is shaped by patterns of
communicative need. In support of this claim, they showed that color naming
systems across languages support more precise communication about warm colors
than cool colors, and that the objects we talk about tend to be warm-colored
rather than cool-colored. Here, we present new analyses that alter this
picture. We show that greater communicative precision for warm than for cool
colors, and greater communicative need, may both be explained by perceptual
structure. However, using an information-theoretic analysis, we also show that
color naming across languages bears signs of communicative need beyond what
would be predicted by perceptual structure alone. We conclude that color naming
is shaped both by perceptual structure, as has traditionally been argued, and
by patterns of communicative need, as argued by Gibson et al. - although for
reasons other than those they advanced
A Rate–Distortion view of human pragmatic reasoning
What computational principles underlie human pragmatic reasoning? A prominent approach to pragmatics is the Rational Speech Act (RSA) framework, which formulates pragmatic reasoning as probabilistic speakers and listeners recursively reasoning about each other. While RSA enjoys broad empirical support, it is not yet clear whether the dynamics of such recursive reasoning may be governed by a general optimization principle. Here, we present a novel analysis of the RSA framework that addresses this question. First, we show that RSA recursion implements an alternating maximization for optimizing a tradeoff between expected utility and communicative effort. On that basis, we study the dynamics of RSA recursion and disconfirm the conjecture that expected utility is guaranteed to improve with recursion depth. Second, we show that RSA can be grounded in Rate-Distortion theory, while maintaining a similar ability to account for human behavior and avoiding a bias of RSA toward random utterance production. This work furthers the mathematical understanding of RSA models, and suggests that general information-theoretic principles may give rise to human pragmatic reasoning
Semantic Categories of Artifacts and Animals Reflect Efficient Coding
It has been argued that semantic categories across languages reflect pressure
for efficient communication. Recently, this idea has been cast in terms of a
general information-theoretic principle of efficiency, the Information
Bottleneck (IB) principle, and it has been shown that this principle accounts
for the emergence and evolution of named color categories across languages,
including soft structure and patterns of inconsistent naming. However, it is
not yet clear to what extent this account generalizes to semantic domains other
than color. Here we show that it generalizes to two qualitatively different
semantic domains: names for containers, and for animals. First, we show that
container naming in Dutch and French is near-optimal in the IB sense, and that
IB broadly accounts for soft categories and inconsistent naming patterns in
both languages. Second, we show that a hierarchy of animal categories derived
from IB captures cross-linguistic tendencies in the growth of animal
taxonomies. Taken together, these findings suggest that fundamental
information-theoretic principles of efficient coding may shape semantic
categories across languages and across domains.Comment: To appear in the proceedings of the 41st Annual Conference of the
Cognitive Science Society (CogSci 2019
The forms and meanings of grammatical markers support efficient communication
Functionalist accounts of language suggest that forms are paired with meanings in ways that support efficient communication. Previous work on grammatical marking suggests that word forms have lengths that enable efficient production, and work on the semantic typology of the lexicon suggests that word meanings represent efficient partitions of semantic space. Here we establish a theoretical link between these two lines of work and present an information-theoretic analysis that captures how communicative pressures influence both form and meaning. We apply our approach to the grammatical features of number, tense, and evidentiality and show that the approach explains both which systems of feature values are attested across languages and the relative lengths of the forms for those feature values. Our approach shows that general information-theoretic principles can capture variation in both form and meaning across languages
Human-Guided Complexity-Controlled Abstractions
Neural networks often learn task-specific latent representations that fail to
generalize to novel settings or tasks. Conversely, humans learn discrete
representations (i.e., concepts or words) at a variety of abstraction levels
(e.g., "bird" vs. "sparrow") and deploy the appropriate abstraction based on
task. Inspired by this, we train neural models to generate a spectrum of
discrete representations, and control the complexity of the representations
(roughly, how many bits are allocated for encoding inputs) by tuning the
entropy of the distribution over representations. In finetuning experiments,
using only a small number of labeled examples for a new task, we show that (1)
tuning the representation to a task-appropriate complexity level supports the
highest finetuning performance, and (2) in a human-participant study, users
were able to identify the appropriate complexity level for a downstream task
using visualizations of discrete representations. Our results indicate a
promising direction for rapid model finetuning by leveraging human insight.Comment: NeurIPS 202
Recommended from our members
Optimal compression in human concept learning
The computational principles that underlie human concept learning have been debated in the literature for decades. Here, we formalize and test a new perspective that is grounded in rate-distortion theory (RDT), the mathematical theory of optimal (lossy) data compression, which has recently been gaining increasing popularity in cognitive science. More specifically, we characterize optimal conceptual systems as solutions to a special type of RDT problem, show how these optimal systems can generalize to unseen examples, and test their predictions for human behavior in three foundational concept-learning experiments. We find converging evidence that optimal compression may account for human concept learning. Our work also lends new insight into the relation between learnability and compressibility; integrates prototype, exemplar, and Bayesian approaches to human concepts within the RDT framework; and offers a potential theoretical link between concept learning and other cognitive functions that have been successfully characterized by efficient compression
Recommended from our members
Language use is only sparsely compositional: The case of English adjective-noun phrases in humans and large language models
Compositionality is considered a key hallmark of human
language. However, most research focuses on item-level compositionality, e.g.,
to what extent the meanings of phrases are composed of the meanings of their
sub-parts, rather than on language-level compositionality, which is the degree
to which possible combinations are utilized in practice during language use.
Here, we propose a novel way to quantify the degree of language-level
compositionality and apply it in the case of English adjective-noun
combinations. Using corpus analyses, large language models, and human
acceptability ratings, we find that (1) English only sparsely utilizes the
compositional potential of adjective–noun combinations; and (2) LLMs struggle to
predict human acceptability judgments of rare combinations. Taken together, our
findings shed new light on the role of compositionality in language and
highlight a challenging area for further improving LLMs