23 research outputs found
From simple to complex categories: how structure and label information guides the acquisition of category knowledge
Categorization is a fundamental ability of human cognition, translating complex streams of information
from the all of different senses into simpler, discrete categories. How do people acquire all of
this category knowledge, particularly the kinds of rich, structured categories we interact with every
day in the real-world? In this thesis, I explore how information from category structure and category
labels influence how people learn categories, particular for the kinds of computational problems
that are relevant to real-world category learning. The three learning problems this thesis covers are:
semi-supervised learning, structure learning and category learning with many features. Each of these
three learning problems presents a different kinds of learning challenge, and through a combination
of behavioural experiments and computational modeling, this thesis illustrates how the interplay between
structure and label information can explain how humans can acquire richer kinds of category
knowledge.Thesis (Ph.D.) (Research by Publication) -- University of Adelaide, School of Psychology, 201
Learning word-referent mappings and concepts from raw inputs
How do children learn correspondences between the language and the world from
noisy, ambiguous, naturalistic input? One hypothesis is via cross-situational
learning: tracking words and their possible referents across multiple
situations allows learners to disambiguate correct word-referent mappings (Yu &
Smith, 2007). However, previous models of cross-situational word learning
operate on highly simplified representations, side-stepping two important
aspects of the actual learning problem. First, how can word-referent mappings
be learned from raw inputs such as images? Second, how can these learned
mappings generalize to novel instances of a known word? In this paper, we
present a neural network model trained from scratch via self-supervision that
takes in raw images and words as inputs, and show that it can learn
word-referent mappings from fully ambiguous scenes and utterances through
cross-situational learning. In addition, the model generalizes to novel word
instances, locates referents of words in a scene, and shows a preference for
mutual exclusivity
Recommended from our members
Do additional features help or harm during category learning?An exploration of the curse of dimensionality in human learners
How does the number of features impact category learning?One view suggests that additional features creates a “curse ofdimensionality” - where having more features causes the sizeof the search space to grow so quickly that discovering goodclassification rules becomes increasingly challenging. The op-posing view suggests that additional features provide a wealthof additional information which learners should be able to useto improve their classification performance. Previous researchexploring this issue appears to have produced conflicting re-sults: some find that learning improves with additional features(Hoffman & Murphy, 2006) while others find that it does not(Minda & Smith, 2001; Edgell et al., 1996). Here we inves-tigate the possibility that category structure may explain thisapparent discrepancy – that more features are useful in cate-gories with family resemblance structure, but are not (and mayeven be harmful) in more rule-based categories. We find whilethe impact of having many features does indeed depend on cat-egory structure, the results can be explained by a single unifiedmodel: one that attends to a single feature on any given trialand uses information learned from that particular feature tomake classification judgments
Fast and flexible: Human program induction in abstract reasoning tasks
The Abstraction and Reasoning Corpus (ARC) is a challenging program induction
dataset that was recently proposed by Chollet (2019). Here, we report the first
set of results collected from a behavioral study of humans solving a subset of
tasks from ARC (40 out of 1000). Although this subset of tasks contains
considerable variation, our results showed that humans were able to infer the
underlying program and generate the correct test output for a novel test input
example, with an average of 80% of tasks solved per participant, and with 65%
of tasks being solved by more than 80% of participants. Additionally, we find
interesting patterns of behavioral consistency and variability within the
action sequences during the generation process, the natural language
descriptions to describe the transformations for each task, and the errors
people made. Our findings suggest that people can quickly and reliably
determine the relevant features and properties of a task to compose a correct
solution. Future modeling work could incorporate these findings, potentially by
connecting the natural language descriptions we collected here to the
underlying semantics of ARC.Comment: 7 pages, 7 figures, 1 tabl
Learning time-varying categories
Many kinds of objects and events in our world have a strongly time-dependent quality. However, most theories about concepts and categories either are insensitive to variation over time or treat it as a nuisance factor that produces irrational order effects during learning. In this article, we present two category learning experiments in which we explored peoples’ ability to learn categories whose structure is strongly time-dependent. We suggest that order effects in categorization may in part reflect a sensitivity to changing environments, and that understanding dynamically changing concepts is an important part of developing a full account of human categorization.Daniel J. Navarro, Amy Perfors, Wai Keen Von
Evolutionary Computation, Optimization and Learning Algorithms for Data Science
A large number of engineering, science and computational problems have yet to be solved in a computationally efficient way. One of the emerging challenges is how evolving technologies grow towards autonomy and intelligent decision making. This leads to collection of large amounts of data from various sensing and measurement technologies, e.g., cameras, smart phones, health sensors, smart electricity meters, and environment sensors. Hence, it is imperative to develop efficient algorithms for generation, analysis, classification, and illustration of data. Meanwhile, data is structured purposefully through different representations, such as large-scale networks and graphs. We focus on data science as a crucial area, specifically focusing on a curse of dimensionality (CoD) which is due to the large amount of generated/sensed/collected data. This motivates researchers to think about optimization and to apply nature-inspired algorithms, such as evolutionary algorithms (EAs) to solve optimization problems. Although these algorithms look un-deterministic, they are robust enough to reach an optimal solution. Researchers do not adopt evolutionary algorithms unless they face a problem which is suffering from placement in local optimal solution, rather than global optimal solution. In this chapter, we first develop a clear and formal definition of the CoD problem, next we focus on feature extraction techniques and categories, then we provide a general overview of meta-heuristic algorithms, its terminology, and desirable properties of evolutionary algorithms
Recommended from our members
Categorising images by generating natural language rules
The ability to generate rules and hypotheses plays a key role in multiple aspects of human cognition including concept learning and explanation. Previous research has framed this ability as a form of inference via probabilistic program induction. However, this modeling approach often requires careful construction of the right grammar and hypothesis space for a particular task, and cannot easily be transferred to other domains. In this work, we present an alternative computational account of rule generation, leveraging advances in multimodal learning and large language models. Taking naturalistic images as input, our computational model is capable of generating candidate rules that are specified in natural language, and verifying them to determine their fit to the data. We show that our model can generate, in a zero-shot manner, plausible rules for visual concepts across multiple domains
Recommended from our members
Learning word-referent mappings and concepts from raw inputs
How do children learn correspondences between the language and the world from noisy, ambiguous, naturalistic input?One hypothesis is via cross-situational learning: tracking words and their possible referents across multiple situationsallows learners to disambiguate correct word-referent mappings (Yu and Smith, 2007). While previous models of cross-situational word learning operate on highly simplified representations, recent advances in multimodal learning have shownpromise as richer models of cross-situational word learning to enable learning the meanings of words from raw inputs.Here, we present a neural network model of cross-situational word learning that leverages some of these ideas and examineits ability to account for a variety of empirical phenomena from the word learning literature