145 research outputs found
Multi-level Adversarial Spatio-temporal Learning for Footstep Pressure based FoG Detection
Freezing of gait (FoG) is one of the most common symptoms of Parkinson's
disease, which is a neurodegenerative disorder of the central nervous system
impacting millions of people around the world. To address the pressing need to
improve the quality of treatment for FoG, devising a computer-aided detection
and quantification tool for FoG has been increasingly important. As a
non-invasive technique for collecting motion patterns, the footstep pressure
sequences obtained from pressure sensitive gait mats provide a great
opportunity for evaluating FoG in the clinic and potentially in the home
environment. In this study, FoG detection is formulated as a sequential
modelling task and a novel deep learning architecture, namely Adversarial
Spatio-temporal Network (ASTN), is proposed to learn FoG patterns across
multiple levels. A novel adversarial training scheme is introduced with a
multi-level subject discriminator to obtain subject-independent FoG
representations, which helps to reduce the over-fitting risk due to the high
inter-subject variance. As a result, robust FoG detection can be achieved for
unseen subjects. The proposed scheme also sheds light on improving
subject-level clinical studies from other scenarios as it can be integrated
with many existing deep architectures. To the best of our knowledge, this is
one of the first studies of footstep pressure-based FoG detection and the
approach of utilizing ASTN is the first deep neural network architecture in
pursuit of subject-independent representations. Experimental results on 393
trials collected from 21 subjects demonstrate encouraging performance of the
proposed ASTN for FoG detection with an AUC 0.85
ANPL: Compiling Natural Programs with Interactive Decomposition
The advents of Large Language Models (LLMs) have shown promise in augmenting
programming using natural interactions. However, while LLMs are proficient in
compiling common usage patterns into a programming language, e.g., Python, it
remains a challenge how to edit and debug an LLM-generated program. We
introduce ANPL, a programming system that allows users to decompose
user-specific tasks. In an ANPL program, a user can directly manipulate sketch,
which specifies the data flow of the generated program. The user annotates the
modules, or hole with natural language descriptions offloading the expensive
task of generating functionalities to the LLM. Given an ANPL program, the ANPL
compiler generates a cohesive Python program that implements the
functionalities in hole, while respecting the dataflows specified in sketch. We
deploy ANPL on the Abstraction and Reasoning Corpus (ARC), a set of unique
tasks that are challenging for state-of-the-art AI systems, showing it
outperforms baseline programming systems that (a) without the ability to
decompose tasks interactively and (b) without the guarantee that the modules
can be correctly composed together. We obtain a dataset consisting of 300/400
ARC tasks that were successfully decomposed and grounded in Python, providing
valuable insights into how humans decompose programmatic tasks. See the dataset
at https://iprc-dip.github.io/DARC
Self-driven Grounding: Large Language Model Agents with Automatical Language-aligned Skill Learning
Large language models (LLMs) show their powerful automatic reasoning and
planning capability with a wealth of semantic knowledge about the human world.
However, the grounding problem still hinders the applications of LLMs in the
real-world environment. Existing studies try to fine-tune the LLM or utilize
pre-defined behavior APIs to bridge the LLMs and the environment, which not
only costs huge human efforts to customize for every single task but also
weakens the generality strengths of LLMs. To autonomously ground the LLM onto
the environment, we proposed the Self-Driven Grounding (SDG) framework to
automatically and progressively ground the LLM with self-driven skill learning.
SDG first employs the LLM to propose the hypothesis of sub-goals to achieve
tasks and then verify the feasibility of the hypothesis via interacting with
the underlying environment. Once verified, SDG can then learn generalized
skills with the guidance of these successfully grounded subgoals. These skills
can be further utilized to accomplish more complex tasks which fail to pass the
verification phase. Verified in the famous instruction following task
set-BabyAI, SDG achieves comparable performance in the most challenging tasks
compared with imitation learning methods that cost millions of demonstrations,
proving the effectiveness of learned skills and showing the feasibility and
efficiency of our framework
Assessing and Understanding Creativity in Large Language Models
In the field of natural language processing, the rapid development of large
language model (LLM) has attracted more and more attention. LLMs have shown a
high level of creativity in various tasks, but the methods for assessing such
creativity are inadequate. The assessment of LLM creativity needs to consider
differences from humans, requiring multi-dimensional measurement while
balancing accuracy and efficiency. This paper aims to establish an efficient
framework for assessing the level of creativity in LLMs. By adapting the
modified Torrance Tests of Creative Thinking, the research evaluates the
creative performance of various LLMs across 7 tasks, emphasizing 4 criteria
including Fluency, Flexibility, Originality, and Elaboration. In this context,
we develop a comprehensive dataset of 700 questions for testing and an
LLM-based evaluation method. In addition, this study presents a novel analysis
of LLMs' responses to diverse prompts and role-play situations. We found that
the creativity of LLMs primarily falls short in originality, while excelling in
elaboration. Besides, the use of prompts and the role-play settings of the
model significantly influence creativity. Additionally, the experimental
results also indicate that collaboration among multiple LLMs can enhance
originality. Notably, our findings reveal a consensus between human evaluations
and LLMs regarding the personality traits that influence creativity. The
findings underscore the significant impact of LLM design on creativity and
bridges artificial intelligence and human creativity, offering insights into
LLMs' creativity and potential applications
Online Prototype Alignment for Few-shot Policy Transfer
Domain adaptation in reinforcement learning (RL) mainly deals with the
changes of observation when transferring the policy to a new environment. Many
traditional approaches of domain adaptation in RL manage to learn a mapping
function between the source and target domain in explicit or implicit ways.
However, they typically require access to abundant data from the target domain.
Besides, they often rely on visual clues to learn the mapping function and may
fail when the source domain looks quite different from the target domain. To
address these problems, we propose a novel framework Online Prototype Alignment
(OPA) to learn the mapping function based on the functional similarity of
elements and is able to achieve the few-shot policy transfer within only
several episodes. The key insight of OPA is to introduce an exploration
mechanism that can interact with the unseen elements of the target domain in an
efficient and purposeful manner, and then connect them with the seen elements
in the source domain according to their functionalities (instead of visual
clues). Experimental results show that when the target domain looks visually
different from the source domain, OPA can achieve better transfer performance
even with much fewer samples from the target domain, outperforming prior
methods.Comment: This paper has been accepted at ICML202
Contrastive Modules with Temporal Attention for Multi-Task Reinforcement Learning
In the field of multi-task reinforcement learning, the modular principle,
which involves specializing functionalities into different modules and
combining them appropriately, has been widely adopted as a promising approach
to prevent the negative transfer problem that performance degradation due to
conflicts between tasks. However, most of the existing multi-task RL methods
only combine shared modules at the task level, ignoring that there may be
conflicts within the task. In addition, these methods do not take into account
that without constraints, some modules may learn similar functions, resulting
in restricting the model's expressiveness and generalization capability of
modular methods. In this paper, we propose the Contrastive Modules with
Temporal Attention(CMTA) method to address these limitations. CMTA constrains
the modules to be different from each other by contrastive learning and
combining shared modules at a finer granularity than the task level with
temporal attention, alleviating the negative transfer within the task and
improving the generalization ability and the performance for multi-task RL. We
conducted the experiment on Meta-World, a multi-task RL benchmark containing
various robotics manipulation tasks. Experimental results show that CMTA
outperforms learning each task individually for the first time and achieves
substantial performance improvements over the baselines.Comment: This paper has been accepted at NeurIPS 2023 as a poste
- …