334 research outputs found
MoDS: Model-oriented Data Selection for Instruction Tuning
Instruction tuning has become the de facto method to equip large language
models (LLMs) with the ability of following user instructions. Usually,
hundreds of thousands or millions of instruction-following pairs are employed
to fine-tune the foundation LLMs. Recently, some studies show that a small
number of high-quality instruction data is enough. However, how to select
appropriate instruction data for a given LLM is still an open problem. To
address this problem, in this paper we present a model-oriented data selection
(MoDS) approach, which selects instruction data based on a new criteria
considering three aspects: quality, coverage and necessity. First, our approach
utilizes a quality evaluation model to filter out the high-quality subset from
the original instruction dataset, and then designs an algorithm to further
select from the high-quality subset a seed instruction dataset with good
coverage. The seed dataset is applied to fine-tune the foundation LLM to obtain
an initial instruction-following LLM. Finally, we develop a necessity
evaluation model to find out the instruction data which are performed badly in
the initial instruction-following LLM and consider them necessary instructions
to further improve the LLMs. In this way, we can get a small high-quality,
broad-coverage and high-necessity subset from the original instruction
datasets. Experimental results show that, the model fine-tuned with 4,000
instruction pairs selected by our approach could perform better than the model
fine-tuned with the full original dataset which includes 214k instruction data
DualSMC: Tunneling Differentiable Filtering and Planning under Continuous POMDPs
A major difficulty of solving continuous POMDPs is to infer the multi-modal
distribution of the unobserved true states and to make the planning algorithm
dependent on the perceived uncertainty. We cast POMDP filtering and planning
problems as two closely related Sequential Monte Carlo (SMC) processes, one
over the real states and the other over the future optimal trajectories, and
combine the merits of these two parts in a new model named the DualSMC network.
In particular, we first introduce an adversarial particle filter that leverages
the adversarial relationship between its internal components. Based on the
filtering results, we then propose a planning algorithm that extends the
previous SMC planning approach [Piche et al., 2018] to continuous POMDPs with
an uncertainty-dependent policy. Crucially, not only can DualSMC handle complex
observations such as image input but also it remains highly interpretable. It
is shown to be effective in three continuous POMDP domains: the floor
positioning domain, the 3D light-dark navigation domain, and a modified Reacher
domain.Comment: IJCAI 202
Unlink to Unlearn: Simplifying Edge Unlearning in GNNs
As concerns over data privacy intensify, unlearning in Graph Neural Networks
(GNNs) has emerged as a prominent research frontier in academia. This concept
is pivotal in enforcing the \textit{right to be forgotten}, which entails the
selective removal of specific data from trained GNNs upon user request. Our
research focuses on edge unlearning, a process of particular relevance to
real-world applications. Current state-of-the-art approaches like GNNDelete can
eliminate the influence of specific edges yet suffer from
\textit{over-forgetting}, which means the unlearning process inadvertently
removes excessive information beyond needed, leading to a significant
performance decline for remaining edges. Our analysis identifies the loss
functions of GNNDelete as the primary source of over-forgetting and also
suggests that loss functions may be redundant for effective edge unlearning.
Building on these insights, we simplify GNNDelete to develop \textbf{Unlink to
Unlearn} (UtU), a novel method that facilitates unlearning exclusively through
unlinking the forget edges from graph structure. Our extensive experiments
demonstrate that UtU delivers privacy protection on par with that of a
retrained model while preserving high accuracy in downstream tasks, by
upholding over 97.3\% of the retrained model's privacy protection capabilities
and 99.8\% of its link prediction accuracy. Meanwhile, UtU requires only
constant computational demands, underscoring its advantage as a highly
lightweight and practical edge unlearning solution.Comment: Accepted by WWW 2024 as a Short Research Pape
An intelligent video fire detection approach based on object detection technology
PresentationFire that is one of the most serious accidents in chemical factories, may lead to considerable product losses, equipment damages and casualties. With the rapid development of computer vision technology, intelligent fire detection has been proposed and applied in various scenarios. This paper presents a new intelligent video fire detection approach based on object detection technology using convolutional neural networks (CNN). First, a CNN model is trained for the fire detection task which is framed as a regression problem to predict bounding boxes and associated probabilities. In the application phase, videos from surveillance cameras are detected frame by frame. Once fire appears in the current frame, the model will output the coordinates of the fire region. Simultaneously, the frame where the fire region is localized will be immediately sent to safety supervisors as a fire alarm. This will help detect fire at the early stage, prevent fire spreading and improve the emergency response
Constraining interacting dark energy models with the halo concentration - mass relation
The interacting dark energy (IDE) model is a promising alternative
cosmological model which has the potential to solve the fine-tuning and
coincidence problems by considering the interaction between dark matter and
dark energy. Previous studies have shown that the energy exchange between the
dark sectors in this model can significantly affect the dark matter halo
properties. In this study, utilising a large set of cosmological -body
simulations, we analyse the redshift evolution of the halo concentration - mass
( - ) relation in the IDE model, and show that the - relation is
a sensitive proxy of the interaction strength parameter , especially at
lower redshifts. Furthermore, we construct parametrized formulae to quantify
the dependence of the - relation on at redshifts ranging from
to . Our parametrized formulae provide a useful tool in constraining
with the observational - relation. As a first attempt, we use
the data from X-ray, gravitational lensing, and galaxy rotational curve
observations and obtain a tight constraint on , i.e. . Our work demonstrates that the halo - relation, which reflects
the halo assembly history, is a powerful probe to constrain the IDE model.Comment: 9 pages, 5 figures, 5 table
- …