213 research outputs found
Prompt Stealing Attacks Against Large Language Models
The increasing reliance on large language models (LLMs) such as ChatGPT in
various fields emphasizes the importance of ``prompt engineering,'' a
technology to improve the quality of model outputs. With companies investing
significantly in expert prompt engineers and educational resources rising to
meet market demand, designing high-quality prompts has become an intriguing
challenge. In this paper, we propose a novel attack against LLMs, named prompt
stealing attacks. Our proposed prompt stealing attack aims to steal these
well-designed prompts based on the generated answers. The prompt stealing
attack contains two primary modules: the parameter extractor and the prompt
reconstruction. The goal of the parameter extractor is to figure out the
properties of the original prompts. We first observe that most prompts fall
into one of three categories: direct prompt, role-based prompt, and in-context
prompt. Our parameter extractor first tries to distinguish the type of prompts
based on the generated answers. Then, it can further predict which role or how
many contexts are used based on the types of prompts. Following the parameter
extractor, the prompt reconstructor can be used to reconstruct the original
prompts based on the generated answers and the extracted features. The final
goal of the prompt reconstructor is to generate the reversed prompts, which are
similar to the original prompts. Our experimental results show the remarkable
performance of our proposed attacks. Our proposed attacks add a new dimension
to the study of prompt engineering and call for more attention to the security
issues on LLMs
Polymer Derived Ceramic for Lithium-ion Storage, and Electrospun Polyelectrolyte Fiber for Heavy Metal Ions Removal
This dissertation includes two major projects. The first project investigated the great potential of polymer-derived ceramics (PDCs) as lithium-ion battery anode materials with good cycling stability and large capacity. SiCNO ceramic nanoparticles were produced by pyrolysis of polysilazane nanoparticles synthesized via an oil-in-oil emulsion crosslinking. The SiCNO nanoparticles had an average particle size of around 9 nm and contained graphitic carbon, Si3N4, and SiO2 domains. The electrochemical behavior of SiCNO nanoparticles anode was investigated to evaluate the Li-ion storage performance and understand its mechanism of Li-ion storage. The lithiation of SiCNO was observed at ~0.385 V versus Li/Li+. The anode had a large capacity of 705 mAh g-1 after 350 cycles with a current density of 0.1 A g-1. Moreover, it showed excellent cyclic stability with a capacity decay of 0.049 mAh g-1 (0.0097%) per cycle. In situ TEM analysis demonstrated that the SiCNO nanoparticles exhibit extraordinary structural stability with only 9.36% linear expansion in the lithiation process. The second project investigated the removal of heavy metals ions from wastewater using electrospun polyelectrolyte fibers of polyacrylic acid (PAA) and polyallylamine hydrochloride (PAH). Polyelectrolyte fiber mats were fabricated by electrospinning followed by thermal crosslinking. The fiber mats were evaluated for their efficiency in removing heavy metals in synthetic metal solutions. 70 %, 98 %, and 92 % removals of Pb2+, Cd2+, and Cu2+, respectively, were observed at pH 7.4. Metal ion-carboxylate complexations were studied by FT-IR spectra, which indicate carboxylate groups from PAA play important role in heavy metal ion removal
NAS-Bench-Graph: Benchmarking Graph Neural Architecture Search
Graph neural architecture search (GraphNAS) has recently aroused considerable
attention in both academia and industry. However, two key challenges seriously
hinder the further research of GraphNAS. First, since there is no consensus for
the experimental setting, the empirical results in different research papers
are often not comparable and even not reproducible, leading to unfair
comparisons. Secondly, GraphNAS often needs extensive computations, which makes
it highly inefficient and inaccessible to researchers without access to
large-scale computation. To solve these challenges, we propose NAS-Bench-Graph,
a tailored benchmark that supports unified, reproducible, and efficient
evaluations for GraphNAS. Specifically, we construct a unified, expressive yet
compact search space, covering 26,206 unique graph neural network (GNN)
architectures and propose a principled evaluation protocol. To avoid
unnecessary repetitive training, we have trained and evaluated all of these
architectures on nine representative graph datasets, recording detailed metrics
including train, validation, and test performance in each epoch, the latency,
the number of parameters, etc. Based on our proposed benchmark, the performance
of GNN architectures can be directly obtained by a look-up table without any
further computation, which enables fair, fully reproducible, and efficient
comparisons. To demonstrate its usage, we make in-depth analyses of our
proposed NAS-Bench-Graph, revealing several interesting findings for GraphNAS.
We also showcase how the benchmark can be easily compatible with GraphNAS open
libraries such as AutoGL and NNI. To the best of our knowledge, our work is the
first benchmark for graph neural architecture search
Graph Meets LLMs: Towards Large Graph Models
Large models have emerged as the most recent groundbreaking achievements in
artificial intelligence, and particularly machine learning. However, when it
comes to graphs, large models have not achieved the same level of success as in
other fields, such as natural language processing and computer vision. In order
to promote applying large models for graphs forward, we present a perspective
paper to discuss the challenges and opportunities associated with developing
large graph models. First, we discuss the desired characteristics of large
graph models. Then, we present detailed discussions from three key
perspectives: representation basis, graph data, and graph models. In each
category, we provide a brief overview of recent advances and highlight the
remaining challenges together with our visions. Finally, we discuss valuable
applications of large graph models. We believe this perspective can encourage
further investigations into large graph models, ultimately pushing us one step
closer towards artificial general intelligence (AGI). We are the first to
comprehensively study large graph models, to the best of our knowledge.Comment: Accepted by NeurIPS 2023 New Frontiers in Graph Learning Workshop.
Comments are welcom
BusReF: Infrared-Visible images registration and fusion focus on reconstructible area using one set of features
In a scenario where multi-modal cameras are operating together, the problem
of working with non-aligned images cannot be avoided. Yet, existing image
fusion algorithms rely heavily on strictly registered input image pairs to
produce more precise fusion results, as a way to improve the performance of
downstream high-level vision tasks. In order to relax this assumption, one can
attempt to register images first. However, the existing methods for registering
multiple modalities have limitations, such as complex structures and reliance
on significant semantic information. This paper aims to address the problem of
image registration and fusion in a single framework, called BusRef. We focus on
Infrared-Visible image registration and fusion task (IVRF). In this framework,
the input unaligned image pairs will pass through three stages: Coarse
registration, Fine registration and Fusion. It will be shown that the unified
approach enables more robust IVRF. We also propose a novel training and
evaluation strategy, involving the use of masks to reduce the influence of
non-reconstructible regions on the loss functions, which greatly improves the
accuracy and robustness of the fusion task. Last but not least, a
gradient-aware fusion network is designed to preserve the complementary
information. The advanced performance of this algorithm is demonstrated b
Quantum super-resolution for imaging two pointlike entangled photon sources
We investigate the resolution for imaging two pointlike entangled sources by
using the method of the moments and the spatial-mode demultiplexing (SPADE),
where the pointlike entangled sources can be generated by injecting single-mode
sources with arbitrary quantum statistics distribution into an optical
parametric amplifier (OPA). We demonstrate that the separation estimation
sensitivity is mainly determined by the photon distribution in each detected
modes and it can be enhanced by either increasing the squeezed parameter of the
OPA or eliminating the relative phase difference of the entangle sources.
Furthermore, in the limiting case of infinitely small source separation, the
usage of entangled sources can have better resolution than those using
incoherent and coherent sources. The results here can find important
applications for the quantum super-resolution imaging and quantum metrology
LLM4DyG: Can Large Language Models Solve Problems on Dynamic Graphs?
In an era marked by the increasing adoption of Large Language Models (LLMs)
for various tasks, there is a growing focus on exploring LLMs' capabilities in
handling web data, particularly graph data. Dynamic graphs, which capture
temporal network evolution patterns, are ubiquitous in real-world web data.
Evaluating LLMs' competence in understanding spatial-temporal information on
dynamic graphs is essential for their adoption in web applications, which
remains unexplored in the literature. In this paper, we bridge the gap via
proposing to evaluate LLMs' spatial-temporal understanding abilities on dynamic
graphs, to the best of our knowledge, for the first time. Specifically, we
propose the LLM4DyG benchmark, which includes nine specially designed tasks
considering the capability evaluation of LLMs from both temporal and spatial
dimensions. Then, we conduct extensive experiments to analyze the impacts of
different data generators, data statistics, prompting techniques, and LLMs on
the model performance. Finally, we propose Disentangled Spatial-Temporal
Thoughts (DST2) for LLMs on dynamic graphs to enhance LLMs' spatial-temporal
understanding abilities. Our main observations are: 1) LLMs have preliminary
spatial-temporal understanding abilities on dynamic graphs, 2) Dynamic graph
tasks show increasing difficulties for LLMs as the graph size and density
increase, while not sensitive to the time span and data generation mechanism,
3) the proposed DST2 prompting method can help to improve LLMs'
spatial-temporal understanding abilities on dynamic graphs for most tasks. The
data and codes will be open-sourced at publication time
Throughput of Hybrid UAV Networks with Scale-Free Topology
Unmanned Aerial Vehicles (UAVs) hold great potential to support a wide range
of applications due to the high maneuverability and flexibility. Compared with
single UAV, UAV swarm carries out tasks efficiently in harsh environment, where
the network resilience is of vital importance to UAV swarm. The network
topology has a fundamental impact on the resilience of UAV network. It is
discovered that scale-free network topology, as a topology that exists widely
in nature, has the ability to enhance the network resilience. Besides,
increasing network throughput can enhance the efficiency of information
interaction, improving the network resilience. Facing these facts, this paper
studies the throughput of UAV Network with scale-free topology. Introducing the
hybrid network structure combining both ad hoc transmission mode and cellular
transmission mode into UAV Network, the throughput of UAV Network is improved
compared with that of pure ad hoc UAV network. Furthermore, this work also
investigates the optimal setting of the hop threshold for the selection of ad
hoc or cellular transmission mode. It is discovered that the optimal hop
threshold is related with the number of UAVs and the parameters of scale-free
topology. This paper may motivate the application of hybrid network structure
into UAV Network.Comment: 15 pages, 7 figure
- …