176 research outputs found
A Novel Adaptive Elite-Based Particle Swarm Optimization Applied to VAR Optimization in Electric Power Systems
Particle swarm optimization (PSO) has been successfully applied to solve many practical engineering problems. However, more efficient strategies are needed to coordinate global and local searches in the solution space when the studied problem is extremely nonlinear and highly dimensional. This work proposes a novel adaptive elite-based PSO approach. The adaptive elite strategies involve the following two tasks: (1) appending the mean search to the original approach and (2) pruning/cloning particles. The mean search, leading to stable convergence, helps the iterative process coordinate between the global and local searches. The mean of the particles and standard deviation of the distances between pairs of particles are utilized to prune distant particles. The best particle is cloned and it replaces the pruned distant particles in the elite strategy. To evaluate the performance and generality of the proposed method, four benchmark functions were tested by traditional PSO, chaotic PSO, differential evolution, and genetic algorithm. Finally, a realistic loss minimization problem in an electric power system is studied to show the robustness of the proposed method
Cascaded Local Implicit Transformer for Arbitrary-Scale Super-Resolution
Implicit neural representation has recently shown a promising ability in
representing images with arbitrary resolutions. In this paper, we present a
Local Implicit Transformer (LIT), which integrates the attention mechanism and
frequency encoding technique into a local implicit image function. We design a
cross-scale local attention block to effectively aggregate local features. To
further improve representative power, we propose a Cascaded LIT (CLIT) that
exploits multi-scale features, along with a cumulative training strategy that
gradually increases the upsampling scales during training. We have conducted
extensive experiments to validate the effectiveness of these components and
analyze various training strategies. The qualitative and quantitative results
demonstrate that LIT and CLIT achieve favorable results and outperform the
prior works in arbitrary super-resolution tasks
MuRAL: Multi-Scale Region-based Active Learning for Object Detection
Obtaining large-scale labeled object detection dataset can be costly and
time-consuming, as it involves annotating images with bounding boxes and class
labels. Thus, some specialized active learning methods have been proposed to
reduce the cost by selecting either coarse-grained samples or fine-grained
instances from unlabeled data for labeling. However, the former approaches
suffer from redundant labeling, while the latter methods generally lead to
training instability and sampling bias. To address these challenges, we propose
a novel approach called Multi-scale Region-based Active Learning (MuRAL) for
object detection. MuRAL identifies informative regions of various scales to
reduce annotation costs for well-learned objects and improve training
performance. The informative region score is designed to consider both the
predicted confidence of instances and the distribution of each object category,
enabling our method to focus more on difficult-to-detect classes. Moreover,
MuRAL employs a scale-aware selection strategy that ensures diverse regions are
selected from different scales for labeling and downstream finetuning, which
enhances training stability. Our proposed method surpasses all existing
coarse-grained and fine-grained baselines on Cityscapes and MS COCO datasets,
and demonstrates significant improvement in difficult category performance
Shilling Black-box Review-based Recommender Systems through Fake Review Generation
Review-Based Recommender Systems (RBRS) have attracted increasing research
interest due to their ability to alleviate well-known cold-start problems. RBRS
utilizes reviews to construct the user and items representations. However, in
this paper, we argue that such a reliance on reviews may instead expose systems
to the risk of being shilled. To explore this possibility, in this paper, we
propose the first generation-based model for shilling attacks against RBRSs.
Specifically, we learn a fake review generator through reinforcement learning,
which maliciously promotes items by forcing prediction shifts after adding
generated reviews to the system. By introducing the auxiliary rewards to
increase text fluency and diversity with the aid of pre-trained language models
and aspect predictors, the generated reviews can be effective for shilling with
high fidelity. Experimental results demonstrate that the proposed framework can
successfully attack three different kinds of RBRSs on the Amazon corpus with
three domains and Yelp corpus. Furthermore, human studies also show that the
generated reviews are fluent and informative. Finally, equipped with Attack
Review Generators (ARGs), RBRSs with adversarial training are much more robust
to malicious reviews
SINC: Self-Supervised In-Context Learning for Vision-Language Tasks
Large Pre-trained Transformers exhibit an intriguing capacity for in-context
learning. Without gradient updates, these models can rapidly construct new
predictors from demonstrations presented in the inputs. Recent works promote
this ability in the vision-language domain by incorporating visual information
into large language models that can already make in-context predictions.
However, these methods could inherit issues in the language domain, such as
template sensitivity and hallucination. Also, the scale of these language
models raises a significant demand for computations, making learning and
operating these models resource-intensive. To this end, we raise a question:
``How can we enable in-context learning without relying on the intrinsic
in-context ability of large language models?". To answer it, we propose a
succinct and general framework, Self-supervised IN-Context learning (SINC),
that introduces a meta-model to learn on self-supervised prompts consisting of
tailored demonstrations. The learned models can be transferred to downstream
tasks for making in-context predictions on-the-fly. Extensive experiments show
that SINC outperforms gradient-based methods in various vision-language tasks
under few-shot settings. Furthermore, the designs of SINC help us investigate
the benefits of in-context learning across different tasks, and the analysis
further reveals the essential components for the emergence of in-context
learning in the vision-language domain.Comment: Accepted by ICCV 2023; Camera Ready Versio
Two highly similar DEAD box proteins, OsRH2 and OsRH34, homologous to eukaryotic initiation factor 4AIII, play roles of the exon junction complex in regulating growth and development in rice
Accession numbers and proteins homologous to eIF4A. (DOCX 18Â kb
Autonomic Dysfunction Because of Severe Tetanus in an Unvaccinated Child
Tetanus is rare in a country with a national vaccination program. When it does occur, the associated autonomic dysfunction is a challenge for physicians. We report here a case of an unvaccinated 5-year-old boy who suffered from tetanus complicated by autonomic dysfunction, which was successfully controlled by the infusion of magnesium sulfate. This is the first case that demonstrated the therapeutic effect of magnesium sulfate in a child with tetanus. This case highlights the importance of implementing a vaccination program
- …