200 research outputs found
Energy Efficient Ant Colony Algorithms for Data Aggregation in Wireless Sensor Networks
In this paper, a family of ant colony algorithms called DAACA for data
aggregation has been presented which contains three phases: the initialization,
packet transmission and operations on pheromones. After initialization, each
node estimates the remaining energy and the amount of pheromones to compute the
probabilities used for dynamically selecting the next hop. After certain rounds
of transmissions, the pheromones adjustment is performed periodically, which
combines the advantages of both global and local pheromones adjustment for
evaporating or depositing pheromones. Four different pheromones adjustment
strategies are designed to achieve the global optimal network lifetime, namely
Basic-DAACA, ES-DAACA, MM-DAACA and ACS-DAACA. Compared with some other data
aggregation algorithms, DAACA shows higher superiority on average degree of
nodes, energy efficiency, prolonging the network lifetime, computation
complexity and success ratio of one hop transmission. At last we analyze the
characteristic of DAACA in the aspects of robustness, fault tolerance and
scalability.Comment: To appear in Journal of Computer and System Science
CLIP-driven Outliers Synthesis for few-shot OOD detection
Few-shot OOD detection focuses on recognizing out-of-distribution (OOD)
images that belong to classes unseen during training, with the use of only a
small number of labeled in-distribution (ID) images. Up to now, a mainstream
strategy is based on large-scale vision-language models, such as CLIP. However,
these methods overlook a crucial issue: the lack of reliable OOD supervision
information, which can lead to biased boundaries between in-distribution (ID)
and OOD. To tackle this problem, we propose CLIP-driven Outliers
Synthesis~(CLIP-OS). Firstly, CLIP-OS enhances patch-level features' perception
by newly proposed patch uniform convolution, and adaptively obtains the
proportion of ID-relevant information by employing CLIP-surgery-discrepancy,
thus achieving separation between ID-relevant and ID-irrelevant. Next, CLIP-OS
synthesizes reliable OOD data by mixing up ID-relevant features from different
classes to provide OOD supervision information. Afterward, CLIP-OS leverages
synthetic OOD samples by unknown-aware prompt learning to enhance the
separability of ID and OOD. Extensive experiments across multiple benchmarks
demonstrate that CLIP-OS achieves superior few-shot OOD detection capability.Comment: 9 pages,5 figure
Test-Time Training for Semantic Segmentation with Output Contrastive Loss
Although deep learning-based segmentation models have achieved impressive
performance on public benchmarks, generalizing well to unseen environments
remains a major challenge. To improve the model's generalization ability to the
new domain during evaluation, the test-time training (TTT) is a challenging
paradigm that adapts the source-pretrained model in an online fashion. Early
efforts on TTT mainly focus on the image classification task. Directly
extending these methods to semantic segmentation easily experiences unstable
adaption due to segmentation's inherent characteristics, such as extreme class
imbalance and complex decision spaces. To stabilize the adaptation process, we
introduce contrastive loss (CL), known for its capability to learn robust and
generalized representations. Nevertheless, the traditional CL operates in the
representation space and cannot directly enhance predictions. In this paper, we
resolve this limitation by adapting the CL to the output space, employing a
high temperature, and simplifying the formulation, resulting in a
straightforward yet effective loss function called Output Contrastive Loss
(OCL). Our comprehensive experiments validate the efficacy of our approach
across diverse evaluation scenarios. Notably, our method excels even when
applied to models initially pre-trained using domain adaptation methods on test
domain data, showcasing its resilience and adaptability.\footnote{Code and more
information could be found at~ \url{https://github.com/dazhangyu123/OCL}
Exploring Unsupervised Cell Recognition with Prior Self-activation Maps
The success of supervised deep learning models on cell recognition tasks
relies on detailed annotations. Many previous works have managed to reduce the
dependency on labels. However, considering the large number of cells contained
in a patch, costly and inefficient labeling is still inevitable. To this end,
we explored label-free methods for cell recognition. Prior self-activation maps
(PSM) are proposed to generate pseudo masks as training targets. To be
specific, an activation network is trained with self-supervised learning. The
gradient information in the shallow layers of the network is aggregated to
generate prior self-activation maps. Afterward, a semantic clustering module is
then introduced as a pipeline to transform PSMs to pixel-level semantic pseudo
masks for downstream tasks. We evaluated our method on two histological
datasets: MoNuSeg (cell segmentation) and BCData (multi-class cell detection).
Compared with other fully-supervised and weakly-supervised methods, our method
can achieve competitive performance without any manual annotations. Our simple
but effective framework can also achieve multi-class cell detection which can
not be done by existing unsupervised methods. The results show the potential of
PSMs that might inspire other research to deal with the hunger for labels in
medical area.Comment: MICCAI 2023. arXiv admin note: substantial text overlap with
arXiv:2210.0786
FedTP: Federated Learning by Transformer Personalization
Federated learning is an emerging learning paradigm where multiple clients
collaboratively train a machine learning model in a privacy-preserving manner.
Personalized federated learning extends this paradigm to overcome heterogeneity
across clients by learning personalized models. Recently, there have been some
initial attempts to apply Transformers to federated learning. However, the
impacts of federated learning algorithms on self-attention have not yet been
studied. This paper investigates this relationship and reveals that federated
averaging algorithms actually have a negative impact on self-attention where
there is data heterogeneity. These impacts limit the capabilities of the
Transformer model in federated learning settings. Based on this, we propose
FedTP, a novel Transformer-based federated learning framework that learns
personalized self-attention for each client while aggregating the other
parameters among the clients. Instead of using a vanilla personalization
mechanism that maintains personalized self-attention layers of each client
locally, we develop a learn-to-personalize mechanism to further encourage the
cooperation among clients and to increase the scablability and generalization
of FedTP. Specifically, the learn-to-personalize is realized by learning a
hypernetwork on the server that outputs the personalized projection matrices of
self-attention layers to generate client-wise queries, keys and values.
Furthermore, we present the generalization bound for FedTP with the
learn-to-personalize mechanism. Notably, FedTP offers a convenient environment
for performing a range of image and language tasks using the same federated
network architecture - all of which benefit from Transformer personalization.
Extensive experiments verify that FedTP with the learn-to-personalize mechanism
yields state-of-the-art performance in non-IID scenarios. Our code is available
online
Extended Wiener-Khinchin theorem for quantum spectral analysis
The classical Wiener-Khinchin theorem (WKT), which can extract spectral
information by classical interferometers through Fourier transform, is a
fundamental theorem used in many disciplines. However, there is still need for
a quantum version of WKT, which could connect correlated biphoton spectral
information by quantum interferometers. Here, we extend the classical WKT to
its quantum counterpart, i.e., extended WKT (e-WKT), which is based on
two-photon quantum interferometry. According to the e-WKT, the
difference-frequency distribution of the biphoton wavefunctions can be
extracted by applying a Fourier transform on the time-domain Hong-Ou-Mandel
interference (HOMI) patterns, while the sum-frequency distribution can be
extracted by applying a Fourier transform on the time-domain NOON state
interference (NOONI) patterns. We also experimentally verified the WKT and
e-WKT in a Mach-Zehnder interference (MZI), a HOMI and a NOONI. This theorem
can be directly applied to quantum spectroscopy, where the spectral correlation
information of biphotons can be obtained from time-domain quantum interferences
by Fourier transform. This may open a new pathway for the study of light-matter
interaction at the single photon level.Comment: 13 pages, 5 figure
Fully Automated Deep Learning-enabled Detection for Hepatic Steatosis on Computed Tomography: A Multicenter International Validation Study
Despite high global prevalence of hepatic steatosis, no automated diagnostics
demonstrated generalizability in detecting steatosis on multiple international
datasets. Traditionally, hepatic steatosis detection relies on clinicians
selecting the region of interest (ROI) on computed tomography (CT) to measure
liver attenuation. ROI selection demands time and expertise, and therefore is
not routinely performed in populations. To automate the process, we validated
an existing artificial intelligence (AI) system for 3D liver segmentation and
used it to purpose a novel method: AI-ROI, which could automatically select the
ROI for attenuation measurements. AI segmentation and AI-ROI method were
evaluated on 1,014 non-contrast enhanced chest CT images from eight
international datasets: LIDC-IDRI, NSCLC-Lung1, RIDER, VESSEL12, RICORD-1A,
RICORD-1B, COVID-19-Italy, and COVID-19-China. AI segmentation achieved a mean
dice coefficient of 0.957. Attenuations measured by AI-ROI showed no
significant differences (p = 0.545) and a reduction of 71% time compared to
expert measurements. The area under the curve (AUC) of the steatosis
classification of AI-ROI is 0.921 (95% CI: 0.883 - 0.959). If performed as a
routine screening method, our AI protocol could potentially allow early
non-invasive, non-pharmacological preventative interventions for hepatic
steatosis. 1,014 expert-annotated liver segmentations of patients with hepatic
steatosis annotations can be downloaded here:
https://drive.google.com/drive/folders/1-g_zJeAaZXYXGqL1OeF6pUjr6KB0igJX
Unleashing the Power of Prompt-driven Nucleus Instance Segmentation
Nucleus instance segmentation in histology images is crucial for a broad
spectrum of clinical applications. Current dominant algorithms rely on
regression of nuclear proxy maps. Distinguishing nucleus instances from the
estimated maps requires carefully curated post-processing, which is error-prone
and parameter-sensitive. Recently, the Segment Anything Model (SAM) has earned
huge attention in medical image segmentation, owing to its impressive
generalization ability and promptable property. Nevertheless, its potential on
nucleus instance segmentation remains largely underexplored. In this paper, we
present a novel prompt-driven framework that consists of a nucleus prompter and
SAM for automatic nucleus instance segmentation. Specifically, the prompter
learns to generate a unique point prompt for each nucleus while the SAM is
fine-tuned to output the corresponding mask for the prompted nucleus.
Furthermore, we propose the inclusion of adjacent nuclei as negative prompts to
enhance the model's capability to identify overlapping nuclei. Without
complicated post-processing, our proposed method sets a new state-of-the-art
performance on three challenging benchmarks. Code is available at
\url{github.com/windygoo/PromptNucSeg}Comment: under revie
- …