179 research outputs found
The chemokine and chemokine receptor superfamilies and their molecular evolution
The human chemokine superfamily currently includes at least 46 ligands, which bind to 18 functionally signaling G-protein-coupled receptors and two decoy or scavenger receptors. The chemokine ligands probably comprise one of the first completely known molecular superfamilies. The genomic organization of the chemokine ligand genes and a comparison of their sequences between species shows that tandem gene duplication has taken place independently in the mouse and human lineages of some chemokine families. This means that care needs to be taken when extrapolating experimental results on some chemokines from mouse to human
PillarNeSt: Embracing Backbone Scaling and Pretraining for Pillar-based 3D Object Detection
This paper shows the effectiveness of 2D backbone scaling and pretraining for
pillar-based 3D object detectors. Pillar-based methods mainly employ randomly
initialized 2D convolution neural network (ConvNet) for feature extraction and
fail to enjoy the benefits from the backbone scaling and pretraining in the
image domain. To show the scaling-up capacity in point clouds, we introduce the
dense ConvNet pretrained on large-scale image datasets (e.g., ImageNet) as the
2D backbone of pillar-based detectors. The ConvNets are adaptively designed
based on the model size according to the specific features of point clouds,
such as sparsity and irregularity. Equipped with the pretrained ConvNets, our
proposed pillar-based detector, termed PillarNeSt, outperforms the existing 3D
object detectors by a large margin on the nuScenes and Argoversev2 datasets.
Our code shall be released upon acceptance
Malignant transformation of diffuse infiltrating glial neoplasm after prolonged stable period initially discovered with hypothalamic hamartoma
We present a case of malignant transformation of diffuse infiltrating glial neoplasm after a prolonged stable period on magnetic resonance imaging (MRI) and spectroscopy (MRS) initially discovered with a hypothalamic hamartoma. Although MRI and MRS suggest the possibility of malignant transformation in future, they cannot precisely predict the timing of rapid growth
Vision Learners Meet Web Image-Text Pairs
Most recent self-supervised learning methods are pre-trained on the
well-curated ImageNet-1K dataset. In this work, given the excellent scalability
of web data, we consider self-supervised pre-training on noisy web sourced
image-text paired data. First, we conduct a benchmark study of representative
self-supervised pre-training methods on large-scale web data in a like-for-like
setting. We compare a range of methods, including single-modal ones that use
masked training objectives and multi-modal ones that use image-text
constrastive training. We observe that existing multi-modal methods do not
outperform their single-modal counterparts on vision transfer learning tasks.
We derive an information-theoretical view to explain these benchmark results,
which provides insight into how to design a novel vision learner. Inspired by
this insight, we present a new visual representation pre-training method,
MUlti-modal Generator~(MUG), that learns from scalable web sourced image-text
data. MUG achieves state-of-the-art transfer performance on a variety of tasks
and demonstrates promising scaling properties. Pre-trained models and code will
be made public upon acceptance.Comment: Project page: https://bzhao.me/MUG
- …