138 research outputs found
ME-PCN: Point Completion Conditioned on Mask Emptiness
Point completion refers to completing the missing geometries of an object
from incomplete observations. Main-stream methods predict the missing shapes by
decoding a global feature learned from the input point cloud, which often leads
to deficient results in preserving topology consistency and surface details. In
this work, we present ME-PCN, a point completion network that leverages
`emptiness' in 3D shape space. Given a single depth scan, previous methods
often encode the occupied partial shapes while ignoring the empty regions (e.g.
holes) in depth maps. In contrast, we argue that these `emptiness' clues
indicate shape boundaries that can be used to improve topology representation
and detail granularity on surfaces. Specifically, our ME-PCN encodes both the
occupied point cloud and the neighboring `empty points'. It estimates
coarse-grained but complete and reasonable surface points in the first stage,
followed by a refinement stage to produce fine-grained surface details.
Comprehensive experiments verify that our ME-PCN presents better qualitative
and quantitative performance against the state-of-the-art. Besides, we further
prove that our `emptiness' design is lightweight and easy to embed in existing
methods, which shows consistent effectiveness in improving the CD and EMD
scores.Comment: Accepted to ICCV 2021; typos correcte
STORM-GAN: Spatio-Temporal Meta-GAN for Cross-City Estimation of Human Mobility Responses to COVID-19
Human mobility estimation is crucial during the COVID-19 pandemic due to its
significant guidance for policymakers to make non-pharmaceutical interventions.
While deep learning approaches outperform conventional estimation techniques on
tasks with abundant training data, the continuously evolving pandemic poses a
significant challenge to solving this problem due to data nonstationarity,
limited observations, and complex social contexts. Prior works on mobility
estimation either focus on a single city or lack the ability to model the
spatio-temporal dependencies across cities and time periods. To address these
issues, we make the first attempt to tackle the cross-city human mobility
estimation problem through a deep meta-generative framework. We propose a
Spatio-Temporal Meta-Generative Adversarial Network (STORM-GAN) model that
estimates dynamic human mobility responses under a set of social and policy
conditions related to COVID-19. Facilitated by a novel spatio-temporal
task-based graph (STTG) embedding, STORM-GAN is capable of learning shared
knowledge from a spatio-temporal distribution of estimation tasks and quickly
adapting to new cities and time periods with limited training samples. The STTG
embedding component is designed to capture the similarities among cities to
mitigate cross-task heterogeneity. Experimental results on real-world data show
that the proposed approach can greatly improve estimation performance and
out-perform baselines.Comment: Accepted at the 22nd IEEE International Conference on Data Mining
(ICDM 2022) Full Pape
Task-Aware Sampling Layer for Point-Wise Analysis
Sampling, grouping, and aggregation are three important components in the
multi-scale analysis of point clouds. In this paper, we present a novel
data-driven sampler learning strategy for point-wise analysis tasks. Unlike the
widely used sampling technique, Farthest Point Sampling (FPS), we propose to
learn sampling and downstream applications jointly. Our key insight is that
uniform sampling methods like FPS are not always optimal for different tasks:
sampling more points around boundary areas can make the point-wise
classification easier for segmentation. Towards this end, we propose a novel
sampler learning strategy that learns sampling point displacement supervised by
task-related ground truth information and can be trained jointly with the
underlying tasks. We further demonstrate our methods in various point-wise
analysis tasks, including semantic part segmentation, point cloud completion,
and keypoint detection. Our experiments show that jointly learning of the
sampler and task brings better performance than using FPS in various
point-based networks.Comment: 14 pages, 13 figures and 14 table
4-{2-[2-(4-Formylphenoxy)ethoxy]ethoxy}benzaldehyde
The title compound, C18H18O5, was obtained by the reaction of 4-hydroxybenzaldehyde with bis(2,2-dichloroethyl) ether in dimethylformamide. In the crystal, the molecule lies on a twofold rotation axis that passes through the central O atom of the aliphatic chain, thus leading to one half-molecule being present per asymmetric unit. The carbonyl, aryl and O—CH2—CH2 groups are almost coplanar, with an r.m.s. deviation of 0.030 Å. The aromatic rings are approximately perpendicular to each other, forming a dihedral angle of 78.31 sh;H⋯O hydrogen bonds and C—H⋯π interactions help to consolidate the three-dimensional network
DELTA: Pre-train a Discriminative Encoder for Legal Case Retrieval via Structural Word Alignment
Recent research demonstrates the effectiveness of using pre-trained language
models for legal case retrieval. Most of the existing works focus on improving
the representation ability for the contextualized embedding of the [CLS] token
and calculate relevance using textual semantic similarity. However, in the
legal domain, textual semantic similarity does not always imply that the cases
are relevant enough. Instead, relevance in legal cases primarily depends on the
similarity of key facts that impact the final judgment. Without proper
treatments, the discriminative ability of learned representations could be
limited since legal cases are lengthy and contain numerous non-key facts. To
this end, we introduce DELTA, a discriminative model designed for legal case
retrieval. The basic idea involves pinpointing key facts in legal cases and
pulling the contextualized embedding of the [CLS] token closer to the key facts
while pushing away from the non-key facts, which can warm up the case embedding
space in an unsupervised manner. To be specific, this study brings the word
alignment mechanism to the contextual masked auto-encoder. First, we leverage
shallow decoders to create information bottlenecks, aiming to enhance the
representation ability. Second, we employ the deep decoder to enable
translation between different structures, with the goal of pinpointing key
facts to enhance discriminative ability. Comprehensive experiments conducted on
publicly available legal benchmarks show that our approach can outperform
existing state-of-the-art methods in legal case retrieval. It provides a new
perspective on the in-depth understanding and processing of legal case
documents.Comment: 11 page
SARS Pandemic Exposure Impaired Early Childhood Development in China
Social and mental stressors associated with the pandemic of a novel infectious disease, e.g., COVID-19 or SARS may promote long-term effects on child development. However, reports aimed at identifying the relationship between pandemics and child health are limited. A retrospective study was conducted to associate the SARS pandemic in 2003 with development milestones or physical examinations among longitudinal measurements of 14,647 children. Experiencing SARS during childhood was associated with delayed milestones, with hazard ratios of 3.17 (95% confidence intervals CI: 2.71, 3.70), 3.98 (3.50, 4.53), 4.96 (4.48, 5.49), or 5.57 (5.00, 6.20) for walking independently, saying a complete sentence, counting 0–10, and undressing him/herself for urination, respectively. These results suggest relevant impacts from COVID-19 on child development should be investigated
FLM-101B: An Open LLM and How to Train It with $100K Budget
Large language models (LLMs) have achieved remarkable success in NLP and
multimodal tasks, among others. Despite these successes, two main challenges
remain in developing LLMs: (i) high computational cost, and (ii) fair and
objective evaluations. In this paper, we report a solution to significantly
reduce LLM training cost through a growth strategy. We demonstrate that a
101B-parameter LLM with 0.31T tokens can be trained with a budget of 100K US
dollars. Inspired by IQ tests, we also consolidate an additional range of
evaluations on top of existing evaluations that focus on knowledge-oriented
abilities. These IQ evaluations include symbolic mapping, rule understanding,
pattern mining, and anti-interference. Such evaluations minimize the potential
impact of memorization. Experimental results show that our model, named
FLM-101B, trained with a budget of 100K US dollars, achieves performance
comparable to powerful and well-known models, e.g., GPT-3 and GLM-130B,
especially on the additional range of IQ evaluations. The checkpoint of
FLM-101B is released at https://huggingface.co/CofeAI/FLM-101B
- …