97 research outputs found
Unique Solution of a Coupled Fractional Differential System Involving Integral Boundary Conditions from Economic Model
We study the existence and uniqueness of the positive solution for the fractional differential system involving the Riemann-Stieltjes integral boundary conditions , , , , , and , where , , and and are the standard Riemann-Liouville derivatives, and are functions of bounded variation, and and denote the Riemann-Stieltjes integral. Our results are based on a generalized fixed point theorem for weakly contractive mappings in partially ordered sets
One-Stage 3D Whole-Body Mesh Recovery with Component Aware Transformer
Whole-body mesh recovery aims to estimate the 3D human body, face, and hands
parameters from a single image. It is challenging to perform this task with a
single network due to resolution issues, i.e., the face and hands are usually
located in extremely small regions. Existing works usually detect hands and
faces, enlarge their resolution to feed in a specific network to predict the
parameter, and finally fuse the results. While this copy-paste pipeline can
capture the fine-grained details of the face and hands, the connections between
different parts cannot be easily recovered in late fusion, leading to
implausible 3D rotation and unnatural pose. In this work, we propose a
one-stage pipeline for expressive whole-body mesh recovery, named OSX, without
separate networks for each part. Specifically, we design a Component Aware
Transformer (CAT) composed of a global body encoder and a local face/hand
decoder. The encoder predicts the body parameters and provides a high-quality
feature map for the decoder, which performs a feature-level upsample-crop
scheme to extract high-resolution part-specific features and adopt
keypoint-guided deformable attention to estimate hand and face precisely. The
whole pipeline is simple yet effective without any manual post-processing and
naturally avoids implausible prediction. Comprehensive experiments demonstrate
the effectiveness of OSX. Lastly, we build a large-scale Upper-Body dataset
(UBody) with high-quality 2D and 3D whole-body annotations. It contains persons
with partially visible bodies in diverse real-life scenarios to bridge the gap
between the basic task and downstream applications.Comment: Accepted to CVPR2023; Top-1 on AGORA SMPLX benchmark; Project Page:
https://osx-ubody.github.io
Binarized Spectral Compressive Imaging
Existing deep learning models for hyperspectral image (HSI) reconstruction
achieve good performance but require powerful hardwares with enormous memory
and computational resources. Consequently, these methods can hardly be deployed
on resource-limited mobile devices. In this paper, we propose a novel method,
Binarized Spectral-Redistribution Network (BiSRNet), for efficient and
practical HSI restoration from compressed measurement in snapshot compressive
imaging (SCI) systems. Firstly, we redesign a compact and easy-to-deploy base
model to be binarized. Then we present the basic unit, Binarized
Spectral-Redistribution Convolution (BiSR-Conv). BiSR-Conv can adaptively
redistribute the HSI representations before binarizing activation and uses a
scalable hyperbolic tangent function to closer approximate the Sign function in
backpropagation. Based on our BiSR-Conv, we customize four binarized
convolutional modules to address the dimension mismatch and propagate
full-precision information throughout the whole network. Finally, our BiSRNet
is derived by using the proposed techniques to binarize the base model.
Comprehensive quantitative and qualitative experiments manifest that our
proposed BiSRNet outperforms state-of-the-art binarization methods and achieves
comparable performance with full-precision algorithms. Code and models are
publicly available at https://github.com/caiyuanhao1998/BiSCI and
https://github.com/caiyuanhao1998/MSTComment: NeurIPS 2023; The first work to study binarized spectral compressive
imaging reconstruction proble
NeFII: Inverse Rendering for Reflectance Decomposition with Near-Field Indirect Illumination
Inverse rendering methods aim to estimate geometry, materials and
illumination from multi-view RGB images. In order to achieve better
decomposition, recent approaches attempt to model indirect illuminations
reflected from different materials via Spherical Gaussians (SG), which,
however, tends to blur the high-frequency reflection details. In this paper, we
propose an end-to-end inverse rendering pipeline that decomposes materials and
illumination from multi-view images, while considering near-field indirect
illumination. In a nutshell, we introduce the Monte Carlo sampling based path
tracing and cache the indirect illumination as neural radiance, enabling a
physics-faithful and easy-to-optimize inverse rendering method. To enhance
efficiency and practicality, we leverage SG to represent the smooth environment
illuminations and apply importance sampling techniques. To supervise indirect
illuminations from unobserved directions, we develop a novel radiance
consistency constraint between implicit neural radiance and path tracing
results of unobserved rays along with the joint optimization of materials and
illuminations, thus significantly improving the decomposition performance.
Extensive experiments demonstrate that our method outperforms the
state-of-the-art on multiple synthetic and real datasets, especially in terms
of inter-reflection decomposition.Comment: Accepted in CVPR 202
Retinexformer: One-stage Retinex-based Transformer for Low-light Image Enhancement
When enhancing low-light images, many deep learning algorithms are based on
the Retinex theory. However, the Retinex model does not consider the
corruptions hidden in the dark or introduced by the light-up process. Besides,
these methods usually require a tedious multi-stage training pipeline and rely
on convolutional neural networks, showing limitations in capturing long-range
dependencies. In this paper, we formulate a simple yet principled One-stage
Retinex-based Framework (ORF). ORF first estimates the illumination information
to light up the low-light image and then restores the corruption to produce the
enhanced image. We design an Illumination-Guided Transformer (IGT) that
utilizes illumination representations to direct the modeling of non-local
interactions of regions with different lighting conditions. By plugging IGT
into ORF, we obtain our algorithm, Retinexformer. Comprehensive quantitative
and qualitative experiments demonstrate that our Retinexformer significantly
outperforms state-of-the-art methods on thirteen benchmarks. The user study and
application on low-light object detection also reveal the latent practical
values of our method. Code, models, and results are available at
https://github.com/caiyuanhao1998/RetinexformerComment: ICCV 2023; The first Transformer-based method for low-light image
enhancemen
Iterative Few-shot Semantic Segmentation from Image Label Text
Few-shot semantic segmentation aims to learn to segment unseen class objects
with the guidance of only a few support images. Most previous methods rely on
the pixel-level label of support images. In this paper, we focus on a more
challenging setting, in which only the image-level labels are available. We
propose a general framework to firstly generate coarse masks with the help of
the powerful vision-language model CLIP, and then iteratively and mutually
refine the mask predictions of support and query images. Extensive experiments
on PASCAL-5i and COCO-20i datasets demonstrate that our method not only
outperforms the state-of-the-art weakly supervised approaches by a significant
margin, but also achieves comparable or better results to recent supervised
methods. Moreover, our method owns an excellent generalization ability for the
images in the wild and uncommon classes. Code will be available at
https://github.com/Whileherham/IMR-HSNet.Comment: ijcai 202
- …