7,455 research outputs found
Lightweight Spatial-Channel Adaptive Coordination of Multilevel Refinement Enhancement Network for Image Reconstruction
Benefiting from the vigorous development of deep learning, many CNN-based
image super-resolution methods have emerged and achieved better results than
traditional algorithms. However, it is difficult for most algorithms to
adaptively adjust the spatial region and channel features at the same time, let
alone the information exchange between them. In addition, the exchange of
information between attention modules is even less visible to researchers. To
solve these problems, we put forward a lightweight spatial-channel adaptive
coordination of multilevel refinement enhancement networks(MREN). Specifically,
we construct a space-channel adaptive coordination block, which enables the
network to learn the spatial region and channel feature information of interest
under different receptive fields. In addition, the information of the
corresponding feature processing level between the spatial part and the channel
part is exchanged with the help of jump connection to achieve the coordination
between the two. We establish a communication bridge between attention modules
through a simple linear combination operation, so as to more accurately and
continuously guide the network to pay attention to the information of interest.
Extensive experiments on several standard test sets have shown that our MREN
achieves superior performance over other advanced algorithms with a very small
number of parameters and very low computational complexity
Searching Efficient 3D Architectures with Sparse Point-Voxel Convolution
Self-driving cars need to understand 3D scenes efficiently and accurately in
order to drive safely. Given the limited hardware resources, existing 3D
perception models are not able to recognize small instances (e.g., pedestrians,
cyclists) very well due to the low-resolution voxelization and aggressive
downsampling. To this end, we propose Sparse Point-Voxel Convolution (SPVConv),
a lightweight 3D module that equips the vanilla Sparse Convolution with the
high-resolution point-based branch. With negligible overhead, this point-based
branch is able to preserve the fine details even from large outdoor scenes. To
explore the spectrum of efficient 3D models, we first define a flexible
architecture design space based on SPVConv, and we then present 3D Neural
Architecture Search (3D-NAS) to search the optimal network architecture over
this diverse design space efficiently and effectively. Experimental results
validate that the resulting SPVNAS model is fast and accurate: it outperforms
the state-of-the-art MinkowskiNet by 3.3%, ranking 1st on the competitive
SemanticKITTI leaderboard. It also achieves 8x computation reduction and 3x
measured speedup over MinkowskiNet with higher accuracy. Finally, we transfer
our method to 3D object detection, and it achieves consistent improvements over
the one-stage detection baseline on KITTI.Comment: ECCV 2020. The first two authors contributed equally to this work.
Project page: http://spvnas.mit.edu
s-LWSR: Super Lightweight Super-Resolution Network
Deep learning (DL) architectures for superresolution (SR) normally contain
tremendous parameters, which has been regarded as the crucial advantage for
obtaining satisfying performance. However, with the widespread use of mobile
phones for taking and retouching photos, this character greatly hampers the
deployment of DL-SR models on the mobile devices. To address this problem, in
this paper, we propose a super lightweight SR network: s-LWSR. There are mainly
three contributions in our work. Firstly, in order to efficiently abstract
features from the low resolution image, we build an information pool to mix
multi-level information from the first half part of the pipeline. Accordingly,
the information pool feeds the second half part with the combination of
hierarchical features from the previous layers. Secondly, we employ a
compression module to further decrease the size of parameters. Intensive
analysis confirms its capacity of trade-off between model complexity and
accuracy. Thirdly, by revealing the specific role of activation in deep models,
we remove several activation layers in our SR model to retain more information
for performance improvement. Extensive experiments show that our s-LWSR, with
limited parameters and operations, can achieve similar performance to other
cumbersome DL-SR methods
- …