1,437 research outputs found
Eastern Enlargement of the EU and its Economic Impact: A CGE Approach
The main objective of this study is to conduct a quantitative assessment of the potential economic effects of the 5th enlargement of the EU including Cyprus, Czech Republic, Estonia, Hungary, Latvia, Lithuania, Malta, Poland, Slovakia, and Slovenia which joined the EU on 1 May 2004 as well as of additional accession of Bulgaria and Romania which is to take place on 1 January 2007 using a multi-region, multi-sector computable general equilibrium (CGE) model. In addition, a full membership of Croatia and Turkey is considered for quantification. Economic effects of eastern enlargements of the EU are expected to be significant for the enlarged Europe, as a bigger and more integrated market boosts economic growth for current and new members alike. On one hand, the wider Europe is to positively affect the economies of third countries such as the CIS (Commonwealth of Independent States) including Russia and the Republics of the former Soviet Union. On the other hand, it is to negatively influence the economies of most of the third countries such as China, Japan, Korea, and North American Free Trade Area (NAFTA
Bi-PointFlowNet: Bidirectional Learning for Point Cloud Based Scene Flow Estimation
Scene flow estimation, which extracts point-wise motion between scenes, is
becoming a crucial task in many computer vision tasks. However, all of the
existing estimation methods utilize only the unidirectional features,
restricting the accuracy and generality. This paper presents a novel scene flow
estimation architecture using bidirectional flow embedding layers. The proposed
bidirectional layer learns features along both forward and backward directions,
enhancing the estimation performance. In addition, hierarchical feature
extraction and warping improve the performance and reduce computational
overhead. Experimental results show that the proposed architecture achieved a
new state-of-the-art record by outperforming other approaches with large margin
in both FlyingThings3D and KITTI benchmarks. Codes are available at
https://github.com/cwc1260/BiFlow.Comment: Accepted as a conference paper at European Conference on Computer
Vision (ECCV) 202
NAS-VAD: Neural Architecture Search for Voice Activity Detection
Various neural network-based approaches have been proposed for more robust
and accurate voice activity detection (VAD). Manual design of such neural
architectures is an error-prone and time-consuming process, which prompted the
development of neural architecture search (NAS) that automatically design and
optimize network architectures. While NAS has been successfully applied to
improve performance in a variety of tasks, it has not yet been exploited in the
VAD domain. In this paper, we present the first work that utilizes NAS
approaches on the VAD task. To effectively search architectures for the VAD
task, we propose a modified macro structure and a new search space with a much
broader range of operations that includes attention operations. The results
show that the network structures found by the propose NAS framework outperform
previous manually designed state-of-the-art VAD models in various noise-added
and real-world-recorded datasets. We also show that the architectures searched
on a particular dataset achieve improved generalization performance on unseen
audio datasets. Our code and models are available at
https://github.com/daniel03c1/NAS_VAD.Comment: Submitted to Interspeech 202
Neural Residual Flow Fields for Efficient Video Representations
Neural fields have emerged as a powerful paradigm for representing various
signals, including videos. However, research on improving the parameter
efficiency of neural fields is still in its early stages. Even though neural
fields that map coordinates to colors can be used to encode video signals, this
scheme does not exploit the spatial and temporal redundancy of video signals.
Inspired by standard video compression algorithms, we propose a neural field
architecture for representing and compressing videos that deliberately removes
data redundancy through the use of motion information across video frames.
Maintaining motion information, which is typically smoother and less complex
than color signals, requires a far fewer number of parameters. Furthermore,
reusing color values through motion information further improves the network
parameter efficiency. In addition, we suggest using more than one reference
frame for video frame reconstruction and separate networks, one for optical
flows and the other for residuals. Experimental results have shown that the
proposed method outperforms the baseline methods by a significant margin. The
code is available in https://github.com/daniel03c1/eff_video_representationComment: Accepted for ACCV 2022, codes are available at
https://github.com/daniel03c1/eff_video_representatio
Compact 3D Gaussian Representation for Radiance Field
Neural Radiance Fields (NeRFs) have demonstrated remarkable potential in
capturing complex 3D scenes with high fidelity. However, one persistent
challenge that hinders the widespread adoption of NeRFs is the computational
bottleneck due to the volumetric rendering. On the other hand, 3D Gaussian
splatting (3DGS) has recently emerged as an alternative representation that
leverages a 3D Gaussisan-based representation and adopts the rasterization
pipeline to render the images rather than volumetric rendering, achieving very
fast rendering speed and promising image quality. However, a significant
drawback arises as 3DGS entails a substantial number of 3D Gaussians to
maintain the high fidelity of the rendered images, which requires a large
amount of memory and storage. To address this critical issue, we place a
specific emphasis on two key objectives: reducing the number of Gaussian points
without sacrificing performance and compressing the Gaussian attributes, such
as view-dependent color and covariance. To this end, we propose a learnable
mask strategy that significantly reduces the number of Gaussians while
preserving high performance. In addition, we propose a compact but effective
representation of view-dependent color by employing a grid-based neural field
rather than relying on spherical harmonics. Finally, we learn codebooks to
compactly represent the geometric attributes of Gaussian by vector
quantization. With model compression techniques such as quantization and
entropy coding, we consistently show over 25 reduced storage and
enhanced rendering speed, while maintaining the quality of the scene
representation, compared to 3DGS. Our work provides a comprehensive framework
for 3D scene representation, achieving high performance, fast training,
compactness, and real-time rendering. Our project page is available at
https://maincold2.github.io/c3dgs/.Comment: Project page: http://maincold2.github.io/c3dgs
Coordinate-Aware Modulation for Neural Fields
Neural fields, mapping low-dimensional input coordinates to corresponding
signals, have shown promising results in representing various signals. Numerous
methodologies have been proposed, and techniques employing MLPs and grid
representations have achieved substantial success. MLPs allow compact and high
expressibility, yet often suffer from spectral bias and slow convergence speed.
On the other hand, methods using grids are free from spectral bias and achieve
fast training speed, however, at the expense of high spatial complexity. In
this work, we propose a novel way for exploiting both MLPs and grid
representations in neural fields. Unlike the prevalent methods that combine
them sequentially (extract features from the grids first and feed them to the
MLP), we inject spectral bias-free grid representations into the intermediate
features in the MLP. More specifically, we suggest a Coordinate-Aware
Modulation (CAM), which modulates the intermediate features using scale and
shift parameters extracted from the grid representations. This can maintain the
strengths of MLPs while mitigating any remaining potential biases, facilitating
the rapid learning of high-frequency components. In addition, we empirically
found that the feature normalizations, which have not been successful in neural
filed literature, proved to be effective when applied in conjunction with the
proposed CAM. Experimental results demonstrate that CAM enhances the
performance of neural representation and improves learning stability across a
range of signals. Especially in the novel view synthesis task, we achieved
state-of-the-art performance with the least number of parameters and fast
training speed for dynamic scenes and the best performance under 1MB memory for
static scenes. CAM also outperforms the best-performing video compression
methods using neural fields by a large margin.Comment: Project page: http://maincold2.github.io/cam
- …