4,424 research outputs found
Seismic Performance and Design of Bridge Foundations in Liquefiable Ground with a Frozen Crust
INE/AUTC 12.3
Convergence of martingale solution to slow-fast systems with jumps modulated by Markovian switching
This paper investigates the convergence of martingale solutions to slow-fast
systems with jumps modulated by Markovian switching on weakly irreducible
class. The key point here is to deals with slow-fast systems and two-time-scale
Markovian switching simultaneously, while averaging on the slow component
requires two invariant measures respectively due to the coexistence of the fast
component and Markovian switching. We first investigate the slow-fast systems
modulated by Markovian chains with single weakly irreducible class, and the
existence and uniqueness of the solution will be proved. Then weak convergence
is presented based on tightness and the exponential ergodicity of the fast
component with the martingale method, where the appropriate perturbed test
functions plays a decisive role in processing. Finally we extend results to the
case of the multiple irreducible class
Deep Neural Network Architectures for Modulation Classification
In this work, we investigate the value of employing deep learning for the
task of wireless signal modulation recognition. Recently in [1], a framework
has been introduced by generating a dataset using GNU radio that mimics the
imperfections in a real wireless channel, and uses 10 different modulation
types. Further, a convolutional neural network (CNN) architecture was developed
and shown to deliver performance that exceeds that of expert-based approaches.
Here, we follow the framework of [1] and find deep neural network architectures
that deliver higher accuracy than the state of the art. We tested the
architecture of [1] and found it to achieve an accuracy of approximately 75% of
correctly recognizing the modulation type. We first tune the CNN architecture
of [1] and find a design with four convolutional layers and two dense layers
that gives an accuracy of approximately 83.8% at high SNR. We then develop
architectures based on the recently introduced ideas of Residual Networks
(ResNet [2]) and Densely Connected Networks (DenseNet [3]) to achieve high SNR
accuracies of approximately 83.5% and 86.6%, respectively. Finally, we
introduce a Convolutional Long Short-term Deep Neural Network (CLDNN [4]) to
achieve an accuracy of approximately 88.5% at high SNR.Comment: 5 pages, 10 figures, In proc. Asilomar Conference on Signals,
Systems, and Computers, Nov. 201
- …