2,125 research outputs found
Self-Supervised Disentanglement of Harmonic and Rhythmic Features in Music Audio Signals
The aim of latent variable disentanglement is to infer the multiple
informative latent representations that lie behind a data generation process
and is a key factor in controllable data generation. In this paper, we propose
a deep neural network-based self-supervised learning method to infer the
disentangled rhythmic and harmonic representations behind music audio
generation. We train a variational autoencoder that generates an audio
mel-spectrogram from two latent features representing the rhythmic and harmonic
content. In the training phase, the variational autoencoder is trained to
reconstruct the input mel-spectrogram given its pitch-shifted version. At each
forward computation in the training phase, a vector rotation operation is
applied to one of the latent features, assuming that the dimensions of the
feature vectors are related to pitch intervals. Therefore, in the trained
variational autoencoder, the rotated latent feature represents the
pitch-related information of the mel-spectrogram, and the unrotated latent
feature represents the pitch-invariant information, i.e., the rhythmic content.
The proposed method was evaluated using a predictor-based disentanglement
metric on the learned features. Furthermore, we demonstrate its application to
the automatic generation of music remixes.Comment: Accepted to DAFx 202
音楽音響信号に対する自動コード推定のための生成・識別統合的アプローチ
京都大学新制・課程博士博士(情報学)甲第23540号情博第770号新制||情||131(附属図書館)京都大学大学院情報学研究科知能情報学専攻(主査)准教授 吉井 和佳, 教授 河原 達也, 教授 西野 恒, 教授 鹿島 久嗣学位規則第4条第1項該当Doctor of InformaticsKyoto UniversityDFA
Switch-based Active Deep Dyna-Q: Efficient Adaptive Planning for Task-Completion Dialogue Policy Learning
Training task-completion dialogue agents with reinforcement learning usually
requires a large number of real user experiences. The Dyna-Q algorithm extends
Q-learning by integrating a world model, and thus can effectively boost
training efficiency using simulated experiences generated by the world model.
The effectiveness of Dyna-Q, however, depends on the quality of the world model
- or implicitly, the pre-specified ratio of real vs. simulated experiences used
for Q-learning. To this end, we extend the recently proposed Deep Dyna-Q (DDQ)
framework by integrating a switcher that automatically determines whether to
use a real or simulated experience for Q-learning. Furthermore, we explore the
use of active learning for improving sample efficiency, by encouraging the
world model to generate simulated experiences in the state-action space where
the agent has not (fully) explored. Our results show that by combining switcher
and active learning, the new framework named as Switch-based Active Deep Dyna-Q
(Switch-DDQ), leads to significant improvement over DDQ and Q-learning
baselines in both simulation and human evaluations.Comment: 8 pages, 9 figures, AAAI 201
Robust Control of Crane with Perturbations
In the presence of persistent perturbations in both unactuated and actuated dynamics of crane systems, an observer-based robust control method is proposed, which achieves the objective of trolley positioning and cargo swing suppression. By dealing with the unactuated and unknown perturbation as an augmented state variable, the system dynamics are transformed into a quasi-chain-of-integrators form based on which a reduced-order augmented-state observer is established to recover the perturbations appearing in the unactuated dynamics. A novel sliding manifold is constructed to improve the robust performance of the control system, and a linear control law is presented to make the state variables stay on the manifold in the presence of perturbations in unactuated dynamics. A Lyapunov function candidate is constructed, and the entire closed-loop system is proved rigorously to be exponentially stable at the equilibrium point. The effectiveness and robustness of the proposed observer-based robust controller are verified by numerical simulation results
- …