49,480 research outputs found
Investigation of Air Transportation Technology at Princeton University, 1989-1990
The Air Transportation Technology Program at Princeton University proceeded along six avenues during the past year: microburst hazards to aircraft; machine-intelligent, fault tolerant flight control; computer aided heuristics for piloted flight; stochastic robustness for flight control systems; neural networks for flight control; and computer aided control system design. These topics are briefly discussed, and an annotated bibliography of publications that appeared between January 1989 and June 1990 is given
Scalable and Sustainable Deep Learning via Randomized Hashing
Current deep learning architectures are growing larger in order to learn from
complex datasets. These architectures require giant matrix multiplication
operations to train millions of parameters. Conversely, there is another
growing trend to bring deep learning to low-power, embedded devices. The matrix
operations, associated with both training and testing of deep networks, are
very expensive from a computational and energy standpoint. We present a novel
hashing based technique to drastically reduce the amount of computation needed
to train and test deep networks. Our approach combines recent ideas from
adaptive dropouts and randomized hashing for maximum inner product search to
select the nodes with the highest activation efficiently. Our new algorithm for
deep learning reduces the overall computational cost of forward and
back-propagation by operating on significantly fewer (sparse) nodes. As a
consequence, our algorithm uses only 5% of the total multiplications, while
keeping on average within 1% of the accuracy of the original model. A unique
property of the proposed hashing based back-propagation is that the updates are
always sparse. Due to the sparse gradient updates, our algorithm is ideally
suited for asynchronous and parallel training leading to near linear speedup
with increasing number of cores. We demonstrate the scalability and
sustainability (energy efficiency) of our proposed algorithm via rigorous
experimental evaluations on several real datasets
Evolving stochastic learning algorithm based on Tsallis entropic index
In this paper, inspired from our previous algorithm, which was based on the theory of Tsallis statistical mechanics, we develop a new evolving stochastic learning algorithm for neural networks. The new algorithm combines deterministic and stochastic search steps by employing a different adaptive stepsize for each network weight, and applies a form of noise that is characterized by the nonextensive entropic index q, regulated by a weight decay term. The behavior of the learning algorithm can be made more stochastic or deterministic depending on the trade off between the temperature T and the q values. This is achieved by introducing a formula that defines a time-dependent relationship between these two important learning parameters. Our experimental study verifies that there are indeed improvements in the convergence speed of this new evolving stochastic learning algorithm, which makes learning faster than using the original Hybrid Learning Scheme (HLS). In addition, experiments are conducted to explore the influence of the entropic index q and temperature T on the convergence speed and stability of the proposed method
Enhancing Stock Movement Prediction with Adversarial Training
This paper contributes a new machine learning solution for stock movement
prediction, which aims to predict whether the price of a stock will be up or
down in the near future. The key novelty is that we propose to employ
adversarial training to improve the generalization of a neural network
prediction model. The rationality of adversarial training here is that the
input features to stock prediction are typically based on stock price, which is
essentially a stochastic variable and continuously changed with time by nature.
As such, normal training with static price-based features (e.g. the close
price) can easily overfit the data, being insufficient to obtain reliable
models. To address this problem, we propose to add perturbations to simulate
the stochasticity of price variable, and train the model to work well under
small yet intentional perturbations. Extensive experiments on two real-world
stock data show that our method outperforms the state-of-the-art solution with
3.11% relative improvements on average w.r.t. accuracy, validating the
usefulness of adversarial training for stock prediction task.Comment: IJCAI 201
- …