258 research outputs found
AWS PredSpot: Machine Learning for Predicting the Price of Spot Instances in AWS Cloud
Elastic Cloud Compute (EC2) is one of the most well-known services provided by Amazon for provisioning cloud computing resources, also known as instances. Besides the classical on-demand scheme, where users purchase compute capacity at a fixed cost, EC2 supports so-called spot instances, which are offered following a bidding scheme, where users can save up to 90% of the cost of the on-demand instance. EC2 spot instances can be a useful alternative for attaining an important reduction in infrastructure cost, but designing bidding policies can be a difficult task, since bidding under their cost will either prevent users from provisioning instances or losing those that they already own. Towards this extent, accurate forecasting of spot instance prices can be of an outstanding interest for designing working bidding policies. In this paper, we propose the use of different machine learning techniques to estimate the future price of EC2 spot instances. These include linear, ridge and lasso regressions, multilayer perceptrons, K-nearest neighbors, extra trees and random forests. The obtained performance varies significantly between instances types, and root mean squared errors ranges between values very close to zero up to values over 60 in some of the most expensive instances. Still, we can see that for most of the instances, forecasting performance is remarkably good, encouraging further research in this field of study
Feature engineering for mid-price prediction with deep learning
Mid-price movement prediction based on limit order book (LOB) data is a
challenging task due to the complexity and dynamics of the LOB. So far, there
have been very limited attempts for extracting relevant features based on LOB
data. In this paper, we address this problem by designing a new set of
handcrafted features and performing an extensive experimental evaluation on
both liquid and illiquid stocks. More specifically, we implement a new set of
econometrical features that capture statistical properties of the underlying
securities for the task of mid-price prediction. Moreover, we develop a new
experimental protocol for online learning that treats the task as a
multi-objective optimization problem and predicts i) the direction of the next
price movement and ii) the number of order book events that occur until the
change takes place. In order to predict the mid-price movement, the features
are fed into nine different deep learning models based on multi-layer
perceptrons (MLP), convolutional neural networks (CNN) and long short-term
memory (LSTM) neural networks. The performance of the proposed method is then
evaluated on liquid and illiquid stocks, which are based on TotalView-ITCH US
and Nordic stocks, respectively. For some stocks, results suggest that the
correct choice of a feature set and a model can lead to the successful
prediction of how long it takes to have a stock price movement
Comparativa de los models clásicos de series temporales con la red neuronal recurrente LSTM: Una aplicación a las acciones del S&P 500
[EN] In the financial literature, there is great interest in the prediction of stock prices. Stock prediction is necessary for the creation of di erent investment strategies, both speculative and hedging ones. The application of neural networks has involved a change in the creation of predictive models. In this paper,
we analyze the capacity of recurrent neural networks, in particular the long short-term recurrent neural
network (LSTM) as opposed to classic time series models such as the Exponential Smooth Time Series
(ETS) and the Arima model (ARIMA). These models have been estimated for 284 stocks from the S&P 500
stock market index, comparing the MAE obtained from their predictions. The results obtained confirm a
significant reduction in prediction errors when LSTM is applied. These results are consistent with other
similar studies applied to stocks included in other stock market indices, as well as other financial assets
such as exchange rates.[ES] En la literatura financiera existe un gran interés por la predicción de precios bursátiles que es necesario
para la creación de diferentes estrategias de inversion, tanto especulativas como de cobertura. La aplicación de las redes neuronales ha supuesto un cambio en la creación de modelos de predicción. En este
trabajo se analiza la capacidad que tienen las redes neuronales recurrentes, en concreto la long shortterm recurrent neural network (LSTM) frente a modelos de series temporales clásicos como el Exponential Smooth Time Series (ETS) y el modelo Arima (ARIMA). Para ello se ha estimado dichos modelos para
284 acciones pertenecientes al índice bursátil S&P 500, comparando el MAE obtenido de sus predicciones, con el modelo LSTM. Los resultados obtenidos confirman una reducción importante de los errores
de predicción. Estos resultados son coincidentes con otros estudios similares aplicados a acciones de
otros índices bursátiles así como a otros activos financieros como los tipos de cambio.Oliver-Muncharaz, J. (2020). Comparing classic time series models and the LSTM recurrent neural network: An application to S&P 500 stocks. Finance, Markets and Valuation. 6(2):137-148. https://doi.org/10.46503/ZVBS2781S1371486
Reducing the price of resource provisioning using EC2 spot instances with prediction models
The increasing demand of computing resources has boosted the use of cloud computing providers. This has raised a new dimension in which the connections between resource usage and costs have to be considered from an organizational perspective. As a part of its EC2 service, Amazon introduced spot instances (SI) as a cheap public infrastructure, but at the price of not ensuring reliability of the service. On the Amazon SI model, hired instances can be abruptly terminated by the service provider when necessary. The interface for managing SI is based on a bidding strategy that depends on non-public Amazon pricing strategies, which makes complicated for users to apply any scheduling or resource provisioning strategy based on such (cheaper) resources. Although it is believed that the use of the EC2 SIs infrastructure can reduce costs for final users, a deep review of literature concludes that their characteristics and possibilities have not yet been deeply explored. In this work we present a framework for the analysis of the EC2 SIs infrastructure that uses the price history of such resources in order to classify the SI availability zones and then generate price prediction models adapted to each class. The proposed models are validated through a formal experimentation process. As a result, these models are applied to generate resource provisioning plans that get the optimal price when using the SI infrastructure in a real scenario. Finally, the recent changes that Amazon has introduced in the SI model and how this work can adapt to these changes is discussed
Automated Crowdturfing Attacks and Defenses in Online Review Systems
Malicious crowdsourcing forums are gaining traction as sources of spreading
misinformation online, but are limited by the costs of hiring and managing
human workers. In this paper, we identify a new class of attacks that leverage
deep learning language models (Recurrent Neural Networks or RNNs) to automate
the generation of fake online reviews for products and services. Not only are
these attacks cheap and therefore more scalable, but they can control rate of
content output to eliminate the signature burstiness that makes crowdsourced
campaigns easy to detect.
Using Yelp reviews as an example platform, we show how a two phased review
generation and customization attack can produce reviews that are
indistinguishable by state-of-the-art statistical detectors. We conduct a
survey-based user study to show these reviews not only evade human detection,
but also score high on "usefulness" metrics by users. Finally, we develop novel
automated defenses against these attacks, by leveraging the lossy
transformation introduced by the RNN training and generation cycle. We consider
countermeasures against our mechanisms, show that they produce unattractive
cost-benefit tradeoffs for attackers, and that they can be further curtailed by
simple constraints imposed by online service providers
TransNets: Learning to Transform for Recommendation
Recently, deep learning methods have been shown to improve the performance of
recommender systems over traditional methods, especially when review text is
available. For example, a recent model, DeepCoNN, uses neural nets to learn one
latent representation for the text of all reviews written by a target user, and
a second latent representation for the text of all reviews for a target item,
and then combines these latent representations to obtain state-of-the-art
performance on recommendation tasks. We show that (unsurprisingly) much of the
predictive value of review text comes from reviews of the target user for the
target item. We then introduce a way in which this information can be used in
recommendation, even when the target user's review for the target item is not
available. Our model, called TransNets, extends the DeepCoNN model by
introducing an additional latent layer representing the target user-target item
pair. We then regularize this layer, at training time, to be similar to another
latent representation of the target user's review of the target item. We show
that TransNets and extensions of it improve substantially over the previous
state-of-the-art.Comment: Accepted for publication in the 11th ACM Conference on Recommender
Systems (RecSys 2017
- …