428,520 research outputs found
A Research on Dimension Reduction Method of Time Series Based on Trend Division
The characteristics of high dimension, complexity and multi granularity of financial time series make it difficult to deal with effectively. In order to solve the problem that the commonly used dimensionality reduction methods cannot reduce the dimensionality of time series with different granularity at the same time, in this paper, a method for dimensionality reduction of time series based on trend division is proposed. This method extracts the extreme value points of time series, identifies the important points in time series quickly and accurately, and compresses them. Experimental results show that, compared with the discrete Fourier transform and wavelet transform, the proposed method can effectively process data of different granularity and different trends on the basis of fully preserving the original information of time series. Moreover, the time complexity is low, the operation is easy, and the proposed method can provide decision support for high-frequency stock trading at the actual level
Mixed-integer Quadratic Programming is in NP
Mixed-integer quadratic programming is the problem of optimizing a quadratic
function over points in a polyhedral set where some of the components are
restricted to be integral. In this paper, we prove that the decision version of
mixed-integer quadratic programming is in NP, thereby showing that it is
NP-complete. This is established by showing that if the decision version of
mixed-integer quadratic programming is feasible, then there exists a solution
of polynomial size. This result generalizes and unifies classical results that
quadratic programming is in NP and integer linear programming is in NP
Generalization Bounds in the Predict-then-Optimize Framework
The predict-then-optimize framework is fundamental in many practical
settings: predict the unknown parameters of an optimization problem, and then
solve the problem using the predicted values of the parameters. A natural loss
function in this environment is to consider the cost of the decisions induced
by the predicted parameters, in contrast to the prediction error of the
parameters. This loss function was recently introduced in Elmachtoub and Grigas
(2017) and referred to as the Smart Predict-then-Optimize (SPO) loss. In this
work, we seek to provide bounds on how well the performance of a prediction
model fit on training data generalizes out-of-sample, in the context of the SPO
loss. Since the SPO loss is non-convex and non-Lipschitz, standard results for
deriving generalization bounds do not apply.
We first derive bounds based on the Natarajan dimension that, in the case of
a polyhedral feasible region, scale at most logarithmically in the number of
extreme points, but, in the case of a general convex feasible region, have
linear dependence on the decision dimension. By exploiting the structure of the
SPO loss function and a key property of the feasible region, which we denote as
the strength property, we can dramatically improve the dependence on the
decision and feature dimensions. Our approach and analysis rely on placing a
margin around problematic predictions that do not yield unique optimal
solutions, and then providing generalization bounds in the context of a
modified margin SPO loss function that is Lipschitz continuous. Finally, we
characterize the strength property and show that the modified SPO loss can be
computed efficiently for both strongly convex bodies and polytopes with an
explicit extreme point representation.Comment: Preliminary version in NeurIPS 201
On the decomposition of Generalized Additive Independence models
The GAI (Generalized Additive Independence) model proposed by Fishburn is a
generalization of the additive utility model, which need not satisfy mutual
preferential independence. Its great generality makes however its application
and study difficult. We consider a significant subclass of GAI models, namely
the discrete 2-additive GAI models, and provide for this class a decomposition
into nonnegative monotone terms. This decomposition allows a reduction from
exponential to quadratic complexity in any optimization problem involving
discrete 2-additive models, making them usable in practice
Towards Machine Wald
The past century has seen a steady increase in the need of estimating and
predicting complex systems and making (possibly critical) decisions with
limited information. Although computers have made possible the numerical
evaluation of sophisticated statistical models, these models are still designed
\emph{by humans} because there is currently no known recipe or algorithm for
dividing the design of a statistical model into a sequence of arithmetic
operations. Indeed enabling computers to \emph{think} as \emph{humans} have the
ability to do when faced with uncertainty is challenging in several major ways:
(1) Finding optimal statistical models remains to be formulated as a well posed
problem when information on the system of interest is incomplete and comes in
the form of a complex combination of sample data, partial knowledge of
constitutive relations and a limited description of the distribution of input
random variables. (2) The space of admissible scenarios along with the space of
relevant information, assumptions, and/or beliefs, tend to be infinite
dimensional, whereas calculus on a computer is necessarily discrete and finite.
With this purpose, this paper explores the foundations of a rigorous framework
for the scientific computation of optimal statistical estimators/models and
reviews their connections with Decision Theory, Machine Learning, Bayesian
Inference, Stochastic Optimization, Robust Optimization, Optimal Uncertainty
Quantification and Information Based Complexity.Comment: 37 page
Ergodic Control and Polyhedral approaches to PageRank Optimization
We study a general class of PageRank optimization problems which consist in
finding an optimal outlink strategy for a web site subject to design
constraints. We consider both a continuous problem, in which one can choose the
intensity of a link, and a discrete one, in which in each page, there are
obligatory links, facultative links and forbidden links. We show that the
continuous problem, as well as its discrete variant when there are no
constraints coupling different pages, can both be modeled by constrained Markov
decision processes with ergodic reward, in which the webmaster determines the
transition probabilities of websurfers. Although the number of actions turns
out to be exponential, we show that an associated polytope of transition
measures has a concise representation, from which we deduce that the continuous
problem is solvable in polynomial time, and that the same is true for the
discrete problem when there are no coupling constraints. We also provide
efficient algorithms, adapted to very large networks. Then, we investigate the
qualitative features of optimal outlink strategies, and identify in particular
assumptions under which there exists a "master" page to which all controlled
pages should point. We report numerical results on fragments of the real web
graph.Comment: 39 page
- …