5,804 research outputs found
Load curve data cleansing and imputation via sparsity and low rank
The smart grid vision is to build an intelligent power network with an
unprecedented level of situational awareness and controllability over its
services and infrastructure. This paper advocates statistical inference methods
to robustify power monitoring tasks against the outlier effects owing to faulty
readings and malicious attacks, as well as against missing data due to privacy
concerns and communication errors. In this context, a novel load cleansing and
imputation scheme is developed leveraging the low intrinsic-dimensionality of
spatiotemporal load profiles and the sparse nature of "bad data.'' A robust
estimator based on principal components pursuit (PCP) is adopted, which effects
a twofold sparsity-promoting regularization through an -norm of the
outliers, and the nuclear norm of the nominal load profiles. Upon recasting the
non-separable nuclear norm into a form amenable to decentralized optimization,
a distributed (D-) PCP algorithm is developed to carry out the imputation and
cleansing tasks using networked devices comprising the so-termed advanced
metering infrastructure. If D-PCP converges and a qualification inequality is
satisfied, the novel distributed estimator provably attains the performance of
its centralized PCP counterpart, which has access to all networkwide data.
Computer simulations and tests with real load curve data corroborate the
convergence and effectiveness of the novel D-PCP algorithm.Comment: 8 figures, submitted to IEEE Transactions on Smart Grid - Special
issue on "Optimization methods and algorithms applied to smart grid
Distributed Random Convex Programming via Constraints Consensus
This paper discusses distributed approaches for the solution of random convex
programs (RCP). RCPs are convex optimization problems with a (usually large)
number N of randomly extracted constraints; they arise in several applicative
areas, especially in the context of decision under uncertainty, see [2],[3]. We
here consider a setup in which instances of the random constraints (the
scenario) are not held by a single centralized processing unit, but are
distributed among different nodes of a network. Each node "sees" only a small
subset of the constraints, and may communicate with neighbors. The objective is
to make all nodes converge to the same solution as the centralized RCP problem.
To this end, we develop two distributed algorithms that are variants of the
constraints consensus algorithm [4],[5]: the active constraints consensus (ACC)
algorithm, and the vertex constraints consensus (VCC) algorithm. We show that
the ACC algorithm computes the overall optimal solution in finite time, and
with almost surely bounded communication at each iteration. The VCC algorithm
is instead tailored for the special case in which the constraint functions are
convex also w.r.t. the uncertain parameters, and it computes the solution in a
number of iterations bounded by the diameter of the communication graph. We
further devise a variant of the VCC algorithm, namely quantized vertex
constraints consensus (qVCC), to cope with the case in which communication
bandwidth among processors is bounded. We discuss several applications of the
proposed distributed techniques, including estimation, classification, and
random model predictive control, and we present a numerical analysis of the
performance of the proposed methods. As a complementary numerical result, we
show that the parallel computation of the scenario solution using ACC algorithm
significantly outperforms its centralized equivalent
Past, Present, and Future of Simultaneous Localization And Mapping: Towards the Robust-Perception Age
Simultaneous Localization and Mapping (SLAM)consists in the concurrent
construction of a model of the environment (the map), and the estimation of the
state of the robot moving within it. The SLAM community has made astonishing
progress over the last 30 years, enabling large-scale real-world applications,
and witnessing a steady transition of this technology to industry. We survey
the current state of SLAM. We start by presenting what is now the de-facto
standard formulation for SLAM. We then review related work, covering a broad
set of topics including robustness and scalability in long-term mapping, metric
and semantic representations for mapping, theoretical performance guarantees,
active SLAM and exploration, and other new frontiers. This paper simultaneously
serves as a position paper and tutorial to those who are users of SLAM. By
looking at the published research with a critical eye, we delineate open
challenges and new research issues, that still deserve careful scientific
investigation. The paper also contains the authors' take on two questions that
often animate discussions during robotics conferences: Do robots need SLAM? and
Is SLAM solved
- …