2,101 research outputs found
Successive Convexification of Non-Convex Optimal Control Problems and Its Convergence Properties
This paper presents an algorithm to solve non-convex optimal control
problems, where non-convexity can arise from nonlinear dynamics, and non-convex
state and control constraints. This paper assumes that the state and control
constraints are already convex or convexified, the proposed algorithm
convexifies the nonlinear dynamics, via a linearization, in a successive
manner. Thus at each succession, a convex optimal control subproblem is solved.
Since the dynamics are linearized and other constraints are convex, after a
discretization, the subproblem can be expressed as a finite dimensional convex
programming subproblem. Since convex optimization problems can be solved very
efficiently, especially with custom solvers, this subproblem can be solved in
time-critical applications, such as real-time path planning for autonomous
vehicles. Several safe-guarding techniques are incorporated into the algorithm,
namely virtual control and trust regions, which add another layer of
algorithmic robustness. A convergence analysis is presented in continuous- time
setting. By doing so, our convergence results will be independent from any
numerical schemes used for discretization. Numerical simulations are performed
for an illustrative trajectory optimization example.Comment: Updates: corrected wordings for LICQ. This is the full version. A
brief version of this paper is published in 2016 IEEE 55th Conference on
Decision and Control (CDC). http://ieeexplore.ieee.org/document/7798816
Minimal structures for the implementation of digital rational lossless systems
Digital lossless transfer matrices and vectors (power-complementary vectors) are discussed for applications in digital filter bank systems, both single rate and multirate. Two structures for the implementation of rational lossless systems are presented. The first structure represents a characterization of single-input, multioutput lossless systems in terms of complex planar rotations, whereas the second structure offers a representation of M-input, M-output lossless systems in terms of unit-norm vectors. This property makes the second structure desirable in applications that involve optimization of the parameters. Modifications of the second structure for implementing single-input, multioutput, and lossless bounded real (LBR) systems are also included. The main importance of the structures is that they are completely general, i.e. they span the entire set of MĂ1 and MĂM lossless systems. This is demonstrated by showing that any such system can be synthesized using these structures. The structures are also minimal in the sense that they use the smallest number of scalar delays and parameters to implement a lossless system of given degree and dimensions. A design example to demonstrate the main results is included
On the Information Loss of the Max-Log Approximation in BICM Systems
We present a comprehensive study of the information rate loss of the max-log
approximation for -ary pulse-amplitude modulation (PAM) in a bit-interleaved
coded modulation (BICM) system. It is widely assumed that the calculation of
L-values using the max-log approximation leads to an information loss. We prove
that this assumption is correct for all -PAM constellations and labelings
with the exception of a symmetric 4-PAM constellation labeled with a Gray code.
We also show that for max-log L-values, the BICM generalized mutual information
(GMI), which is an achievable rate for a standard BICM decoder, is too
pessimistic. In particular, it is proved that the so-called "harmonized" GMI,
which can be seen as the sum of bit-level GMIs, is achievable without any
modifications to the decoder. We then study how bit-level channel
symmetrization and mixing affect the mutual information (MI) and the GMI for
max-log L-values. Our results show that these operations, which are often used
when analyzing BICM systems, preserve the GMI. However, this is not necessarily
the case when the MI is considered. Necessary and sufficient conditions under
which these operations preserve the MI are provided
Lecture Notes on Network Information Theory
These lecture notes have been converted to a book titled Network Information
Theory published recently by Cambridge University Press. This book provides a
significantly expanded exposition of the material in the lecture notes as well
as problems and bibliographic notes at the end of each chapter. The authors are
currently preparing a set of slides based on the book that will be posted in
the second half of 2012. More information about the book can be found at
http://www.cambridge.org/9781107008731/. The previous (and obsolete) version of
the lecture notes can be found at http://arxiv.org/abs/1001.3404v4/
An Overview of Integral Quadratic Constraints for Delayed Nonlinear and Parameter-Varying Systems
A general framework is presented for analyzing the stability and performance
of nonlinear and linear parameter varying (LPV) time delayed systems. First,
the input/output behavior of the time delay operator is bounded in the
frequency domain by integral quadratic constraints (IQCs). A constant delay is
a linear, time-invariant system and this leads to a simple, intuitive
interpretation for these frequency domain constraints. This simple
interpretation is used to derive new IQCs for both constant and varying delays.
Second, the performance of nonlinear and LPV delayed systems is bounded using
dissipation inequalities that incorporate IQCs. This step makes use of recent
results that show, under mild technical conditions, that an IQC has an
equivalent representation as a finite-horizon time-domain constraint. Numerical
examples are provided to demonstrate the effectiveness of the method for both
class of systems
Generalized â2 synthesis
A framework for optimal controller design with generalized â2 objectives is presented. The allowable disturbances are constrained to be in a pre-specified set; the design objective is to ensure that the resulting output errors do not belong to another pre-specified set. The solution takes the form of an affine matrix inequality (AMI), which is both a necessary
and sufficient condition for the posed problem to have a solution. In the simplest case, the resulting optimization reduces to standard ââ synthesis
- âŠ