119 research outputs found
Separable and Low-Rank Continuous Games
In this paper, we study nonzero-sum separable games, which are continuous
games whose payoffs take a sum-of-products form. Included in this subclass are
all finite games and polynomial games. We investigate the structure of
equilibria in separable games. We show that these games admit finitely
supported Nash equilibria. Motivated by the bounds on the supports of mixed
equilibria in two-player finite games in terms of the ranks of the payoff
matrices, we define the notion of the rank of an n-player continuous game and
use this to provide bounds on the cardinality of the support of equilibrium
strategies. We present a general characterization theorem that states that a
continuous game has finite rank if and only if it is separable. Using our rank
results, we present an efficient algorithm for computing approximate equilibria
of two-player separable games with fixed strategy spaces in time polynomial in
the rank of the game
A Robust Solution Procedure for Hyperelastic Solids with Large Boundary Deformation
Compressible Mooney-Rivlin theory has been used to model hyperelastic solids,
such as rubber and porous polymers, and more recently for the modeling of soft
tissues for biomedical tissues, undergoing large elastic deformations. We
propose a solution procedure for Lagrangian finite element discretization of a
static nonlinear compressible Mooney-Rivlin hyperelastic solid. We consider the
case in which the boundary condition is a large prescribed deformation, so that
mesh tangling becomes an obstacle for straightforward algorithms. Our solution
procedure involves a largely geometric procedure to untangle the mesh: solution
of a sequence of linear systems to obtain initial guesses for interior nodal
positions for which no element is inverted. After the mesh is untangled, we
take Newton iterations to converge to a mechanical equilibrium. The Newton
iterations are safeguarded by a line search similar to one used in
optimization. Our computational results indicate that the algorithm is up to 70
times faster than a straightforward Newton continuation procedure and is also
more robust (i.e., able to tolerate much larger deformations). For a few
extremely large deformations, the deformed mesh could only be computed through
the use of an expensive Newton continuation method while using a tight
convergence tolerance and taking very small steps.Comment: Revision of earlier version of paper. Submitted for publication in
Engineering with Computers on 9 September 2010. Accepted for publication on
20 May 2011. Published online 11 June 2011. The final publication is
available at http://www.springerlink.co
Perspective Cuts for the ACOPF with Generators
International audienceThe alternating current optimal power flow problem is a fundamental problem in the management of smart grids. In this paper we consider a variant which includes activation/deactivation of generators at some of the grid sites. We formulate the problem as a mathematical program, prove its NP-hardness w.r.t. ac-tivation/deactivation, and derive two perspective reformulations
Analysis of and workarounds for element reversal for a finite element-based algorithm for warping triangular and tetrahedral meshes
We consider an algorithm called FEMWARP for warping triangular and
tetrahedral finite element meshes that computes the warping using the finite
element method itself. The algorithm takes as input a two- or three-dimensional
domain defined by a boundary mesh (segments in one dimension or triangles in
two dimensions) that has a volume mesh (triangles in two dimensions or
tetrahedra in three dimensions) in its interior. It also takes as input a
prescribed movement of the boundary mesh. It computes as output updated
positions of the vertices of the volume mesh. The first step of the algorithm
is to determine from the initial mesh a set of local weights for each interior
vertex that describes each interior vertex in terms of the positions of its
neighbors. These weights are computed using a finite element stiffness matrix.
After a boundary transformation is applied, a linear system of equations based
upon the weights is solved to determine the final positions of the interior
vertices. The FEMWARP algorithm has been considered in the previous literature
(e.g., in a 2001 paper by Baker). FEMWARP has been succesful in computing
deformed meshes for certain applications. However, sometimes FEMWARP reverses
elements; this is our main concern in this paper. We analyze the causes for
this undesirable behavior and propose several techniques to make the method
more robust against reversals. The most successful of the proposed methods
includes combining FEMWARP with an optimization-based untangler.Comment: Revision of earlier version of paper. Submitted for publication in
BIT Numerical Mathematics on 27 April 2010. Accepted for publication on 7
September 2010. Published online on 9 October 2010. The final publication is
available at http://www.springerlink.co
Proceedings of the second "international Traveling Workshop on Interactions between Sparse models and Technology" (iTWIST'14)
The implicit objective of the biennial "international - Traveling Workshop on
Interactions between Sparse models and Technology" (iTWIST) is to foster
collaboration between international scientific teams by disseminating ideas
through both specific oral/poster presentations and free discussions. For its
second edition, the iTWIST workshop took place in the medieval and picturesque
town of Namur in Belgium, from Wednesday August 27th till Friday August 29th,
2014. The workshop was conveniently located in "The Arsenal" building within
walking distance of both hotels and town center. iTWIST'14 has gathered about
70 international participants and has featured 9 invited talks, 10 oral
presentations, and 14 posters on the following themes, all related to the
theory, application and generalization of the "sparsity paradigm":
Sparsity-driven data sensing and processing; Union of low dimensional
subspaces; Beyond linear and convex inverse problem; Matrix/manifold/graph
sensing/processing; Blind inverse problems and dictionary learning; Sparsity
and computational neuroscience; Information theory, geometry and randomness;
Complexity/accuracy tradeoffs in numerical methods; Sparsity? What's next?;
Sparse machine learning and inference.Comment: 69 pages, 24 extended abstracts, iTWIST'14 website:
http://sites.google.com/site/itwist1
An FPTAS for optimizing a class of low-rank functions over a polytope
We present a fully polynomial time approximation scheme (FPTAS) for optimizing a very general class of non-linear functions of low rank over a polytope. Our approximation scheme relies on constructing an approximate Pareto-optimal front of the linear functions which constitute the given low-rank function. In contrast to existing results in the literature, our approximation scheme does not require the assumption of quasi-concavity on the objective function. For the special case of quasi-concave function minimization, we give an alternative FPTAS, which always returns a solution which is an extreme point of the polytope. Our technique can also be used to obtain an FPTAS for combinatorial optimization problems with non-linear objective functions, for example when the objective is a product of a fixed number of linear functions. We also show that it is not possible to approximate the minimum of a general concave function over the unit hypercube to within any factor, unless P = NP. We prove this by showing a similar hardness of approximation result for supermodular function minimization, a result that may be of independent interest
Central Path Curvature and Iteration-Complexity for Redundant Klee—Minty Cubes
We consider a family of linear optimization problems over the n-dimensional Klee—Minty cube and show that the central path may visit all of its vertices in the same order as simplex methods do. This is achieved by carefully adding an exponential number of redundant constraints that forces the central path to take at least 2n − 2 sharp turns. This fact sug-gests that any feasible path-following interior-point method will take at least O(2n) iterations to solve this problem, whereas in practice typically only a few iterations (e.g., 50) suffices to obtain a high-quality solution. Thus, the construction potentially exhibits the worst-case iteration-complexity known to date which almost matches the theoretical iteration-complexity bound for this type of methods. In addition, this construction gives a counterexample to a conjecture that the total central path curvature is O(n)
- …