17 research outputs found
Energy Storage Optimization for Grid Reliability
Large scale renewable energy integration is being planned for multiple power
grids around the world. To achieve secure and stable grid operations,
additional resources/reserves are needed to mitigate the inherent intermittency
of renewable energy sources (RES). In this paper, we present formulations to
understand the effect of fast storage reserves in improving grid reliability
under different cost functions. Our formulations and solution schemes not only
aim to minimize imbalance but also maintain state-of-charge (SoC) of storage.
In particular, we show that accounting for system response due to inertia and
local governor response enables a more realistic quantification of storage
requirements for damping net load fluctuations. The storage requirement is
significantly lower than values determined when such traditional response are
not accounted for. We demonstrate the performance of our designed policies
through studies using real data from the Elia TSO in Belgium and BPA agency in
the USA. The numerical results enable us to benchmark the marginal effect on
reliability due to increasing storage size under different system responses and
associated cost functions
Model-free Consensus Maximization for Non-Rigid Shapes
Many computer vision methods use consensus maximization to relate
measurements containing outliers with the correct transformation model. In the
context of rigid shapes, this is typically done using Random Sampling and
Consensus (RANSAC) by estimating an analytical model that agrees with the
largest number of measurements (inliers). However, small parameter models may
not be always available. In this paper, we formulate the model-free consensus
maximization as an Integer Program in a graph using `rules' on measurements. We
then provide a method to solve it optimally using the Branch and Bound (BnB)
paradigm. We focus its application on non-rigid shapes, where we apply the
method to remove outlier 3D correspondences and achieve performance superior to
the state of the art. Our method works with outlier ratio as high as 80\%. We
further derive a similar formulation for 3D template to image matching,
achieving similar or better performance compared to the state of the art.Comment: ECCV1
Effects of antiplatelet therapy on stroke risk by brain imaging features of intracerebral haemorrhage and cerebral small vessel diseases: subgroup analyses of the RESTART randomised, open-label trial
Background
Findings from the RESTART trial suggest that starting antiplatelet therapy might reduce the risk of recurrent symptomatic intracerebral haemorrhage compared with avoiding antiplatelet therapy. Brain imaging features of intracerebral haemorrhage and cerebral small vessel diseases (such as cerebral microbleeds) are associated with greater risks of recurrent intracerebral haemorrhage. We did subgroup analyses of the RESTART trial to explore whether these brain imaging features modify the effects of antiplatelet therapy
Communications to the Editor--Exponential Forecasting: Some New Variations
Exponential forecasting or smoothing models are presented which take into account all combinations of trend and cyclical effects in additive and multiplicative form. Although other authors have hinted that the proposed models existed, no one has ever presented them in comprehensive form. This paper presents the nine possible models in graphical form and summarizes them by one summary formula.
The Variable Reduction Method for Nonlinear Programming
A first-order method for solving the problem: minimize f(x) subject to Ax - b \geqq 0 is presented. The method contains ideas based on variable reduction with anti-zig-zagging and acceleration devices based on the Variable Metric Method. Proof of convergence to a Kuhn-Tucker Point, and statement of the rate of convergence when the strict second order sufficiency conditions hold are given.
The Sequential Unconstrained Minimization Technique for Nonlinear Programing, a Primal-Dual Method
This article is based on an idea proposed by C. W. Carroll for transforming a mathematical programming problem into a sequence of unconstrained minimization problems. It describes the theoretical validation of Carroll's proposal for the convex programming problem. A number of important new results are derived that were not originally envisaged: The method generates primal-feasible and dual-feasible points, the primal objective is monotonically decreased, and a subproblem of the original programming problem is solved with each unconstrained minimization. Briefly surveyed is computational experience with a newly developed algorithm that makes the technique competitive with known methodology. (A subsequent article describing the computational algorithm is in preparation.)
Extensions of SUMT for Nonlinear Programming: Equality Constraints and Extrapolation
This paper extends the Sequential Unconstrained Minimization Technique for nonlinear programming to include problems where the constraints are a mixture of inequalities and equalities. Theorems indicating the dual nature of the method and the solution of a subproblem are given. Finally, the theoretical basis for a k-point extrapolation procedure is established.
Computational Algorithm for the Sequential Unconstrained Minimization Technique for Nonlinear Programming
In a previous article [Fiacco, A. V., G. P. McCormick. 1964. The sequential unconstrained minimization technique for nonlinear programming, a primal-dual method. Management Sci. 10(2) 360-366.] the authors gave the theoretical validation of the sequential unconstrained minimization technique for solving the convex programming problem. The technique is based on an idea proposed by C. W. Carroll [Carroll, C. W. 1961. The created response surface technique for optimizing nonlinear restrained systems. Oper. Res. 9(2) 169-184; Carroll, C. W. 1959. An operations research approach to the economic optimization of a Kraft Pulping Process. Doctoral dissertation, The Institute of Paper Chemistry, Appleton, Wisc.]. The method has been implemented via an algorithm based on a second-order gradient technique that has proved extremely efficient on a considerable number of test problems of varying complexity. This paper explores the computational aspects of the method. Included are discussions of parameter selection, convergence criteria, and methods of minimizing an unconstrained function. It is shown that the problem variables, on the trajectory of minima of the sequence of unconstrained functions, can be developed as functions of a single parameter. This provides the theoretical basis for an extrapolation technique that significantly accelerates convergence in actual computations. The detailed computer solution of a small example is given to illustrate the typical convergence characteristics of the method. The speed and accuracy of the computational procedure are believed to be competitive with other known techniques for solving the convex programming problem.