1,672 research outputs found
Regularizing Portfolio Optimization
The optimization of large portfolios displays an inherent instability to
estimation error. This poses a fundamental problem, because solutions that are
not stable under sample fluctuations may look optimal for a given sample, but
are, in effect, very far from optimal with respect to the average risk. In this
paper, we approach the problem from the point of view of statistical learning
theory. The occurrence of the instability is intimately related to over-fitting
which can be avoided using known regularization methods. We show how
regularized portfolio optimization with the expected shortfall as a risk
measure is related to support vector regression. The budget constraint dictates
a modification. We present the resulting optimization problem and discuss the
solution. The L2 norm of the weight vector is used as a regularizer, which
corresponds to a diversification "pressure". This means that diversification,
besides counteracting downward fluctuations in some assets by upward
fluctuations in others, is also crucial because it improves the stability of
the solution. The approach we provide here allows for the simultaneous
treatment of optimization and diversification in one framework that enables the
investor to trade-off between the two, depending on the size of the available
data set
Second Order Differences of Cyclic Data and Applications in Variational Denoising
In many image and signal processing applications, as interferometric
synthetic aperture radar (SAR), electroencephalogram (EEG) data analysis or
color image restoration in HSV or LCh spaces the data has its range on the
one-dimensional sphere . Although the minimization of total
variation (TV) regularized functionals is among the most popular methods for
edge-preserving image restoration such methods were only very recently applied
to cyclic structures. However, as for Euclidean data, TV regularized
variational methods suffer from the so called staircasing effect. This effect
can be avoided by involving higher order derivatives into the functional.
This is the first paper which uses higher order differences of cyclic data in
regularization terms of energy functionals for image restoration. We introduce
absolute higher order differences for -valued data in a sound way
which is independent of the chosen representation system on the circle. Our
absolute cyclic first order difference is just the geodesic distance between
points. Similar to the geodesic distances the absolute cyclic second order
differences have only values in [0,{\pi}]. We update the cyclic variational TV
approach by our new cyclic second order differences. To minimize the
corresponding functional we apply a cyclic proximal point method which was
recently successfully proposed for Hadamard manifolds. Choosing appropriate
cycles this algorithm can be implemented in an efficient way. The main steps
require the evaluation of proximal mappings of our cyclic differences for which
we provide analytical expressions. Under certain conditions we prove the
convergence of our algorithm. Various numerical examples with artificial as
well as real-world data demonstrate the advantageous performance of our
algorithm.Comment: 32 pages, 16 figures, shortened version of submitted manuscrip
On some knot energies involving Menger curvature
We investigate knot-theoretic properties of geometrically defined curvature
energies such as integral Menger curvature. Elementary radii-functions, such as
the circumradius of three points, generate a family of knot energies
guaranteeing self-avoidance and a varying degree of higher regularity of finite
energy curves. All of these energies turn out to be charge, minimizable in
given isotopy classes, tight and strong. Almost all distinguish between knots
and unknots, and some of them can be shown to be uniquely minimized by round
circles. Bounds on the stick number and the average crossing number, some
non-trivial global lower bounds, and unique minimization by circles upon
compaction complete the picture.Comment: 31 pages, 4 figures; version 2 with minor changes and modification
A Second Order TV-type Approach for Inpainting and Denoising Higher Dimensional Combined Cyclic and Vector Space Data
In this paper we consider denoising and inpainting problems for higher
dimensional combined cyclic and linear space valued data. These kind of data
appear when dealing with nonlinear color spaces such as HSV, and they can be
obtained by changing the space domain of, e.g., an optical flow field to polar
coordinates. For such nonlinear data spaces, we develop algorithms for the
solution of the corresponding second order total variation (TV) type problems
for denoising, inpainting as well as the combination of both. We provide a
convergence analysis and we apply the algorithms to concrete problems.Comment: revised submitted versio
L1TV computes the flat norm for boundaries
We show that the recently introduced L1TV functional can be used to
explicitly compute the flat norm for co-dimension one boundaries. While this
observation alone is very useful, other important implications for image
analysis and shape statistics include a method for denoising sets which are not
boundaries or which have higher co-dimension and the fact that using the flat
norm to compute distances not only gives a distance, but also an informative
decomposition of the distance. This decomposition is made to depend on scale
using the "flat norm with scale" which we define in direct analogy to the L1TV
functional. We illustrate the results and implications with examples and
figures
ํฉ์ฑ๊ณฑ ์ปค๋ ์ ๊ทํ๋ฅผ ์ํ ๊ณ ๋ฅธ ๊ฐ๋๋ถ์ฐ๋ฐฉ๋ฒ
ํ์๋
ผ๋ฌธ(๋ฐ์ฌ) -- ์์ธ๋ํ๊ต๋ํ์ : ์์ฐ๊ณผํ๋ํ ์๋ฆฌ๊ณผํ๋ถ, 2022. 8. ๊ฐ๋ช
์ฃผ.In this thesis, we propose new convolutional kernel regularization methods. Along with the development of deep learning, there have been attempts to effectively regularize a convolutional layer, which is an important basic module of deep neural networks. Convolutional neural networks (CNN) are excellent at abstracting input data, but deepening causes gradient vanishing or explosion issues and produces redundant features. An approach to solve these issues is to directly regularize convolutional kernel weights of CNN. Its basic idea is to convert a convolutional kernel weight into a matrix and make the row or column vectors of the matrix orthogonal. However, this approach has some shortcomings. Firstly, it requires appropriate manipulation because overcomlete issue occurs when the number of vectors is larger than the dimension of vectors. As a method to deal with this issue, we define the concept of evenly dispersed state and propose PH0 and MST regularizations using this. Secondly, prior regularizations which enforce the Gram matrix of a matrix to be an identity matrix might not be an optimal approach for orthogonality of the matrix. We point out that these rather reduces the update of angles between some two vectors when two vectors are adjacent. Therefore, to complement for this issue, we propose EADK and EADC regularizations which update directly the angle. Through various experiments, we demonstrate that EADK and EADC regularizations outperform prior methods in some neural network architectures and, in particular, EADK has fast learning time.์ด ๋
ผ๋ฌธ์์๋ ํฉ์ฑ๊ณฑ์ปค๋์ ๋ํ ์๋ก์ด ์ ๊ทํ ๋ฐฉ๋ฒ๋ค์ ์ ์ํ๋ค. ๋ฅ๋ฌ๋์ ๋ฐ๋ฌ๊ณผ ๋๋ถ์ด ์ ๊ฒฝ๋ง์ ๊ฐ์ฅ ๊ธฐ๋ณธ์ ์ธ ๋ชจ๋์ธ ํฉ์ฑ๊ณฑ ๋ ์ด์ด๋ฅผ ํจ๊ณผ์ ์ผ๋ก ์ ๊ทํ ํ๋ ค๋ ์๋๋ค์ด ์์ด ์๋ค. ํฉ์ฑ๊ณฑ์ ๊ฒฝ๋ง๋ ์ธํ๋ฐ์ดํฐ๋ฅผ ์ถ์ํํ๋๋ฐ ํ์ํ์ง๋ง ๋คํธ์ํฌ์ ๊น์ด๊ฐ ๊น์ด์ง๋ฉด ๊ทธ๋ ๋์ธํธ ์๋ฉธ์ด๋ ํญ๋ฐ ๋ฌธ์ ๋ฅผ ์ผ์ผํค๊ณ ์ค๋ณต๋ ํผ์ณ๋ค์ ๋ง๋ ๋ค. ์ด๋ฌํ ๋ฌธ์ ๋ค์ ํด๊ฒฐํ๊ธฐ ์ํ ์ ๊ทผ๋ฒ ์ค ํ๋๋ ์ง์ ํฉ์ฑ๊ณฑ ์ ๊ฒฝ๋ง์ ํฉ์ฑ๊ณฑ์ปค๋์ ์ง์ ์ ๊ทํ ํ๋ ๊ฒ์ด๋ค. ์ด ๋ฐฉ๋ฒ์ ํฉ์ฑ๊ณฑ์ปค๋์ ์ด๋ค ํ๋ ฌ๋ก ๋ณํํ๊ณ ํ๋ ฌ์ ํ ๋๋ ์ด๋ค์ ๋ฒกํฐ๋ค์ ์ง๊ต์ํค๋ ๊ฒ์ด๋ค. ๊ทธ๋ฌ๋ ์ด๋ฌํ ์ ๊ทผ๋ฒ์ ๋ช๊ฐ์ง ๋จ์ ์ด ์๋ค. ์ฒซ์งธ๋ก, ๋ฒกํฐ์ ์๊ฐ ๋ฒกํฐ์ ์ฐจ์๋ณด๋ค ๋ง์ ๋๋ ๋ชจ๋ ๋ฒกํฐ๋ฅผ ์ง๊ตํ ์ํฌ ์ ์๊ฒ ๋๋ฏ๋ก ์ ์ ํ ๊ธฐ๋ฒ๋ค์ ํ์๋ก ํ๋ค. ์ด ๋ฌธ์ ๋ฅผ ๋ค๋ฃจ๊ธฐ ์ํ ํ ๊ฐ์ง ๋ฐฉ๋ฒ์ผ๋ก ์ฐ๋ฆฌ๋ ๋ถ์ฐ ์ํ๋ผ๋ ๊ฐ๋
์ ์ ์ํ๊ณ ์ด ๊ฐ๋
์ ํ์ฉํ PH0์ MST ์ ๊ทํ๋ฒ์ ์ ์ํ๋ค. ๋์งธ๋ก, ๊ทธ๋ํ๋ ฌ์ ํญ๋ฑํ๋ ฌ๋ก ๊ทผ์ฌ์ํค๋ ๋ฐฉ๋ฒ์ ์ฌ์ฉํ๋ ๊ธฐ์กด ์ ๊ทํ๋ฒ์ด ๋ฒกํฐ๋ค์ ์ง๊ตํ์ํค๋ ์ต์ ์ ๋ฐฉ๋ฒ์ด ์๋ ์ ์๋ค๋ ์ ์ด๋ค. ์ฆ, ๊ธฐ์กด์ ์ ๊ทํ๋ฒ์ด ๋ ๋ฒกํฐ๊ฐ ๊ฐ๊น์ธ ๋๋ ์คํ๋ ค ๊ฐ๋์ ์
๋ฐ์ดํธ๋ฅผ ์ค์ด๊ฒ ๋๋ค.๋ฐ๋ผ์ ์ด๋ฅผ ๋ณด์ํ๊ธฐ ์ํ์ฌ ์ฐ๋ฆฌ๋ ๊ฐ๋๋ฅผ ์ง์ ์
๋ฐ์ดํธํ๋ EADK์ EADC ์ ๊ทํ๋ฒ์ ์ ์ํ๋ค. ๊ทธ๋ฆฌ๊ณ ๋ค์ํ ์คํ์ ํตํด EADK์ EADC ์ ๊ทํ๋ฒ์ด ๋ค์์ ์ ๊ฒฝ๋ง๊ตฌ์กฐ์์ ๊ธฐ์กด์ ๋ฐฉ๋ฒ๋ค๋ณด๋ค ์ฐ์ํ ์ฑ๋ฅ์ ๋ณด์ด๊ณ ํนํ EADK๋ ๋น ๋ฅธ ํ์ต์๊ฐ์ ๊ฐ์ง๋ค๋ ๊ฒ์ ํ์ธํ๋ค.Abstract i
1 Introduction 1
2 Preliminaries 4
2.1 Two Ways of Understanding CNN Layers as Matrix Operations 5
2.1.1 Kernel Matrix 6
2.1.2 Convolution Matrix 7
2.2 Soft Orthogonality 11
2.2.1 SO Regularization 11
2.2.2 DSO Regularization 12
2.3 Mutual Coherence 13
2.3.1 MC Regularization 13
2.4 Spectral Restricted Isometry Property 13
2.4.1 Restricted Isometry Property 13
2.4.2 SRIP Regularization 15
2.5 Orthogonal Convolutional Neural Networks 18
2.5.1 OCNN Regularizaiton 18
3 Topological Dispersing Regularizations 22
3.1 Evenly Dispersed State 23
3.1.1 Dispersing Vectors on Sphere 23
3.1.2 Evenly Dispersed State in the Real Projective Spaces 25
3.2 Persistent Homology Regularization 33
3.2.1 Cech and Vietoris-Rips Complexes 35
3.2.2 Persistent Homology 36
3.2.3 PH0 Regularization 38
3.3 Minimum Spanning Tree Regularization 39
3.3.1 Minimum Spanning Tree 39
3.3.2 MST Regularization 41
4 Evenly Angle Dispersing Regularizations 42
4.1 Analysis of Soft Orthogonality 43
4.1.1 Analysis of Soft Orthogonality 43
4.2 Evenly Angle Dispersing Regularizations 47
4.2.1 Evenly Angle Dispersing Regularization with Kernel Matrix 47
4.2.2 Evenly Angle Dispersing Regularization with Convolution Matrix 52
5 Algorithms & Experiments 54
5.1 Algorithms 55
5.1.1 PH0 and MST 55
5.1.2 EADK 57
5.1.3 EADC 58
5.2 Experiments 59
5.2.1 Analysis for Angle Dispersing 59
5.2.2 Experimental Setups 62
5.2.3 Classification Accuracy 68
5.2.4 Additional Experiments 76
6 Conclusion 78
The bibliography 80
Abstract (in Korean) 85๋ฐ
- โฆ