2 research outputs found

    Surrounding the solution of a Linear System of Equations from all sides

    Full text link
    Suppose A∈RnΓ—nA \in \mathbb{R}^{n \times n} is invertible and we are looking for the solution of Ax=bAx = b. Given an initial guess x1∈Rx_1 \in \mathbb{R}, we show that by reflecting through hyperplanes generated by the rows of AA, we can generate an infinite sequence (xk)k=1∞(x_k)_{k=1}^{\infty} such that all elements have the same distance to the solution, i.e. βˆ₯xkβˆ’xβˆ₯=βˆ₯x1βˆ’xβˆ₯\|x_k - x\| = \|x_1 - x\|. If the hyperplanes are chosen at random, averages over the sequence converge and Eβˆ₯xβˆ’1mβˆ‘k=1mxkβˆ₯≀1+βˆ₯Aβˆ₯Fβˆ₯Aβˆ’1βˆ₯mβ‹…βˆ₯xβˆ’x1βˆ₯. \mathbb{E} \left\| x - \frac{1}{m} \sum_{k=1}^{m}{ x_k} \right\| \leq \frac{1 + \|A\|_F \|A^{-1}\|}{\sqrt{m}} \cdot\|x-x_1\|. The bound does not depend on the dimension of the matrix. This introduces a purely geometric way of attacking the problem: are there fast ways of estimating the location of the center of a sphere from knowing many points on the sphere? Our convergence rate (coinciding with that of the Random Kaczmarz method) comes from averaging, can one do better

    A Weighted Randomized Kaczmarz Method for Solving Linear Systems

    Full text link
    The Kaczmarz method for solving a linear system Ax=bAx = b interprets such a system as a collection of equations ⟨ai,x⟩=bi\left\langle a_i, x\right\rangle = b_i, where aia_i is the iβˆ’i-th row of AA, then picks such an equation and corrects xk+1=xk+Ξ»aix_{k+1} = x_k + \lambda a_i where Ξ»\lambda is chosen so that the iβˆ’i-th equation is satisfied. Convergence rates are difficult to establish. Assuming the rows to be normalized, βˆ₯aiβˆ₯β„“2=1\|a_i\|_{\ell^2}=1, Strohmer \& Vershynin established that if the order of equations is chosen at random, EΒ βˆ₯xkβˆ’xβˆ₯β„“2\mathbb{E}~ \|x_k - x\|_{\ell^2} converges exponentially. We prove that if the iβˆ’i-th row is selected with likelihood proportional to ∣⟨ai,xkβŸ©βˆ’bi∣p\left|\left\langle a_i, x_k \right\rangle - b_i\right|^{p}, where 0<p<∞0<p<\infty, then EΒ βˆ₯xkβˆ’xβˆ₯β„“2\mathbb{E}~\|x_k - x\|_{\ell^2} converges faster than the purely random method. As pβ†’βˆžp \rightarrow \infty, the method de-randomizes and explains, among other things, why the maximal correction method works well. We empirically observe that the method computes approximations of small singular vectors of AA as a byproduct
    corecore