117 research outputs found

    Polynomial Time and Private Learning of Unbounded Gaussian Mixture Models

    Full text link
    We study the problem of privately estimating the parameters of dd-dimensional Gaussian Mixture Models (GMMs) with kk components. For this, we develop a technique to reduce the problem to its non-private counterpart. This allows us to privatize existing non-private algorithms in a blackbox manner, while incurring only a small overhead in the sample complexity and running time. As the main application of our framework, we develop an (ฮต,ฮด)(\varepsilon, \delta)-differentially private algorithm to learn GMMs using the non-private algorithm of Moitra and Valiant [MV10] as a blackbox. Consequently, this gives the first sample complexity upper bound and first polynomial time algorithm for privately learning GMMs without any boundedness assumptions on the parameters. As part of our analysis, we prove a tight (up to a constant factor) lower bound on the total variation distance of high-dimensional Gaussians which can be of independent interest.Comment: Accepted in ICML 202

    Private Distribution Learning with Public Data: The View from Sample Compression

    Full text link
    We study the problem of private distribution learning with access to public data. In this setup, which we refer to as public-private learning, the learner is given public and private samples drawn from an unknown distribution pp belonging to a class Q\mathcal Q, with the goal of outputting an estimate of pp while adhering to privacy constraints (here, pure differential privacy) only with respect to the private samples. We show that the public-private learnability of a class Q\mathcal Q is connected to the existence of a sample compression scheme for Q\mathcal Q, as well as to an intermediate notion we refer to as list learning. Leveraging this connection: (1) approximately recovers previous results on Gaussians over Rd\mathbb R^d; and (2) leads to new ones, including sample complexity upper bounds for arbitrary kk-mixtures of Gaussians over Rd\mathbb R^d, results for agnostic and distribution-shift resistant learners, as well as closure properties for public-private learnability under taking mixtures and products of distributions. Finally, via the connection to list learning, we show that for Gaussians in Rd\mathbb R^d, at least dd public samples are necessary for private learnability, which is close to the known upper bound of d+1d+1 public samples.Comment: 31 page

    Mixtures of Gaussians are Privately Learnable with a Polynomial Number of Samples

    Full text link
    We study the problem of estimating mixtures of Gaussians under the constraint of differential privacy (DP). Our main result is that O~(k2d4logโก(1/ฮด)/ฮฑ2ฮต)\tilde{O}(k^2 d^4 \log(1/\delta) / \alpha^2 \varepsilon) samples are sufficient to estimate a mixture of kk Gaussians up to total variation distance ฮฑ\alpha while satisfying (ฮต,ฮด)(\varepsilon, \delta)-DP. This is the first finite sample complexity upper bound for the problem that does not make any structural assumptions on the GMMs. To solve the problem, we devise a new framework which may be useful for other tasks. On a high level, we show that if a class of distributions (such as Gaussians) is (1) list decodable and (2) admits a "locally small'' cover (Bun et al., 2021) with respect to total variation distance, then the class of its mixtures is privately learnable. The proof circumvents a known barrier indicating that, unlike Gaussians, GMMs do not admit a locally small cover (Aden-Ali et al., 2021b)

    Private hypothesis selection

    Full text link
    We provide a differentially private algorithm for hypothesis selection. Given samples from an unknown probability distribution P and a set of m probability distributions H, the goal is to output, in a ฮต-differentially private manner, a distribution from H whose total variation distance to P is comparable to that of the best such distribution (which we denote by ฮฑ). The sample complexity of our basic algorithm is O(log m/ฮฑ^2 + log m/ฮฑฮต), representing a minimal cost for privacy when compared to the non-private algorithm. We also can handle infinite hypothesis classes H by relaxing to (ฮต, ฮด)-differential privacy. We apply our hypothesis selection algorithm to give learning algorithms for a number of natural distribution classes, including Gaussians, product distributions, sums of independent random variables, piecewise polynomials, and mixture classes. Our hypothesis selection procedure allows us to generically convert a cover for a class to a learning algorithm, complementing known learning lower bounds which are in terms of the size of the packing number of the class. As the covering and packing numbers are often closely related, for constant ฮฑ, our algorithms achieve the optimal sample complexity for many classes of interest. Finally, we describe an application to private distribution-free PAC learning.https://arxiv.org/abs/1905.1322

    A Polynomial Time, Pure Differentially Private Estimator for Binary Product Distributions

    Full text link
    We present the first ฮต\varepsilon-differentially private, computationally efficient algorithm that estimates the means of product distributions over {0,1}d\{0,1\}^d accurately in total-variation distance, whilst attaining the optimal sample complexity to within polylogarithmic factors. The prior work had either solved this problem efficiently and optimally under weaker notions of privacy, or had solved it optimally while having exponential running times

    ๋ฏผ๊ฐํ•œ ์ •๋ณด๋ฅผ ๋ณดํ˜ธํ•  ์ˆ˜ ์žˆ๋Š” ํ”„๋ผ์ด๋ฒ„์‹œ ๋ณด์กด ๊ธฐ๊ณ„ํ•™์Šต ๊ธฐ์ˆ  ๊ฐœ๋ฐœ

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ(๋ฐ•์‚ฌ) -- ์„œ์šธ๋Œ€ํ•™๊ต๋Œ€ํ•™์› : ๊ณต๊ณผ๋Œ€ํ•™ ์‚ฐ์—…๊ณตํ•™๊ณผ, 2022. 8. ์ด์žฌ์šฑ.์ตœ๊ทผ ์ธ๊ณต์ง€๋Šฅ์˜ ์„ฑ๊ณต์—๋Š” ์—ฌ๋Ÿฌ ๊ฐ€์ง€ ์š”์ธ์ด ์žˆ์œผ๋‚˜, ์ƒˆ๋กœ์šด ์•Œ๊ณ ๋ฆฌ์ฆ˜์˜ ๊ฐœ๋ฐœ๊ณผ ์ •์ œ๋œ ๋ฐ์ดํ„ฐ ์–‘์˜ ๊ธฐํ•˜๊ธ‰์ˆ˜์ ์ธ ์ฆ๊ฐ€๋กœ ์ธํ•œ ์˜ํ–ฅ์ด ํฌ๋‹ค. ๋”ฐ๋ผ์„œ ๊ธฐ๊ณ„ํ•™์Šต ๋ชจ๋ธ๊ณผ ๋ฐ์ดํ„ฐ๋Š” ์‹ค์žฌ์  ๊ฐ€์น˜๋ฅผ ๊ฐ€์ง€๊ฒŒ ๋˜๋ฉฐ, ํ˜„์‹ค ์„ธ๊ณ„์—์„œ ๊ฐœ์ธ ๋˜๋Š” ๊ธฐ์—…์€ ํ•™์Šต๋œ ๋ชจ๋ธ ๋˜๋Š” ํ•™์Šต์— ์‚ฌ์šฉํ•  ๋ฐ์ดํ„ฐ๋ฅผ ์ œ๊ณตํ•จ์œผ๋กœ์จ ์ด์ต์„ ์–ป์„ ์ˆ˜ ์žˆ๋‹ค. ๊ทธ๋Ÿฌ๋‚˜, ๋ฐ์ดํ„ฐ ๋˜๋Š” ๋ชจ๋ธ์˜ ๊ณต์œ ๋Š” ๊ฐœ์ธ์˜ ๋ฏผ๊ฐ ์ •๋ณด๋ฅผ ์œ ์ถœํ•จ์œผ๋กœ์จ ํ”„๋ผ์ด๋ฒ„์‹œ์˜ ์นจํ•ด๋กœ ์ด์–ด์งˆ ์ˆ˜ ์žˆ๋‹ค๋Š” ์‚ฌ์‹ค์ด ๋ฐํ˜€์ง€๊ณ  ์žˆ๋‹ค. ๋ณธ ๋…ผ๋ฌธ์˜ ๋ชฉํ‘œ๋Š” ๋ฏผ๊ฐ ์ •๋ณด๋ฅผ ๋ณดํ˜ธํ•  ์ˆ˜ ์žˆ๋Š” ํ”„๋ผ์ด๋ฒ„์‹œ ๋ณด์กด ๊ธฐ๊ณ„ํ•™์Šต ๋ฐฉ๋ฒ•๋ก ์„ ๊ฐœ๋ฐœํ•˜๋Š” ๊ฒƒ์ด๋‹ค. ์ด๋ฅผ ์œ„ํ•ด ์ตœ๊ทผ ํ™œ๋ฐœํžˆ ์—ฐ๊ตฌ๋˜๊ณ  ์žˆ๋Š” ๋‘ ๊ฐ€์ง€ ํ”„๋ผ์ด๋ฒ„์‹œ ๋ณด์กด ๊ธฐ์ˆ , ์ฆ‰ ๋™ํ˜• ์•”ํ˜ธ์™€ ์ฐจ๋ถ„ ํ”„๋ผ์ด๋ฒ„์‹œ๋ฅผ ์‚ฌ์šฉํ•œ๋‹ค. ๋จผ์ €, ๋™ํ˜• ์•”ํ˜ธ๋Š” ์•”ํ˜ธํ™”๋œ ๋ฐ์ดํ„ฐ์— ๋Œ€ํ•ด ๊ธฐ๊ณ„ํ•™์Šต ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ์ ์šฉ ๊ฐ€๋Šฅํ•˜๊ฒŒ ํ•จ์œผ๋กœ์จ ๋ฐ์ดํ„ฐ์˜ ํ”„๋ผ์ด๋ฒ„์‹œ๋ฅผ ๋ณดํ˜ธํ•  ์ˆ˜ ์žˆ๋‹ค. ๊ทธ๋Ÿฌ๋‚˜ ๋™ํ˜• ์•”ํ˜ธ๋ฅผ ํ™œ์šฉํ•œ ์—ฐ์‚ฐ์€ ๊ธฐ์กด์˜ ์—ฐ์‚ฐ์— ๋น„ํ•ด ๋งค์šฐ ํฐ ์—ฐ์‚ฐ ์‹œ๊ฐ„์„ ์š”๊ตฌํ•˜๋ฏ€๋กœ ํšจ์œจ์ ์ธ ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ๊ตฌ์„ฑํ•˜๋Š” ๊ฒƒ์ด ์ค‘์š”ํ•˜๋‹ค. ํšจ์œจ์ ์ธ ์—ฐ์‚ฐ์„ ์œ„ํ•ด ์šฐ๋ฆฌ๋Š” ๋‘ ๊ฐ€์ง€ ์ ‘๊ทผ๋ฒ•์„ ์‚ฌ์šฉํ•œ๋‹ค. ์ฒซ ๋ฒˆ์งธ๋Š” ํ•™์Šต ๋‹จ๊ณ„์—์„œ์˜ ์—ฐ์‚ฐ๋Ÿ‰์„ ์ค„์ด๋Š” ๊ฒƒ์ด๋‹ค. ํ•™์Šต ๋‹จ๊ณ„์—์„œ๋ถ€ํ„ฐ ๋™ํ˜• ์•”ํ˜ธ๋ฅผ ์ ์šฉํ•˜๋ฉด ํ•™์Šต ๋ฐ์ดํ„ฐ์˜ ํ”„๋ผ์ด๋ฒ„์‹œ๋ฅผ ํ•จ๊ป˜ ๋ณดํ˜ธํ•  ์ˆ˜ ์žˆ์œผ๋ฏ€๋กœ ์ถ”๋ก  ๋‹จ๊ณ„์—์„œ๋งŒ ๋™ํ˜• ์•”ํ˜ธ๋ฅผ ์ ์šฉํ•˜๋Š” ๊ฒƒ์— ๋น„ํ•ด ํ”„๋ผ์ด๋ฒ„์‹œ์˜ ๋ฒ”์œ„๊ฐ€ ๋„“์–ด์ง€์ง€๋งŒ, ๊ทธ๋งŒํผ ์—ฐ์‚ฐ๋Ÿ‰์ด ๋Š˜์–ด๋‚œ๋‹ค. ๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” ์ผ๋ถ€ ๊ฐ€์žฅ ์ค‘์š”ํ•œ ์ •๋ณด๋งŒ์„ ์•”ํ˜ธํ™”ํ•จ์œผ๋กœ์จ ํ•™์Šต ๋‹จ๊ณ„๋ฅผ ํšจ์œจ์ ์œผ๋กœ ํ•˜๋Š” ๋ฐฉ๋ฒ•๋ก ์„ ์ œ์•ˆํ•œ๋‹ค. ๊ตฌ์ฒด์ ์œผ๋กœ, ์ผ๋ถ€ ๋ฏผ๊ฐ ๋ณ€์ˆ˜๊ฐ€ ์•”ํ˜ธํ™”๋˜์–ด ์žˆ์„ ๋•Œ ์—ฐ์‚ฐ๋Ÿ‰์„ ๋งค์šฐ ์ค„์ผ ์ˆ˜ ์žˆ๋Š” ๋ฆฟ์ง€ ํšŒ๊ท€ ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ๊ฐœ๋ฐœํ•œ๋‹ค. ๋˜ํ•œ ๊ฐœ๋ฐœ๋œ ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ํ™•์žฅ์‹œ์ผœ ๋™ํ˜• ์•”ํ˜ธ ์นœํ™”์ ์ด์ง€ ์•Š์€ ํŒŒ๋ผ๋ฏธํ„ฐ ํƒ์ƒ‰ ๊ณผ์ •์„ ์ตœ๋Œ€ํ•œ ์ œ๊ฑฐํ•  ์ˆ˜ ์žˆ๋Š” ์ƒˆ๋กœ์šด ๋กœ์ง€์Šคํ‹ฑ ํšŒ๊ท€ ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ํ•จ๊ป˜ ์ œ์•ˆํ•œ๋‹ค. ํšจ์œจ์ ์ธ ์—ฐ์‚ฐ์„ ์œ„ํ•œ ๋‘ ๋ฒˆ์งธ ์ ‘๊ทผ๋ฒ•์€ ๋™ํ˜• ์•”ํ˜ธ๋ฅผ ๊ธฐ๊ณ„ํ•™์Šต์˜ ์ถ”๋ก  ๋‹จ๊ณ„์—์„œ๋งŒ ์‚ฌ์šฉํ•˜๋Š” ๊ฒƒ์ด๋‹ค. ์ด๋ฅผ ํ†ตํ•ด ์‹œํ—˜ ๋ฐ์ดํ„ฐ์˜ ์ง์ ‘์ ์ธ ๋…ธ์ถœ์„ ๋ง‰์„ ์ˆ˜ ์žˆ๋‹ค. ๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” ์„œํฌํŠธ ๋ฒกํ„ฐ ๊ตฐ์ง‘ํ™” ๋ชจ๋ธ์— ๋Œ€ํ•œ ๋™ํ˜• ์•”ํ˜ธ ์นœํ™”์  ์ถ”๋ก  ๋ฐฉ๋ฒ•์„ ์ œ์•ˆํ•œ๋‹ค. ๋™ํ˜• ์•”ํ˜ธ๋Š” ์—ฌ๋Ÿฌ ๊ฐ€์ง€ ์œ„ํ˜‘์— ๋Œ€ํ•ด์„œ ๋ฐ์ดํ„ฐ์™€ ๋ชจ๋ธ ์ •๋ณด๋ฅผ ๋ณดํ˜ธํ•  ์ˆ˜ ์žˆ์œผ๋‚˜, ํ•™์Šต๋œ ๋ชจ๋ธ์„ ํ†ตํ•ด ์ƒˆ๋กœ์šด ๋ฐ์ดํ„ฐ์— ๋Œ€ํ•œ ์ถ”๋ก  ์„œ๋น„์Šค๋ฅผ ์ œ๊ณตํ•  ๋•Œ ์ถ”๋ก  ๊ฒฐ๊ณผ๋กœ๋ถ€ํ„ฐ ๋ชจ๋ธ๊ณผ ํ•™์Šต ๋ฐ์ดํ„ฐ๋ฅผ ๋ณดํ˜ธํ•˜์ง€ ๋ชปํ•œ๋‹ค. ์—ฐ๊ตฌ๋ฅผ ํ†ตํ•ด ๊ณต๊ฒฉ์ž๊ฐ€ ์ž์‹ ์ด ๊ฐ€์ง„ ๋ฐ์ดํ„ฐ์™€ ๊ทธ ๋ฐ์ดํ„ฐ์— ๋Œ€ํ•œ ์ถ”๋ก  ๊ฒฐ๊ณผ๋งŒ์„ ์ด์šฉํ•˜์—ฌ ์ด์šฉํ•˜์—ฌ ๋ชจ๋ธ๊ณผ ํ•™์Šต ๋ฐ์ดํ„ฐ์— ๋Œ€ํ•œ ์ •๋ณด๋ฅผ ์ถ”์ถœํ•  ์ˆ˜ ์žˆ์Œ์ด ๋ฐํ˜€์ง€๊ณ  ์žˆ๋‹ค. ์˜ˆ๋ฅผ ๋“ค์–ด, ๊ณต๊ฒฉ์ž๋Š” ํŠน์ • ๋ฐ์ดํ„ฐ๊ฐ€ ํ•™์Šต ๋ฐ์ดํ„ฐ์— ํฌํ•จ๋˜์–ด ์žˆ๋Š”์ง€ ์•„๋‹Œ์ง€๋ฅผ ์ถ”๋ก ํ•  ์ˆ˜ ์žˆ๋‹ค. ์ฐจ๋ถ„ ํ”„๋ผ์ด๋ฒ„์‹œ๋Š” ํ•™์Šต๋œ ๋ชจ๋ธ์— ๋Œ€ํ•œ ํŠน์ • ๋ฐ์ดํ„ฐ ์ƒ˜ํ”Œ์˜ ์˜ํ–ฅ์„ ์ค„์ž„์œผ๋กœ์จ ์ด๋Ÿฌํ•œ ๊ณต๊ฒฉ์— ๋Œ€ํ•œ ๋ฐฉ์–ด๋ฅผ ๋ณด์žฅํ•˜๋Š” ํ”„๋ผ์ด๋ฒ„์‹œ ๊ธฐ์ˆ ์ด๋‹ค. ์ฐจ๋ถ„ ํ”„๋ผ์ด๋ฒ„์‹œ๋Š” ํ”„๋ผ์ด๋ฒ„์‹œ์˜ ์ˆ˜์ค€์„ ์ •๋Ÿ‰์ ์œผ๋กœ ํ‘œํ˜„ํ•จ์œผ๋กœ์จ ์›ํ•˜๋Š” ๋งŒํผ์˜ ํ”„๋ผ์ด๋ฒ„์‹œ๋ฅผ ์ถฉ์กฑ์‹œํ‚ฌ ์ˆ˜ ์žˆ์ง€๋งŒ, ํ”„๋ผ์ด๋ฒ„์‹œ๋ฅผ ์ถฉ์กฑ์‹œํ‚ค๊ธฐ ์œ„ํ•ด์„œ๋Š” ์•Œ๊ณ ๋ฆฌ์ฆ˜์— ๊ทธ๋งŒํผ์˜ ๋ฌด์ž‘์œ„์„ฑ์„ ๋”ํ•ด์•ผ ํ•˜๋ฏ€๋กœ ๋ชจ๋ธ์˜ ์„ฑ๋Šฅ์„ ๋–จ์–ด๋œจ๋ฆฐ๋‹ค. ๋”ฐ๋ผ์„œ, ๋ณธ๋ฌธ์—์„œ๋Š” ๋ชจ์Šค ์ด๋ก ์„ ์ด์šฉํ•˜์—ฌ ์ฐจ๋ถ„ ํ”„๋ผ์ด๋ฒ„์‹œ ๊ตฐ์ง‘ํ™” ๋ฐฉ๋ฒ•๋ก ์˜ ํ”„๋ผ์ด๋ฒ„์‹œ๋ฅผ ์œ ์ง€ํ•˜๋ฉด์„œ๋„ ๊ทธ ์„ฑ๋Šฅ์„ ๋Œ์–ด์˜ฌ๋ฆฌ๋Š” ์ƒˆ๋กœ์šด ๋ฐฉ๋ฒ•๋ก ์„ ์ œ์•ˆํ•œ๋‹ค. ๋ณธ ๋…ผ๋ฌธ์—์„œ ๊ฐœ๋ฐœํ•˜๋Š” ํ”„๋ผ์ด๋ฒ„์‹œ ๋ณด์กด ๊ธฐ๊ณ„ํ•™์Šต ๋ฐฉ๋ฒ•๋ก ์€ ๊ฐ๊ธฐ ๋‹ค๋ฅธ ์ˆ˜์ค€์—์„œ ํ”„๋ผ์ด๋ฒ„์‹œ๋ฅผ ๋ณดํ˜ธํ•˜๋ฉฐ, ๋”ฐ๋ผ์„œ ์ƒํ˜ธ ๋ณด์™„์ ์ด๋‹ค. ์ œ์•ˆ๋œ ๋ฐฉ๋ฒ•๋ก ๋“ค์€ ํ•˜๋‚˜์˜ ํ†ตํ•ฉ ์‹œ์Šคํ…œ์„ ๊ตฌ์ถ•ํ•˜์—ฌ ๊ธฐ๊ณ„ํ•™์Šต์ด ๊ฐœ์ธ์˜ ๋ฏผ๊ฐ ์ •๋ณด๋กค ๋ณดํ˜ธํ•ด์•ผ ํ•˜๋Š” ์—ฌ๋Ÿฌ ๋ถ„์•ผ์—์„œ ๋”์šฑ ๋„๋ฆฌ ์‚ฌ์šฉ๋  ์ˆ˜ ์žˆ๋„๋ก ํ•˜๋Š” ๊ธฐ๋Œ€ ํšจ๊ณผ๋ฅผ ๊ฐ€์ง„๋‹ค.Recent development of artificial intelligence systems has been driven by various factors such as the development of new algorithms and the the explosive increase in the amount of available data. In the real-world scenarios, individuals or corporations benefit by providing data for training a machine learning model or the trained model. However, it has been revealed that sharing of data or the model can lead to invasion of personal privacy by leaking personal sensitive information. In this dissertation, we focus on developing privacy-preserving machine learning methods which can protect sensitive information. Homomorphic encryption can protect the privacy of data and the models because machine learning algorithms can be applied to encrypted data, but requires much larger computation time than conventional operations. For efficient computation, we take two approaches. The first is to reduce the amount of computation in the training phase. We present an efficient training algorithm by encrypting only few important information. In specific, we develop a ridge regression algorithm that greatly reduces the amount of computation when one or two sensitive variables are encrypted. Furthermore, we extend the method to apply it to classification problems by developing a new logistic regression algorithm that can maximally exclude searching of hyper-parameters that are not suitable for machine learning with homomorphic encryption. Another approach is to apply homomorphic encryption only when the trained model is used for inference, which prevents direct exposure of the test data and the model information. We propose a homomorphic-encryption-friendly algorithm for inference of support based clustering. Though homomorphic encryption can prevent various threats to data and the model information, it cannot defend against secondary attacks through inference APIs. It has been reported that an adversary can extract information about the training data only with his or her input and the corresponding output of the model. For instance, the adversary can determine whether specific data is included in the training data or not. Differential privacy is a mathematical concept which guarantees defense against those attacks by reducing the impact of specific data samples on the trained model. Differential privacy has the advantage of being able to quantitatively express the degree of privacy, but it reduces the utility of the model by adding randomness to the algorithm. Therefore, we propose a novel method which can improve the utility while maintaining the privacy of differentially private clustering algorithms by utilizing Morse theory. The privacy-preserving machine learning methods proposed in this paper can complement each other to prevent different levels of attacks. We expect that our methods can construct an integrated system and be applied to various domains where machine learning involves sensitive personal information.Chapter 1 Introduction 1 1.1 Motivation of the Dissertation 1 1.2 Aims of the Dissertation 7 1.3 Organization of the Dissertation 10 Chapter 2 Preliminaries 11 2.1 Homomorphic Encryption 11 2.2 Differential Privacy 14 Chapter 3 Efficient Homomorphic Encryption Framework for Ridge Regression 18 3.1 Problem Statement 18 3.2 Framework 22 3.3 Proposed Method 25 3.3.1 Regression with one Encrypted Sensitive Variable 25 3.3.2 Regression with two Encrypted Sensitive Variables 30 3.3.3 Adversarial Perturbation Against Attribute Inference Attack 35 3.3.4 Algorithm for Ridge Regression 36 3.3.5 Algorithm for Adversarial Perturbation 37 3.4 Experiments 40 3.4.1 Experimental Setting 40 3.4.2 Experimental Results 42 3.5 Chapter Summary 47 Chapter 4 Parameter-free Homomorphic-encryption-friendly Logistic Regression 53 4.1 Problem Statement 53 4.2 Proposed Method 56 4.2.1 Motivation 56 4.2.2 Framework 58 4.3 Theoretical Results 63 4.4 Experiments 68 4.4.1 Experimental Setting 68 4.4.2 Experimental Results 70 4.5 Chapter Summary 75 Chapter 5 Homomorphic-encryption-friendly Evaluation for Support Vector Clustering 76 5.1 Problem Statement 76 5.2 Background 78 5.2.1 CKKS scheme 78 5.2.2 SVC 80 5.3 Proposed Method 82 5.4 Experiments 86 5.4.1 Experimental Setting 86 5.4.2 Experimental Results 87 5.5 Chapter Summary 89 Chapter 6 Differentially Private Mixture of Gaussians Clustering with Morse Theory 95 6.1 Problem Statement 95 6.2 Background 98 6.2.1 Mixture of Gaussians 98 6.2.2 Morse Theory 99 6.2.3 Dynamical System Perspective 101 6.3 Proposed Method 104 6.3.1 Differentially private clustering 105 6.3.2 Transition equilibrium vectors and the weighted graph 108 6.3.3 Hierarchical merging of sub-clusters 111 6.4 Theoretical Results 112 6.5 Experiments 117 6.5.1 Experimental Setting 117 6.5.2 Experimental Results 119 6.6 Chapter Summary 122 Chapter 7 Conclusion 124 7.1 Conclusion 124 7.2 Future Direction 126 Bibliography 128 ๊ตญ๋ฌธ์ดˆ๋ก 154๋ฐ•
    • โ€ฆ
    corecore