3,695 research outputs found

    ํฌ๋ผ์šฐ๋“œ์†Œ์‹ฑ ์‹œ์Šคํ…œ์—์„œ์˜ ๋น ๋ฅด๊ณ  ์‹ ๋ขฐ์„ฑ ๋†’์€ ์ถ”๋ก  ์•Œ๊ณ ๋ฆฌ์ฆ˜

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ (๋ฐ•์‚ฌ) -- ์„œ์šธ๋Œ€ํ•™๊ต ๋Œ€ํ•™์› : ๊ณต๊ณผ๋Œ€ํ•™ ์ „๊ธฐยท์ปดํ“จํ„ฐ๊ณตํ•™๋ถ€, 2021. 2. ์ •๊ต๋ฏผ.As the need for large scale labeled data grows in various fields, the appearance of web-based crowdsourcing systems gives a promising solution to exploiting the wisdom of crowds efficiently in a short time with a relatively low budget. Despite their efficiency, crowdsourcing systems have an inherent problem in that responses from workers can be unreliable since workers are low-paid and have low responsibility. Although simple majority voting can be a natural solution, various research studies have sought to aggregate noisy responses to obtain greater reliability in results. In this dissertation, we propose novel iterative massage-passing style algorithms to infer the groundtruths from noisy answers, which can be directly applied to real crowdsourcing systems. While EM-based algorithms get the limelight in crowdsourcing systems due to their useful inference techniques, our proposed algorithms draw faster and more reliable answers through an iterative scheme based on the idea of low-rank matrix approximations. We show that the performance of our proposed iterative algorithms are order-optimal, which outperforms majority voting and EM-based algorithms. Unlike other researches solving simple binary-choice questions (yes & no), our studies cover more complex task types which contain multiple-choice questions, short-answer questions, K-approval voting, and real-valued vector regression.๋‹ค์–‘ํ•œ ๋ถ„์•ผ์—์„œ ๋ผ๋ฒจ๋œ ๋น…๋ฐ์ดํ„ฐ๋ฅผ ํ•„์š”๋กœ ํ•˜๋Š” ํ˜„์žฌ, ์›น ๊ธฐ๋ฐ˜ ํฌ๋ผ์šฐ๋“œ์†Œ์‹ฑ ์„œ๋น„์Šค๋“ค์ด ์ถœ๋ฒ”ํ•˜๋ฉฐ ์ƒ๋Œ€์ ์œผ๋กœ ์ ์€ ์˜ˆ์‚ฐ๊ณผ ์งง์€ ์‹œ๊ฐ„์—๋„ ํšจ์œจ์ ์œผ๋กœ ์‚ฌ๋žŒ๋“ค์˜ ์ง€ํ˜œ๋ฅผ ํ™œ์šฉํ•  ์ˆ˜ ์žˆ๋Š” ๋ฐฉ๋ฒ•๋“ค์ด ์ œ์‹œ๋˜๊ณ  ์žˆ๋‹ค. ์ด๋Ÿฌํ•œ ๋ฐฉ๋ฒ•๋“ค์˜ ํšจ์œจ์„ฑ์—๋„ ๋ถˆ๊ตฌํ•˜๊ณ , ํฌ๋ผ์šฐ๋“œ์†Œ์‹ฑ ์‹œ์Šคํ…œ์˜ ์„ ์ฒœ์ ์ธ ๋ฌธ์ œ์ ์€ ์ผ์„ ๋งก์€ ์‚ฌ๋žŒ๋“ค์˜ ์ ์€ ๋ณด์ƒ ๋ฐ ์ฑ…์ž„๊ฐ ๊ฒฐ์—ฌ๋กœ ์ธํ•ด ๊ทธ๋“ค์˜ ์‘๋‹ต์„ ์™„์ „ํžˆ ์‹ ๋ขฐํ•  ์ˆ˜ ์—†๋‹ค๋Š” ์ ์— ์žˆ๋‹ค. ์ด์— ๋‹ค์ˆ˜๊ฒฐ ๋ฐฉ์‹์ด ์ž์—ฐ์Šค๋Ÿฌ์šด ํ•ด๋ฒ•์œผ๋กœ ์‚ฌ์šฉ๋˜์ง€๋งŒ, ๋ณด๋‹ค ์‹ ๋ขฐ ๋†’์€ ๋‹ต์„ ์–ป์–ด๋‚ด๊ธฐ ์œ„ํ•ด ๋งŽ์€ ์—ฐ๊ตฌ๋“ค์ด ์ง„ํ–‰๋˜๊ณ  ์žˆ๋‹ค. ๋ณธ ๋ฐ•์‚ฌํ•™์œ„ ๋…ผ๋ฌธ์—์„œ๋Š” ํฌ๋ผ์šฐ๋“œ์†Œ์‹ฑ ์‹œ์Šคํ…œ์—์„œ ์ˆ˜๋งŽ์€ ์‚ฌ๋žŒ๋“ค๋กœ๋ถ€ํ„ฐ ๋ฐ›์€ ์‘๋‹ต๋“ค์„ ๋ชจ์•„ ์‹ ๋ขฐ์„ฑ ๋†’์€ ์‘๋‹ต์„ ์ถ”๋ก ํ•˜๋Š” ๋ฐ˜๋ณต์  ๋ฉ”์„ธ์ง€์ „๋‹ฌ ํ˜•ํƒœ์˜ ์•Œ๊ณ ๋ฆฌ์ฆ˜๋“ค์„ ์ œ์‹œํ•œ๋‹ค. ๋ณธ ์•Œ๊ณ ๋ฆฌ์ฆ˜๋“ค์€ ๋‚ฎ์€๋žญํฌ๊ทผ์‚ฌ์— ๊ธฐ๋ฐ˜ํ•œ ๋ฐ˜๋ณต ์ถ”๋ก  ๋ฐฉ๋ฒ•์œผ๋กœ, ๊ธฐ์กด์— ๊ฐ๊ด‘๋ฐ›๋˜ EM ์•Œ๊ณ ๋ฆฌ์ฆ˜๋“ค์— ๋น„ํ•ด ๋” ๋น ๋ฅด๊ณ  ์‹ ๋ขฐ์ ์ธ ์ •๋‹ต์„ ์ถ”๋ก ํ•ด๋‚ธ๋‹ค. ๋”๋ถˆ์–ด ๋ณธ ์•Œ๊ณ ๋ฆฌ์ฆ˜๋“ค์˜ ์ถ”๋ก  ์ •ํ™•๋„๊ฐ€ ์ตœ์ ์— ๋งค์šฐ ๊ทผ์ ‘ํ•˜๋ฉฐ ๋‹ค์ˆ˜๊ฒฐ ๋ฐฉ์‹ ๋ฐ EM ์•Œ๊ณ ๋ฆฌ์ฆ˜๋“ค์˜ ์ •ํ™•๋„๋ฅผ ์ƒํšŒํ•œ๋‹ค๋Š” ๊ฒƒ์„ ์ด๋ก ์  ์ฆ๋ช… ๋ฐ ์‹คํ—˜์  ๊ฒฐ๊ณผ๋ฅผ ํ†ตํ•ด ์ œ์‹œํ•œ๋‹ค. ๋ณธ ์—ฐ๊ตฌ๋Š” ์‹ค์ œ ํฌ๋ผ์šฐ๋“œ์†Œ์‹ฑ์—์„œ ๋Œ€๋‹ค์ˆ˜์˜ ์‘๋‹ต ์œ ํ˜•์„ ์ฐจ์ง€ํ•˜๋Š” ๊ฐ๊ด€์‹ ์‘๋‹ต, ์ฃผ๊ด€์‹ ์‘๋‹ต, ๋ณต์ˆ˜ ์„ ํƒ ์‘๋‹ต, ๋ฐ ์‹ค์ˆ˜ ๊ฐ’ ์‘๋‹ต์˜ ์ถ”๋ก  ๋ฌธ์ œ๋ฅผ ๋‹ค๋ฃจ๋ฉฐ, ๊ธฐ์กด ์–‘์žํƒ์ผ ์‘๋‹ต ์ถ”๋ก  ๋ฌธ์ œ๋งŒ์„ ๋‹ค๋ฃจ๋Š” ๊ธฐ์กด ์—ฐ๊ตฌ๋“ค๊ณผ ํฐ ์ฐจ๋ณ„์„ฑ์„ ๊ฐ€์ง„๋‹ค.1 Introduction 1 2 Background 9 2.1 Crowdsourcing Systems for Binary-choice Questions 9 2.1.1 Majority Voting 10 2.1.2 Expectation Maximization 11 2.1.3 Message Passing 11 3 Crowdsourcing Systems for Multiple-choice Questions 12 3.1 Related Work 13 3.2 Problem Setup 16 3.3 Inference Algorithm 17 3.3.1 Task Allocation 17 3.3.2 Multiple Iterative Algorithm 18 3.3.3 Task Allocation for General Setting 20 3.4 Applications 23 3.5 Analysis of Algorithms 25 3.5.1 Quality of Workers 25 3.5.2 Bound on the Average Error Probability 27 3.5.3 Proof of the Error Bounds 29 3.5.4 Proof of Sub-Gaussianity 32 3.6 Experimental Results 36 3.6.1 Comparison with Other Algorithms 37 3.6.2 Adaptive Scenario 38 3.6.3 Simulations on a Set of Various D Values 41 3.7 Conclusion 43 4 Crowdsourcing Systems for Multiple-choice Questions with K-Approval Voting 45 4.1 Related Work 47 4.2 Problem Setup 49 4.2.1 Problem Definition 49 4.2.2 Worker Model for Various (D, K) 50 4.3 Inference Algorithm 51 4.4 Analysis of Algorithms 53 4.4.1 Worker Model 55 4.4.2 Quality of Workers 56 4.4.3 Bound on the Average Error Probability 58 4.4.4 Proof of the Error Bounds 59 4.4.5 Proof of Sub-Gaussianity 62 4.4.6 Phase Transition 67 4.5 Experimental Results 68 4.5.1 Performance on the Average Error with q and l 68 4.5.2 Relationship between Reliability and y-message 69 4.5.3 Performance on the Average Error with Various (D, K) Pairs 69 4.6 Conclusion 72 5 Crowdsourcing Systems for Real-valued Vector Regression 73 5.1 Related Work 75 5.2 Problem Setup 77 5.3 Inference Algorithm 78 5.3.1 Task Message 79 5.3.2 Worker Message 80 5.4 Analysis of Algorithms 81 5.4.1 Worker Model 81 5.4.2 Oracle Estimator 84 5.4.3 Bound on the Average Error Probability 86 5.5 Experimental Results 91 5.5.1 Real Crowdsourcing Data 91 5.5.2 Verification of the Error Bounds with Synthetic data 96 5.6 Conclusion 98 6 Conclusions 99Docto

    Optimal Inference in Crowdsourced Classification via Belief Propagation

    Full text link
    Crowdsourcing systems are popular for solving large-scale labelling tasks with low-paid workers. We study the problem of recovering the true labels from the possibly erroneous crowdsourced labels under the popular Dawid-Skene model. To address this inference problem, several algorithms have recently been proposed, but the best known guarantee is still significantly larger than the fundamental limit. We close this gap by introducing a tighter lower bound on the fundamental limit and proving that Belief Propagation (BP) exactly matches this lower bound. The guaranteed optimality of BP is the strongest in the sense that it is information-theoretically impossible for any other algorithm to correctly label a larger fraction of the tasks. Experimental results suggest that BP is close to optimal for all regimes considered and improves upon competing state-of-the-art algorithms.Comment: This article is partially based on preliminary results published in the proceeding of the 33rd International Conference on Machine Learning (ICML 2016

    Iterative Bayesian Learning for Crowdsourced Regression

    Full text link
    Crowdsourcing platforms emerged as popular venues for purchasing human intelligence at low cost for large volume of tasks. As many low-paid workers are prone to give noisy answers, a common practice is to add redundancy by assigning multiple workers to each task and then simply average out these answers. However, to fully harness the wisdom of the crowd, one needs to learn the heterogeneous quality of each worker. We resolve this fundamental challenge in crowdsourced regression tasks, i.e., the answer takes continuous labels, where identifying good or bad workers becomes much more non-trivial compared to a classification setting of discrete labels. In particular, we introduce a Bayesian iterative scheme and show that it provably achieves the optimal mean squared error. Our evaluations on synthetic and real-world datasets support our theoretical results and show the superiority of the proposed scheme
    • โ€ฆ
    corecore