1,965 research outputs found
Accurator: Nichesourcing for Cultural Heritage
With more and more cultural heritage data being published online, their
usefulness in this open context depends on the quality and diversity of
descriptive metadata for collection objects. In many cases, existing metadata
is not adequate for a variety of retrieval and research tasks and more specific
annotations are necessary. However, eliciting such annotations is a challenge
since it often requires domain-specific knowledge. Where crowdsourcing can be
successfully used for eliciting simple annotations, identifying people with the
required expertise might prove troublesome for tasks requiring more complex or
domain-specific knowledge. Nichesourcing addresses this problem, by tapping
into the expert knowledge available in niche communities. This paper presents
Accurator, a methodology for conducting nichesourcing campaigns for cultural
heritage institutions, by addressing communities, organizing events and
tailoring a web-based annotation tool to a domain of choice. The contribution
of this paper is threefold: 1) a nichesourcing methodology, 2) an annotation
tool for experts and 3) validation of the methodology and tool in three case
studies. The three domains of the case studies are birds on art, bible prints
and fashion images. We compare the quality and quantity of obtained annotations
in the three case studies, showing that the nichesourcing methodology in
combination with the image annotation tool can be used to collect high quality
annotations in a variety of domains and annotation tasks. A user evaluation
indicates the tool is suited and usable for domain specific annotation tasks
Digital Image Access & Retrieval
The 33th Annual Clinic on Library Applications of Data Processing, held at the University of Illinois at Urbana-Champaign in March of 1996, addressed the theme of "Digital Image Access & Retrieval." The papers from this conference cover a wide range of topics concerning digital imaging technology for visual resource collections. Papers covered three general areas: (1) systems, planning, and implementation; (2) automatic and semi-automatic indexing; and (3) preservation with the bulk of the conference focusing on indexing and retrieval.published or submitted for publicatio
SUPERT: Towards New Frontiers in Unsupervised Evaluation Metrics for Multi-Document Summarization
We study unsupervised multi-document summarization evaluation metrics, which
require neither human-written reference summaries nor human annotations (e.g.
preferences, ratings, etc.). We propose SUPERT, which rates the quality of a
summary by measuring its semantic similarity with a pseudo reference summary,
i.e. selected salient sentences from the source documents, using contextualized
embeddings and soft token alignment techniques. Compared to the
state-of-the-art unsupervised evaluation metrics, SUPERT correlates better with
human ratings by 18-39%. Furthermore, we use SUPERT as rewards to guide a
neural-based reinforcement learning summarizer, yielding favorable performance
compared to the state-of-the-art unsupervised summarizers. All source code is
available at https://github.com/yg211/acl20-ref-free-eval.Comment: ACL 202
Automatic Multi-Label Image Classification Model for Construction Site Images
ํ์๋
ผ๋ฌธ(์์ฌ)--์์ธ๋ํ๊ต ๋ํ์ :๊ณต๊ณผ๋ํ ๊ฑด์ถํ๊ณผ,2019. 8. ๋ฐ๋ฌธ์.์ต๊ทผ ์ด๋ฏธ์ง ๋ถ์ ๊ธฐ์ ์ด ๋ฐ์ ํจ์ ๋ฐ๋ผ ๊ฑด์ค ํ์ฅ์์ ๋ค์ํ ๋ฐฉ๋ฉด์์ ํ์ฅ์์ ์์ง๋ ์ฌ์ง์ ํ์ฉํ์ฌ ๊ฑด์ค ํ๋ก์ ํธ๋ฅผ ๊ด๋ฆฌํ๊ณ ์ ํ๋ ์๋๊ฐ ์ด๋ฃจ์ด์ง๊ณ ์๋ค. ํนํ ์ดฌ์ ์ฅ๋น์ ๋ฐ์ ๋์ ๊ฑด์ค ํ์ฅ์์ ์์ฐ๋๋ ์ฌ์ง์ ์๊ฐ ๊ธ์ฆํ์ฌ ๊ฑด์ค ํ์ฅ ์ฌ์ง์ ์ ์ฌ์ ์ธ ํ์ฉ๋๋ ๋์ฑ ๋ ๋์์ง๊ณ ์๋ค. ํ์ง๋ง ์ด๋ ๊ฒ ์์ฐ๋๋ ๋ง์ ์์ ์ฌ์ง์ ๋๋ถ๋ถ ์ ๋๋ก ๋ถ๋ฅ๋์ง ์์ ์ํ๋ก ๋ณด๊ด๋๊ณ ์๊ธฐ ๋๋ฌธ์ ํ์ฅ ์ฌ์ง์ผ๋ก๋ถํฐ ํ์ํ ํ๋ก์ ํธ ์ ๋ณด๋ฅผ ์ถ์ถํ๋ ๊ฒ์ ๋งค์ฐ ์ด๋ ค์ด ์ค์ ์ด๋ค. ํ์ฌ ํ์ฅ์์ ์ฌ์ง์ ๋ถ๋ฅํ๋ ๋ฐฉ์์ ์ฌ์ฉ์๊ฐ ์ง์ ๊ฐ๋ณ ์ฌ์ง์ ๊ฒํ ํ ๋ค ๋ถ๋ฅํ๊ธฐ ๋๋ฌธ์ ๋ง์ ์๊ฐ๊ณผ ๋
ธ๋ ฅ์ด ์๊ตฌ๋๊ณ , ์ด๋ฏธ์ง ๋ถ๋ฅ๋ฅผ ์ํ ํน์ง์ ์ง์ ์ ์ผ๋ก ์ถ์ถํ๋ ๊ธฐ์กด์ ์ด๋ฏธ์ง ๋ถ์ ๊ธฐ์ ์ญ์ ๋ณต์กํ ๊ฑด์ค ํ์ฅ ์ฌ์ง์ ํน์ง์ ๋ฒ์ฉ์ ์ผ๋ก ํ์ตํ๋ ๋ฐ๋ ํ๊ณ๊ฐ ์๋ค.
์ด์ ๋ณธ ์ฐ๊ตฌ์์๋ ๊ฑด์ค ํ์ฅ ์ฌ์ง์ ๋ชจ์ต์ด ๋งค์ฐ ๋ค์ํ๊ณ , ๋์ ์ผ๋ก ๋ณํ๋ ๊ฒ์ ๋์ํ๊ธฐ ์ํด ์ด๋ฏธ์ง ๋ถ๋ฅ์์ ๋์ ์ฑ๋ฅ์ ๋ณด์ด๊ณ ํฉ์ฑ๊ณฑ ์ ๊ฒฝ๋ง(Deep Convolutional Neural Network) ์๊ณ ๋ฆฌ์ฆ์ ์ ์ฉํ์ฌ ๊ฐ๋ณ ๊ฑด์ค ํ์ฅ ์ฌ์ง์ ์ ํฉํ ํค์๋๋ฅผ ์๋์ผ๋ก ํ ๋นํ ์ ์๋ ๋ชจ๋ธ์ ๊ฐ๋ฐํ๊ณ ์ ํ๋ค. ํฉ์ฑ๊ณฑ ์ ๊ฒฝ๋ง ๋ชจ๋ธ์ ๋ชจ๋ธ ๊ตฌ์กฐ๊ฐ ๊น์ด์ง์ ๋ฐ๋ผ ๋์ ์ฐจ์์ ํญ์์ฑ(invariant) ํน์ง๋ ํจ๊ณผ์ ์ผ๋ก ํ์ตํ ์ ์๋ ํน์ง์ด ์๊ธฐ ๋๋ฌธ์ ๋ณต์กํ ๊ฑด์ค ํ์ฅ ์ฌ์ง ๋ถ๋ฅ ๋ฌธ์ ์ ์ ํฉํ๋ค.
๋ฐ๋ผ์ ๋ณธ ์ฐ๊ตฌ์์๋ ํฉ์ฑ๊ณฑ ์ ๊ฒฝ๋ง ๋ชจ๋ธ์ ํ ๋๋ก ํ์ฅ์์ ํ์ํ ์ฌ์ง์ ๋น ๋ฅด๊ณ ์ ํํ๊ฒ ์ฐพ์ ์ ์๋๋ก ๊ฐ ์ฌ์ง์ ์ ํฉํ ํค์๋๋ฅผ ์๋์ผ๋ก ํ ๋นํ๋ ๋ชจ๋ธ์ ๊ฐ๋ฐํ์๋ค. ํนํ, ๊ฑด์ค ํ์ฅ ์ฌ์ง์ ๋๋ถ๋ถ์ด ํ๋ ์ด์์ ๋ ์ด๋ธ๊ณผ ์ฐ๊ด์ด ์๋ค๋ ์ ์ ๊ธฐ๋ฐํ์ฌ ๋ค์ค ๋ ์ด๋ธ ๋ถ๋ฅ ๋ชจ๋ธ์ ์ ์ฉํ์๋ค. ์ด๋ฅผ ํตํด ์ผ์ฐจ์ ์ผ๋ก๋ ๊ฑด์ค ์ฌ์ง์์ ํ๋ก์ ํธ์ ๊ด๋ จ๋ ๋ค์ํ ์ ๋ณด๋ฅผ ์ถ์ถํ์ฌ ๊ฑด์ค ํ์ฅ ์ฌ์ง์ ํ์ฉ๋๋ฅผ ๊ฐ์ ํ๊ณ , ๋์๊ฐ ์ฌ์ง ๋ฐ์ดํฐ๋ฅผ ํ์ฉํ์ฌ ํจ์จ์ ์ธ ๊ฑด์ค ๊ด๋ฆฌ๋ฅผ ๋๋ชจํ๊ณ ์ ํ๋ค.
๋ณธ ์ฐ๊ตฌ์ ์งํ ์์๋ ๋ค์๊ณผ ๊ฐ๋ค. ์ฐ์ ๋ชจ๋ธ์ ํ์ต์ํค๊ธฐ ์ํด์ ์ค์ ๊ฑด์ค ํ์ฅ ๋ฐ ์คํ์์ค ๊ฒ์์์ง์ ํตํ์ฌ ์ด 6๊ฐ ๊ณต์ข
์ ์ฌ์ง์ ์์งํ๊ณ , ํ์ ๋ถ๋ฅ ๋ฒ์๋ฅผ ํฌํจํ ์ด 10๊ฐ ๋ ์ด๋ธ์ ๋ฐ์ดํฐ์
์ ๊ตฌ์ฑํ์ฌ ํ์ต์ ์งํํ๋ค. ๋ํ ๊ตฌ์ฒด์ ์ธ ๋ชจ๋ธ ์ ํ์ ์ํด ๋ํ์ ์ธ ํฉ์ฑ๊ณฑ ์ ๊ฒฝ๋ง ๋ชจ๋ธ์ ๋น๊ต ๊ฒํ ํ์ฌ ๊ฐ์ฅ ์ฐ์ํ ์ฑ๋ฅ์ ๋ณด์ธ ResNet 18์ ์ต์ข
๋ชจ๋ธ๋ก ์ ํํ๋ค. ์คํ ๊ฒฐ๊ณผ ํ๊ท 91%์ ์ ํ๋๋ฅผ ๋ณด์ด๋ฉฐ ๊ฑด์ค ํ์ฅ ์ฌ์ง์ ์๋์ผ๋ก ๋ถ๋ฅํ ์ ์๋ ๊ฐ๋ฅ์ฑ์ ํ์ธํ์๋ค.
๋ํ ๋ณธ ์ฐ๊ตฌ๋ ์ต๊ทผ ํ ๋ถ์ผ ์ด๋ฏธ์ง ๋ถ์์์ ์ข์ ์ฑ๊ณผ๋ฅผ ๋ณด์ธ ํฉ์ฑ๊ณฑ ์ ๊ฒฝ๋ง์ ํ์ฉํ์ฌ ๊ฑด์ค ํ์ฅ ์ฌ์ง์ ์๋์ผ๋ก ๋ถ๋ฅํ ์ ์๋ค๋ ๊ฐ๋ฅ์ฑ์ ํ์ธํ๋ค๋ ์ ๊ณผ, ๊ฑด์ค ํ์ฅ ์ฌ์ง ๋ถ๋ฅ ๋ฌธ์ ์ ๋ค์ค ๋ ์ด๋ธ ๋ถ๋ฅ๋ฅผ ์ ์ฉํ ์ฒซ ์ฐ๊ตฌ๋ผ๋ ์ ์์ ์์๊ฐ ์๋ค. ์ค์ ํ์ฅ์์๋ ์ฌ์ง์ ์๋์ผ๋ก ๋ถ๋ฅํ ์ ์๊ฒ ๋จ์ ๋ฐ๋ผ ๊ธฐ์กด์ ๋ฒ๊ฑฐ๋ก์ด ์๋ ์ฌ์ง ๋ถ๋ฅ ์์
์ ์ค์ด๊ณ , ๊ฑด์ค ํ์ฅ ์ฌ์ง์ ํ์ฉ๋๋ฅผ ๋์ผ ์ ์์ ๊ฒ์ผ๋ก ๊ธฐ๋๋๋ค.
ํ์ง๋ง ๋ณธ ์ฐ๊ตฌ๋ ๊ฐ ๋ ์ด๋ธ ๊ฐ์ ์ฐ๊ด์ฑ์ด๋ ์์กด์ฑ์ ๊ณ ๋ คํ์ง ์๊ธฐ ๋๋ฌธ์ ์ถํ ์ฐ๊ตฌ์์๋ ๊ฐ ์ฌ์ง ๊ฐ์ ๊ณ์ธต์ ๊ด๊ณ๋ฅผ ๋ชจ๋ธ์ ์ถ๊ฐ์ ์ผ๋ก ํ์ต์์ผ ์ ํ๋๋ฅผ ๋์ด๊ณ , ํ์ต ๋ ์ด๋ธ๋ ๋ ๋ฎ์ ๋จ๊ณ์ ํค์๋๊น์ง ํฌํจํ์ฌ ํ์ฅ ์ฌ์ง์ผ๋ก๋ถํฐ ๋ณด๋ค ๋ค์ํ ์ ๋ณด๋ฅผ ์ป์ ์ ์๋๋ก ๋ชจ๋ธ์ ๊ฐ์ ํ๋ ๊ฒ์ ๋ชฉํ๋ก ํ๊ณ ์๋ค.Activity recognition in construction performs as the prerequisite step in the process for various tasks and thus is critical for successful project management. In the last several years, the computer vision community has blossomed, taking advantage of the exploding amount of construction images and deploying the visual analytics technology for cumbersome construction tasks. However, the current annotation practice itself, which is a critical preliminary step for prompt image retrieval and image understanding, is remained as both time-consuming and labor-intensive. Because previous attempts to make the process more efficient were inappropriate to handle dynamic nature of construction images and showed limited performance in classifying construction activities, this research aims to develop a model which is not only robust to a wide range of appearances but also multi-composition of construction activity images. The proposed model adopts a deep convolutional neural network model to learn high dimensional feature with less human-engineering and annotate multi-labels of semantic information in the images. The result showed that our model was capable of distinguishing different trades of activities at different stages of the activity. The average accuracy of 83% and a maximum accuracy of 91% holds promise in an actual implementation of automated activity recognition for construction operations. Ultimately, it demonstrated a potential method to provide automated and reliable procedure to monitor construction activity.Chapter 1. Introduction 1
1.1. Research Background 1
1.2. Research Objectives and Scope 5
1.3. Research Outline 7
Chapter 2. Preliminary Study 9
2.1. Challenges with Construction Activity Image Classification Task 10
2.2. Applications of Traditional Vision-based Algorithms in Construction Domain 13
2.3. Convolutional Neural Network-based Image Classification in Construction Domain 18
2.4. Summary 21
Chapter 3. Development of Construction Image Classification Model 22
3.1. Customized Construction Image Dataset Preparation 23
3.1.1. Construction Activity Classification System 23
3.1.2. Dataset Collection 24
3.1.3. Data Pre-Processing 25
3.2. Construction Image Classification Model Framework 27
3.2.1. Multi-label Image Classification 27
3.2.2. Base CNN Model Selection 28
3.2.3. Proposed ResNet Model Architecture 29
3.3. Model Training and Validation 33
3.3.1. Transfer Learning 33
3.3.2. Loss Computation and Model Optimization 33
3.3.3. Model Performance Indicator 35
3.4. Summary 37
Chapter 4. Experiment Results and Discussion 38
4.1. Experiment Results 38
4.2. Analysis of Experiment Results 42
4.3. Summary 44
Chapter 5. Conclusion 45
5.1. Research Summary 45
5.2. Research Contributions 46
5.3. Limitations and Further Study 47
References 49
Appendix 57
Abstract in Korean 63Maste
- โฆ