1,112 research outputs found
Anatomy-guided domain adaptation for 3D in-bed human pose estimation
3D human pose estimation is a key component of clinical monitoring systems.
The clinical applicability of deep pose estimation models, however, is limited
by their poor generalization under domain shifts along with their need for
sufficient labeled training data. As a remedy, we present a novel domain
adaptation method, adapting a model from a labeled source to a shifted
unlabeled target domain. Our method comprises two complementary adaptation
strategies based on prior knowledge about human anatomy. First, we guide the
learning process in the target domain by constraining predictions to the space
of anatomically plausible poses. To this end, we embed the prior knowledge into
an anatomical loss function that penalizes asymmetric limb lengths, implausible
bone lengths, and implausible joint angles. Second, we propose to filter pseudo
labels for self-training according to their anatomical plausibility and
incorporate the concept into the Mean Teacher paradigm. We unify both
strategies in a point cloud-based framework applicable to unsupervised and
source-free domain adaptation. Evaluation is performed for in-bed pose
estimation under two adaptation scenarios, using the public SLP dataset and a
newly created dataset. Our method consistently outperforms various
state-of-the-art domain adaptation methods, surpasses the baseline model by
31%/66%, and reduces the domain gap by 65%/82%. Source code is available at
https://github.com/multimodallearning/da-3dhpe-anatomy.Comment: submitted to Medical Image Analysi
Adversarial content manipulation for analyzing and improving model robustness
The recent rapid progress in machine learning systems has opened up many real-world applications --- from recommendation engines on web platforms to safety critical systems like autonomous vehicles. A model deployed in the real-world will often encounter inputs far from its training distribution. For example, a self-driving car might come across a black stop sign in the wild. To ensure safe operation, it is vital to quantify the robustness of machine learning models to such out-of-distribution data before releasing them into the real-world. However, the standard paradigm of benchmarking machine learning models with fixed size test sets drawn from the same distribution as the training data is insufficient to identify these corner cases efficiently. In principle, if we could generate all valid variations of an input and measure the model response, we could quantify and guarantee model robustness locally. Yet, doing this with real world data is not scalable. In this thesis, we propose an alternative, using generative models to create synthetic data variations at scale and test robustness of target models to these variations. We explore methods to generate semantic data variations in a controlled fashion across visual and text modalities. We build generative models capable of performing controlled manipulation of data like changing visual context, editing appearance of an object in images or changing writing style of text. Leveraging these generative models we propose tools to study robustness of computer vision systems to input variations and systematically identify failure modes. In the text domain, we deploy these generative models to improve diversity of image captioning systems and perform writing style manipulation to obfuscate private attributes of the user. Our studies quantifying model robustness explore two kinds of input manipulations, model-agnostic and model-targeted. The model-agnostic manipulations leverage human knowledge to choose the kinds of changes without considering the target model being tested. This includes automatically editing images to remove objects not directly relevant to the task and create variations in visual context. Alternatively, in the model-targeted approach the input variations performed are directly adversarially guided by the target model. For example, we adversarially manipulate the appearance of an object in the image to fool an object detector, guided by the gradients of the detector. Using these methods, we measure and improve the robustness of various computer vision systems -- specifically image classification, segmentation, object detection and visual question answering systems -- to semantic input variations.Der schnelle Fortschritt von Methoden des maschinellen Lernens hat viele neue Anwendungen ermรถglicht โ von Recommender-Systemen bis hin zu sicherheitskritischen Systemen wie autonomen Fahrzeugen. In der realen Welt werden diese Systeme oft mit Eingaben auรerhalb der Verteilung der Trainingsdaten konfrontiert. Zum Beispiel kรถnnte ein autonomes Fahrzeug einem schwarzen Stoppschild begegnen. Um sicheren Betrieb zu gewรคhrleisten, ist es entscheidend, die Robustheit dieser Systeme zu quantifizieren, bevor sie in der Praxis eingesetzt werden. Aktuell werden diese Modelle auf festen Eingaben von derselben Verteilung wie die Trainingsdaten evaluiert. Allerdings ist diese Strategie unzureichend, um solche Ausnahmefรคlle zu identifizieren. Prinzipiell kรถnnte die Robustheit โlokalโ bestimmt werden, indem wir alle zulรคssigen Variationen einer Eingabe generieren und die Ausgabe des Systems รผberprรผfen. Jedoch skaliert dieser Ansatz schlecht zu echten Daten. In dieser Arbeit benutzen wir generative Modelle, um synthetische Variationen von Eingaben zu erstellen und so die Robustheit eines Modells zu รผberprรผfen. Wir erforschen Methoden, die es uns erlauben, kontrolliert semantische รnderungen an Bild- und Textdaten vorzunehmen. Wir lernen generative Modelle, die kontrollierte Manipulation von Daten ermรถglichen, zum Beispiel den visuellen Kontext zu รคndern, die Erscheinung eines Objekts zu bearbeiten oder den Schreibstil von Text zu รคndern. Basierend auf diesen Modellen entwickeln wir neue Methoden, um die Robustheit von Bilderkennungssystemen bezรผglich Variationen in den Eingaben zu untersuchen und Fehlverhalten zu identifizieren. Im Gebiet von Textdaten verwenden wir diese Modelle, um die Diversitรคt von sogenannten Automatische Bildbeschriftung-Modellen zu verbessern und Schreibtstil-Manipulation zu erlauben, um private Attribute des Benutzers zu verschleiern. Um die Robustheit von Modellen zu quantifizieren, werden zwei Arten von Eingabemanipulationen untersucht: Modell-agnostische und Modell-spezifische Manipulationen. Modell-agnostische Manipulationen basieren auf menschlichem Wissen, um bestimmte รnderungen auszuwรคhlen, ohne das entsprechende Modell miteinzubeziehen. Dies beinhaltet das Entfernen von fรผr die Aufgabe irrelevanten Objekten aus Bildern oder Variationen des visuellen Kontextes. In dem alternativen Modell-spezifischen Ansatz werden รnderungen vorgenommen, die fรผr das Modell mรถglichst ungรผnstig sind. Zum Beispiel รคndern wir die Erscheinung eines Objekts um ein Modell der Objekterkennung tรคuschen. Dies ist durch den Gradienten des Modells mรถglich. Mithilfe dieser Werkzeuge kรถnnen wir die Robustheit von Systemen zur Bildklassifizierung oder -segmentierung, Objekterkennung und Visuelle Fragenbeantwortung quantifizieren und verbessern
A Survey on Continual Semantic Segmentation: Theory, Challenge, Method and Application
Continual learning, also known as incremental learning or life-long learning,
stands at the forefront of deep learning and AI systems. It breaks through the
obstacle of one-way training on close sets and enables continuous adaptive
learning on open-set conditions. In the recent decade, continual learning has
been explored and applied in multiple fields especially in computer vision
covering classification, detection and segmentation tasks. Continual semantic
segmentation (CSS), of which the dense prediction peculiarity makes it a
challenging, intricate and burgeoning task. In this paper, we present a review
of CSS, committing to building a comprehensive survey on problem formulations,
primary challenges, universal datasets, neoteric theories and multifarious
applications. Concretely, we begin by elucidating the problem definitions and
primary challenges. Based on an in-depth investigation of relevant approaches,
we sort out and categorize current CSS models into two main branches including
\textit{data-replay} and \textit{data-free} sets. In each branch, the
corresponding approaches are similarity-based clustered and thoroughly
analyzed, following qualitative comparison and quantitative reproductions on
relevant datasets. Besides, we also introduce four CSS specialities with
diverse application scenarios and development tendencies. Furthermore, we
develop a benchmark for CSS encompassing representative references, evaluation
results and reproductions, which is available
at~\url{https://github.com/YBIO/SurveyCSS}. We hope this survey can serve as a
reference-worthy and stimulating contribution to the advancement of the
life-long learning field, while also providing valuable perspectives for
related fields.Comment: 20 pages, 12 figures. Undergoing Revie
Recommended from our members
Learning from aggregated data
Data aggregation is ubiquitous in modern life. Due to various reasons like privacy, scalability, robustness, etc., ground truth data is often subjected to aggregation before being released to the public, or utilised by researchers and analysts. Learning from aggregated data is a challenging problem that requires significant algorithmic innovation, since naive application of standard techniques to aggregated data is vulnerable to the ecological fallacy. In this work, we explore three different versions of this setting.
First, we tackle the problem of using generalised linear models when features/covariates are fully observed but the targets are only available as histograms- a common scenario in the healthcare domain where many datasets contain both non-sensitive attributes like age, sex, zip-code, etc., as well as privacy sensitive attributes like healthcare records. We introduce an efficient algorithm that uses alternating data imputation and GLM estimation steps to learn predictive models in this setting.
Next, we look at the problem of learning sparse linear models when both features and targets are in aggregated form, specified as empirical estimates of group-wise means computed over different sub-groups of the population. We show that if the true sub-populations are heterogeneous enough, the optimal sparse parameter can be recovered within an arbitrarily small tolerance even in the presence of noise, provided the empirical estimates are obtained from a sufficiently large number of observations.
Third, we tackle the scenario of predictive modelling with data that is subjected to spatio-temporal aggregation. We show that by formulating the problem in the frequency domain, we can bypass the mathematical and representational challenges that arise due to non-uniform aggregation, misaligned sampling periods and aliasing. We introduce a novel algorithm that uses restricted Fourier transforms to estimate a linear model which, when applied to spatio-temporally aggregated data, has a generalisation error that is provably close to the optimal performance by the best possible linear model that can be learned from the non-aggregated data set.
We then focus our attention on the complementary problem that involves designing aggregation strategies that can allow learning, as well as developing algorithmic techniques that can use only the aggregates to train a model that works on individual samples. We motivate our methods by using the example of Gaussian regression, and subsequently extend our techniques to subsume binary classifiers and generalised linear models. We deonstrate the effectiveness of our techniques with empirical evaluation on data from healthcare and telecommunication.
Finally, we present a concrete example of our methods applied to a real life practical problem. Specifically, we consider an application in the domain of online advertising where the complexity of bidding strategies require accurate estimates of most probable cost-per-click or CPC incurred by advertisers, but the data used for training these CPC prediction models are only available as aggregated invoices supplied by an ad publisher on a daily or hourly basis. We introduce a novel learning framework that can use aggregates computed at varying levels of granularity for building individual-level predictive models. We generalise our modelling and algorithmic framework to handle data from diverse domains, and extend our techniques to cover arbitrary aggregation paradigms like sliding windows and overlapping/non-uniform aggregation. We show empirical evidence for the efficacy of our techniques with experiments on both synthetic data and real data from the online advertising domain as well as healthcare to demonstrate the wider applicability of our framework.Electrical and Computer Engineerin
Federated Domain Generalization: A Survey
Machine learning typically relies on the assumption that training and testing
distributions are identical and that data is centrally stored for training and
testing. However, in real-world scenarios, distributions may differ
significantly and data is often distributed across different devices,
organizations, or edge nodes. Consequently, it is imperative to develop models
that can effectively generalize to unseen distributions where data is
distributed across different domains. In response to this challenge, there has
been a surge of interest in federated domain generalization (FDG) in recent
years. FDG combines the strengths of federated learning (FL) and domain
generalization (DG) techniques to enable multiple source domains to
collaboratively learn a model capable of directly generalizing to unseen
domains while preserving data privacy. However, generalizing the federated
model under domain shifts is a technically challenging problem that has
received scant attention in the research area so far. This paper presents the
first survey of recent advances in this area. Initially, we discuss the
development process from traditional machine learning to domain adaptation and
domain generalization, leading to FDG as well as provide the corresponding
formal definition. Then, we categorize recent methodologies into four classes:
federated domain alignment, data manipulation, learning strategies, and
aggregation optimization, and present suitable algorithms in detail for each
category. Next, we introduce commonly used datasets, applications, evaluations,
and benchmarks. Finally, we conclude this survey by providing some potential
research topics for the future
SoK: Chasing Accuracy and Privacy, and Catching Both in Differentially Private Histogram Publication
Histograms and synthetic data are of key importance in data analysis. However, researchers have shown that even aggregated data such as histograms, containing no obvious sensitive attributes, can result in privacy leakage. To enable data analysis, a strong notion of privacy is required to avoid risking unintended privacy violations.Such a strong notion of privacy is differential privacy, a statistical notion of privacy that makes privacy leakage quantifiable. The caveat regarding differential privacy is that while it has strong guarantees for privacy, privacy comes at a cost of accuracy. Despite this trade-off being a central and important issue in the adoption of differential privacy, there exists a gap in the literature regarding providing an understanding of the trade-off and how to address it appropriately. Through a systematic literature review (SLR), we investigate the state-of-the-art within accuracy improving differentially private algorithms for histogram and synthetic data publishing. Our contribution is two-fold: 1) we identify trends and connections in the contributions to the field of differential privacy for histograms and synthetic data and 2) we provide an understanding of the privacy/accuracy trade-off challenge by crystallizing different dimensions to accuracy improvement. Accordingly, we position and visualize the ideas in relation to each other and external work, and deconstruct each algorithm to examine the building blocks separately with the aim of pinpointing which dimension of accuracy improvement each technique/approach is targeting. Hence, this systematization of knowledge (SoK) provides an understanding of in which dimensions and how accuracy improvement can be pursued without sacrificing privacy
3D ์ ํฌ์ฆ ์ธ์์ ์ํ ์ธ์กฐ ๋ฐ์ดํฐ์ ์ด์ฉ
ํ์๋
ผ๋ฌธ(๋ฐ์ฌ) -- ์์ธ๋ํ๊ต๋ํ์ : ์ตํฉ๊ณผํ๊ธฐ์ ๋ํ์ ์ตํฉ๊ณผํ๋ถ(์ง๋ฅํ์ตํฉ์์คํ
์ ๊ณต), 2021.8. ์ํ์ด.3D hand pose estimation (HPE) based on RGB images has been studied for a long time. Relevant methods have focused mainly on optimization of neural framework for graphically connected finger joints. Training RGB-based HPE models has not been easy to train because of the scarcity on RGB hand pose datasets; unlike human body pose datasets, the finger joints that span hand postures are structured delicately and exquisitely. Such structure makes accurately annotating each joint with unique 3D world coordinates difficult, which is why many conventional methods rely on synthetic data samples to cover large variations of hand postures.
Synthetic dataset consists of very precise annotations of ground truths, and further allows control over the variety of data samples, yielding a learning model to be trained with a large pose space. Most of the studies, however, have performed frame-by-frame estimation based on independent static images. Synthetic visual data can provide practically infinite diversity and rich labels, while avoiding ethical issues with privacy and bias. However, for many tasks, current models trained on synthetic data generalize poorly to real data. The task of 3D human hand pose estimation is a particularly interesting example of this synthetic-to-real problem, because learning-based approaches perform reasonably well given real training data, yet labeled 3D poses are extremely difficult to obtain in the wild, limiting scalability.
In this dissertation, we attempt to not only consider the appearance of a hand but incorporate the temporal movement information of a hand in motion into the learning framework for better 3D hand pose estimation performance, which leads to the necessity of a large scale dataset with sequential RGB hand images.
We propose a novel method that generates a synthetic dataset that mimics natural human hand movements by re-engineering annotations of an extant static hand pose dataset into pose-flows. With the generated dataset, we train a newly proposed recurrent framework, exploiting visuo-temporal features from sequential images of synthetic hands in motion and emphasizing temporal smoothness of estimations with a temporal consistency constraint. Our novel training strategy of detaching the recurrent layer of the framework during domain finetuning from synthetic to real allows preservation of the visuo-temporal features learned from sequential synthetic hand images. Hand poses that are sequentially estimated consequently produce natural and smooth hand movements which lead to more robust estimations. We show that utilizing temporal information for
3D hand pose estimation significantly enhances general pose estimations by outperforming state-of-the-art methods in experiments on hand pose estimation benchmarks.
Since a fixed set of dataset provides a finite distribution of data samples, the generalization of a learning pose estimation network is limited in terms of pose, RGB and viewpoint spaces. We further propose to augment the data automatically such that the augmented pose sampling is performed in favor of training pose estimators generalization performance. Such auto-augmentation of poses is performed within a learning feature space in order to avoid computational burden of generating synthetic sample for every iteration of updates. The proposed
effort can be considered as generating and utilizing synthetic samples for network training in the feature space. This allows training efficiency by requiring less number of real data samples, enhanced generalization power over multiple dataset domains and estimation performance caused by efficient augmentation.2D ์ด๋ฏธ์ง์์ ์ฌ๋์ ์ ๋ชจ์๊ณผ ํฌ์ฆ๋ฅผ ์ธ์ํ๊ณ ๊ตฌํํ๋ ์ฐ๊ตฌ๋ ๊ฐ ์๊ฐ๋ฝ ์กฐ์ธํธ๋ค์ 3D ์์น๋ฅผ ๊ฒ์ถํ๋ ๊ฒ์ ๋ชฉํ๋กํ๋ค. ์ ํฌ์ฆ๋ ์๊ฐ๋ฝ ์กฐ์ธํธ๋ค๋ก ๊ตฌ์ฑ๋์ด ์๊ณ ์๋ชฉ ๊ด์ ๋ถํฐ MCP, PIP, DIP ์กฐ์ธํธ๋ค๋ก ์ฌ๋ ์์ ๊ตฌ์ฑํ๋ ์ ์ฒด์ ์์๋ค์ ์๋ฏธํ๋ค. ์ ํฌ์ฆ ์ ๋ณด๋ ๋ค์ํ ๋ถ์ผ์์ ํ์ฉ๋ ์ ์๊ณ ์ ์ ์ค์ณ ๊ฐ์ง ์ฐ๊ตฌ ๋ถ์ผ์์ ์ ํฌ์ฆ ์ ๋ณด๊ฐ ๋งค์ฐ ํ๋ฅญํ ์
๋ ฅ ํน์ง ๊ฐ์ผ๋ก ์ฌ์ฉ๋๋ค.
์ฌ๋์ ์ ํฌ์ฆ ๊ฒ์ถ ์ฐ๊ตฌ๋ฅผ ์ค์ ์์คํ
์ ์ ์ฉํ๊ธฐ ์ํด์๋ ๋์ ์ ํ๋, ์ค์๊ฐ์ฑ, ๋ค์ํ ๊ธฐ๊ธฐ์ ์ฌ์ฉ ๊ฐ๋ฅํ๋๋ก ๊ฐ๋ฒผ์ด ๋ชจ๋ธ์ด ํ์ํ๊ณ , ์ด๊ฒ์ ๊ฐ๋ฅ์ผ ํ๊ธฐ ์ํด์ ํ์ตํ ์ธ๊ณต์ ๊ฒฝ๋ง ๋ชจ๋ธ์ ํ์ตํ๋๋ฐ์๋ ๋ง์ ๋ฐ์ดํฐ๊ฐ ํ์๋ก ํ๋ค. ํ์ง๋ง ์ฌ๋ ์ ํฌ์ฆ๋ฅผ ์ธก์ ํ๋ ๊ธฐ๊ณ๋ค์ด ๊ฝค ๋ถ์์ ํ๊ณ , ์ด ๊ธฐ๊ณ๋ค์ ์ฅ์ฐฉํ๊ณ ์๋ ์ด๋ฏธ์ง๋ ์ฌ๋ ์ ํผ๋ถ ์๊ณผ๋ ๋ง์ด ๋ฌ๋ผ ํ์ต์ ์ฌ์ฉํ๊ธฐ๊ฐ ์ ์ ํ์ง ์๋ค. ๊ทธ๋ฌ๊ธฐ ๋๋ฌธ์ ๋ณธ ๋
ผ๋ฌธ์์๋ ์ด๋ฌํ ๋ฌธ์ ๋ฅผ ํด๊ฒฐํ๊ธฐ ์ํด ์ธ๊ณต์ ์ผ๋ก ๋ง๋ค์ด๋ธ ๋ฐ์ดํฐ๋ฅผ ์ฌ๊ฐ๊ณต ๋ฐ ์ฆ๋ํ์ฌ ํ์ต์ ์ฌ์ฉํ๊ณ , ๊ทธ๊ฒ์ ํตํด ๋ ์ข์ ํ์ต์ฑ๊ณผ๋ฅผ ์ด๋ฃจ๋ ค๊ณ ํ๋ค.
์ธ๊ณต์ ์ผ๋ก ๋ง๋ค์ด๋ธ ์ฌ๋ ์ ์ด๋ฏธ์ง ๋ฐ์ดํฐ๋ค์ ์ค์ ์ฌ๋ ์ ํผ๋ถ์๊ณผ๋ ๋น์ทํ ์ง์ธ์ ๋ํ
์ผํ ํ
์ค์ณ๊ฐ ๋ง์ด ๋ฌ๋ผ, ์ค์ ๋ก ์ธ๊ณต ๋ฐ์ดํฐ๋ฅผ ํ์ตํ ๋ชจ๋ธ์ ์ค์ ์ ๋ฐ์ดํฐ์์ ์ฑ๋ฅ์ด ํ์ ํ ๋ง์ด ๋จ์ด์ง๋ค. ์ด ๋ ๋ฐ์ดํ์ ๋๋ฉ์ธ์ ์ค์ด๊ธฐ ์ํด์ ์ฒซ๋ฒ์งธ๋ก๋ ์ฌ๋์์ ๊ตฌ์กฐ๋ฅผ ๋จผ์ ํ์ต ์ํค๊ธฐ์ํด, ์ ๋ชจ์
์ ์ฌ๊ฐ๊ณตํ์ฌ ๊ทธ ์์ง์ ๊ตฌ์กฐ๋ฅผ ํ์คํ ์๊ฐ์ ์ ๋ณด๋ฅผ ๋บ ๋๋จธ์ง๋ง ์ค์ ์ ์ด๋ฏธ์ง ๋ฐ์ดํฐ์ ํ์ตํ์๊ณ ํฌ๊ฒ ํจ๊ณผ๋ฅผ ๋ด์๋ค.
์ด๋ ์ค์ ์ฌ๋ ์๋ชจ์
์ ๋ชจ๋ฐฉํ๋ ๋ฐฉ๋ฒ๋ก ์ ์ ์ํ์๋ค.
๋๋ฒ์งธ๋ก๋ ๋ ๋๋ฉ์ธ์ด ๋ค๋ฅธ ๋ฐ์ดํฐ๋ฅผ ๋คํธ์ํฌ ํผ์ณ ๊ณต๊ฐ์์ align์์ผฐ๋ค. ๊ทธ๋ฟ๋ง์๋๋ผ ์ธ๊ณต ํฌ์ฆ๋ฅผ ํน์ ๋ฐ์ดํฐ๋ค๋ก augmentํ์ง ์๊ณ ๋คํธ์ํฌ๊ฐ ๋ง์ด ๋ณด์ง ๋ชปํ ํฌ์ฆ๊ฐ ๋ง๋ค์ด์ง๋๋ก ํ๋์ ํ๋ฅ ๋ชจ๋ธ๋ก์ ์ค์ ํ์ฌ ๊ทธ๊ฒ์์ ์ํ๋งํ๋ ๊ตฌ์กฐ๋ฅผ ์ ์ํ์๋ค.
๋ณธ ๋
ผ๋ฌธ์์๋ ์ธ๊ณต ๋ฐ์ดํฐ๋ฅผ ๋ ํจ๊ณผ์ ์ผ๋ก ์ฌ์ฉํ์ฌ annotation์ด ์ด๋ ค์ด ์ค์ ๋ฐ์ดํฐ๋ฅผ ๋ ๋ชจ์ผ๋ ์๊ณ ์ค๋ฌ์ ์์ด ์ธ๊ณต ๋ฐ์ดํฐ๋ค์ ๋ ํจ๊ณผ์ ์ผ๋ก ๋ง๋ค์ด ๋ด๋ ๊ฒ ๋ฟ๋ง ์๋๋ผ, ๋ ์์ ํ๊ณ ์ง์ญ์ ํน์ง๊ณผ ์๊ฐ์ ํน์ง์ ํ์ฉํด์ ํฌ์ฆ์ ์ฑ๋ฅ์ ๊ฐ์ ํ๋ ๋ฐฉ๋ฒ๋ค์ ์ ์ํ๋ค. ๋ํ, ๋คํธ์ํฌ๊ฐ ์ค์ค๋ก ํ์ํ ๋ฐ์ดํฐ๋ฅผ ์ฐพ์์ ํ์ตํ ์ ์๋ ์๋ ๋ฐ์ดํฐ ์ฆ๋ ๋ฐฉ๋ฒ๋ก ๋ ํจ๊ป ์ ์ํ์๋ค. ์ด๋ ๊ฒ ์ ์๋ ๋ฐฉ๋ฒ์ ๊ฒฐํฉํด์ ๋ ๋์ ์ ํฌ์ฆ์ ์ฑ๋ฅ์ ํฅ์ ํ ์ ์๋ค.1. Introduction 1
2. Related Works 14
3. Preliminaries: 3D Hand Mesh Model 27
4. SeqHAND: RGB-sequence-based 3D Hand Pose and Shape Estimation 31
5. Hand Pose Auto-Augment 66
6. Conclusion 85
Abstract (Korea) 101
๊ฐ์ฌ์ ๊ธ 103๋ฐ
- โฆ