16,040 research outputs found

    Human semantic parsing for person re-identification

    Get PDF
    Person re-identification is a challenging task mainly dueto factors such as background clutter, pose, illuminationand camera point of view variations. These elements hinder the process of extracting robust and discriminative representations, hence preventing different identities from being successfully distinguished. To improve the representation learning, usually local features from human body partsare extracted. However, the common practice for such aprocess has been based on bounding box part detection.In this paper, we propose to adopt human semantic parsing which, due to its pixel-level accuracy and capabilityof modeling arbitrary contours, is naturally a better alternative. Our proposed SPReID integrates human semanticparsing in person re-identification and not only considerably outperforms its counter baseline, but achieves stateof-the-art performance. We also show that, by employinga simple yet effective training strategy, standard populardeep convolutional architectures such as Inception-V3 andResNet-152, with no modification, while operating solelyon full image, can dramatically outperform current stateof-the-art. Our proposed methods improve state-of-the-artperson re-identification on: Market-1501 [48] by ~17% inmAP and ~6% in rank-1, CUHK03 [24] by ~4% in rank-1and DukeMTMC-reID [50] by ~24% in mAP and ~10% inrank-1.Computer Vision FoundationWOS:000457843601020Scopus - Affiliation ID: 60105072Conference Proceedings Citation Index- ScienceProceedings PaperHaziran2018YÖK - 2017-1

    UniHCP: A Unified Model for Human-Centric Perceptions

    Full text link
    Human-centric perceptions (e.g., pose estimation, human parsing, pedestrian detection, person re-identification, etc.) play a key role in industrial applications of visual models. While specific human-centric tasks have their own relevant semantic aspect to focus on, they also share the same underlying semantic structure of the human body. However, few works have attempted to exploit such homogeneity and design a general-propose model for human-centric tasks. In this work, we revisit a broad range of human-centric tasks and unify them in a minimalist manner. We propose UniHCP, a Unified Model for Human-Centric Perceptions, which unifies a wide range of human-centric tasks in a simplified end-to-end manner with the plain vision transformer architecture. With large-scale joint training on 33 human-centric datasets, UniHCP can outperform strong baselines on several in-domain and downstream tasks by direct evaluation. When adapted to a specific task, UniHCP achieves new SOTAs on a wide range of human-centric tasks, e.g., 69.8 mIoU on CIHP for human parsing, 86.18 mA on PA-100K for attribute prediction, 90.3 mAP on Market1501 for ReID, and 85.8 JI on CrowdHuman for pedestrian detection, performing better than specialized models tailored for each task.Comment: Accepted for publication at the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2023 (CVPR 2023
    • …
    corecore