214 research outputs found

    Feedback-prop: Convolutional Neural Network Inference under Partial Evidence

    Full text link
    We propose an inference procedure for deep convolutional neural networks (CNNs) when partial evidence is available. Our method consists of a general feedback-based propagation approach (feedback-prop) that boosts the prediction accuracy for an arbitrary set of unknown target labels when the values for a non-overlapping arbitrary set of target labels are known. We show that existing models trained in a multi-label or multi-task setting can readily take advantage of feedback-prop without any retraining or fine-tuning. Our feedback-prop inference procedure is general, simple, reliable, and works on different challenging visual recognition tasks. We present two variants of feedback-prop based on layer-wise and residual iterative updates. We experiment using several multi-task models and show that feedback-prop is effective in all of them. Our results unveil a previously unreported but interesting dynamic property of deep CNNs. We also present an associated technical approach that takes advantage of this property for inference under partial evidence in general visual recognition tasks.Comment: Accepted to CVPR 201

    Men Also Like Shopping: Reducing Gender Bias Amplification using Corpus-level Constraints

    Full text link
    Language is increasingly being used to define rich visual recognition problems with supporting image collections sourced from the web. Structured prediction models are used in these tasks to take advantage of correlations between co-occurring labels and visual input but risk inadvertently encoding social biases found in web corpora. In this work, we study data and models associated with multilabel object classification and visual semantic role labeling. We find that (a) datasets for these tasks contain significant gender bias and (b) models trained on these datasets further amplify existing bias. For example, the activity cooking is over 33% more likely to involve females than males in a training set, and a trained model further amplifies the disparity to 68% at test time. We propose to inject corpus-level constraints for calibrating existing structured prediction models and design an algorithm based on Lagrangian relaxation for collective inference. Our method results in almost no performance loss for the underlying recognition task but decreases the magnitude of bias amplification by 47.5% and 40.5% for multilabel classification and visual semantic role labeling, respectively.Comment: 11 pages, published in EMNLP 201

    Clinicopathological characteristics and prognosis of synchronous brain metastases from non-small cell lung cancer compared with metachronous brain metastases

    Get PDF
    PurposeBrain metastasis (BM) from non-small cell lung cancer (NSCLC) is a serious complication severely affecting patients’ prognoses. We aimed to compare the clinicopathological features and prognosis of synchronous and metachronous BM from NSCLC.MethodsClinical data of 461 patients with brain metastases from NSCLC who visited the Cancer Hospital of China Medical University from 2005 to 2017 were retrospectively collected. We analyzed the pathophysiological characteristics of synchronous and metachronous BM from NSCLC and survival rates of the patients. Propensity score matching analysis was used to reduce bias between groups. In addition, we used the Kaplan-Meier method for survival analysis, log-rank test to compare survival rates, and Cox proportional hazards regression model for multivariate prognosis analysis.ResultsAmong 461 patients with BM, the number of people who met the inclusion criteria was 400 cases, and after 1:2 propensity score matching,130 had synchronous BM and 260 had metachronous BM. The survival time was longer for metachronous BM in driver mutation-negative patients with squamous cell carcinoma than synchronous BM. Conversely, metachronous and synchronous BM with gene mutations and adenocarcinoma showed no differences in survival time. Multivariate analysis showed that metachronous BM was an independent prognostic factor for overall survival. Furthermore, the pathological type squamous cell carcinoma and Karnofsky Performance Status score <80 were independent risk factors affecting overall survival.ConclusionBM status is an independent factor influencing patient outcome. Moreover, synchronous and metachronous BM from NSCLC differ in gene mutation profile, pathological type, and disease progression and hence require different treatments

    Variation of Gender Biases in Visual Recognition Models Before and After Finetuning

    Full text link
    We introduce a framework to measure how biases change before and after fine-tuning a large scale visual recognition model for a downstream task. Deep learning models trained on increasing amounts of data are known to encode societal biases. Many computer vision systems today rely on models typically pretrained on large scale datasets. While bias mitigation techniques have been developed for tuning models for downstream tasks, it is currently unclear what are the effects of biases already encoded in a pretrained model. Our framework incorporates sets of canonical images representing individual and pairs of concepts to highlight changes in biases for an array of off-the-shelf pretrained models across model sizes, dataset sizes, and training objectives. Through our analyses, we find that (1) supervised models trained on datasets such as ImageNet-21k are more likely to retain their pretraining biases regardless of the target dataset compared to self-supervised models. We also find that (2) models finetuned on larger scale datasets are more likely to introduce new biased associations. Our results also suggest that (3) biases can transfer to finetuned models and the finetuning objective and dataset can impact the extent of transferred biases.Comment: 10 pages, 3 Figure

    XVTP3D: Cross-view Trajectory Prediction Using Shared 3D Queries for Autonomous Driving

    Full text link
    Trajectory prediction with uncertainty is a critical and challenging task for autonomous driving. Nowadays, we can easily access sensor data represented in multiple views. However, cross-view consistency has not been evaluated by the existing models, which might lead to divergences between the multimodal predictions from different views. It is not practical and effective when the network does not comprehend the 3D scene, which could cause the downstream module in a dilemma. Instead, we predicts multimodal trajectories while maintaining cross-view consistency. We presented a cross-view trajectory prediction method using shared 3D Queries (XVTP3D). We employ a set of 3D queries shared across views to generate multi-goals that are cross-view consistent. We also proposed a random mask method and coarse-to-fine cross-attention to capture robust cross-view features. As far as we know, this is the first work that introduces the outstanding top-down paradigm in BEV detection field to a trajectory prediction problem. The results of experiments on two publicly available datasets show that XVTP3D achieved state-of-the-art performance with consistent cross-view predictions.Comment: 11 pages, 6 figures, accepted by IJCAI 2

    Gender Biases in Automatic Evaluation Metrics for Image Captioning

    Full text link
    Model-based evaluation metrics (e.g., CLIPScore and GPTScore) have demonstrated decent correlations with human judgments in various language generation tasks. However, their impact on fairness remains largely unexplored. It is widely recognized that pretrained models can inadvertently encode societal biases, thus employing these models for evaluation purposes may inadvertently perpetuate and amplify biases. For example, an evaluation metric may favor the caption "a woman is calculating an account book" over "a man is calculating an account book," even if the image only shows male accountants. In this paper, we conduct a systematic study of gender biases in model-based automatic evaluation metrics for image captioning tasks. We start by curating a dataset comprising profession, activity, and object concepts associated with stereotypical gender associations. Then, we demonstrate the negative consequences of using these biased metrics, including the inability to differentiate between biased and unbiased generations, as well as the propagation of biases to generation models through reinforcement learning. Finally, we present a simple and effective way to mitigate the metric bias without hurting the correlations with human judgments. Our dataset and framework lay the foundation for understanding the potential harm of model-based evaluation metrics, and facilitate future works to develop more inclusive evaluation metrics.Comment: Accepted to EMNLP 202
    • …
    corecore