Out-of-distribution (OOD) inputs can compromise the performance and safety of
real world machine learning systems. While many methods exist for OOD detection
and work well on small scale datasets with lower resolution and few classes,
few methods have been developed for large-scale OOD detection. Existing
large-scale methods generally depend on maximum classification probability,
such as the state-of-the-art grouped softmax method. In this work, we develop a
novel approach that calculates the probability of the predicted class label
based on label distributions learned during the training process. Our method
performs better than current state-of-the-art methods with only a negligible
increase in compute cost. We evaluate our method against contemporary methods
across 14 datasets and achieve a statistically significant improvement with
respect to AUROC (84.2 vs 82.4) and AUPR (96.2 vs 93.7)