# Your Classifier Can Be Secretly A Likelihood-Based Ood Detector Anonymous authors Paper under double-blind review ## Abstract The ability to detect out-of-distribution (OOD) inputs is critical to guarantee the reliability of classification models deployed in an open environment. A fundamental challenge in OOD detection is that a discriminative classifier is typically trained to estimate the posterior probability p(y|z) for class y given an input z, but lacks the explicit likelihood estimation of p(z) ideally needed for OOD detection. While numerous OOD scoring functions have been proposed for classification models, these estimate scores are often heuristic-driven and cannot be rigorously interpreted as likelihood. To bridge the gap, we propose Intrinsic Likelihood (INK), which offers rigorous likelihood interpretation to modern discriminativebased classifiers. Specifically, our proposed INK score operates on the constrained latent embeddings of a discriminative classifier, which are modeled as a mixture of hyperspherical embeddings with constant norm. We draw a novel connection between the hyperspherical distribution and the intrinsic likelihood, which can be effectively optimized in modern neural networks. Extensive experiments on the OpenOOD benchmark empirically demonstrate that INK establishes a new state-of-the-art in a variety of OOD detection setups, including both far-OOD and near-OOD. ## 1 Introduction The problem of out-of-distribution (OOD) detection has garnered significant attention in reliable machine learning. OOD detection refers to the task of identifying test samples that do not belong to the same distribution as the training data, which can occur due to various factors such as the emergence of unseen classes in the model's operating environment. The ability to detect OOD samples is crucial for ensuring the reliability and safety of machine learning models (Amodei et al., 2016), especially in applications where the consequences of misidentifying OOD samples can be severe. To facilitate the separation between indistribution (ID) and OOD data, a rich line of works (Yang et al., 2021b) have developed OOD indicator functions, which, at the core, compute a scalar score for each input signifying the degree of OOD-ness. The design of these scoring functions has emerged as a critical cornerstone in achieving effective OOD detection. By definition, OOD data inherently diverges from ID data by means of their data density distributions, rendering likelihood or log-likelihood an ideal scoring function for detection. However, a fundamental dilemma in OOD detection is that a discriminative classifier is typically trained to estimate the posterior probability p(y|z) for class y given an input z, but lacks the explicit likelihood estimation of p(z) needed for OOD detection. While many OOD scoring functions have been proposed for classification models, they are either heuristic-driven and cannot be rigorously interpreted as log-likelihood, or impose strong assumptions on the density function that can fail to hold in practice. Against this backdrop, the key research question of this paper thus centers on: RQ: How to design an OOD scoring function that can offer rigorous likelihood interpretation to modern discriminative-based classifiers, without resorting to separate generative models? To bridge the gap, we introduce a novel framework INtrinsic liKelihood (INK) for classifier-based OOD detection, which can be theoretically reasoned from a log-likelihood perspective. Specifically, our proposed ![1_image_0.png](1_image_0.png) Intrinsic Likelihood Figure 1: Overview of intrinsic likelihood framework for OOD detection. The neural network is trained on the in-distribution (ID) data, which lies on the unit hypersphere in the latent space. The intrinsic likelihood score is theoretically equivalent to the log-likelihood log p(z), which suits OOD detection. INK score operates on the constrained latent embeddings of a discriminative classifier, which are modeled as a mixture of hyperspherical embeddings with constant norm (see Figure 1). Under our probabilistic model, each class has a mean vector in the hypersphere, where the class-conditional density function decays from the center exponentially by the similarity between an embedding to the center. The proposed INK score is formalized by drawing a novel connection between the hyperspherical distribution and the density function under our probabilistic model. Theoretically, we show that the INK score is equivalent to the log-likelihood score log p(z) (up to some constant difference), and thus can act rigorously as a density-based score for OOD detection. Though simple and elegant in hindsight, establishing this connection was non-trivial. We lay out the full theoretical derivation in Section 3.1, along with how to effectively optimize for the intrinsic likelihood in the context of discriminative neural networks in Section 3.2. We show that our theoretical guarantee indeed translates into competitive empirical performance (Section 4). We extensively evaluate our method on the latest OpenOOD benchmarks (Zhang et al., 2023a), containing CIFAR and ImageNet-1k as ID datasets. Our method exhibits significant performance improvements when compared to the state-of-the-art method ASH (Djurisic et al., 2023), which employs a heuristic-driven activation shaping approach and lacks a rigorous likelihood interpretation. Moreover, we show that our method is effective under both far-OOD and near-OOD scenarios, and generalizes to different model architectures including ViT (Dosovitskiy et al., 2021). The overall evaluation demonstrates notable strengths of our approach in both empirical performance and theoretical underpinnings. ## 2 Preliminaries Let X and Y = {1*, . . . , C*} represent the input space and ID label space, respectively. The joint ID distribution, represented as PXIYI , is a joint distribution defined over *X × Y*. During testing time, there are some unknown OOD joint distributions DXOYO defined over *X ×Y*c, where Y cis the complementary set of Y. We also denote p(x) as the density of the ID marginal distribution PXI . According to Fang et al. (2022), OOD detection can be formally defined as follows: Definition 2.1 (**OOD Detection**). Given labeled ID data Din = {(x1, y1), ...,(xN , yN )}*, which is drawn* from PXIYI independent and identically distributed, the aim of OOD detection is to learn a predictor g by using Din such that for any test data x: 1) if x *is drawn from* PXI , then g can classify x into correct ID classes, and 2) if x is drawn from PXO , then g can detect x *as OOD data.* Definition 2.2 (**Likelihood-based OOD Detector**). Given ID data density function pθ(·) *estimated by a* model fθ and a pre-defined threshold λ, then for any data x ∈ X , $$g(\mathbf{x})=I D,\;\;i f\,p_{\theta}(\mathbf{x})\geq\lambda;\;\;o t h e r w i s e,\;\;g(\mathbf{x})=O O D.$$ .) $\frac{1}{2}$ . g (x) = ID, if pθ(x) ≥ λ; *otherwise*, g (x) = OOD. (1) The performance of OOD detection thus heavily relies on the estimation of data density pθ(x). *However, in* supervised learning, the classifiers directly estimate the posterior probability pθ(y|x) *instead of the likelihood* pθ(x)*, which makes likelihood-based OOD detection non-trivial.* Our focus on discriminative models differs from unsupervised OOD detection using generative-based models, which do not offer classification capability. For completeness, we review these studies extensively in the related work (Section 6). ## 3 Methodology In this section, we introduce our proposed framework, INtrinsic liKelihood (INK), for OOD detection. We first provide an overview of the key challenge and motivation. Then, we introduce the notion of intrinsic likelihood and theoretically justify its feasibility as an indicator function of OOD in Section 3.1. Finally, in Section 3.2, we discuss how to optimize neural networks to achieve the desired intrinsic likelihood for OOD detection. A classifier is typically trained via maximum likelihood estimation (MLE) on the training dataset Din = {(xi, yi)} N i=1, in order to maximize the predictive probability for the corresponding class: $$\operatorname{argmax}_{\theta}\prod_{i=1}^{N}p(y=y_{i}\mid\mathbf{z}_{i}),\tag{1}$$ $$\left(2\right)$$ $$\left({\mathrm{3}}\right)$$ where i is the index of the sample, ziis the latent representation of xi, and N is the size of the training set. By the Bayes' rule, the posterior probability can be rewritten as the following under uniform prior: $$p(y\mid\mathbf{z})={\frac{p(\mathbf{z}\mid y)p(y)}{p(\mathbf{z})}}={\frac{p(\mathbf{z}\mid y)}{\sum_{j=1}^{C}p(\mathbf{z}\mid y=j)}}.$$ Thus, one can define S(z) = log PC j=1 p(z | y = j) as a measure of how well a given embedding matches the distribution of ID samples, which is proportional to the log-likelihood log p(z). Despite the connection, the challenge lies in how to concretely define the class-conditional latent distribution p(z | y), as well as how to efficiently optimize neural networks to achieve such latent distribution while training a classifier. In what follows, we address these two challenges. ## 3.1 Intrinsic Likelihood To address the challenge of defining class-conditional latent distributions, we model the representations by the von Mises-Fisher (vMF) distribution in the hypersphere. A hypersphere is a topological space that is equivalent to a standard (d − 1)-sphere, which represents the set of points in d-dimensional Euclidean space that are equidistant from the center (see Figure 1). The unit hypersphere is a special sphere with a unit radius and is denoted as S d−1:= {z ∈ R d|∥z∥2 = 1} in the (d − 1)-dimensional space. The formal definition of vMF distribution is given below: Definition 3.1 (von Mises-Fisher Distribution (Fisher, 1953)). *For a unit vector* z ∈ R din class c, the conditional probability density function of the vMF distribution is defined as $$p(\mathbf{z}\mid y=c)=Z_{d}(\kappa)\exp(\kappa\boldsymbol{\mu}_{c}^{\top}\mathbf{z}),$$ c z), (4) where µc ∈ R d denotes the mean direction of the class c, κ ≥ 0 denotes the concentration of the distribution around µc , and Zd(κ) *denotes the normalization factor.* $$\left({\boldsymbol{4}}\right)$$ Benefits of the probabilistic model. We model the latent representations as vMF distribution because it is a simple and expressive probability distribution in directional statistics. Compared with multivariate Gaussian distribution, vMF distribution avoids estimating large covariance matrices for high-dimensional data that is shown to be costly and unstable in literature (Chen et al., 2017). Meanwhile, choosing the vMF distribution allows us to have the features live on the unit hypersphere, which leads to several benefits. For example, fixed-norm vectors are known to improve training stability in modern machine learning where dot products are ubiquitous (Wang & Isola, 2020), and are beneficial for learning a good representation space. For this reason, hyperspherical representations have been popularly adopted in recent contrastive learning literature (Chen et al., 2020a; Khosla et al., 2020). Under this probabilistic model, an embedding z is assigned to the class c with the following probability $$p(y=c\mid\mathbf{z};\{\kappa,\boldsymbol{\mu}_{j}\}_{j=1}^{C})=\frac{Z_{d}(\kappa)\exp(\kappa\boldsymbol{\mu}_{c}^{\top}\mathbf{z})}{\sum_{j=1}^{C}Z_{d}(\kappa)\exp(\kappa\boldsymbol{\mu}_{j}^{\top}\mathbf{z})}$$ $$=\frac{\exp(\boldsymbol{\mu}_{c}^{\top}\mathbf{z}/\tau)}{\sum_{j=1}^{C}\exp(\boldsymbol{\mu}_{j}^{\top}\mathbf{z}/\tau)},$$ $$\mathbf{\dot{\theta}}$$ where τ = 1/κ denotes a temperature parameter. Now, we are ready to draw a novel connection between the hyperspherical distribution and the intrinsic likelihood, which is formalized below. Definition 3.2 (**Intrinsic Likelihood Score**). *We define the intrinsic likelihood function over* z ∈ R din terms of the log partition function (or denominator in Eq. 5): $$S(\mathbf{z};\{\tau,\boldsymbol{\mu}_{j}\}_{j=1}^{C})=\tau\cdot\log\sum_{j=1}^{C}\exp(\boldsymbol{\mu}_{j}^{\top}\mathbf{z}/\tau).$$ $$\left(6\right)$$ Theorem 3.1 shows that our scoring function in Eq. 6 can be functionally equivalent to the log-likelihood score, for OOD detection purposes. Theorem 3.1. Under uniform class prior, the intrinsic likelihood score S(z) *is a logarithmic function of the* density p(z)*, with a constant difference. The two measurements return the same level set for OOD detection.* Proof. The likelihood p(z) is the probability of observing a given embedding z under the vMF distribution. p(z) can be calculated as a summation of class-conditional likelihood p(z | y = j), weighted by the class prior: $$p({\bf z})=\sum_{j=1}^{C}p({\bf z}\mid y=j)p(y=j)\tag{7}$$ $$=\frac{Z_{d}(\kappa)}{C}\sum_{j=1}^{C}\exp(\mathbf{\mu}_{j}^{\top}{\bf z}/\tau),$$ where the right-hand side is obtained by plugging in Eq. 4 and class prior 1/C. By connecting the intrinsic likelihood S(z) in Eq. 6 and the density function p(z) in Eq. 7, we can show the following equivalence between the two: $$S(\mathbf{z};\{\tau,\mu_{j}\}_{j=1}^{C})=\tau\cdot\underbrace{\log p(\mathbf{z})}_{\log\mathrm{likelihood~of~\mathbf{z}}}+\underbrace{\tau\cdot\log\frac{C}{Z_{d}(1/\tau)}}_{\mathrm{const.~independent~of~\mathbf{z}}}$$ $$({\boldsymbol{8}})$$ Note that the second term is independent of z, and can be treated as a constant. Thus, S(z) is functionally equivalent to the log-likelihood score, for OOD detection. In principle, one could also derive the equivalence under imbalanced class priors, which we discuss in detail in Section 5. Intrinsic likelihood for OOD detection. Given the above equivalence between INK score and loglikelihood, we propose the following OOD detector: $$g_{\lambda}(\mathbf{x})=\mathbb{I}\{S(\mathbf{z})\geq\lambda\},$$ $$({\mathfrak{g}})$$ (x) = 1{S(z) ≥ λ}, (9) where samples with higher scores are classified as ID, and vice versa. The threshold λ is chosen based on the ID score at a certain percentile (e.g., 95%). ## 3.2 Optimizing Intrinsic Likelihood In Deep Neural Networks So far, we have established that intrinsic likelihood derived under the vMF distribution is desirable for OOD detection. Now, we discuss how to optimize neural networks for intrinsic likelihood. Maximum likelihood estimation. We aim to train the neural network, such that the hyperspherical embedding conforms to a mixture of vMF distributions defined in Eq. 5. Specifically, we consider a deep neural network h : *X 7→* R d which encodes an input x ∈ X to a normalized embedding z := h(x). To optimize for the vMF distribution, one can directly perform standard maximum likelihood estimation on the training dataset {(xi, yi)} N i=1: $$\mbox{argmax}_{\theta}\prod_{i=1}^{N}p(y=y_{i}\mid\mbox{\bf z}_{i};\{\kappa,\mu_{j}\}_{j=1}^{C}),\tag{10}$$ where i is the index of the sample, j is the index of class, and N is the size of the training set. In effect, this loss encourages each ID sample to have a high probability assigned to the correct class in the mixtures of the vMF distributions. While hyperspherical learning algorithms have been studied (Mettes et al., 2019; Du et al., 2022a; Ming et al., 2023), *none of the works explored its principled connection to intrinsic likelihood* for test-time OOD detection, which is our distinct focus. We discuss the differences further in Section *3.3.* Moreover, we show new insights below that minimizing this loss function effectively shapes the intrinsic likelihood to aid OOD detection. How does the loss function shape intrinsic likelihood? We show that the learning objective can induce a higher intrinsic likelihood for ID data. By taking negative-log likelihood, the objective function in Eq. 10 is equivalent to minimizing the following loss: $${\cal L}=\frac{1}{N}\sum_{i=1}^{N}-\log p(y=y_{i}\mid{\bf z}_{i};\{\tau,\mathbf{\mu}_{j}\}_{j=1}^{C}).\tag{1}$$ We can rewrite the loss by expanding Eq. 11 : $$\mathcal{L}=\frac{1}{N}\sum_{i=1}^{N}\left(-\log\frac{\exp(\boldsymbol{\mu}_{y_{i}}^{\top}\mathbf{z}_{i}/\tau)}{\sum_{j=1}^{C}\exp(\boldsymbol{\mu}_{j}^{\top}\mathbf{z}_{i}/\tau)}\right)$$ $$=\frac{1}{N}\sum_{i=1}^{N}\left(\frac{1}{\tau}s(\mathbf{z}_{i},y_{i})+\log\sum_{j=1}^{C}\exp(-s(\mathbf{z}_{i},j)/\tau)\right),$$ in the limit $\tau$ the above $\tau$ $$(11)$$ $$\quad(12)$$ where s(zi, y) = −µ ⊤ y zi. During optimization, the loss reduces s(zi, y) for the correct class while increasing it for other classes. This can be justified by analyzing the gradient of the loss function with respect to the model parameters θ, for an embedding z and its associated class label y: $$\partial\mathcal{L}(\mathbf{z},y;\theta)=\frac{1}{\tau}\frac{\partial s(\mathbf{z},y)}{\partial\theta}$$ $$-\frac{1}{\tau}\sum_{j=1}^{C}\frac{\partial s(\mathbf{z},j)}{\partial\theta}\cdot\frac{\exp(-s(\mathbf{z},j)/\tau)}{\sum_{c=1}^{C}\exp(-s(\mathbf{z},c)/\tau)}$$ $$=\frac{1}{\tau}\underbrace{\frac{\partial s(\mathbf{z},y)}{\partial\theta}(1-p(Y=y\mid\mathbf{z}))}_{\text{i for y}}-\underbrace{\sum_{j\neq y}\frac{\partial s(\mathbf{z},j)}{\partial\theta}p(Y=j\mid z)}_{\text{T for j}\neq y}\tag{14}$$ | | Training-time loss function | Test-time scoring function | Likelihood interpretation | |------------|-------------------------------|-----------------------------------|-----------------------------| | SSD+ | SupCon | Mahalanobis (parametric) | ✗ | | KNN+ | SupCon | KNN (non-parametric) | ✗ | | CIDER | vMF | KNN (non-parametric) | ✗ | | SIREN | vMF | KNN or max p(y|z) | ✗ | | INK (Ours) | vMF | Intrinsic likelihood (parametric) | ✓ | Table 1: Summary of key distinction among different hyperspherical approaches. The sample-wise intrinsic likelihood S(z) of ID data is S(z) = τ ·log PC c=1 exp(−s(z, c)/τ ), which is dominated by the term −s(z, y) with ground truth label. Hence, the training overall induces a higher intrinsic likelihood for ID data. By further connecting to our Theorem 3.1, this higher intrinsic likelihood can directly translate into high (↑) log p(z), for training data distribution. ## 3.3 Differences W.R.T. Existing Approaches Table 1 summarizes the key distinctions among prevalent hyperspherical methodologies, specifically in terms of their training-time loss functions and test-time scoring functions. Both SSD+ (Sehwag et al., 2021) and KNN+ (Sun et al., 2022b) employ the SupCon (Khosla et al., 2020) loss during training, which does not explicitly model the latent representations as vMF distributions. Instead of promoting instance-to-prototype similarity, SupCon promotes instance-to-instance similarity among positive pairs. SupCon's loss formulation thus does not directly correspond to the vMF distribution. Geometrically speaking, our framework directly operates on the vMF distribution, a key property that enables log-likelihood interpretation. Our approach also differs significantly from CIDER (Ming et al., 2023) and SIREN (Du et al., 2022a) in terms of OOD scoring function, which employ either KNN distance or maximum conditional probability max p(y|z). In contrast, our method establishes the novel connection between the hyperspherical representations and intrinsic likelihood for OOD detection, which enjoys rigorous theoretical interpretation from a density perspective. We show empirically in Section 4 that our proposed OOD score achieves competitive performance on different OOD detection benchmarks, and is much more computationally efficient by eliminating the need for time-consuming KNN searches. ## 4 Experiments 4.1 Setup Benchmarks. Our evaluation encompasses both far-OOD and near-OOD detection performance, following the same setup as the latest OpenOOD benchmark1(Zhang et al., 2023a). The benchmark includes CIFAR (Krizhevsky, 2009) and ImageNet-1k (Deng et al., 2009) as ID datasets. Due to space constraints, we primarily focus on the most challenging ImageNet benchmark in the main paper, and report the full results for CIFAR datasets in the Appendix E. For the ImageNet benchmark, iNaturalist (Van Horn et al., 2018), Textures (Cimpoi et al., 2014), and OpenImage-O (Wang et al., 2022) are employed as far-OOD datasets. In addition, we also evaluate on the latest NINCO (Bitterwolf et al., 2023) as the near-OOD dataset, which circumvents the issue of contaminated classes found in other near-OOD datasets (Vaze et al., 2022). We adhere to the same data split used in OpenOOD (Zhang et al., 2023a), which involves the removal of images with semantic overlap. Evaluation metrics and implementation details. We utilize the following metrics, which are widely used to assess the efficacy of OOD detection methods: (1) false positive rate (FPR@95) of OOD samples when the true positive rate of ID samples is set at 95%, (2) area under the receiver operating characteristic curve (AUROC), and (3) ID classification accuracy (ID ACC). Due to the space limit, we provide the implementation Table 2: OOD detection performance for ImageNet (ID) with ResNet-50, using OpenOOD benchmark (Zhang et al., **2023a).** INK achieves competitive performance compared to state-of-the-art methods. Statistical significance is estimated on 3 random independent runs. | | Far-OOD Datasets | | Near-OOD Dataset | | | | | | | | |-------------|---------------------------------------------|------------|--------------------|------------|------------|------------|------------|------------|------------|------------| | Method | iNaturalist | Textures | OpenImage-O | Average | NINCO | | | | | | | | FPR ↓ | AUROC ↑ | FPR ↓ | AUROC ↑ | FPR ↓ | AUROC ↑ | FPR ↓ | AUROC ↑ | FPR ↓ | AUROC ↑ | | | Methods using cross-entropy loss | | | | | | | | | | | MSP | 43.34 | 88.94 | 60.88 | 80.13 | 50.13 | 86.69 | 51.45 | 85.25 | 56.87 | 79.95 | | ODIN | 36.07 | 90.63 | 49.44 | 89.06 | 46.55 | 86.58 | 44.02 | 88.76 | 68.13 | 77.77 | | Mahalanobis | 73.80 | 63.67 | 42.77 | 69.27 | 72.15 | 51.96 | 62.91 | 61.63 | 78.83 | 62.38 | | Energy | 31.34 | 92.54 | 45.76 | 88.26 | 38.08 | 88.96 | 38.39 | 89.92 | 60.58 | 79.70 | | ViM | 30.67 | 92.65 | 10.49 | 86.52 | 32.82 | 90.78 | 24.66 | 89.98 | 62.28 | 78.63 | | ReAct | 16.73 | 96.34 | 29.63 | 91.87 | 32.57 | 93.41 | 26.31 | 93.87 | 55.91 | 81.73 | | DICE | 33.38 | 91.17 | 44.25 | 88.24 | 47.83 | 86.77 | 41.82 | 88.72 | 66.97 | 76.01 | | SHE | 34.03 | 89.56 | 35.30 | 90.50 | 54.98 | 83.33 | 41.44 | 87.80 | 69.75 | 76.48 | | ASH | 14.10 | 97.06 | 15.27 | 93.26 | 29.20 | 93.29 | 19.52 | 94.54 | 52.99 | 83.44 | | | Methods using hyperspherical representation | | | | | | | | | | | SSD+ | 39.82 | 88.41 | 71.90 | 84.86 | 59.05 | 81.65 | 56.92 | 84.97 | 61.74 | 80.61 | | KNN+ | 18.29 | 94.90 | 19.90 | 93.56 | 27.40 | 89.05 | 21.87 | 92.50 | 53.35 | 84.15 | | CIDER | 20.79 | 94.39 | 21.86 | 92.54 | 30.06 | 89.07 | 24.24 | 92.00 | 57.42 | 82.36 | | INK (ours) | 10.04±0.67 | 97.49±0.08 | 7.52±2.08 | 98.62±0.51 | 28.26±0.90 | 94.24±0.23 | 15.27±0.27 | 96.78±0.12 | 48.15±1.33 | 86.42±0.13 | | Score | Score compute time (µs↓) | | | | |---------------|----------------------------|-----------------|----------------------|------------------------| | | CIFAR-10 | CIFAR-100 | ImageNet-1k (α = 1%) | ImageNet-1k (α = 100%) | | KNN+ or CIDER | 3.5619 ± 0.2646 | 5.4886 ± 0.1854 | 3.6157 ± 0.1186 | 336.1780 ± 4.3871 | | INK (ours) | 0.0076 ± 0.0003 | 0.0157 ± 0.0007 | 0.2745 ± 0.0145 | 0.2745 ± 0.0145 | Table 3: **Average and standard deviation of inference time (per sample) for each ID dataset.** The INK score is computationally more efficient compared to the KNN search. details in the Appendix B. For INK score, we use test-time temperature τtest = 0.05 based on validation (see Appendix C). We further show the effect of the temperature in our ablation study (Section 4.3). ## 4.2 Main Results Intrinsic likelihood score establishes state-of-the-art performance. We extensively compare the performance of our approach with state-of-the-art methods. The compared methods are categorized as either hyperspherical-based or non-hyperspherical-based. All the non-hyperspherical-based methods, including MSP (Hendrycks & Gimpel, 2017), ODIN (Liang et al., 2018), Mahalanobis (Lee et al., 2018b), Energy (Liu et al., 2020), ViM (Wang et al., 2022), ReAct (Sun et al., 2021), DICE (Sun & Li, 2022), SHE (Zhang et al., 2023b), and ASH (Djurisic et al., 2023) use the softmax cross-entropy (CE) loss for model training. Hyperspherical-based methods, on the other hand, include SSD+ (Sehwag et al., 2021), KNN+ (Sun et al., 2022b), and CIDER (Ming et al., 2023). It is worth noting that we have adopted the recommended configurations proposed by prior works. Further details of each baseline method are included in Appendix B. As shown in Table 2, our method INK displays competitive OOD detection performance compared to existing ones on the far-OOD group of the ImageNet benchmark. Here we highlight two observations: (1) INK also outperforms the current state-of-the-art cross-entropy loss method ASH (Djurisic et al., 2023), which employs a heuristic-driven activation shaping approach and lacks a rigorous likelihood interpretation. In contrast, INK is derived from a more principled likelihood-based perspective while being empirically strong. (2) INK score outperforms the current state-of-the-art hyperspherical-based methods CIDER by 8.97% and KNN+ by 6.60% in FPR@95. Compared to CIDER, our method enjoys a significant performance boost by leveraging the likelihood interpretation within hyperspherical space, and a drastic reduction in computational cost (more in Section 4.3). Near-OOD detection. We also evaluate the performance of our method in the more challenging near-OOD scenario. As shown in Table 2, INK demonstrates a competitive performance, outperforming CIDER by 9.27% in FPR@95. This improvement underscores the versatility of the INK score, proving its effectiveness in near-OOD detection scenarios where out-of-distribution samples bear a close semantic resemblance to the ID samples. Moreover, our method shows an enhanced performance over ASH in near-OOD detection, with an improvement of 4.84% in FPR@95. This advancement also signifies progress in bridging the performance gap between hyperspherical-based methods and cross-entropy loss methods. CIFAR benchmark. We evaluate the performance of our method on the CIFAR benchmark. Consistent with the results observed in the ImageNet benchmark, the INK score displays competitive performance in handling both far-OOD and near-OOD groups, compared to the current state-of-the-art methods. Due to space constraints, further details are included in Appendix E. ## 4.3 Additional Abalations And Discussions Computational cost. The current state-of-the-art hyperspherical representation methods, such as KNN+ and CIDER, rely on a non-parametric KNN score, which requires a nearest neighbor search in the embedding space. Although recent studies on approximated nearest neighbor search have shown promise (Aumüller et al., 2020), the computation required grows with the size of the embedding pool N. In contrast, INK only incurs complexity O(C), where the number of classes C ≪ N is independent of ID sample size. Thus, our method can significantly reduce the computation overhead. To demonstrate this, we conduct an experiment to compare the average compute time between the KNN score and the INK score. In this experiment, we assess the performance of the model on CIFAR-10, CIFAR-100, and ImageNet-1k benchmarks using default values of k in Sun et al. (2022b). In particular, we measure the time taken to calculate the score per sample, excluding the time to extract the sample embedding from the neural network. KNN-based methods utilize the Faiss library (Johnson et al., 2019), which is optimized for efficient nearest-neighbor search. The size of the embedding pool is 50,000 for both CIFAR-10 and CIFAR-100, 12K for ImageNet (α = 1%), and 1.2M for ImageNet (α = 100%), where α denotes the fraction of training data used for the nearest neighbor search. Our results, shown in Table 3, demonstrate that INK is significantly faster than the KNN-based methods, making it computationally desirable for OOD detection tasks. Relationship with energy score. We discuss the relationship of INK with prior work by Liu et al. Liu et al. (2020), which attempted to interpret the energy score derived from a multi-class classifier from a likelihood-based view. The key distinction between the two methods lies in the geometry of the embedding space, from which the score is derived. Specifically, an energy score is defined as the negative log of the denominator in the softmax function: $$E(\mathbf{x};f_{\boldsymbol{\theta}})=-\tau\cdot\log\sum_{j=1}^{C}\exp(f_{\boldsymbol{\theta}}^{j}(\mathbf{x})/\tau),$$ ![7_image_0.png](7_image_0.png) iNaturalist Textures OpenImage-O NINCO Datasets 0 10 20 30 40 50 60 70 Method Energy Ours where f j θ (x) denotes the logit corresponding to the class j. Unfortunately, the energy score derived from the Euclidean space logits fθ(x) can be *unconstrained*, and does not necessarily reflect the loglikelihood (see Theorem A.1 and proof in Appendix). In contrast, our framework explicitly models the latent representations in the discriminative classifier as *constrained hyperspherical space*, which enables a rigorous interpretation from a log-likelihood perspective (*c.f.* Theorem 3.1). In our formulation (Eq. 5), the classifier's logit f j θ (x) is equivalent to µ ⊤ j z, where z is the penultimate layer embedding of input x and µj can be interpreted as the weight vector connecting the embedding and output logit. Different from Liu et al. (2020), the embedding z and the weight connection µj are explicitly constrained together to form a meaningful probabilistic distribution during our optimization. Figure 2: **Comparison of Energy vs. INK on Ima-**Greedy (w=0) ARGS-greedy (k=10, w=1.5) geNet (ID) with ResNet-50, using OpenOOD benchmark Zhang et al. **(2023a)**. INK consistently outperforms Energy on far-OOD and near-OOD datasets. ![8_image_0.png](8_image_0.png) Figure 3: **Density plots illustrating the distribution of energy score (left) and INK (right).** Lighter and darker blue shades represent the distributions of scores for the ID and OOD datasets, respectively. 0.0 0.2 0.4 0.6 0.8 1.0 Test temperature 85.5 86.0 86.5 87.0 87.5 88.0 88.5 Figure 2 illustrates a comparison of the performance on the ImageNet benchmark. Our method reduces the average FPR@95 by 23.12%, which validates that the proposed intrinsic likelihood score is significantly more effective for OOD detection. The performance improvement can also be reasoned theoretically, as we show in Theorem 3.1. Overall, our method enjoys both strong empirical performance and theoretical justification. Additionally, for a more detailed understanding, we have included a visualization of ID-OOD score distributions in Figure 3. We illustrate a comparison between the distributions of energy and INK scores for ImageNet (ID) vs. Textures (OOD). The noticeable separability of using INK score enhances the performance in OOD detection. Ablation study on test-time temperature τ . We conduct an ablation study on the test-time temperature parameter to investigate its impact on OOD detection performance, as shown in Figure 4. The curve highlights that the optimal value of τ roughly matches the training-time temperature. This also matches our theory, where INK under the same temperature τ in testing allows recovering the log-likelihood log p(z) (up to some constant) learned in training time. Ablation study on different model architecture. To explore the impact of model architecture on the performance, we additionally evaluate the Vision Transformer (ViT-B-16) (Dosovitskiy et al., 2021). As shown in Table 7, our INK score maintains competitive performance with the ViT model. This underscores the adaptability of our method across various model architectures. Further details are included in Appendix F. ![8_image_1.png](8_image_1.png) Average AUROC (%) Figure 4: **Ablation on test-time temperatures.** The results are averaged across the far-OOD test sets and **3 random training runs**, based on ResNet-34 trained on CIFAR-100. ## 5 Generalization To Class-Imbalanced Datasets In the proof of Theorem 3.1, we adopt the commonly assumed scenario prevalent in OOD detection literature, where datasets are class-balanced with a uniform prior distribution over classes, i.e., p(y = c) = 1/C. However, we acknowledge that this assumption may not always hold in real-world scenarios. To further demonstrate the generality and effectiveness of our method, we extend it to account for non-uniform class priors. We perform standard maximum likelihood estimation on the training dataset {(xi, yi)} N i=1: $$\operatorname{argmax}_{\theta}\prod_{i=1}^{N}p(y=y_{i}\mid\mathbf{z}_{i};\{\kappa,\boldsymbol{\mu}_{j}\}_{j=1}^{C}),\tag{1}$$ $$\left(15\right)$$ where i is the index of the sample, j is the index of the class, and N is the size of the training set. The posterior probability is given by: $$p(y=c\mid\mathbf{z};\{\kappa,\mu_{j}\}_{j=1}^{C})={\frac{p(y=c)\exp(\mu_{c}^{\top}\mathbf{z}/\tau)}{\sum_{j=1}^{C}p(y=j)\exp(\mu_{j}^{\top}\mathbf{z}/\tau)}},$$ which accounts for the class prior probability, allowing for non-uniform class distributions. We define the generalized intrinsic likelihood as: $$S(\mathbf{z};\{\tau,\mu_{j}\}_{j=1}^{C})=\tau\cdot\log\sum_{j=1}^{C}p(y=j)\exp(\mu_{j}^{\top}\mathbf{z}/\tau),$$ $$(16)$$ $$(17)$$ which maintains functional equivalence to the log-likelihood for OOD detection, even when the class priors are not uniform. We further verify this class-imbalanced setting empirically by eliminating half of the training examples from the first 5 classes in CIFAR-10 and the first 10 classes in CIFAR-100. Table 4 compares the FPR@95 between ASH and the generalized intrinsic likelihood, which shows that our method significantly outperforms state-of-the-art in both far-OOD and near-OOD detection tasks. Table 4: **Comparison of OOD Performance with Imbalanced Class Prior.** Generalized INK achieves competitive performance compared to state-of-the-art method ASH across different datasets. CIFAR-10 Imbalanced (ResNet-18) CIFAR-100 Imbalanced (ResNet-34) | Dataset | ASH | Ours | |-------------------|-------|--------| | Far-OOD MNIST | 48.29 | 37.72 | | SVHN | 33.19 | 10.47 | | Texture | 55.31 | 34.29 | | Places365 | 67.81 | 62.47 | | Far-OOD Average | 51.15 | 36.24 | | Near-OOD CIFAR-10 | 74.39 | 76.36 | | TIN | 65.74 | 53.60 | | Near-OOD Average | 70.07 | 64.98 | | Dataset | ASH | Ours | |--------------------|-------|--------| | Far-OOD MNIST | 25.60 | 23.97 | | SVHN | 52.73 | 9.12 | | Texture | 42.00 | 18.69 | | Places365 | 40.28 | 26.12 | | Far-OOD Average | 40.15 | 19.48 | | Near-OOD CIFAR-100 | 53.44 | 40.33 | | TIN | 48.38 | 28.47 | | Near-OOD Average | 50.91 | 34.40 | ## 6 Related Works OOD detection for classification models. OOD detection is an essential component for reliable machine learning systems that operate in the open world. Recent studies in OOD detection focus on post-hoc strategies and train-time regularization (Yang et al., 2021b). In particular, post-hoc strategies focus on deriving test-time OOD scoring functions, including confidence-based scores (Hendrycks & Gimpel, 2017; Liang et al., 2018), distance-based score (Lee et al., 2018b; Tack et al., 2020; Sehwag et al., 2021; Sastry & Oore, 2020; Ren et al., 2021; Sun et al., 2022b; Du et al., 2022a; Ming et al., 2022; Ren et al., 2023), energy-based score (Liu et al., 2020; Wang et al., 2021; Morteza & Li, 2022; Zhang et al., 2023b; Jiang et al., 2023; Lafon et al., 2023; Peng et al., 2024), gradient-based score (Huang et al., 2021; Behpour et al., 2023; Chen et al., 2023; Liu et al., 2023a), and multimodal score (Ming et al., 2022; Miyai et al., 2023). Recent high-performing methods focus on suppressing activations of pre-trained models (Sun et al., 2021; Sun & Li, 2022; Djurisic et al., 2023; Xu et al., 2023b). On the other hand, train-time regularization approaches (Bevandić et al., 2018; Malinin & Gales, 2018; Lee et al., 2018a; Geifman & El-Yaniv, 2019; Hendrycks et al., 2019; Hein et al., 2019; Meinke & Hein, 2020; Van Amersfoort et al., 2020; Yang et al., 2021a; Du et al., 2022b; Wei et al., 2022; Katz-Samuels et al., 2022; Du et al., 2022a; Huang et al., 2022; Colombo et al., 2022; Yu et al., 2023; Tao et al., 2023; Ming et al., 2023; Wang et al., 2023; Xu et al., 2023a) design an algorithm to encourage the model to produce desired properties. Different from prior works, we regularize a model to produce hyperspherical representations aligned with the vMF distribution and propose intrinsic likelihood as a test-time scoring function, which has not been explored in the past. Moreover, *our method offers theoretical justification based* on the likelihood view, which is lacking in previous approaches. Likelihood-based OOD detection. Deep generative models have been explored for unsupervised OOD detection (Nalisnick et al., 2019; Ren et al., 2019; Abati et al., 2019; Choi et al., 2019; Serrà et al., 2020; Xiao et al., 2020; Wang et al., 2020; Schirrmeister et al., 2020; Sastry & Oore, 2020; Kirichenko et al., 2020; Li et al., 2022; Cai & Li, 2023; Liu et al., 2023b). While these approaches directly estimate the likelihood of ID data, the unsupervised nature *prohibits classification ability*. Moreover, issues have been highlighted that deep generative models can frequently assign spuriously higher likelihoods to OOD than ID samples (Ren et al., 2019), and the performance can often lag behind the discriminative approaches (Yang et al., 2021b). In contrast, our method circumvents this issue by virtue of its supervised learning objective, leveraging ground truth semantic labels for each ID data point. Consequently, the latent embeddings are shaped and guided by image semantics without necessarily overfitting on background elements. Hyperspherical representation. Hyperspherical representation under vMF distribution has been extensively used in various machine learning applications, such as supervised classification (Kobayashi, 2021; Scott et al., 2021; Govindarajan et al., 2023), face verification (Hasnat et al., 2017; Conti et al., 2022), generative modeling (Davidson et al., 2018), segmentation (Hwang et al., 2019; Sun et al., 2022a), and clustering (Gopal & Yang, 2014). In addition, some researchers have utilized the vMF distribution for anomaly detection by employing generative models and using it as the prior for zero-shot learning (Chen et al., 2020b) and document analysis (Zhuang et al., 2017). Recent studies devise vMF-based learning for OOD detection, which offers a direct and clear geometrical interpretation of hyperspherical embeddings. Specifically, Du et al. (2022a) explores modeling representations into vMF distribution for object-level OOD detection, while CIDER (Ming et al., 2023) shows that additional regularization for large angular distances among different class prototypes is crucial for achieving strong OOD detection performance. Our work adopts a vMF-based learning approach for OOD detection and builds upon these recent studies by regularizing a model to produce representations that align with the vMF distribution and proposing a new intrinsic likelihood scoring function for test-time OOD detection. ## 7 Conclusion In this paper, we introduce intrinsic likelihood for detecting out-of-distribution samples. Our approach is based on density estimation in the hyperspherical space, which enjoys a sound interpretation from a log-likelihood perspective. Through extensive experiments on common OOD detection benchmarks, we demonstrate the state-of-the-art performance of our approach for both far-OOD and near-OOD detection. Moreover, our method outperforms KNN-based approaches in terms of computational efficiency. Overall, our proposed intrinsic likelihood score provides a promising solution for both effective and efficient OOD detection. ## 8 Broader Impacts OOD detection can be used to improve the reliability of machine learning models, which can have a significant positive impact on society. In situations where saving features is a concern, the use of the KNN score may not be feasible as it requires access to a certain amount of labeled data. In contrast, our INK score does not require access to a pool of ID data. This makes it a valuable tool in various applications and industries where sensitive information needs to be protected. It is important to note that the usefulness and effectiveness of OOD detection may vary across different applications and domains, and careful evaluation and validation are necessary before its deployment. ## References Davide Abati, Angelo Porrello, Simone Calderara, and Rita Cucchiara. Latent Space Autoregression for Novelty Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision and Pattern Recognition, 2019. Dario Amodei, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete problems in ai safety. *arXiv preprint arXiv:1606.06565*, 2016. Martin Aumüller, Erik Bernhardsson, and Alexander Faithfull. Ann-benchmarks: A benchmarking tool for approximate nearest neighbor algorithms. *Information Systems*, 87, 2020. Sima Behpour, Thang Doan, Xin Li, Wenbin He, Liang Gou, and Liu Ren. Gradorth: A simple yet efficient out-of-distribution detection with orthogonal projection of gradients. In Advances in Neural Information Processing Systems, 2023. Petra Bevandić, Ivan Krešo, Marin Oršić, and Siniša Šegvić. Discriminative out-of-distribution detection for semantic segmentation. *arXiv preprint arXiv:1808.07703*, 2018. Julian Bitterwolf, Maximilian Mueller, and Matthias Hein. In or out? fixing imagenet out-of-distribution detection evaluation. In *International Conference on Machine Learning*, 2023. Mu Cai and Yixuan Li. Out-of-distribution detection via frequency-regularized generative models. In Proceedings of IEEE/CVF Winter Conference on Applications of Computer Vision, 2023. Jinggang Chen, Junjie Li, Xiaoyang Qu, Jianzong Wang, Jiguang Wan, and Jing Xiao. GAIA: Delving into gradient-based attribution abnormality for out-of-distribution detection. In *Thirty-seventh Conference on* Neural Information Processing Systems, 2023. Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations. In *International conference on machine learning*, pp. 1597–1607. PMLR, 2020a. Xingyu Chen, Xuguang Lan, Fuchun Sun, and Nanning Zheng. A boundary based out-of-distribution classifier for generalized zero-shot learning. In *European Conference on Computer Vision*, 2020b. Xixian Chen, Michael R Lyu, and Irwin King. Toward efficient and accurate covariance matrix estimation on compressed data. In *International Conference on Machine Learning*, pp. 767–776. PMLR, 2017. Hyunsun Choi, Eric Jang, and Alexander A. Alemi. Waic, but why? generative ensembles for robust anomaly detection, 2019. Mircea Cimpoi, Subhransu Maji, Iasonas Kokkinos, Sammy Mohamed, and Andrea Vedaldi. Describing textures in the wild. In *IEEE Conference on Computer Vision and Pattern Recognition*, 2014. Pierre Colombo, Eduardo Dadalto Câmara Gomes, Guillaume Staerman, Nathan Noiry, and Pablo Piantanida. Beyond mahalanobis distance for textual OOD detection. In *Advances in Neural Information Processing* Systems, 2022. Jean-Rémy Conti, Nathan Noiry, Stephan Clemencon, Vincent Despiegel, and Stéphane Gentric. Mitigating gender bias in face recognition using the von mises-Fisher mixture model. In *Proceedings of the 39th* International Conference on Machine Learning, 2022. Tim R. Davidson, Luca Falorsi, Nicola De Cao, Thomas Kipf, and Jakub M. Tomczak. Hyperspherical variational auto-encoders. In *34th Conference on Uncertainty in Artificial Intelligence*, 2018. Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In *2009 IEEE Conference on Computer Vision and Pattern Recognition*, 2009. Li Deng. The mnist database of handwritten digit images for machine learning research [best of the web]. Signal Processing Magazine, IEEE, 29:141–142, 11 2012. doi: 10.1109/MSP.2012.2211477. Andrija Djurisic, Nebojsa Bozanic, Arjun Ashok, and Rosanne Liu. Extremely simple activation shaping for out-of-distribution detection. In *The Eleventh International Conference on Learning Representations*, 2023. Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby. An image is worth 16x16 words: Transformers for image recognition at scale. In International Conference on Learning Representations, 2021. Xuefeng Du, Gabriel Gozum, Yifei Ming, and Yixuan Li. Siren: Shaping representations for detecting out-of-distribution objects. In *Advances in Neural Information Processing Systems*, 2022a. Xuefeng Du, Zhaoning Wang, Mu Cai, and Yixuan Li. Vos: Learning what you don't know by virtual outlier synthesis. *Proceedings of the International Conference on Learning Representations*, 2022b. Zhen Fang, Yixuan Li, Jie Lu, Jiahua Dong, Bo Han, and Feng Liu. Is out-of-distribution detection learnable? In *Advances in Neural Information Processing Systems*, 2022. Ronald Aylmer Fisher. Dispersion on a sphere. *Proceedings of the Royal Society of London. Series A.* Mathematical and Physical Sciences, 217(1130):295–305, 1953. Yonatan Geifman and Ran El-Yaniv. SelectiveNet: A deep neural network with an integrated reject option. In *Proceedings of the 36th International Conference on Machine Learning*, 2019. Siddharth Gopal and Yiming Yang. Von mises-fisher clustering models. In Proceedings of the 31st International Conference on Machine Learning, 2014. Hariprasath Govindarajan, Per Sidén, Jacob Roll, and Fredrik Lindsten. DINO as a von mises-fisher mixture model. In *The Eleventh International Conference on Learning Representations*, 2023. Md. Abul Hasnat, Julien Bohné, Jonathan Milgram, Stéphane Gentric, and Liming Chen. von mises-fisher mixture model-based deep learning: Application to face verification. *arXiv preprint arXiv:1706.04264*, 2017. Matthias Hein, Maksym Andriushchenko, and Julian Bitterwolf. Why relu networks yield high-confidence predictions far away from the training data and how to mitigate the problem. In 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019. Dan Hendrycks and Kevin Gimpel. A baseline for detecting misclassified and out-of-distribution examples in neural networks. *Proceedings of International Conference on Learning Representations*, 2017. Dan Hendrycks, Mantas Mazeika, and Thomas Dietterich. Deep anomaly detection with outlier exposure. Proceedings of the International Conference on Learning Representations, 2019. Rui Huang, Andrew Geng, and Yixuan Li. On the importance of gradients for detecting distributional shifts in the wild. In *Advances in Neural Information Processing Systems*, 2021. Wenjian Huang, Hao Wang, Jiahao Xia, Chengyan Wang, and Jianguo Zhang. Density-driven regularization for out-of-distribution detection. In *Advances in Neural Information Processing Systems*, 2022. Jyh-Jing Hwang, Stella X Yu, Jianbo Shi, Maxwell D Collins, Tien-Ju Yang, Xiao Zhang, and Liang-Chieh Chen. Segsort: Segmentation by discriminative sorting of segments. In *Proceedings of the IEEE International* Conference on Computer Vision, 2019. Xue Jiang, Feng Liu, Zhen Fang, Hong Chen, Tongliang Liu, Feng Zheng, and Bo Han. Detecting out-ofdistribution data through in-distribution class prior. In Proceedings of the 40th International Conference on Machine Learning, 2023. Jeff Johnson, Matthijs Douze, and Hervé Jégou. Billion-scale similarity search with GPUs. IEEE Transactions on Big Data, 7(3):535–547, 2019. Julian Katz-Samuels, Julia B Nakhleh, Robert Nowak, and Yixuan Li. Training OOD detectors in their natural habitats. In *Proceedings of the 39th International Conference on Machine Learning*, 2022. Prannay Khosla, Piotr Teterwak, Chen Wang, Aaron Sarna, Yonglong Tian, Phillip Isola, Aaron Maschinot, Ce Liu, and Dilip Krishnan. Supervised contrastive learning. In Advances in Neural Information Processing Systems, 2020. Polina Kirichenko, Pavel Izmailov, and Andrew G Wilson. Why normalizing flows fail to detect out-ofdistribution data. In *Advances in Neural Information Processing Systems*, 2020. Takumi Kobayashi. T-vmf similarity for regularizing intra-class feature distribution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021. A. Krizhevsky. Learning multiple layers of features from tiny images. Master's thesis, Computer Science Department, University of Toronto, 2009. Marc Lafon, Elias Ramzi, Clément Rambour, and Nicolas Thome. Hybrid energy based model in the feature space for out-of-distribution detection. In *International Conference on Machine Learning*, 2023. Kimin Lee, Honglak Lee, Kibok Lee, and Jinwoo Shin. Training confidence-calibrated classifiers for detecting out-of-distribution samples. In *International Conference on Learning Representations*, 2018a. Kimin Lee, Kibok Lee, Honglak Lee, and Jinwoo Shin. A simple unified framework for detecting out-ofdistribution samples and adversarial attacks. In *Advances in Neural Information Processing Systems*, 2018b. Junnan Li, Caiming Xiong, and Steven Hoi. Mopro: Webly supervised learning with momentum prototypes. In *International Conference on Learning Representations*, 2020. Yewen Li, Chaojie Wang, Xiaobo Xia, Tongliang Liu, and Bo An. Out-of-distribution detection with an adaptive likelihood ratio on informative hierarchical vae. Advances in Neural Information Processing Systems, 35:7383–7396, 2022. Shiyu Liang, Yixuan Li, and R. Srikant. Enhancing the reliability of out-of-distribution image detection in neural networks. In *International Conference on Learning Representations*, 2018. Weitang Liu, Xiaoyun Wang, John Owens, and Yixuan Li. Energy-based out-of-distribution detection. Advances in Neural Information Processing Systems, 2020. Yibing Liu, Chris Xing Tian, Haoliang Li, Lei Ma, and Shiqi Wang. Neuron activation coverage: Rethinking out-of-distribution detection and generalization. *arXiv preprint arXiv:2306.02879*, 2023a. Zhenzhen Liu, Jin Peng Zhou, Yufan Wang, and Kilian Q Weinberger. Unsupervised out-of-distribution detection with diffusion inpainting. In *International Conference on Machine Learning*, 2023b. Andrey Malinin and Mark Gales. Predictive uncertainty estimation via prior networks. In Advances in Neural Information Processing Systems, 2018. Alexander Meinke and Matthias Hein. Towards neural networks that provably know when they don't know. Proceedings of International Conference on Learning Representations, 2020. Pascal Mettes, Elise van der Pol, and Cees Snoek. Hyperspherical prototype networks. *Advances in Neural* Information Processing Systems, 32, 2019. Yifei Ming, Ziyang Cai, Jiuxiang Gu, Yiyou Sun, Wei Li, and Yixuan Li. Delving into out-of-distribution detection with vision-language representations. In *Advances in Neural Information Processing Systems*, 2022. Yifei Ming, Yiyou Sun, Ousmane Dia, and Yixuan Li. How to exploit hyperspherical embeddings for out-of-distribution detection? In *The Eleventh International Conference on Learning Representations*, 2023. Atsuyuki Miyai, Qing Yu, Go Irie, and Kiyoharu Aizawa. Locoop: Few-shot out-of-distribution detection via prompt learning. In *Thirty-Seventh Conference on Neural Information Processing Systems*, 2023. Peyman Morteza and Yixuan Li. Provable guarantees for understanding out-of-distribution detection. Proceedings of the AAAI Conference on Artificial Intelligence, 36, 2022. Eric Nalisnick, Akihiro Matsukawa, Yee Whye Teh, Dilan Gorur, and Balaji Lakshminarayanan. Do deep generative models know what they don't know? In *International Conference on Learning Representations*, 2019. Yuval Netzer, Tao Wang, Adam Coates, Alessandro Bissacco, Bo Wu, and Andrew Y. Ng. Reading digits in natural images with unsupervised feature learning. In NIPS Workshop on Deep Learning and Unsupervised Feature Learning 2011, 2011. Bo Peng, Yadan Luo, Yonggang Zhang, Yixuan Li, and Zhen Fang. Conjnorm: Tractable density estimation for out-of-distribution detection. In *The Twelfth International Conference on Learning Representations*, 2024. Jie Ren, Peter J. Liu, Emily Fertig, Jasper Snoek, Ryan Poplin, Mark Depristo, Joshua Dillon, and Balaji Lakshminarayanan. Likelihood ratios for out-of-distribution detection. In Advances in Neural Information Processing Systems, 2019. Jie Ren, Stanislav Fort, Jeremiah Liu, Abhijit Guha Roy, Shreyas Padhy, and Balaji Lakshminarayanan. A simple fix to mahalanobis distance for improving near-ood detection, 2021. Jie Ren, Jiaming Luo, Yao Zhao, Kundan Krishna, Mohammad Saleh, Balaji Lakshminarayanan, and Peter J Liu. Out-of-distribution detection and selective generation for conditional language models. In The Eleventh International Conference on Learning Representations, 2023. Chandramouli Shama Sastry and Sageev Oore. Detecting out-of-distribution examples with Gram matrices. In *Proceedings of the 37th International Conference on Machine Learning*, volume 119 of *Proceedings of* Machine Learning Research, 2020. Robin Schirrmeister, Yuxuan Zhou, Tonio Ball, and Dan Zhang. Understanding anomaly detection with deep invertible networks through hierarchies of distributions and features. In *Advances in Neural Information* Processing Systems, 2020. T. R. Scott, A. C. Gallagher, and M. C. Mozer. von mises–fisher loss: An exploration of embedding geometries for supervised learning. In *2021 IEEE/CVF International Conference on Computer Vision (ICCV)*, 2021. Vikash Sehwag, Mung Chiang, and Prateek Mittal. Ssd: A unified framework for self-supervised outlier detection. In *International Conference on Learning Representations*, 2021. Joan Serrà, David Álvarez, Vicenç Gómez, Olga Slizovskaia, José F. Núñez, and Jordi Luque. Input complexity and out-of-distribution detection with likelihood-based generative models. In *International Conference on* Learning Representations, 2020. Yihong Sun, Adam Kortylewski, and Alan Yuille. Amodal segmentation through out-of-task and outof-distribution generalization with a bayesian model. In *Proceedings of the IEEE/CVF Conference on* Computer Vision and Pattern Recognition (CVPR), 2022a. Yiyou Sun and Yixuan Li. Dice: Leveraging sparsification for out-of-distribution detection. In *European* Conference on Computer Vision, 2022. Yiyou Sun, Chuan Guo, and Yixuan Li. React: Out-of-distribution detection with rectified activations. In Advances in Neural Information Processing Systems, 2021. Yiyou Sun, Yifei Ming, Xiaojin Zhu, and Yixuan Li. Out-of-distribution detection with deep nearest neighbors. In *2022 International Conference on Machine Learning*, 2022b. Jihoon Tack, Sangwoo Mo, Jongheon Jeong, and Jinwoo Shin. Csi: Novelty detection via contrastive learning on distributionally shifted instances. In *Advances in Neural Information Processing Systems*, 2020. Leitian Tao, Xuefeng Du, Jerry Zhu, and Yixuan Li. Non-parametric outlier synthesis. In *The Eleventh* International Conference on Learning Representations, 2023. Joost Van Amersfoort, Lewis Smith, Yee Whye Teh, and Yarin Gal. Uncertainty estimation using a single deep deterministic neural network. In *Proceedings of the 37th International Conference on Machine Learning*, volume 119 of *Proceedings of Machine Learning Research*, 2020. Grant Van Horn, Oisin Mac Aodha, Yang Song, Yin Cui, Chen Sun, Alex Shepard, Hartwig Adam, Pietro Perona, and Serge Belongie. The inaturalist species classification and detection dataset. In 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 8769–8778, 2018. doi: 10.1109/CVPR.2018. 00914. Sagar Vaze, Kai Han, Andrea Vedaldi, and Andrew Zisserman. Open-set recognition: A good closedset classifier is all you need. In *International Conference on Learning Representations*, 2022. URL https://openreview.net/forum?id=5hLP5JY9S2d. Haoqi Wang, Zhizhong Li, Litong Feng, and Wayne Zhang. Vim: Out-of-distribution with virtual-logit matching. In *Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition*, 2022. Haoran Wang, Weitang Liu, Alex Bocchieri, and Yixuan Li. Can multi-label classification networks know what they don't know? In *Advances in Neural Information Processing Systems*, 2021. Qizhou Wang, Zhen Fang, Yonggang Zhang, Feng Liu, Yixuan Li, and Bo Han. Learning to augment distributions for out-of-distribution detection. In Thirty-seventh Conference on Neural Information Processing Systems, 2023. Tongzhou Wang and Phillip Isola. Understanding contrastive representation learning through alignment and uniformity on the hypersphere. In *International conference on machine learning*, pp. 9929–9939. PMLR, 2020. Ziyu Wang, Bin Dai, David Wipf, and Jun Zhu. Further analysis of outlier detection with deep generative models. In *Advances in Neural Information Processing Systems*, 2020. Hongxin Wei, Renchunzi Xie, Hao Cheng, Lei Feng, Bo An, and Yixuan Li. Mitigating neural network overconfidence with logit normalization. In *Proceedings of the 39th International Conference on Machine* Learning, volume 162 of *Proceedings of Machine Learning Research*, 2022. Zhisheng Xiao, Qing Yan, and Yali Amit. Likelihood regret: An out-of-distribution detection score for variational auto-encoder. In *Advances in Neural Information Processing Systems*, 2020. Kai Xu, Rongyu Chen, Gianni Franchi, and Angela Yao. Scaling for training time and post-hoc out-ofdistribution detection enhancement, 2023a. Mingyu Xu, Zheng Lian, Bin Liu, and Jianhua Tao. VRA: Variational rectified activation for out-of-distribution detection. In *Thirty-seventh Conference on Neural Information Processing Systems*, 2023b. Jingkang Yang, Haoqi Wang, Litong Feng, Xiaopeng Yan, Huabin Zheng, Wayne Zhang, and Ziwei Liu. Semantically coherent out-of-distribution detection. In Proceedings of the IEEE International Conference on Computer Vision, 2021a. Jingkang Yang, Kaiyang Zhou, Yixuan Li, and Ziwei Liu. Generalized out-of-distribution detection: A survey. arXiv preprint arXiv:2110.11334, 2021b. Shuyang Yu, Junyuan Hong, Haotao Wang, Zhangyang Wang, and Jiayu Zhou. Turning the curse of heterogeneity in federated learning into a blessing for out-of-distribution detection. In *The Eleventh* International Conference on Learning Representations, 2023. Jingyang Zhang, Jingkang Yang, Pengyun Wang, Haoqi Wang, Yueqian Lin, Haoran Zhang, Yiyou Sun, Xuefeng Du, Kaiyang Zhou, Wayne Zhang, Yixuan Li, Ziwei Liu, Yiran Chen, and Hai Li. Openood v1.5: Enhanced benchmark for out-of-distribution detection. *arXiv preprint arXiv:2306.09301*, 2023a. Jinsong Zhang, Qiang Fu, Xu Chen, Lun Du, Zelin Li, Gang Wang, xiaoguang Liu, Shi Han, and Dongmei Zhang. Out-of-distribution detection based on in-distribution data patterns memorization with modern hopfield energy. In *The Eleventh International Conference on Learning Representations*, 2023b. Bolei Zhou, Aditya Khosla, Agata Lapedriza, Antonio Torralba, and Aude Oliva. Places: An image database for deep scene understanding, 2016. Honglei Zhuang, Chi Wang, Fangbo Tao, Lance Kaplan, and Jiawei Han. Identifying semantically deviating outlier documents. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, 2017. ## A Derivation Theorem A.1 (**Misalignment between energy score and likelihood**). The energy score defined in Equation 4.3 *does not necessarily reflect the negative log probability:* $$E(\mathbf{x})$$ E(x; fθ) *̸∝ −* log p (x). To prove this, one can rewrite the posterior probability pθ(y = c|x) = pθ(x, y = c)/pθ(x). Note that one can always rescale the logits fθ(x) by a function log ϕ(x), without changing the posterior probability: pθ(y = c|x) = pθ(x, y = c)ϕ(x)/pθ(x)ϕ(x). In this case, the negative log of the denominator in softmax (i.e., the energy score) is − log[pθ(x)ϕ(x)], which is no longer truthfully reflecting the − log p(x). ## B Experimental Details Software and hardware. We conduct our experiments on NVIDIA RTX A6000 GPUs (48GB VRAM). We use Ubuntu 22.04.2 LTS as the operating system and install the NVIDIA CUDA Toolkit version 11.6 and cuDNN 8.9. All experiments are implemented in Python 3.8 using the PyTorch 1.8.1 framework. Training cross-entropy models. For methods using cross-entropy loss, such as MSP (Hendrycks & Gimpel, 2017), ODIN (Liang et al., 2018), Mahalanobis (Lee et al., 2018b), Energy (Liu et al., 2020), ViM (Wang et al., 2022), ReAct (Sun et al., 2021), DICE (Sun & Li, 2022), SHE (Zhang et al., 2023b), and ASH (Djurisic et al., 2023), the models are trained using stochastic gradient descent with momentum 0.9 and weight decay 10−4. The initial learning rate is 0.1 and decays by a factor of 10 at epochs 100, 150, and 180. We train the models for 200 epochs on CIFAR. For the ImageNet benchmark, we adopt Torchvision's pre-trained ResNet-50 model with ImageNet-1k weights. Training contrastive models. For methods using SupCon loss (Khosla et al., 2020), such as KNN+ (Khosla et al., 2020) and SSD+ (Sehwag et al., 2021), we adopt the same training scheme as in (Ming et al., 2023). For CIFAR, we use stochastic gradient descent with a momentum of 0.9 and a weight decay of 10−4. The initial learning rate to 0.5 and follows a cosine annealing schedule. We train the models for 500 epochs. The training-time temperature τ is set to be 0.1. For the ImageNet model, we use the checkpoint provided in Sun et al. (2022b). For methods using vMF loss, we train on CIFAR using stochastic gradient descent with momentum 0.9 and weight decay 10−4. The initial learning rate is 0.5 and follows a cosine annealing schedule. We use a batch size of 512 and train the model for 500 epochs. The training-time temperature τ is set to be 0.1. We adopt the exponential-moving-average (EMA) for the prototype update Li et al. (2020), with a momentum of 0.5 for CIFAR-100. For ImageNet-1k, we fine-tune a pre-trained ResNet-50 model in Khosla et al. (2020) for 100 epochs, with an initial learning rate of 0.01 and cosine annealing schedule. ## Evaluation Configurations. We outline the configurations for methods that require appropriate hyperparameter selection, which has already been extensively studied in the literature. - MSP uses the maximum softmax probability obtained from the logits. The method does not require any specific configuration. - ODIN uses temperature scaling to calibrate the softmax score. We set the temperature T to 1,000. - Mahalanobis utilizes class conditional Gaussian distributions based on the low- and upper-level features obtained from a model. These distributions are then used to calculate the Mahalanobis distance. - Energy derives the energy score, which includes the temperature parameter. We set the temperature to the default value of T = 1. - ViM generates an additional logit from the residual of the feature against the principal space. We set the dimension of principal space to D = 256 for ResNet-18 and ResNet-34 and D = 1024 for ResNet-50. - ReAct improves the energy score by rectifying activations at an upper limit, which is set based on the p-th percentile of the activations estimated on the in-distribution (ID) data. We set p to the default value of p = 90. - DICE utilizes logit sparsification to enhance the vanilla energy score, which is set based on the p-th percentile of the unit contributions estimated on the ID data. We set the sparsity parameter p = 0.7. - SHE stores the mean direction of the penultimate layer features from correctly classified training samples. During inference, the Hopfield energy score is calculated as the dot product between the sample embedding and the class mean of the predicted class. - KNN+ and CIDER utilize the KNN score, which requires selecting a number of nearest neighbors k. Following the settings in Ming et al. (2023), we set k to 300 and 1,000 for CIFAR-100 and the ImageNet benchmark, respectively. - ASH proposes three heuristic-based approaches for activation shaping: Pruning, Binarizing, and Scaling. In line with the OpenOOD framework, we adopt the Binarizing method (ASH-B), specifically with a parameter setting of p = 90. ## C Validation Method For Selecting Test-Time Temperature To select the test-time temperature for our proposed INK score, we follow the validation method outlined in Hendrycks et al. (2019). We generate a validation distribution by corrupting in-distribution data with speckle noise, creating speckle-noised anomalies that simulate out-of-distribution data. After that, we compute the performance of INK at different test-time temperatures on the validation set and select the one that achieves the highest AUROC. ## D Id Classification Accuracy Table 5 presents the in-distribution classification accuracy for each training dataset. We evaluate the classification accuracy by performing linear probing on normalized features, following the approach in Khosla et al. (2020). Using vMF loss shows competitive ID classification accuracy compared to standard cross-entropy loss, indicating it does not compromise the model's capability to distinguish samples between in-distribution classes. Table 5: ID classification accuracy on CIFAR-100 and ImageNet (%). | Method | ID classification accuracy ↑ CIFAR-100 ImageNet-1k | | |---------------|------------------------------------------------------|-------| | Cross-entropy | 75.03 | 76.15 | | vMF | 74.28 | 76.55 | ## E Results On Cifar Benchmark In this section, we present additional results and analysis on the CIFAR-100 benchmarks. Specifically, we utilize CIFAR-10 and TIN as the near-OOD datasets, while MNIST (Deng, 2012), SVHN (Netzer et al., 2011), Textures (Cimpoi et al., 2014), and Places (Zhou et al., 2016) serve as our far-OOD datasets. As shown in Table 6, INK displays competitive performance compared to existing state-of-the-art methods. INK achieves average FPR@95 values of 44.51% and 61.83% for the far-OOD and near-OOD groups. Compared to KNN+ and CIDER, our method achieves competitive performance while reducing the inference time by more than 300 times (Section 4.3). Table 6: **OOD detection performance for CIFAR-100 (ID) with ResNet-34.** INK achieves competitive performance compared to state-of-the-art methods. | Method | Far-OOD Datasets | Near-OOD Dataset | | | | | | | | | | | | | | | |--------------------------------------------------------------------------------------------------------------------------------------------------|--------------------|--------------------|-----------|---------|----------|-------|---------|-------|-------|-------|-------|-------|-------|-------|-------|-------| | MNIST | SVHN | Textures | Places365 | Average | CIFAR-10 | TIN | Average | | | | | | | | | | | FPR ↓ AUROC ↑ FPR ↓ AUROC ↑ FPR ↓ AUROC ↑ FPR ↓ AUROC ↑ FPR ↓ AUROC ↑ FPR ↓ AUROC ↑ FPR ↓ AUROC ↑ FPR ↓ AUROC ↑ Methods using cross-entropy loss | | | | | | | | | | | | | | | | | | MSP | 65.59 | 72.26 | 60.33 | 74.38 | 71.97 | 72.12 | 58.68 | 77.07 | 64.14 | 73.96 | 62.82 | 75.89 | 54.87 | 79.23 | 58.84 | 77.56 | | ODIN | 49.66 | 82.27 | 61.34 | 75.56 | 70.58 | 74.30 | 59.71 | 78.67 | 60.32 | 77.70 | 62.74 | 77.04 | 56.67 | 80.63 | 59.70 | 78.84 | | Mahalanobis | 81.44 | 57.61 | 54.07 | 81.79 | 52.73 | 87.46 | 83.37 | 60.84 | 67.90 | 71.93 | 88.04 | 54.39 | 79.83 | 63.20 | 83.94 | 58.80 | | Energy | 57.12 | 77.17 | 49.97 | 80.90 | 70.74 | 72.40 | 58.66 | 78.34 | 59.12 | 77.20 | 64.14 | 77.51 | 54.48 | 81.26 | 59.31 | 79.39 | | ViM | 59.28 | 71.20 | 38.34 | 88.67 | 46.42 | 88.20 | 63.41 | 75.73 | 51.86 | 80.95 | 72.01 | 73.31 | 55.71 | 80.11 | 63.86 | 76.71 | | ReAct | 60.37 | 76.35 | 49.61 | 81.71 | 59.58 | 76.28 | 58.67 | 78.58 | 57.06 | 78.23 | 63.71 | 77.24 | 54.71 | 81.26 | 59.21 | 79.25 | | DICE | 51.62 | 81.48 | 36.59 | 88.63 | 69.66 | 73.52 | 60.27 | 77.46 | 54.54 | 80.27 | 60.53 | 77.40 | 58.01 | 79.36 | 59.27 | 78.38 | | SHE | 52.42 | 79.46 | 58.78 | 77.27 | 78.68 | 66.80 | 68.64 | 70.36 | 64.63 | 73.47 | 65.31 | 75.10 | 60.60 | 76.17 | 62.95 | 75.63 | | ASH | 55.88 | 78.93 | 41.59 | 87.20 | 55.22 | 80.91 | 63.80 | 75.96 | 54.12 | 80.75 | 64.62 | 75.68 | 59.27 | 78.94 | 61.95 | 77.31 | | Methods using hyperspherical representation | | | | | | | | | | | | | | | | | | KNN+ | 62.34 | 75.27 | 27.31 | 92.29 | 45.41 | 85.70 | 52.60 | 81.84 | 46.91 | 83.78 | 69.22 | 76.05 | 49.38 | 83.36 | 59.30 | 79.70 | | SSD+ | 58.57 | 75.15 | 44.86 | 80.69 | 68.37 | 76.45 | 57.06 | 77.42 | 57.22 | 77.43 | 72.27 | 72.89 | 56.41 | 79.41 | 64.34 | 76.15 | | CIDER | 63.79 | 70.24 | 23.82 | 95.95 | 46.08 | 86.73 | 59.64 | 78.15 | 48.33 | 82.77 | 74.59 | 73.33 | 51.89 | 83.08 | 63.24 | 78.20 | | Ours | 58.60 | 74.30 | 21.41 | 96.39 | 39.80 | 90.75 | 58.24 | 79.77 | 44.51 | 85.30 | 72.77 | 74.35 | 50.90 | 84.19 | 61.83 | 79.27 | | Far-OOD Datasets | Near-OOD Dataset | | | | | | | | | | |---------------------------------------------|--------------------|----------|-------------|---------|---------|-------|---------|-------|---------|-------| | Method | iNaturalist | Textures | OpenImage-O | Average | NINCO | | | | | | | FPR ↓ | AUROC ↑ | FPR ↓ | AUROC ↑ | FPR ↓ | AUROC ↑ | FPR ↓ | AUROC ↑ | FPR ↓ | AUROC ↑ | | | Methods using cross-entropy loss | | | | | | | | | | | | MSP | 42.42 | 88.19 | 56.44 | 85.06 | 56.11 | 84.87 | 51.99 | 86.04 | 77.35 | 78.11 | | ODIN | 81.14 | 79.55 | 85.62 | 77.16 | 91.09 | 71.46 | 85.95 | 76.72 | 92.62 | 65.14 | | Mahalanobis | 20.66 | 96.01 | 38.90 | 89.41 | 30.35 | 92.38 | 29.64 | 92.60 | 48.76 | 86.52 | | Energy | 83.58 | 79.30 | 83.65 | 81.17 | 88.79 | 76.48 | 85.34 | 78.65 | 94.16 | 66.02 | | ViM | 17.59 | 95.72 | 40.41 | 90.61 | 29.59 | 92.18 | 29.20 | 92.50 | 57.45 | 84.64 | | ReAct | 48.22 | 86.11 | 55.87 | 86.66 | 57.68 | 84.29 | 53.26 | 85.69 | 78.50 | 75.43 | | DICE | 47.92 | 82.51 | 54.79 | 82.21 | 52.57 | 82.23 | 51.76 | 82.32 | 81.09 | 71.67 | | SHE | 22.17 | 93.57 | 25.65 | 92.65 | 33.59 | 91.04 | 27.14 | 92.75 | 56.01 | 84.18 | | ASH | 97.02 | 50.63 | 98.49 | 48.53 | 94.80 | 55.52 | 96.10 | 51.56 | 95.40 | 52.52 | | Methods using hyperspherical representation | | | | | | | | | | | | CIDER | 26.40 | 93.27 | 48.23 | 85.94 | 34.84 | 89.87 | 36.49 | 89.69 | 47.02 | 85.22 | | INK (ours) | 28.45 | 92.90 | 50.57 | 85.16 | 36.80 | 89.28 | 38.61 | 89.11 | 48.86 | 84.55 | Table 7: **OOD detection performance for ImageNet (ID) with ViT-B-16.** INK achieves competitive performance compared to state-of-the-art methods. ## F Results On Vision Transformer Architecture This section details the performance evaluation of different methods for out-of-distribution (OOD) detection using the Vision Transformer (ViT) architecture, focusing on the ImageNet-1k dataset. Our approach involves fine-tuning a pre-trained ViT-B-16 model over 100 epochs. The initial learning rate is set at 0.01, and we employ a cosine annealing schedule for optimization. The performance of various methods is compared in Table 7. Notably, INK demonstrates competitive performance in both far-OOD and near-OOD groups, with an FPR@95 rate of 38.64% and 48.80%. This result highlights the robustness and versatility of our proposed method when applied to diverse architectural models. ## G Visualization Analysis For Large-Scale Dataset Figure 5 presents the UMAP visualization of the learned embeddings derived from a subset of ImageNet-1k classes and larger-scale out-of-distribution (OOD) datasets. These visualizations indeed reveal compact representations, where each sample appears to be effectively drawn in towards its corresponding class prototype. A notable separation between in-distribution (ID) and OOD classes is observed. ## H Limitation A key strength of our method is its grounded interpretation from a likelihood perspective. This approach, however, necessitates training the model under the vMF distribution. While our method achieves accuracy comparable to state-of-the-art models in certain scenarios, it's important to note that it is less common than models trained using traditional cross-entropy loss. This may limit its immediate applicability in environments ![20_image_0.png](20_image_0.png) Figure 5: UMAP visualization of a subset of ImageNet classes and OOD datasets. The class prototypes are designated by a star symbol *, while the OOD embeddings are distinguished by pink color. where standard cross-entropy loss models are the norm. However, we believe that the trustworthiness of the models is worth the effort of training them in a certain way.