OODformer: Out-Of-Distribution Detection Transformer

Rajat Koner, Poulami Sinhamahapatra, Karsten Roscher, Stephan Günnemann, Volker Tresp

Publikation: KonferenzbeitragPapierBegutachtung

5 Zitate (Scopus)

Abstract

A serious problem in image classification is that a trained model might perform well for input data that originates from the same distribution as the data available for model training, but performs much worse for out-of-distribution (OOD) samples. In real-world safety-critical applications, in particular, it is important to be aware if a new data point is OOD. To date, OOD detection is typically addressed using either confidence scores, autoencoder based reconstruction, or contrastive learning. However, the global image context has not yet been explored to discriminate the non-local objectness between in-distribution and OOD samples. This paper proposes a first-of-its-kind OOD detection architecture named OODformer that leverages the contextualization capabilities of the transformer. Incorporating the transformer as the principal feature extractor allows us to exploit the object concepts and their discriminatory attributes along with their co-occurrence via visual attention. Based on contextualised embedding, we demonstrate OOD detection using both class-conditioned latent space similarity and a network confidence score. Our approach shows improved generalizability across various datasets. We have achieved a new state-of-the-art result on CIFAR-10/-100 and ImageNet30. Code is available at: https://github.com/rajatkoner08/oodformer.

OriginalspracheEnglisch
PublikationsstatusVeröffentlicht - 2021
Veranstaltung32nd British Machine Vision Conference, BMVC 2021 - Virtual, Online
Dauer: 22 Nov. 202125 Nov. 2021

Konferenz

Konferenz32nd British Machine Vision Conference, BMVC 2021
OrtVirtual, Online
Zeitraum22/11/2125/11/21

Fingerprint

Untersuchen Sie die Forschungsthemen von „OODformer: Out-Of-Distribution Detection Transformer“. Zusammen bilden sie einen einzigartigen Fingerprint.

Dieses zitieren