Unsupervised Out-of-Distribution Detection Using Few in-Distribution Samples

ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)(2023)

引用 0|浏览0
暂无评分
摘要
This paper tackles the out-of-distribution (OOD) detection problem for natural language classifiers. While the previous OOD detection methods require large-scale in-distribution (ID) training data, we attack this problem from the few-shot perspective in an unsupervised manner where the training relies on only a few samples from ID data. First, we develop various baselines for Few-shot OOD (FSOOD) detection in text classification based on the three well-known few-shot learning approaches (well-explored in the vision domain), i.e., meta-learning, metric learning, and data augmentation (DA). Then, we introduce the concept of demonstration-based data augmentation with meta and metric-learning approaches to reap the combined benefit of both approaches. A pre-trained transformer is fine-tuned on a few available ID samples in all developed methods. In tandem with this fine-tuning, an OOD detector is fitted over the ID training samples to reject the data from the unknown classes using two kinds of distance metrics, namely Mahalanobis distance and Cosine similarity. At last, we present an extensive evaluation of three ID datasets and three OOD datasets. We also perform an ablation study to analyze the impact of various components of our method.
更多
查看译文
关键词
available ID samples,demonstration-based data augmentation,distance metrics,Few-shot OOD detection,few-shot perspective,fine-tuning,ID data,ID datasets,ID training samples,in-distribution samples,in-distribution training data,meta-learning,metric learning,metric-learning approaches,natural language classifiers,OOD datasets,OOD detector,out-of-distribution detection problem,pre-trained transformer,previous OOD detection methods,text classification,unsupervised manner
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要