Details of Research Outputs

Status已发表Published
TitleTransformer-based contrastive learning framework for image anomaly detection
Creator
Date Issued2023-10-01
Source PublicationInternational Journal of Machine Learning and Cybernetics
ISSN1868-8071
Volume14Issue:10Pages:3413-3426
Abstract

Anomaly detection refers to the problem of uncovering patterns in a given data set that do not conform to the expected behavior. Recently, owing to the continuous development of deep representation learning, a large number of anomaly detection approaches based on deep learning models have been developed and achieved promising performance. In this work, an image anomaly detection approach based on contrastive learning framework is proposed. Rather than adopting ResNet or other CNN-based deep neural networks as in most of the previous deep learning-based image anomaly detection approaches to learn representations from training samples, a contrastive learning framework is developed for anomaly detection in which Transformer is adopted for extracting better representations. Then, we develop a triple contrastive loss function and embed it into the proposed contrastive learning framework to alleviate the problem of catastrophic collapse that is often encountered in many anomaly detection approaches. Furthermore, a nonlinear Projector is integrated with our model to improve the performance of anomaly detection. The effectiveness of our image anomaly detection approach is validated through experiments on multiple benchmark data sets. According to the experimental results, our approach can obtain better or comparative performance in comparison with state-of-the-art anomaly detection approaches.

KeywordAnomaly detection Contrastive learning Transformer Triple contrastive loss
DOI10.1007/s13042-023-01840-7
URLView source
Indexed BySCIE
Language英语English
WOS Research AreaComputer Science
WOS SubjectComputer Science, Artificial Intelligence
WOS IDWOS:000980786000001
Scopus ID2-s2.0-85158147574
Citation statistics
Cited Times:4[WOS]   [WOS Record]     [Related Records in WOS]
Document TypeJournal article
Identifierhttp://repository.uic.edu.cn/handle/39GCC9TT/10900
CollectionBeijing Normal-Hong Kong Baptist University
Corresponding AuthorFan, Wentao
Affiliation
1.Department of Computer Science,Beijing Normal University-Hong Kong Baptist University United International College (BNU-HKBU UIC),Zhuhai,Guangdong,China
2.Guangdong Provincial Key Laboratory of Interdisciplinary Research and Application for Data Science,BNU-HKBU United International College,Zhuhai,China
3.Department of Computer Science and Technology,Huaqiao University,Xiamen,China
First Author AffilicationBeijing Normal-Hong Kong Baptist University
Corresponding Author AffilicationBeijing Normal-Hong Kong Baptist University
Recommended Citation
GB/T 7714
Fan, Wentao,Shangguan, Weimin,Chen, Yewang. Transformer-based contrastive learning framework for image anomaly detection[J]. International Journal of Machine Learning and Cybernetics, 2023, 14(10): 3413-3426.
APA Fan, Wentao, Shangguan, Weimin, & Chen, Yewang. (2023). Transformer-based contrastive learning framework for image anomaly detection. International Journal of Machine Learning and Cybernetics, 14(10), 3413-3426.
MLA Fan, Wentao,et al."Transformer-based contrastive learning framework for image anomaly detection". International Journal of Machine Learning and Cybernetics 14.10(2023): 3413-3426.
Files in This Item:
There are no files associated with this item.
Related Services
Usage statistics
Google Scholar
Similar articles in Google Scholar
[Fan, Wentao]'s Articles
[Shangguan, Weimin]'s Articles
[Chen, Yewang]'s Articles
Baidu academic
Similar articles in Baidu academic
[Fan, Wentao]'s Articles
[Shangguan, Weimin]'s Articles
[Chen, Yewang]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Fan, Wentao]'s Articles
[Shangguan, Weimin]'s Articles
[Chen, Yewang]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.