TY - JOUR
T1 - AONet
T2 - Attention network with optional activation for unsupervised video anomaly detection
AU - Rakhmonov, Akhrorjon Akhmadjon Ugli
AU - Subramanian, Barathi
AU - Amirian Varnousefaderani, Bahar
AU - Kim, Jeonghong
N1 - Publisher Copyright:
1225-6463/$ © 2024 ETRI.
PY - 2024/10
Y1 - 2024/10
N2 - Anomaly detection in video surveillance is crucial but challenging due to the rarity of irregular events and ambiguity of defining anomalies. We propose a method called AONet that utilizes a spatiotemporal module to extract spatiotemporal features efficiently, as well as a residual autoencoder equipped with an attention network for effective future frame prediction in video anomaly detection. AONet utilizes a novel activation function called OptAF that combines the strengths of the ReLU, leaky ReLU, and sigmoid functions. Furthermore, the proposed method employs a combination of robust loss functions to address various aspects of prediction errors and enhance training effectiveness. The performance of the proposed method is evaluated on three widely used benchmark datasets. The results indicate that the proposed method outperforms existing state-of-the-art methods and demonstrates comparable performance, achieving area under the curve values of 97.0%, 86.9%, and 73.8% on the UCSD Ped2, CUHK Avenue, and ShanghaiTech Campus datasets, respectively. Additionally, the high speed of the proposed method enables its application to real-time tasks.
AB - Anomaly detection in video surveillance is crucial but challenging due to the rarity of irregular events and ambiguity of defining anomalies. We propose a method called AONet that utilizes a spatiotemporal module to extract spatiotemporal features efficiently, as well as a residual autoencoder equipped with an attention network for effective future frame prediction in video anomaly detection. AONet utilizes a novel activation function called OptAF that combines the strengths of the ReLU, leaky ReLU, and sigmoid functions. Furthermore, the proposed method employs a combination of robust loss functions to address various aspects of prediction errors and enhance training effectiveness. The performance of the proposed method is evaluated on three widely used benchmark datasets. The results indicate that the proposed method outperforms existing state-of-the-art methods and demonstrates comparable performance, achieving area under the curve values of 97.0%, 86.9%, and 73.8% on the UCSD Ped2, CUHK Avenue, and ShanghaiTech Campus datasets, respectively. Additionally, the high speed of the proposed method enables its application to real-time tasks.
KW - activation function
KW - convolutional neural network
KW - loss function
KW - unsupervised learning
KW - video anomaly detection
UR - http://www.scopus.com/inward/record.url?scp=85208021093&partnerID=8YFLogxK
U2 - 10.4218/etrij.2024-0115
DO - 10.4218/etrij.2024-0115
M3 - Article
AN - SCOPUS:85208021093
SN - 1225-6463
VL - 46
SP - 890
EP - 903
JO - ETRI Journal
JF - ETRI Journal
IS - 5
ER -