Redirigiendo al acceso original de articulo en 23 segundos...
Inicio  /  Future Internet  /  Vol: 16 Par: 3 (2024)  /  Artículo
ARTÍCULO
TITULO

Edge-Enhanced TempoFuseNet: A Two-Stream Framework for Intelligent Multiclass Video Anomaly Recognition in 5G and IoT Environments

Gulshan Saleem    
Usama Ijaz Bajwa    
Rana Hammad Raza and Fan Zhang    

Resumen

Surveillance video analytics encounters unprecedented challenges in 5G and IoT environments, including complex intra-class variations, short-term and long-term temporal dynamics, and variable video quality. This study introduces Edge-Enhanced TempoFuseNet, a cutting-edge framework that strategically reduces spatial resolution to allow the processing of low-resolution images. A dual upscaling methodology based on bicubic interpolation and an encoder?bank?decoder configuration is used for anomaly classification. The two-stream architecture combines the power of a pre-trained Convolutional Neural Network (CNN) for spatial feature extraction from RGB imagery in the spatial stream, while the temporal stream focuses on learning short-term temporal characteristics, reducing the computational burden of optical flow. To analyze long-term temporal patterns, the extracted features from both streams are combined and routed through a Gated Recurrent Unit (GRU) layer. The proposed framework (TempoFuseNet) outperforms the encoder?bank?decoder model in terms of performance metrics, achieving a multiclass macro average accuracy of 92.28%, an F1-score of 69.29%, and a false positive rate of 4.41%. This study presents a significant advancement in the field of video anomaly recognition and provides a comprehensive solution to the complex challenges posed by real-world surveillance scenarios in the context of 5G and IoT.

 Artículos similares

       
 
Yanjie Sun, Mingguang Wu, Xiaoyan Liu and Liangchen Zhou    
High-precision dynamic traffic noise maps can describe the spatial and temporal distributions of noise and are necessary for actual noise prevention. Existing monitoring point-based methods suffer from limited spatial adaptability, and prediction model-b... ver más

 
Shijing Han, Xiaorui Dong, Xiangyang Hao and Shufeng Miao    
Surveillance systems focus on the image itself, mainly from the perspective of computer vision, which lacks integration with geographic information. It is difficult to obtain the location, size, and other spatial information of moving objects from survei... ver más

 
Haochen Zou, Keyan Cao and Chong Jiang    
Urban road traffic spatio-temporal characters reflect how citizens move and how goods are transported, which is crucial for trip planning, traffic management, and urban design. Video surveillance camera plays an important role in intelligent transport sy... ver más

 
Zhou Lei and Yiyong Huang    
Video captioning is a popular task which automatically generates a natural-language sentence to describe video content. Previous video captioning works mainly use the encoder?decoder framework and exploit special techniques such as attention mechanisms t... ver más
Revista: Future Internet

 
Kunlin Liu, Ping Wang, Wenbo Zhou, Zhenyu Zhang, Yanhao Ge, Honggu Liu, Weiming Zhang and Nenghai Yu    
Deepfake aims to swap a face of an image with someone else?s likeness in a reasonable manner. Existing methods usually perform deepfake frame by frame, thus ignoring video consistency and producing incoherent results. To address such a problem, we propos... ver más
Revista: Future Internet