TY - GEN
T1 - Action Knowledge Graph for Violence Detection Using Audiovisual Features
AU - Khan, Mustaqeem
AU - Saad, Muhammad
AU - Khan, Abbas
AU - Gueaieb, Wail
AU - Saddik, Abdulmotaleb El
AU - Masi, Giulia De
AU - Karray, Fakhri
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Detecting violent content in video frames is a crucial aspect of violence detection. Combining visual and audio cues is often the most effective way to identify violent behavior, as they complement each other. However, studies that examine the fusion of these cues in violence detection are computationally expensive and limited. To address this problem, we investigated various methods for integrating visual and audio information and proposed a Fused Vision-based Action Knowledge Graph (FV-AKG) for violence detection using audiovisual information. The authors have designed a network with three parallel branches named integrated, specialized, and scoring that capture and integrate the distinct relationships between audio and video samples. Our proposed FV-AKG captures the long-range dependencies based on similarity priors in the integrated branch, while proximity priors are used for local positional relationships in the specialized branch. In addition, the scoring branch indicates how close the predictions are to reality. We used two key operations during model training: Aggregation and update, each with its learnable weights. In the aggregation operation, long-range dependencies are compiled from global vertices, whereas in the update function, nonlinear transforms are used to compute new representations. We thoroughly investigated the possibilities of temporal context modeling using graphs and found that FV-AKG is the best option for real-Time violence detection. Our experiments showed that FV-AKG outperforms the current top State-of-The-Art (SoTA) methods on the XD-Violence datasets.
AB - Detecting violent content in video frames is a crucial aspect of violence detection. Combining visual and audio cues is often the most effective way to identify violent behavior, as they complement each other. However, studies that examine the fusion of these cues in violence detection are computationally expensive and limited. To address this problem, we investigated various methods for integrating visual and audio information and proposed a Fused Vision-based Action Knowledge Graph (FV-AKG) for violence detection using audiovisual information. The authors have designed a network with three parallel branches named integrated, specialized, and scoring that capture and integrate the distinct relationships between audio and video samples. Our proposed FV-AKG captures the long-range dependencies based on similarity priors in the integrated branch, while proximity priors are used for local positional relationships in the specialized branch. In addition, the scoring branch indicates how close the predictions are to reality. We used two key operations during model training: Aggregation and update, each with its learnable weights. In the aggregation operation, long-range dependencies are compiled from global vertices, whereas in the update function, nonlinear transforms are used to compute new representations. We thoroughly investigated the possibilities of temporal context modeling using graphs and found that FV-AKG is the best option for real-Time violence detection. Our experiments showed that FV-AKG outperforms the current top State-of-The-Art (SoTA) methods on the XD-Violence datasets.
UR - http://www.scopus.com/inward/record.url?scp=85186992509&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85186992509&partnerID=8YFLogxK
U2 - 10.1109/ICCE59016.2024.10444158
DO - 10.1109/ICCE59016.2024.10444158
M3 - Conference contribution
AN - SCOPUS:85186992509
T3 - Digest of Technical Papers - IEEE International Conference on Consumer Electronics
BT - 2024 IEEE International Conference on Consumer Electronics, ICCE 2024
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2024 IEEE International Conference on Consumer Electronics, ICCE 2024
Y2 - 6 January 2024 through 8 January 2024
ER -