跳至主導覽 跳至搜尋 跳過主要內容

TransHFC: Joints Hypergraph Filtering Convolution and Transformer Framework for Temporal Forgery Localization

研究成果: Article同行評審

摘要

The authenticity of audio-visual content is being challenged by advanced multimedia editing technologies inspired by Artificial Intelligence-Generated Content (AIGC). Temporal forgery localization aims to detect suspicious contents by locating forged segments. So far, most of the existing methods are based on Convolutional Neural Networks (CNNs) or Transformers, yet neither of them has fully considered the complex relationships within forged audio-visual content. To address this issue, in this paper, we propose a novel method, named TransHFC, which innovatively introduces hypergraphs to model group relationships among segments while considering point-to-point relationships through Transformers. Through its dual hypergraph filtering convolution branch, TransHFC captures both temporal and spatial level group relationships, enhancing the representation of forged segment features. Furthermore, we propose a new hypergraph filtering convolution Auto-Encoder that uses a multi-frequency filter bank for adaptive signal capture. This design compensates for the limitation of a single hypergraph filter. Our extensive experiments on Lav-DF, TVIL, Psynd, and HAD datasets demonstrate that TransHFC achieves state-of-the-art performance.

原文English
頁(從 - 到)9261-9275
頁數15
期刊IEEE Transactions on Circuits and Systems for Video Technology
35
發行號9
DOIs
出版狀態Published - 2025

指紋

深入研究「TransHFC: Joints Hypergraph Filtering Convolution and Transformer Framework for Temporal Forgery Localization」主題。共同形成了獨特的指紋。

引用此