吉林大学学报(工学版) ›› 2025, Vol. 55 ›› Issue (6): 2114-2121.doi: 10.13229/j.cnki.jdxbgxb.20231027

• 计算机科学与技术 • 上一篇    下一篇

多尺度感知与边界引导的图像篡改检测方法

陈海鹏(),张世博,吕颖达()   

  1. 吉林大学 计算机科学与技术学院,长春 130012
  • 收稿日期:2023-09-25 出版日期:2025-06-01 发布日期:2025-07-23
  • 通讯作者: 吕颖达 E-mail:chenhp@jlu.edu.cn;ydlv@jlu.edu.cn
  • 作者简介:陈海鹏(1978-),男,教授,博士.研究方向:图像处理与模式识别.E-mail:chenhp@jlu.edu.cn
  • 基金资助:
    国家自然科学基金面上项目(62276112);国家自然科学基金区域联合基金子项目(U19A2057);吉林省科技发展计划重点研发项目(20230201088GX);安徽高校协同创新项目(GXXT-2022-044)

Multi⁃scale context⁃aware and boundary⁃guided image manipulation detection method

Hai-peng CHEN(),Shi-bo ZHANG,Ying-da LYU()   

  1. College of Computer Science and Technology,Jilin University,Changchun 130012,China
  • Received:2023-09-25 Online:2025-06-01 Published:2025-07-23
  • Contact: Ying-da LYU E-mail:chenhp@jlu.edu.cn;ydlv@jlu.edu.cn

摘要:

针对传统图像篡改检测方法存在边界模糊、提取特征尺度单一、忽略背景信息等问题,本文提出多尺度感知与边界引导的图像篡改检测方法。首先,使用改进的金字塔视觉变压器提取篡改图像的空间细节和基础特征。其次,通过边缘感知模块探索与伪造区域边缘相关的信息,生成边缘预测图。再次,利用边缘引导模块突出所提取特征中的关键通道,减少冗余通道的干扰。然后,通过多尺度上下文感知模块,从多个感受野学习伪造区域丰富的上下文信息。最后,利用特征融合模块交替关注篡改图像的前景和背景,精确分割伪造区域。将本文方法在5个常用的公开图像篡改检测数据集上进行定量和定性对比,实验结果表明,本文方法可以有效检测伪造区域,并且优于其他方法。

关键词: 计算机应用, 图像篡改检测, 多尺度上下文感知, 边界引导

Abstract:

Aiming at the problems of traditional image manipulation detection methods, such as fuzzy boundaries, single scale of extracted features, and ignoring background information, this paper proposes an image manipulation detection method based on multi-scale context-aware and boundary-guided. First, spatial details and base features of manipulated images are extracted using an improved pyramid vision transformer. Second, information related to the edge of the falsified region is explored by an edge context-aware module to generate an edge prediction map. Again, the edge guidance module is utilized to highlight the key channels in the extracted features and reduce the interference of redundant channels. Then, the rich contextual information of the manipulated region is learned from multiple sensory fields through the multi-scale context-aware module. Finally, the feature fusion module is utilized to accurately segment the manipulated region by focusing alternately on the foreground and background of the manipulated images. Comparing this paper's method quantitatively and qualitatively on five commonly used public image manipulation detection datasets, the experimental results show that this paper's method can effectively detect manipulated regions and outperforms other methods.

Key words: computer application, image manipulation detection, multi-scale context-aware, boundary guidance

中图分类号: 

  • TP391

图1

本文提出的网络整体结构图"

图2

边缘感知模块结构图"

图3

边缘引导模块结构图"

图4

多尺度上下文感知模块结构图"

图5

特征融合模块结构图"

表1

数据集细节描述"

数据集规模类别后处理操作
训练集测试集拼接复制-粘贴移除
CASIA135 123921
NIST14404160
Columbia15180
COVER167525
IMD2020171 610400

表2

本文方法与其他方法在5个数据集的定量比较"

方法CASIANISTColumbiaCOVERIMD2020Mean
AUCF1AUCF1AUCF1AUCF1AUCF1AUCF1
ManTra180.7960.2670.9590.6380.7360.2430.7770.2830.7730.2490.8080.336
SPAN190.7090.2130.7790.2520.7410.4630.7910.3250.7550.313
MVSS200.8470.3180.9810.7680.8080.4170.8080.2840.8020.3960.8490.437
GSRNet80.8360.3400.9670.6400.9000.4330.7880.2180.8730.408
DenseFCN210.8090.2030.9790.8120.7610.2570.7540.1850.7150.2720.8040.346
LocateNet220.7540.2730.9860.7380.7180.4110.8130.2820.8180.426
EMTNet100.8560.4590.9870.8250.8320.5610.8120.3530.8720.550
Ours0.8480.6470.9650.8980.7700.6090.7760.3720.8010.5070.8320.607

图6

本文方法与其他方法的定性比较"

表3

不同组件在CASIA数据集的定量结果"

方法组件AUCF1
EAMEFMMCAMFFM
a.Baseline0.8000.595
b.Baseline+EAM0.8160.617
c.Baseline+EAM+EFM0.8350.634
d.Baseline+EAM+EFM+MCAM0.8410.639
e.Baseline+EAM+EFM+MCAM+FFM0.8480.647

表4

F1分数在NIST数据集上的鲁棒性比较"

后处理方法参数值方法
本文MVSSSPANManTraGSRNetDenseFCNLocateNetEMTNet
高斯噪声标准差=30.8720.7640.1640.0600.6020.8020.7240.822
标准差=90.8540.7580.1650.0510.6030.7490.7070.809
标准差=150.8390.7510.1650.0500.5970.6900.6820.808
JPEG 压缩质量因子=500.8730.7610.2500.1860.5670.8110.6910.812
质量因子=750.8870.7700.2520.2260.5690.8120.7360.825
质量因子=1000.8910.7660.2520.4570.5720.8110.7380.825
高斯模糊卷积核=30.8800.7680.2490.1980.6020.8020.7370.824
卷积核=90.8560.7420.2390.1650.5830.7490.7160.811
卷积核=150.8150.7110.2310.1610.5730.6900.6450.782
[1] Shi Z, Chen H, Zhang D. Transformer-auxiliary neural networks for image manipulation localization by operator inductions[J]. IEEE Transactions on Circuits and Systems for Video Technology, 2023, 33(9): 4907-4920.
[2] 钟辉, 康恒, 吕颖达, 等. 基于注意力卷积神经网络的图像篡改定位算法[J]. 吉林大学学报: 工学版, 2021, 51(5): 1838-1844.
Zhong Hui, Kang Heng, Ying-da Lyu, et al. Image manipulation localization algorithm based on channel attention convolutional neural networks[J]. Journal of Jilin University (Engineering and Technology Edition), 2021, 51(5): 1838-1844.
[3] 石泽男, 陈海鹏, 张冬, 等. 预训练驱动的多模态边界感知视觉Transformer[J]. 软件学报, 2023, 34(5): 2051-2067.
Shi Ze-nan, Chen Hai-peng, Zhang Dong, et al. Pretraining-driven multimodal boundary-aware vision transformer[J]. Journal of Software, 2023, 34(5): 2051-2067.
[4] Xu D, Shen X, Lyu Y, et al. MC-Net: Learning mutually complementary features for image manipulation localization[J]. International Journal of Intelligent Systems, 2022, 37(5): 3072-3089.
[5] Mahdian B, Saic S. Using noise inconsistencies for blind image forensics[J]. Image and Vision Computing, 2009, 27(10): 1497-1503.
[6] Lin Z, He J, Tang X, et al. Fast, automatic and fine-grained tampered JPEG image detection via DCT coefficient analysis[J]. Pattern Recognition, 2009, 42(11): 2492-2501.
[7] Popescu A C, Farid H. Exposing digital forgeries in color filter array interpolated images[J]. IEEE Transactions on Signal Processing, 2005, 53(10): 3948-3959.
[8] Zhou P, Chen B C, Han X, et al. Generate, segment, and refine: towards generic manipulation segmentation[C]//Proceedings of the AAAI Conference on Artificial Intelligence, New York, USA, 2020: 13058-13065.
[9] Wang J, Wu Z, Chen J, et al. Objectformer for image manipulation detection and localization[C]∥Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, USA, 2022: 2364-2373.
[10] Lin X, Wang S, Deng J, et al. Image manipulationdetection by multiple tampering traces and edge artifact enhancement[J]. Pattern Recognition, 2023, 133:109026-109036.
[11] Wang W, Xie E, Li X, et al. PVT v2: improved baselines with pyramid vision transformer[J]. Computational Visual Media, 2022, 8(3): 415-424.
[12] 胡林辉, 陈保营, 谭舜泉, 等. 基于Convnext-Upernet的图像篡改检测定位模型[J/OL]. [2023-09-10].
Hu Lin-hui, Chen Bao-ying, Tan Shun-quan, et al. Convnext-Upernet based deep-learning model for image forgery detection and localization[J/OL]. [2023-09-10].
[13] Dong J, Wang W, Tan T. Casia image tampering detection evaluation database[C]∥2013 IEEE China Summit and International Conference on Signal and Information Processing, Beijing, China, 2013: 422-426.
[14] Guan H, Kozak M, Robertson E, et al. MFC datasets: Large-scale benchmark datasets for media for ensic challenge evaluation[C]∥2019 IEEE Winter Applications of Computer Vision Workshops (WACVW), Waikoloa, USA, 2019: 63-72.
[15] Hsu Y F, Chang S F. Detecting image splicing usinggeometry invariants and camera characteristics consistency[C]∥2006 IEEE International Conference on Multimedia and Expo, Toronto, Canada, 2006: 549-552.
[16] Wen B, Zhu Y, Subramanian R, et al. COVERAGE: a novel database for copy-move forgery detection[C]∥2016 IEEE International Conference on Image Processing (ICIP), Phoenix, USA, 2016: 161-165.
[17] Novozamsky A, Mahdian B, Saic S. IMD2020: a large-scale annotated dataset tailored for detecting manipulated images[C]∥Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision Workshops, Snowmass Village, USA, 2020: 71-80.
[18] Wu Y, AbdAlmageed W, Natarajan P. Mantra-net: Manipulation tracing network for detection and localization of image forgeries with anomalous features[C]∥Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, USA, 2019: 9543-9552.
[19] Hu X, Zhang Z, Jiang Z, et al. SPAN: spatial pyramid attention network for image manipulation localization[C]∥Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 2020: 312-328.
[20] Chen X, Dong C, Ji J, et al. Image manipulation detection by multi-view multi-scale supervision[C]∥Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, Canada, 2021: 14165-14173.
[21] Zhuang P, Li H, Tan S, et al. Image tampering localization using a dense fully convolutional network[J].IEEE Transactions on Information Forensics and Security, 2021, 16: 2986-2999.
[22] Zhuo L, Tan S, Li B, et al. Self-adversarial training incorporating forgery attention for image forgery localization[J]. IEEE Transactions on Information Forensics and Security, 2022, 17: 819-834.
[1] 王健,贾晨威. 面向智能网联车辆的轨迹预测模型[J]. 吉林大学学报(工学版), 2025, 55(6): 1963-1972.
[2] 车翔玖,孙雨鹏. 基于相似度随机游走聚合的图节点分类算法[J]. 吉林大学学报(工学版), 2025, 55(6): 2069-2075.
[3] 刘萍萍,商文理,解小宇,杨晓康. 基于细粒度分析的不均衡图像分类算法[J]. 吉林大学学报(工学版), 2025, 55(6): 2122-2130.
[4] 周丰丰,郭喆,范雨思. 面向不平衡多组学癌症数据的特征表征算法[J]. 吉林大学学报(工学版), 2025, 55(6): 2089-2096.
[5] 赵宏伟,周伟民. 基于数据增强的半监督单目深度估计框架[J]. 吉林大学学报(工学版), 2025, 55(6): 2082-2088.
[6] 申自浩,高永生,王辉,刘沛骞,刘琨. 面向车联网隐私保护的深度确定性策略梯度缓存方法[J]. 吉林大学学报(工学版), 2025, 55(5): 1638-1647.
[7] 王友卫,刘奥,凤丽洲. 基于知识蒸馏和评论时间的文本情感分类新方法[J]. 吉林大学学报(工学版), 2025, 55(5): 1664-1674.
[8] 赵宏伟,周明珠,刘萍萍,周求湛. 基于置信学习和协同训练的医学图像分割方法[J]. 吉林大学学报(工学版), 2025, 55(5): 1675-1681.
[9] 侯越,郭劲松,林伟,张迪,武月,张鑫. 分割可跨越车道分界线的多视角视频车速提取方法[J]. 吉林大学学报(工学版), 2025, 55(5): 1692-1704.
[10] 王军,司昌馥,王凯鹏,付强. 融合集成学习技术和PSO-GA算法的特征提取技术的入侵检测方法[J]. 吉林大学学报(工学版), 2025, 55(4): 1396-1405.
[11] 徐涛,孔帅迪,刘才华,李时. 异构机密计算综述[J]. 吉林大学学报(工学版), 2025, 55(3): 755-770.
[12] 赵孟雪,车翔玖,徐欢,刘全乐. 基于先验知识优化的医学图像候选区域生成方法[J]. 吉林大学学报(工学版), 2025, 55(2): 722-730.
[13] 蔡晓东,周青松,张言言,雪韵. 基于动静态和关系特征全局捕获的社交推荐模型[J]. 吉林大学学报(工学版), 2025, 55(2): 700-708.
[14] 车翔玖,武宇宁,刘全乐. 基于因果特征学习的有权同构图分类算法[J]. 吉林大学学报(工学版), 2025, 55(2): 681-686.
[15] 郭晓然,王铁君,闫悦. 基于局部注意力和本地远程监督的实体关系抽取方法[J]. 吉林大学学报(工学版), 2025, 55(1): 307-315.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
[1] 吉林大学学报(工学版)编辑部. 吉林大学学报(工学版)2006年总目录[J]. 吉林大学学报(工学版), 2006, 36(06): 0 -0008 .
[2] 李少华,黄慧,郭婷婷,蒲建业 . 颗粒在横向紊动射流中运动轨迹的数值模拟[J]. 吉林大学学报(工学版), 2006, 36(06): 1029 -1033 .
[3] 李建桥;文立阁;张秀芝;谭力 . 基于UG的灭茬刀辊参数化设计方法[J]. 吉林大学学报(工学版), 2008, 38(03): 595 -0598 .
[4] 姚智胜,邵春福,熊志华,岳昊 . 基于主成分分析和支持向量机的道路网短时交通流量预测 [J]. 吉林大学学报(工学版), 2008, 38(01): 48 -52 .
[5] 赵宏伟,刘萍萍,臧雪柏,张仁俊,李军令,王莉丽 . 基于逻辑推理和表象推理的类人
机器智能——思维计算
[J]. 吉林大学学报(工学版), 2006, 36(增刊2): 130 -133 .
[6] 张健红,左春柽,钱生君,齐培正 . 基于轴承间隙模型的摆锤碰撞试验台仿真研究[J]. 吉林大学学报(工学版), 2007, 37(03): 553 -0557 .
[7] 赵丁选,冯石柱,巩明德,邓乐 .

遥操作工程机器人改进力反馈控制方法

[J]. 吉林大学学报(工学版), 2008, 38(03): 575 -0579 .
[8] 王庆年,金启前,初 亮,王伟华. 传动系参数和控制参数对并联混合动力轿车性能的影响[J]. 吉林大学学报(工学版), 2005, 35(03): 243 -248 .
[9] 杜忠泽,黄俊霞,符寒光,王经涛,赵西成. 65Mn钢大塑性变形后的组织与力学性能[J]. 吉林大学学报(工学版), 2006, 36(02): 143 -0147 .
[10] 于锡峥;郑建华;高怀宝;刘正常 . 地月系L1和L2点间转移轨道设计[J]. 吉林大学学报(工学版), 2008, 38(03): 741 -0745 .