吉林大学学报(工学版) ›› 2025, Vol. 55 ›› Issue (12): 4063-4071.doi: 10.13229/j.cnki.jdxbgxb.20250014

• 计算机科学与技术 • 上一篇    

基于边界不确定性学习的图像篡改定位方法

陈海鹏1,2(),刘宏昕1,2,康辉1,2,刘雪洁1()   

  1. 1.吉林大学 计算机科学与技术学院,长春 130012
    2.吉林大学 符号计算与知识工程教育部重点实验室,长春 130012
  • 收稿日期:2025-01-06 出版日期:2025-12-01 发布日期:2026-02-03
  • 通讯作者: 刘雪洁 E-mail:chenhp@jlu.edu.cn;xuejie@jlu.edu.cn
  • 作者简介:陈海鹏(1978-),男,教授,博士.研究方向:图像处理与模式识别.E-mail:chenhp@jlu.edu.cn
  • 基金资助:
    国家重点研发计划项目(2022YFB4500600);吉林省科技发展计划重点研发项目(20230201088GX)

Image manipulation localization method based on boundary uncertainty learning

Hai-peng CHEN1,2(),Hong-xin LIU1,2,Hui KANG1,2,Xue-jie LIU1()   

  1. 1.College of Computer Science and Technology,Jilin University,Changchun 130012,China
    2.Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education,Jilin University,Changchun 130012,China
  • Received:2025-01-06 Online:2025-12-01 Published:2026-02-03
  • Contact: Xue-jie LIU E-mail:chenhp@jlu.edu.cn;xuejie@jlu.edu.cn

摘要:

针对目前图像篡改定位方法提取特征尺度单一、小篡改区域易与背景混淆造成误检漏检、预测结果不确定性高等问题,提出了基于边界不确定性学习的图像篡改定位方法。首先,使用金字塔视觉变压器提取篡改图像基础特征。其次,利用多级交互粗定位分支生成粗定位图。再次,利用小目标感知细化分支提高小篡改区域感知定位能力。随后,利用多尺度特征融合模块实现多尺度特征的充分交互与融合。最后,提出基于熵的边界不确定性感知损失进行辅助监督,极大地降低了预测结果的不确定性。在5个常用公开图像篡改数据集上分别进行域内和跨域实验,结果表明,本文方法可精准定位篡改区域,并优于其他方法。

关键词: 计算机应用, 图像篡改定位, 多尺度特征交互, 小篡改区域感知细化, 边界不确定性学习

Abstract:

The limitations of current image manipulation localization methods, such as the extraction of features at a single scale, the misdetection and omission of small tampered regions caused by background confusion, and the uncertainty existing in prediction results, are addressed. An image manipulation localization method based on edge uncertainty learning is proposed. The base features of the tampered image are extracted by means of a pyramid vision transformer. A coarse localization map is then generated through multi-level interactive coarse localization branches. To enhance the detection of small tampered regions, a small target-aware refinement branch is employed. Multi-scale feature fusion is achieved with the use of a dedicated module, which enables the full interaction and integration of features across different scales. Additionally, entropy-based perceptual loss is introduced to supervise boundary uncertainty, thus significantly reducing the uncertainty of the prediction results. The proposed method is evaluated on five widely-used public image tampering datasets in both in-domain and cross-domain experiments. It is demonstrated by the results that the method can effectively localize tampered regions and outperform existing approaches.

Key words: computer application, image manipulation localization, multi-scale feature interaction, refined perception of small tampering areas, boundary uncertainty learning

中图分类号: 

  • TP391

图1

本文方法的整体网络结构"

图2

多注意粗定位模块结构"

图3

全注意力模块结构"

图4

特征感知增强模块结构"

图5

多尺度特征融合模块结构"

表1

数据集描述"

数据集分割设置篡改操作类型后处理操作
训练测试拼接复制-粘贴移除
CASIA155 123921
NIST1616404160
COVER177525
Columbia1813050
IMD2020191 610400

表2

本文方法与其他方法的域内实验定量结果比较"

方法CASIAVNIST16ColumbiaCOVERIMD2020Mean
AUCF1AUCF1AUCF1AUCF1AUCF1AUCF1
ManTra-Net200.6480.2230.7950.4620.8240.7770.2830.7850.2650.7660.308
SPAN210.7090.2130.9610.5820.9360.8150.7910.3250.8490.484
DenseFCN220.6310.2090.9540.7040.8810.7100.7540.1850.7230.2860.7890.419
SATFL230.6970.2460.9370.6130.8920.8040.7670.3470.7960.3000.8180.462
MVSS-Net100.7480.3900.9810.8270.7190.7030.8080.2840.8170.4110.8150.523
TANet20.7390.4930.9550.9010.9820.9600.7560.4250.7660.3830.8400.633
DMFF-Net240.7910.3860.9760.8430.9450.8370.7270.2820.7840.3280.8450.535
DAE-Net110.8050.4940.9840.8710.9730.8720.7410.3300.8260.3380.8660.581
PBUL-Net0.8520.6620.9810.9060.9820.9610.8190.5250.7960.4880.8860.708

表3

与其他方法在未知数据集上的定量结果比较"

方法CASIAVNIST16Columbia
AUCF1AUCF1AUCF1
ManTra-Net210.6480.2230.4750.0950.4810.386
DenseFCN220.6310.2090.6040.0510.5860.317
MVSS-Net230.7480.3900.6430.2460.6970.471
TANet20.7390.4930.6380.2480.7100.440
DAE-Net120.8050.4940.6630.2990.7260.486
PBUL-Net0.8520.6620.7020.3440.8000.690

图6

本文方法与其他方法的定性比较"

表4

不同组件在CASIA数据集的定量实验"

网络组件AUCF1
MFFMXFAMFAEMEUBAL
a0.8100.605
b0.8190.618
c0.8320.629
d0.8380.641
e0.8480.653
f0.8520.662

图7

鲁棒性实验结果"

[1] 钟辉, 康恒, 吕颖达, 等. 基于注意力卷积神经网络的图像篡改定位算法[J]. 吉林大学学报: 工学版, 2021, 51(5): 1838-1844.
Zhong Hui, Kang Heng, Ying-da Lyu, et al. Image manipulation localization algorithm based on channel attention convolutional neural networks[J]. Journal of Jilin University (Engineering and Technology Edition), 2021, 51(5): 1838-1844.
[2] Shi Z, Chen H, Zhang D. Transformer-auxiliary neural networks for image manipulation localization by operator inductions[J]. IEEE Transactions on Circuits and Systems for Video Technology, 2023, 33(9): 4907-4920.
[3] 石泽男, 陈海鹏, 张冬, 等. 预训练驱动的多模态边界感知视觉 Transformer[J]. 软件学报, 2023, 34(5): 2051-2067.
Shi Ze-nan, Chen Hai-peng, Zhang Dong, et al. Pretraining-driven multimodal boundary-aware vision transformer[J]. Journal of Software, 2023, 34(5): 2051-2067.
[4] Liu Y, Zhu X, Zhao X, et al. Adversarial learning for constrained image splicing detection and localization based on atrous convolution[J]. IEEE Transactions on Information Forensics and Security, 2019, 14(10): 2551-2566.
[5] Chen B, Tan W, Coatrieux G, et al. A serial image copy-move forgery localization scheme with source/target distinguishment[J]. IEEE Transactions on Multimedia, 2020, 23: 3506-3517.
[6] Zhang Y, Fu Z, Qi S, et al. Localization of inpainting forgery with feature enhancement network[J]. IEEE Transactions on Big Data, 2022, 9(3): 936-948.
[7] Popescu A C, Farid H. Exposing digital forgeries in color filter array interpolated images[J]. IEEE Transactions on Signal Processing, 2005, 53(10): 3948-3959.
[8] Bianchi T, Piva A. Image forgery localization via block-grained analysis of JPEG artifacts[J]. IEEE Transactions on Information Forensics and Security, 2012, 7(3): 1003-1017.
[9] Mahdian B, Saic S. Using noise inconsistencies for blind image forensics[J]. Image and Vision Computing, 2009, 27(10): 1497-1503.
[10] Chen X, Dong C, Ji J, et al. Image manipulation detection by multi-view multi-scale supervision[C]//Proceedings of the IEEE/CVF. International Conference on Computer Vision. Montreal, QC, Canada, 2021: 14165-14173.
[11] Shi C, Wang C, Zhou X, et al. DAE-net: Dual attention mechanism and edge supervision network for image manipulation detection and localization[J]. IEEE Transactions on Instrumentation and Measurement, 2024(73): No.5028112.
[12] Wang W, Xie E, Li X, et al. Pvt v2: Improved baselines with pyramid vision transformer[J]. Computational Visual Media, 2022, 8(3): 415-424.
[13] Park T, Liu M Y, Wang T C, et al. Semantic image synthesis with spatially-adaptive normalization[C]∥Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Long Beach, CA: USA, 2019: 2337-2346.
[14] Pang Y, Zhao X, Xiang T Z, et al. Zoom in and out: A mixed-scale triplet network for camouflaged object detection[C]∥Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. New Orleans, LA: USA, 2022: 2160-2170.
[15] Dong J, Wang W, Tan T. Casia image tampering detection evaluation database[C]∥2013 IEEE. China Summit and International Conference on Signal and Information Processing.Beijing: China, 2013: 422-426.
[16] Guan H, Kozak M, Robertson E, et al. MFC datasets: Large-scale benchmark datasets for media for ensic challenge evaluation[C]∥2019 IEEE.Winter Applications of Computer Vision Workshops (WACVW). Waikoloa, HI: USA, 2019: 63-72.
[17] Wen B, Zhu Y, Subramanian R, et al. COVERAGE—a novel database for copy-move forgery detection[C]∥2016 IEEE. International Conference on Image Processing (ICIP).Phoenix, AZ: USA, 2016: 161-165.
[18] Hsu Y F, Chang S F. Detecting image splicing usinggeometry invariants and camera characteristics consistency[C]∥2006 IEEE. International Conference on Multimedia and Expo. Toronto, ON: Canada, 2006: 549-552.
[19] Novozamsky A, Mahdian B, Saic S. IMD2020 a largescale annotated dataset tailored for detecting manipulated images[C]∥Proceedings of the IEEE/CVF. Winter Conference on Applications of Computer Vision Workshops. Snowmass Village, CO: USA, 2020: 71-80.
[20] Wu Y, AbdAlmageed W, Natarajan P. Mantra-net manipulation tracing network for detection and localization of image forgeries with anomalous features[C]∥Proceedings of the IEEE/CVF. Conference on Computer Vision and Pattern Recognition. Long Beach, CA: USA, 2019: 9543-9552.
[21] Hu X, Zhang Z, Jiang Z, et al. SPAN spatial pyramid attention network for image manipulation localization[C]∥Proceedings of the European Conference on Computer Vision (ECCV). Glasgow: UK, 2020: 312-328.
[22] Zhuang P, Li H, Tan S, et al. Image tampering localization using a dense fully convolutional network[J]. IEEE Transactions on Information Forensics and Security, 2021, 16: 2986-2999.
[23] Zhuo L, Tan S, Li B, et al. Self-adversarial training incorporating forgery attention for image forgery localization[J]. IEEE Transactions on Information Forensics and Security, 2022, 17: 819-834.
[24] Xia X, Su L C, Wang S P, et al. DMFF-net: Double-stream multilevel feature fusion network for image forgery localization[J]. Engineering Applications of Artificial Intelligence, 2024, 127: No.107200.
[1] 艾青林,刘元宵,杨佳豪. 基于MFF-STDC网络的室外复杂环境小目标语义分割方法[J]. 吉林大学学报(工学版), 2025, 55(8): 2681-2692.
[2] 张宇飞,王丽敏,赵建平,贾智尧,李明洋. 基于中心选择大逃杀优化算法的机器人逆运动学求解[J]. 吉林大学学报(工学版), 2025, 55(8): 2703-2710.
[3] 朴燕,康继元. RAUGAN:基于循环生成对抗网络的红外图像彩色化方法[J]. 吉林大学学报(工学版), 2025, 55(8): 2722-2731.
[4] 刘琼昕,王甜甜,王亚男. 非支配排序粒子群遗传算法解决车辆位置路由问题[J]. 吉林大学学报(工学版), 2025, 55(7): 2464-2474.
[5] 车翔玖,李良. 融合全局与局部细粒度特征的图相似度度量算法[J]. 吉林大学学报(工学版), 2025, 55(7): 2365-2371.
[6] 李文辉,杨晨. 基于对比学习文本感知的小样本遥感图像分类[J]. 吉林大学学报(工学版), 2025, 55(7): 2393-2401.
[7] 庄珊娜,王君帅,白晶,杜京瑾,王正友. 基于三维卷积与自注意力机制的视频行人重识别[J]. 吉林大学学报(工学版), 2025, 55(7): 2409-2417.
[8] 赵宏伟,周伟民. 基于数据增强的半监督单目深度估计框架[J]. 吉林大学学报(工学版), 2025, 55(6): 2082-2088.
[9] 王健,贾晨威. 面向智能网联车辆的轨迹预测模型[J]. 吉林大学学报(工学版), 2025, 55(6): 1963-1972.
[10] 陈海鹏,张世博,吕颖达. 多尺度感知与边界引导的图像篡改检测方法[J]. 吉林大学学报(工学版), 2025, 55(6): 2114-2121.
[11] 周丰丰,郭喆,范雨思. 面向不平衡多组学癌症数据的特征表征算法[J]. 吉林大学学报(工学版), 2025, 55(6): 2089-2096.
[12] 车翔玖,孙雨鹏. 基于相似度随机游走聚合的图节点分类算法[J]. 吉林大学学报(工学版), 2025, 55(6): 2069-2075.
[13] 刘萍萍,商文理,解小宇,杨晓康. 基于细粒度分析的不均衡图像分类算法[J]. 吉林大学学报(工学版), 2025, 55(6): 2122-2130.
[14] 王友卫,刘奥,凤丽洲. 基于知识蒸馏和评论时间的文本情感分类新方法[J]. 吉林大学学报(工学版), 2025, 55(5): 1664-1674.
[15] 赵宏伟,周明珠,刘萍萍,周求湛. 基于置信学习和协同训练的医学图像分割方法[J]. 吉林大学学报(工学版), 2025, 55(5): 1675-1681.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!