4.7 Article

Robust Texture-Aware Computer-Generated Image Forensic: Benchmark and Algorithm

Journal

IEEE TRANSACTIONS ON IMAGE PROCESSING
Volume 30, Issue -, Pages 8439-8453

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TIP.2021.3114989

Keywords

Benchmark testing; Feature extraction; Three-dimensional displays; Image forensics; Task analysis; Motion pictures; Games; Computer-generated image benchmark; texture aware; computer-generated images forensic

Funding

  1. National Key Research and Development Program of China [2020AAA0106800]
  2. National Natural Science Foundation of China [61902401, 61972071, 61906052, 62036011, 61721004, 61972394, U2033210, U1936204, U1803119, 61876100]
  3. Chinese Academy of Science (CAS) Key Research Program of Frontier Sciences [QYZDJ-SSWJSC040]
  4. Youth Innovation Promotion Association, CAS

Ask authors/readers for more resources

This study introduces a Large-Scale CG images Benchmark (LSCGB) and proposes a simple yet strong baseline model for forensic tasks. The benchmark has the advantages of large-scale, high diversity, and small bias. The baseline model improves forensic accuracy through a texture-aware network.
With advances in rendering techniques and generative adversarial networks, computer-generated (CG) images tend to be indistinguishable from photographic (PG) images. Revisiting previous works towards CG image forensic, we observed that existing datasets are constructed years ago and limited in both quantity and diversity. Besides, current algorithms only consider the global visual features for forensic, ignoring finer differences between CG and PG images. To mitigate these problems, we first contribute a Large-Scale CG images Benchmark (LSCGB), and then propose a simple yet strong baseline model to address the forensic task. On the one hand, the introduced benchmark has three superior properties, 1) large-scale: the benchmark contains 71168 CG and 71168 PG images with the corresponding expert-annotated labels. It is orders of magnitude bigger than previous datasets. 2) high diversity: we collect CG images from 4 different scenes generated by various rendering techniques. The PG images are varied in terms of image content, camera models, and photographer styles. 3) small bias: we carefully filter the collected images to ensure that the distributions of color, brightness, tone and saturation between CG and PG images are close. Furthermore, inspired by an empirical study on texture difference between CG and PG images, an effective texture-aware network is proposed to improve forensic accuracy. Concretely, we first strengthen texture information of multilevel features extracted from a backbone. Then, the relations among feature channels are explored by learning its gram matrix. Each feature channel represents a specific texture pattern. The gram matrix is thus able to embed the finer texture differences. Experimental results demonstrate that this baseline surpasses the existing methods. The benchmark is publically available at https://github.com/wmbai/LSCGB.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available