DETECTING STABLE DIFFUSION GENERATED IMAGES USING FREQUENCY ARTIFACTS: A CASE STUDY ON DISNEY-STYLE ART
Junbin Zhang, Yixiao Wang, Hamid Reza Tohidypour, Panos Nasiopoulos
-
SPS
IEEE Members: $11.00
Non-members: $15.00
The use of Stable Diffusion models to generate realistic images has become a popular topic in recent years. However, this technology has also raised concerns about the potential harm it may cause to the copyright holders, particularly in the realm of art where these synthesized images can closely resemble the original work. As these synthesized images are hard for humans to distinguish from authentic ones, it is of great importance to develop methods that may identify them. In this paper, we propose a deep learning-based approach to detect synthesized images using information in the frequency domain. Since there exists no well-established dataset of images synthesized by stable diffusion models, in order to train and evaluate our network we generated a representative dataset consisting of carefully selecting human-created authentic images and synthesized animation images generated by the Stable Diffusion models. We chose to use Disney-style animated content for our case study, given its significance in the realm of intellectual property protection. Experimental results demonstrated that our proposed model outperforms humans and other state-of-the-art methods, achieving an accuracy rate of 99.46%.