期刊文献+

AIGC视觉内容生成与溯源研究进展 被引量:1

Review on the progress of the AIGC visual contentgeneration and traceability
原文传递
导出
摘要 随着数字媒体与创意产业的快速发展,人工智能生成内容(artificial intelligence generated content, AIGC)技术以其在视觉内容生成中的创新应用而逐渐受到关注。本文旨在围绕AIGC视觉内容生成与溯源研究进展深入研讨。首先,针对图像生成技术进行探讨,从基于生成式对抗网络的传统方法出发,系统地分析了基于生成式对抗网络、自回归模型和扩散概率模型的最新进展。接着,深入探讨可控图像生成技术,突出了通过布局、线稿等附加信息以及基于视觉参考的方法来为创作者提供精确控制的技术现状。随着图像生成技术的革新和应用,生成图像的安全性问题逐渐浮现。而预先审核和过滤的技术手段已难以满足实际需求,故亟需实现生成内容的溯源来进行监管。因此,本文进而对生成图像溯源技术进行研讨,并聚焦水印技术在确保生成内容可靠性和安全性方面的应用。依据水印嵌入的流程节点,首先将现有的水印相关的生成图像溯源方法归为无水印嵌入的生成图像溯源、水印前置嵌入的生成图像溯源、水印后置嵌入的生成图像溯源以及联合生成的生成图像溯源并进行详细分析,然后介绍针对生成图像的水印攻击研究现状,最后对生成图像溯源技术进行总结和展望。鉴于视觉内容生成在质量和安全上的挑战,旨在为研究者提供一个视觉内容生成与溯源的系统研究视角,以促进数字媒体创作环境的安全与可信,并引导未来相关技术的发展方向。 In the contemporary digital era,which is characterized by rapid technological advancements,multimedia con⁃tent creation,particularly in visual content generation,has become an integral part of modern societal development.Theexponential growth of digital media and the creative industry has attracted attention to artificial intelligence generated con⁃tent(AIGC)technology.The groundbreaking applications of AIGC in visual content generation not only have equippedmultimedia creators with novel tools and capabilities but also have delivered substantial benefits across diverse domains,which span from the realms of cinema and gaming to the immersive landscapes of virtual reality.This review comprehensiveintroduces the profound advancements within AIGC technology.Our particular emphasis is on the domain of visual contentgeneration and its critical facet of traceability.Initially,our discussions trace the evolutionary path of image generationtechnology,from its inception within generative adversarial networks(GANs)to the latest advancements in Transformerauto-regressive models and diffusion probability models.This progression unveils a remarkable leap in the quality and capa⁃bility of image generation,which underscores the rapid evolution of this field.This evolution has transitioned from itsnascent stages to an era characterized by explosive growth.First,we delve into the development of GANs,encompassingtheir evolution from text-conditioned methods to sophisticated techniques for style control and the development of largescale models.This type of technology pioneered the text-to-image generation.GANs can further improve their performanceby expanding network parameters and dataset size due to their strong scalability.Furthermore,we explore the emergence ofTransformer-based auto-regressive models,such as DALL·E and CogView,which have heralded a new epoch in thedomain of image generation.The basic strategy of autoregressive models is to first use the Transformer structure to predictthe feature sequence of images based on other conditional feature sequences such as text and sketches.Then,it uses a spe⁃cially trained decoding network to decode these feature sequences into a complete image.They can generate realisticimages based on the large-scale parameters.In addition,our discourse delves into the burgeoning interest surrounding dif⁃fusion probability models,which are renowned for their stable training methods and their ability to yield high-quality out⁃puts.The diffusion models first adopt an iterative and random process to simulate the gradual transformation of the observeddata into a known noise distribution.Then,they reconstruct the original data in the opposite direction from the noise distri⁃bution.This random process based on stochastic approach provides a more stable training process,while it also demon⁃strates impressive results in terms of generated quality and diversity.As the development of AIGC technology continues toadvance,it encounters challenges,such as the enhancement in content quality and the need of precise control to align withspecific requisites.Within this context,this review conducts a thorough exploration of controllable image generation tech⁃nology,which is a pivotal research domain that strives to furnish meticulous control over the generated content.Thisachievement is facilitated through the integration of supplementary elements,such as intricate layouts,detailed sketches,and precise visual references.This approach empowers creators to preserve their artistic autonomy while upholding exact⁃ing standards of quality.One notable facet that has garnered considerable academic attention is the utilization of visual ref⁃erences as a mechanism to enable the generation of diverse styles and personalized outcomes by incorporating user-providedvisual elements.This review underscores the profound potential inherent in these methodologies,which illustrates theirtransformative role across domains such as digital art and interactive media.The development of these technologies intro⁃duces new horizons in digital creativity.However,it presents profound challenges,particularly in the domain of imageauthenticity and the potential for malevolent misuse.These risks are exemplified by the creation of deep fakes or the prolif⁃eration of fake news.These challenges extend far beyond mere technical intricacies;they encompass substantial risks per⁃taining to individual privacy,security,and the broader societal implications of eroding public trust and social stability.Inresponse to these formidable challenges,watermark-related image traceability technology has emerged as an indispensablesolution.This technology harnesses the power of watermarking techniques to authenticate and verify AI-generated images,which safeguards their integrity.Within the pages of this review,we meticulously categorize these watermarking techniquesinto distinct types:watermark-free embedding,watermark pre-embedding,watermark post-embedding,and joint genera⁃tion methods.First,we introduce the watermark-free embedding methods,which treat the generated traces left duringmodel generation as fingerprints.The inherent fingerprint information is used to achieve model attribution of generatedimages and achieve traceability purposes.Second,the watermark pre-embedding methods aim to embed the watermark intoinput training data such as noise and image.Another aim is to use the embedded watermark data to train the generationmodel,which can also introduce traceability information in the generated image.Third,the watermark post-embedding methods divide the process of generating watermark images into two stages:image generation and watermark embedding.Watermark embedding is performed after image generation.Finally,the joint generation methods aim to achieve adaptiveembedding of watermark information during the image generation process,minimize damage to the image generation pro⁃cess when fusing with image features,and ultimately generate images carrying watermarks.Each of these approaches playsa pivotal role in the verification of traceability across diverse scenarios,which offers a robust defense against potential mis⁃uses of AI-generated imagery.In conclusion,while AIGC technology offers promising new opportunities in visual contentcreation,it simultaneously causes significant challenges regarding the security and integrity of generated content.This com⁃prehensive review covers the breadth of AIGC technology,which starts from an overview of existing image generation tech⁃nologies,such as GANs,auto-regressive models,and diffusion probability models.It then categorizes and analyzes con⁃trollable image generation technology from the perspectives of additional conditions and visual examples.In addition,thereview focuses on watermark-related image traceability technology,discusses various watermark embedding techniques andthe current state of watermark attacks on generated images,and provides an extensive overview and future outlook of gen⁃eration image traceability technology.The aim is to offer researchers a detailed,systematic,and comprehensive perspec⁃tive on the advancements in AIGC visual content generation and traceability.This study deepens the understanding of cur⁃rent research trends,challenges,and future directions in this rapidly evolving field.
作者 刘安安 苏育挺 王岚君 李斌 钱振兴 张卫明 周琳娜 张新鹏 张勇东 黄继武 俞能海 Liu Anan;Su Yuting;Wang Lanjun;Li Bin;Qian Zhenxing;Zhang Weiming;Zhou Linna;Zhang Xinpeng;Zhang Yongdong;Huang Jiwu;Yu Nenghai(School of Electrical and Information Engineering,Tianjin University,Tianjin 300072,China;College of Electronics and Information Engineering,Shenzhen University,Shenzhen 518060,China;School of Computer Science,Fudan University,Shanghai 200438,China;School of Information Science and Technology,University of Science and Technology of China,Hefei 230026,China;School of Cyberspace Security,Beijing University of Posts and Telecommunications,Beijing 100876,China;School of Cyber Science and Technology,University of Science and Technology of China,Hefei 230027,China)
出处 《中国图象图形学报》 CSCD 北大核心 2024年第6期1535-1554,共20页 Journal of Image and Graphics
基金 国家自然科学基金项目(U21B2024,U20B2047,U2336206,U20B2051,U23B2022,62371330,62202329,62172053)。
关键词 人工智能内容生成(AIGC) 视觉内容生成 可控图像生成 生成内容安全 生成图像溯源 artificial intelligence generated content(AIGC) visual content generation controllable image generation security of generated content traceability of generated images
  • 相关文献

参考文献1

二级参考文献3

共引文献6

同被引文献3

引证文献1

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部