Skip to content
Help Build a Better ComfyUI Knowledge Base Become a Patron

AIGC Latest News

Stay up to date with latest AIGC news and updates.

ByteDance Releases InfiniteYou: Flexible Photo Recrafting While Preserving User Identity

ByteDance Releases InfiniteYou: Flexible Photo Recrafting While Preserving User Identity

ByteDance introduces InfiniteYou (InfU), an innovative framework based on Diffusion Transformers that enables flexible photo recrafting while preserving user identity, addressing limitations in existing methods regarding identity similarity, text-image alignment, and generation quality

StdGEN: Semantic-Decomposed 3D Character Generation from Single Images

StdGEN: Semantic-Decomposed 3D Character Generation from Single Images

Tsinghua University and Tencent AI Lab jointly introduce StdGEN, an innovative pipeline that generates high-quality semantically-decomposed 3D characters from single images, enabling separation of body, clothing, and hair

Stability AI Releases Stable Virtual Camera: Technology to Transform 2D Photos into 3D Videos

Stability AI Releases Stable Virtual Camera: Technology to Transform 2D Photos into 3D Videos

Stability AI launches new AI model Stable Virtual Camera, capable of converting ordinary photos into 3D videos with authentic depth and perspective effects, providing creators with intuitive camera control

Tencent Releases Hunyuan3D 2.0 - Innovative 3D Asset Generation System

Tencent Releases Hunyuan3D 2.0 - Innovative 3D Asset Generation System

Tencent launches Hunyuan3D 2.0 system with a two-stage process for generating high-quality 3D models, featuring multiple open-source model series that support creating high-resolution 3D assets from text and images

Kuaishou Launches ReCamMaster Monocular Video Reframing Technology

Kuaishou Launches ReCamMaster Monocular Video Reframing Technology

Kuaishou Technology has launched ReCamMaster, a generative video technology that allows users to create new camera perspectives and motion paths from a single video.

Open-Sora 2.0 Released: Commercial-Grade Video Generation at Low Cost

Open-Sora 2.0 Released: Commercial-Grade Video Generation at Low Cost

LuChen Technology releases Open-Sora 2.0 open-source video generation model, achieving performance close to top commercial models with just $200,000 in training costs

Ali Tongyi Lab Releases VACE: All-in-One Video Creation and Editing Model

Ali Tongyi Lab Releases VACE: All-in-One Video Creation and Editing Model

Ali Tongyi Lab launches multifunctional video creation and editing model VACE, integrating various video processing tasks into a single framework to lower the barrier of video creation

Microsoft Releases ART Multi-layer Transparent Image Generation Technology

Microsoft Releases ART Multi-layer Transparent Image Generation Technology

Microsoft Research introduces intelligent layered generation based on global text prompts, supporting creation of transparent images with 50+ independent layers

Tencent Open-Sources HunyuanVideo-I2V Image-to-Video Model

Tencent Open-Sources HunyuanVideo-I2V Image-to-Video Model

Tencent's Hunyuan team releases open-source model for generating 5-second videos from single images, featuring smart motion generation and custom effects

Alibaba Open-Sources ViDoRAG Intelligent Document Analysis Tool

Alibaba Open-Sources ViDoRAG Intelligent Document Analysis Tool

Alibaba introduces a document analysis system capable of understanding both text and images, improving processing efficiency for complex documents by over 10%

THUDM Open-Sources CogView4 - Native Chinese-Supported DiT Text-to-Image Model

THUDM Open-Sources CogView4 - Native Chinese-Supported DiT Text-to-Image Model

THUDM releases CogView4 open-source image generation model with native Chinese support, leading in multiple benchmark tests

Sesame Unveils CSM Voice Model for Natural Conversations

Sesame Unveils CSM Voice Model for Natural Conversations

Sesame Research introduces dual-Transformer conversational voice model CSM, achieving human-like interaction with open-source core architecture

Alibaba's Wan2.1 Video Generation Model Officially Open-Sourced

Alibaba's Wan2.1 Video Generation Model Officially Open-Sourced

Alibaba has officially open-sourced its latest video generation model, Wan2.1, which can run with only 8GB of video memory, supporting high-definition video generation, dynamic subtitles, and multi-language dubbing, surpassing models like Sora with a total score of 86.22% on the VBench leaderboard

Alibaba Releases ComfyUI Copilot: AI-Driven Intelligent Workflow Assistant

Alibaba Releases ComfyUI Copilot: AI-Driven Intelligent Workflow Assistant

Alibaba International Digital Commerce Group (AIDC-AI) releases the ComfyUI Copilot plugin, simplifying the user experience of ComfyUI through natural language interaction and AI-driven functionality, supporting Chinese interaction, and offering intelligent node recommendations and other features

Alibaba's WanX 2.1 Video Generation Model to be Open-Sourced

Alibaba's WanX 2.1 Video Generation Model to be Open-Sourced

Alibaba has announced that its latest video generation model, WanX 2.1, will be open-sourced in the second quarter of 2025, supporting high-definition video generation, dynamic subtitles, and multi-language dubbing, ranking first on the VBench leaderboard with a total score of 84.7%

Google Releases PaliGemma 2 Mix: An Open-Source Visual Language Model Supporting Multiple Tasks

Google Releases PaliGemma 2 Mix: An Open-Source Visual Language Model Supporting Multiple Tasks

Google introduces the new PaliGemma 2 mix model, supporting various visual tasks including image description, OCR, object detection, and providing 3B, 10B, and 28B scale versions

Skywork Opensources SkyReels-V1: A Video Generation Model Focused on AI Short Drama Creation

Skywork Opensources SkyReels-V1: A Video Generation Model Focused on AI Short Drama Creation

Skywork has open-sourced its latest video generation model, SkyReels-V1, which supports text-to-video and image-to-video generation, featuring cinematic lighting effects and natural motion representation, and is now available for commercial use.

Light-A-Video - A Video Relighting Technology Without Training

Light-A-Video - A Video Relighting Technology Without Training

Researchers have proposed a new video relighting method, Light-A-Video, which achieves temporally smooth video relighting effects through Consistent Light Attention (CLA) and Progressive Light Fusion (PLF).

StepFun releases Step-Video-T2V: A 300 Billion Parameter Text-to-Video Model

StepFun releases Step-Video-T2V: A 300 Billion Parameter Text-to-Video Model

StepFun has released the open-source text-to-video model Step-Video-T2V, which has 300 billion parameters, supports the generation of high-quality videos up to 204 frames, and provides an online experience platform

Kuaishou Introduces CineMaster: Breakthrough in 3D-Aware Video Generation

Kuaishou Introduces CineMaster: Breakthrough in 3D-Aware Video Generation

Kuaishou officially releases CineMaster text-to-video generation framework, enabling high-quality video content creation through 3D-aware technology

Alibaba Open Sources InspireMusic: An Innovative Framework for Music, Song and Audio Generation

Alibaba Open Sources InspireMusic: An Innovative Framework for Music, Song and Audio Generation

Alibaba's latest open-source project InspireMusic, a unified audio generation framework based on FunAudioLLM, supporting music creation, song generation and various audio synthesis tasks.

Alibaba Open Sources ACE++: Zero-Training Character-Consistent Image Generation

Alibaba Open Sources ACE++: Zero-Training Character-Consistent Image Generation

Alibaba Research Institute open sources image generation tool ACE++, supporting character-consistent image generation from single input through context-aware content filling technology, offering online experience and three specialized models.

ByteDance Releases OmniHuman: Next-Generation Human Animation Framework

ByteDance Releases OmniHuman: Next-Generation Human Animation Framework

ByteDance research team releases OmniHuman-1 human animation framework, capable of generating high-quality human video animations from a single image and motion signals.

DeepSeek Open-Sources Janus-Pro-7B: Multimodal AI Model

DeepSeek Open-Sources Janus-Pro-7B: Multimodal AI Model

Tencent Releases Hunyuan3D 2.0: Open-Source High-Quality 3D Generation Model and End-to-End Creation Engine

Tencent Releases Hunyuan3D 2.0: Open-Source High-Quality 3D Generation Model and End-to-End Creation Engine

Tencent releases Hunyuan3D 2.0, open-sourcing the complete DiT model and launching a one-stop 3D creation engine with innovative features including skeletal animation and sketch-to-3D, revolutionizing metaverse and game content creation

ComfyUI Project Two-Year Anniversary

ComfyUI Project Two-Year Anniversary

ComfyUI celebrates its second birthday, evolving from a personal project into the world's most popular generative AI visual tool. Happy birthday, ComfyUI!

NVIDIA Open Sources Sana - An AI Model for Efficient 4K Image Generation

NVIDIA Open Sources Sana - An AI Model for Efficient 4K Image Generation

NVIDIA releases new Sana model capable of quickly generating images up to 4K resolution on consumer laptop GPUs, with ComfyUI integration support

ByteDance Open Sources LatentSync - High-Precision Lip Sync Technology Based on Diffusion Model

ByteDance Open Sources LatentSync - High-Precision Lip Sync Technology Based on Diffusion Model

ByteDance releases open-source lip sync tool LatentSync, based on audio-conditioned latent space diffusion model, enabling precise lip synchronization for both real people and animated characters while solving frame jittering issues common in traditional methods

VMix: ByteDance Introduces Innovative Aesthetic Enhancement Technology for Text-to-Image Diffusion Models

VMix: ByteDance Introduces Innovative Aesthetic Enhancement Technology for Text-to-Image Diffusion Models

ByteDance and University of Science and Technology of China jointly launch VMix adapter, enhancing AI-generated image aesthetics through cross-attention mixing control technology, seamlessly integrating with existing models without retraining

Tencent Open Sources StereoCrafter: One-Click 2D to 3D Video Conversion

Tencent Open Sources StereoCrafter: One-Click 2D to 3D Video Conversion

StereoCrafter, jointly developed by Tencent AI Lab and ARC Lab, is now open source. It can convert any 2D video into high-quality stereoscopic 3D video, supporting various 3D display devices including Apple Vision Pro

LuminaBrush: AI Lighting Editor Released by ControlNet Creator

LuminaBrush: AI Lighting Editor Released by ControlNet Creator

lllyasviel, creator of ControlNet and IC-Light, releases LuminaBrush - a new AI tool that enables precise lighting control through a two-stage framework with intuitive brush interactions

Genesis: Breakthrough Universal Physics Engine and Generative AI Platform Released

Genesis: Breakthrough Universal Physics Engine and Generative AI Platform Released

Genesis project releases a new universal physics engine and generative AI platform, integrating physics simulation, robot control and generative AI capabilities to provide comprehensive solutions for robotics and physical AI applications

Odyssey Releases Explorer - Breakthrough Generative World Model

Odyssey Releases Explorer - Breakthrough Generative World Model

Odyssey launches Explorer, the first generative world model that can convert 2D images into complete 3D worlds, supporting dynamic effects and mainstream creative tool editing, bringing revolutionary changes to film, gaming and other fields

Mining Virus Found in ComfyUI Impact-Pack Plugin - Urgent Action Required

Mining Virus Found in ComfyUI Impact-Pack Plugin - Urgent Action Required

Popular ComfyUI plugin Impact-Pack discovered to contain mining malware through Ultralytics package, affecting numerous users. This article details the virus situation and solutions.

Luma Launches Photon - A Revolutionary AI Image Generation Model with Exceptional Value

Luma Launches Photon - A Revolutionary AI Image Generation Model with Exceptional Value

Luma introduces revolutionary Photon and Photon Flash image generation models, redefining AI creation with outstanding value and superior image quality

Tencent Open Sources HunyuanVideo - A New Era in Video Generation

Tencent Open Sources HunyuanVideo - A New Era in Video Generation

Tencent officially open sources HunyuanVideo, the industry's largest video generation model with 13 billion parameters, achieving leading performance in video quality and motion stability

Stability AI Releases Stable Diffusion 3.5 Large ControlNet Models

Stability AI Releases Stable Diffusion 3.5 Large ControlNet Models

Stability AI launches three new ControlNet models for Stable Diffusion 3.5 Large, including Blur, Canny, and Depth, providing more precise control capabilities for image generation

NVIDIA Releases Edify 3D - A Revolutionary 3D Asset Generation Technology

NVIDIA Releases Edify 3D - A Revolutionary 3D Asset Generation Technology

NVIDIA has launched the new Edify 3D technology, capable of generating high-quality 3D assets in just two minutes, including detailed geometry, clear topology, high-resolution textures, and PBR materials.

Lightricks Releases Real-Time Video Generation Model LTX-Video

Lightricks Releases Real-Time Video Generation Model LTX-Video

Lightricks launches the LTX-Video real-time video generation model based on DiT, supporting real-time generation of high-quality videos, and has been open-sourced on GitHub and Hugging Face.

InstantX Releases FLUX.1-dev IP-Adapter Model

InstantX Releases FLUX.1-dev IP-Adapter Model

InstantX team has released the IP-Adapter model based on FLUX.1-dev, bringing more powerful image reference capabilities to the FLUX model

FLUX Official Tools Suite Released

FLUX Official Tools Suite Released

Black Forest Labs releases a variety of FLUX official tools, including local redraw, ControlNet, and image style conversion

IC-Light V2 Release: Enhanced Image Editing Capabilities

IC-Light V2 Release: Enhanced Image Editing Capabilities

IC-Light V2 is a Flux-based image editing model that supports various stylized image processing, including oil paintings and anime styles. This article details the new features and applications of IC-Light V2.

Stable Diffusion 3.5 Launches: A New Era in AI Image Generation

Stable Diffusion 3.5 Launches: A New Era in AI Image Generation

Stability AI releases Stable Diffusion 3.5, offering multiple powerful model variants, supporting commercial use, and leading the market in image quality and prompt adherence.

ComfyUI V1 Released: Cross-Platform Desktop App with One-Click Installation

ComfyUI V1 Released: Cross-Platform Desktop App with One-Click Installation

ComfyUI has released Version 1, offering a cross-platform desktop application with one-click installation, a revamped user interface, and numerous feature improvements, significantly enhancing the user experience.

Kuaishou and PKU Jointly Release Pyramidal Flow Matching Video Generation Model

Kuaishou and PKU Jointly Release Pyramidal Flow Matching Video Generation Model

Kuaishou Technology and Peking University jointly developed the Pyramidal Flow Matching model, an autoregressive video generation technology based on flow matching, capable of producing high-quality, long-duration video content.

Jasperai Releases Flux.1-dev ControlNet Model Series

Jasperai Releases Flux.1-dev ControlNet Model Series

Jasperai introduces a series of ControlNet models for Flux.1-dev, including surface normals, depth maps, and super-resolution models, providing more precise control for AI image generation.

ComfyUI-PuLID-Flux: Implementing PuLID-Flux in ComfyUI

ComfyUI-PuLID-Flux: Implementing PuLID-Flux in ComfyUI

ComfyUI-PuLID-Flux is an open-source project that integrates PuLID-Flux into ComfyUI, offering powerful image generation and editing capabilities.

Meta Introduces Movie Gen: AI Models for Video and Audio Generation

Meta Introduces Movie Gen: AI Models for Video and Audio Generation

Meta introduces the Movie Gen series of models, including video generation, audio generation, and personalized video editing capabilities, opening new frontiers in AI content creation.

ostris Releases OpenFLUX.1: A Commercially Usable De-distilled Version of FLUX.1-schnell

ostris Releases OpenFLUX.1: A Commercially Usable De-distilled Version of FLUX.1-schnell

ostris has released OpenFLUX.1, a de-distilled version of the FLUX.1-schnell model that can be fine-tuned, uses normal CFG values, and retains commercial licensing.

Black Forest Lab Launches FLUX 1.1 [Pro]

Black Forest Lab Launches FLUX 1.1 [Pro]

Black Forest Labs introduces FLUX1.1 [pro], a generative AI model featuring ultra-fast generation, superior image quality, and 2K ultra-high resolution generation, opening unprecedented opportunities for creators, developers, and businesses.

THUDM Open Sources New Image Generation Models: CogView3 and CogView-3Plus

THUDM Open Sources New Image Generation Models: CogView3 and CogView-3Plus

THUDM has open-sourced their latest image generation models CogView3 and CogView-3Plus-3B, showcasing exceptional performance and efficiency. CogView3 utilizes cascaded diffusion technology, while CogView-3Plus-3B is a lightweight model based on the DiT architecture, bringing significant breakthroughs to the field of text-to-image generation.