Skip to content

PaperCodex

Subscribe

Diffusion Models

FramePack: Generate Long, High-Quality Videos on a Laptop—Without Cloud Costs or Drifting Artifacts

FramePack: Generate Long, High-Quality Videos on a Laptop—Without Cloud Costs or Drifting Artifacts 16308

Creating long, coherent, and visually rich videos with AI has long been bottlenecked by computational complexity, memory constraints, and error…

12/27/2025Diffusion Models, Next-frame Prediction, Video Generation
InvSR: High-Quality Image Super-Resolution in 1–5 Steps Using Diffusion Inversion

InvSR: High-Quality Image Super-Resolution in 1–5 Steps Using Diffusion Inversion 1341

Image super-resolution (SR) remains a critical capability across computer vision applications—from upscaling smartphone photos to enhancing AI-generated content (AIGC). However,…

12/26/2025AIGC Enhancement, Diffusion Models, Image Super-resolution
OOTDiffusion: High-Fidelity, Controllable Virtual Try-On Without Garment Warping

OOTDiffusion: High-Fidelity, Controllable Virtual Try-On Without Garment Warping 6482

OOTDiffusion represents a significant leap forward in image-based virtual try-on (VTON) technology. Built on the foundation of pretrained latent diffusion…

12/26/2025Diffusion Models, Image Generation, Virtual Try-on
Show-1: High-Quality, Efficient Text-to-Video Generation with Precise Prompt Alignment

Show-1: High-Quality, Efficient Text-to-Video Generation with Precise Prompt Alignment 1133

Text-to-video generation has rapidly evolved, yet technical teams still face a persistent trade-off: high-quality outputs often come at prohibitive computational…

12/22/2025Diffusion Models, Text-to-Video Generation, Video Synthesis
DragDiffusion: Precise, Interactive Image Editing for Real and AI-Generated Photos Using Diffusion Models

DragDiffusion: Precise, Interactive Image Editing for Real and AI-Generated Photos Using Diffusion Models 1234

DragDiffusion is an open-source framework that brings pixel-precise, point-based image manipulation to both real-world photographs and AI-generated images—without requiring users…

12/19/2025Diffusion Models, Image Editing, Interactive Manipulation
VSA: Accelerate Video Diffusion Models by 2.5× with Trainable Sparse Attention—No Quality Tradeoff

VSA: Accelerate Video Diffusion Models by 2.5× with Trainable Sparse Attention—No Quality Tradeoff 2780

Video generation using diffusion transformers (DiTs) is rapidly advancing—but at a steep computational cost. Full 3D attention in these models…

12/19/2025Diffusion Models, Sparse Attention, Video Generation
DreamCraft3D: Generate Photorealistic, View-Consistent 3D Assets from a Single Image

DreamCraft3D: Generate Photorealistic, View-Consistent 3D Assets from a Single Image 2989

Creating high-quality 3D assets has traditionally required expert modeling skills, extensive manual labor, or expensive capture setups—barriers that limit accessibility…

12/18/20253D Generation, Diffusion Models, View-Consistent Rendering
Copyright © 2026 PaperCodex.
  • Facebook
  • YouTube
  • Twitter

PaperCodex