1、GAN/Image Synthesis(图像生成)
- L-MAGIC: Language Model Assisted Generation of Images with Coherence
- CapsFusion: Rethinking Image-Text Data at Scale
- C3Net: Compound Conditioned ControlNet for Multimodal Content Generation
- Scaling Laws of Synthetic Images for Model Training ... for Now
- An edit friendly ddpm noise space: inversion and manipulations
⭐code
🏠project - CoDi-2: In-Context Interleaved and Interactive Any-to-Any Generation
⭐code
🏠project - CapHuman: Capture Your Moments in Parallel Universes
⭐code
🏠project - Text-Conditioned Generative Model of 3D Strand-based Human Hairstyles
🏠project - IMPRINT: Generative Object Compositing by Learning Identity-Preserving Representation
- TexTile: A Differentiable Metric for Texture Tileability
🏠project - SD-DiT: Unleashing the Power of Self-supervised Discrimination in Diffusion Transformer
- PhotoMaker: Customizing Realistic Human Photos via Stacked ID Embedding
⭐code
🏠project - MobileCLIP: Fast Image-Text Models through Multi-Modal Reinforced Training
⭐code - Text-Image Alignment for Diffusion-Based Perception
- AEROBLADE: Training-Free Detection of Latent Diffusion Images Using Autoencoder Reconstruction Error
- FaceChain-SuDe: Building Derived Class to Inherit Category Attributes for One-shot Subject-Driven Generation
⭐code - It's All About Your Sketch: Democratising Sketch Control in Diffusion Models
⭐code - Codebook Transfer with Part-of-Speech for Vector-Quantized Image Modeling
- ProMark: Proactive Diffusion Watermarking for Causal Attribution
- DetDiffusion: Synergizing Generative and Perceptive Models for Enhanced Data Generation and Perception
- GAN
- StyLitGAN: Image-Based Relighting via Latent Control
⭐code
🏠project - StyleCineGAN: Landscape Cinemagraph Generation using a Pre-trained StyleGAN
- What You See is What You GAN: Rendering Every Pixel for High-Fidelity Geometry in 3D GANs
🏠project - Diversity-aware Channel Pruning for StyleGAN Compression
⭐code - Adversarial Score Distillation: When score distillation meets GAN
⭐code
🏠project
- StyLitGAN: Image-Based Relighting via Latent Control
- 扩散
- Fixed Point Diffusion Models
🏠project - Diffusion Models Without Attention
- Image Neural Field Diffusion Models
- Functional Diffusion
🏠project - Smooth Diffusion: Crafting Smooth Latent Spaces in Diffusion Models
- Learned Representation-Guided Diffusion Models for Large-Image Generation
- ACT-Diffusion: Efficient Adversarial Consistency Training for One-step Diffusion Models
- Using Human Feedback to Fine-tune Diffusion Models without Any Reward Model
- LightIt: Illumination Modeling and Control for Diffusion Models
- Towards More Accurate Diffusion Model Acceleration with A Timestep Tuner
- MMA-Diffusion: MultiModal Attack on Diffusion Models
- CommonCanvas: Open Diffusion Models Trained on Creative-Commons Images
- Can Protective Perturbation Safeguard Personal Data from Being Exploited by Stable Diffusion?
- Self-correcting LLM-controlled Diffusion Models
- Attention-Driven Training-Free Efficiency Enhancement of Diffusion Models
- SODA: Bottleneck Diffusion Models for Representation Learning
- PRDP: Proximal Reward Difference Prediction for Large-Scale Reward Finetuning of Diffusion Models
- Don't drop your samples! Coherence-aware training benefits Conditional diffusion
🏠project - Improving Training Efficiency of Diffusion Models via Multi-Stage Framework and Tailored Multi-Decoder Architecture
- DiffLoc: Diffusion Model for Outdoor LiDAR Localization
👍摘要 - EasyDrag: Efficient Point-based Manipulation on Diffusion Models
- Distilling ODE Solvers of Diffusion Models into Smaller Steps
- Cache Me if You Can: Accelerating Diffusion Models through Block Caching
🏠project - Beyond Textual Constraints: Learning Novel Diffusion Conditions with Fewer Examples
- AAMDM: Accelerated Auto-regressive Motion Diffusion Model
- DeepCache: Accelerating Diffusion Models for Free
🏠project - Diffusion Model Alignment Using Direct Preference Optimization
- Perturbing Attention Gives You More Bang for the Buck: Subtle Imaging Perturbations That Efficiently Fool Customized Diffusion Models
- Analyzing and Improving the Training Dynamics of Diffusion Models
- Residual Learning in Diffusion Models
- FreeU: Free Lunch in Diffusion U-Net
⭐code
🏠project - VideoCrafter2: Overcoming Data Limitations for High-Quality Video Diffusion Models
⭐code
🏠project - Diff-BGM: A Diffusion Model for Video Background Music Generation视频背景音乐生成的扩散模型
- Your Student is Better Than Expected: Adaptive Teacher-Student Collaboration for Text-Conditional Diffusion Models
- Shadow Generation for Composite Image Using Diffusion Model
⭐code - Alchemist: Parametric Control of Material Properties with Diffusion Models
- Orthogonal Adaptation for Modular Customization of Diffusion Models
🏠project扩散模型 - Observation-Guided Diffusion Probabilistic Models
⭐code - TFMQ-DM: Temporal Feature Maintenance Quantization for Diffusion Models
⭐code - Visual Anagrams: Generating Multi-View Optical Illusions with Diffusion Models
- SPAD: Spatially Aware Multi-View Diffusers
🏠project - Structure-Guided Adversarial Training of Diffusion Models
- One-step Diffusion with Distribution Matching Distillation
🏠project - Rethinking the Spatial Inconsistency in Classifier-Free Diffusion Guidance
⭐code - Intelligent Grimm -- Open-ended Visual Storytelling via Latent Diffusion Models
⭐code
🏠project - X-Adapter: Adding Universal Compatibility of Plugins for Upgraded Diffusion Model
🏠project - Readout Guidance: Learning Control from Diffusion Features
🏠project - PointInfinity: Resolution-Invariant Point Diffusion Models
🏠project - Unsupervised Keypoints from Pretrained Diffusion Models
⭐code - Amodal Completion via Progressive Mixed Context Diffusion
⭐code
🏠project - SkillDiffuser: Interpretable Hierarchical Planning via Skill Abstractions in Diffusion-Based Task Execution
🏠project - DREAM: Diffusion Rectification and Estimation-Adaptive Models
- Towards Memorization-Free Diffusion Models
- Efficient Dataset Distillation via Minimax Diffusion
⭐code - MatFuse: Controllable Material Generation with Diffusion Models
⭐code
🏠project - Accelerating Diffusion Sampling with Optimized Time Steps
- Boosting Diffusion Models with Moving Average Sampling in Frequency Domain
- One-dimensional Adapter to Rule Them All: Concepts, Diffusion Models and Erasing Applications
⭐code
🏠project - Balancing Act: Distribution-Guided Debiasing in Diffusion Models
⭐code - Shadow Generation for Composite Image Using Diffusion model
⭐code - MACE: Mass Concept Erasure in Diffusion Models
⭐code - DistriFusion: Distributed Parallel Inference for High-Resolution Diffusion Models
🏠project
🏠project
⭐code - Tackling the Singularities at the Endpoints of Time Intervals in Diffusion Models
⭐code
🏠project - DEADiff: An Efficient Stylization Diffusion Model with Disentangled Representations
⭐code
🏠project - SVGDreamer: Text Guided SVG Generation with Diffusion Model
⭐code
🏠project
👍SVGDreamer: 北航&港大发布全新文本引导的矢量图形可微渲染方法 - Relation Rectification in Diffusion Model
⭐code
🏠project
- Fixed Point Diffusion Models
- 图像合成/生成
- 图像合成
- One-Shot Structure-Aware Stylized Image Synthesis
- AnyScene: Customized Image Synthesis with Composited Foreground
- Improving Subject-Driven Image Synthesis with Subject-Agnostic Guidance
- ViewFusion: Towards Multi-View Consistency via Interpolated Denoising
⭐code - PLACE: Adaptive Layout-Semantic Fusion for Semantic Image Synthesis
- Diff-Plugin: Revitalizing Details for Diffusion-based Low-level Tasks
- Rethinking the Objectives of Vector-Quantized Tokenizers for Image Synthesis
- Unmixing Before Fusion: A Generalized Paradigm for Multi-Source-based Hyperspectral Image Synthesis
- [Unlocking Pretrained Image Backbones for Semantic Image Synthesis]
- Unlocking Pre-trained Image Backbones for Semantic Image Synthesis
- Revisiting Non-Autoregressive Transformers for Efficient Image Synthesis
- 场景-文本图像合成
- 图像生成
- [ElasticDiffusion: Training-free Arbitrary Size Image Generation]
- ElasticDiffusion: Training-free Arbitrary Size Image Generation through Global-Local Content Separation
🏠project - SCoFT: Self-Contrastive Fine-Tuning for Equitable Image Generation
- AnyDoor: Zero-shot Object-level Image Customization
🏠project - Taming Stable Diffusion for Text to 360 Panorama Image Generation
⭐code
⭐code - Active Open-Vocabulary Recognition: Let Intelligent Moving Mitigate CLIP Limitations
- Generative Image Dynamics
🏠project - Clockwork Diffusion: Efficient Generation With Model-Step Distillation
- UniGS: Unified Representation for Image Generation and Segmentation
⭐code图像生成 - Exact Fusion via Feature Distribution Matching for Few-shot Image Generation
- FreeCustom: Tuning-Free Customized Image Generation for Multi-Concept Composition
⭐code
🏠project - Adversarial Text to Continuous Image Generation
- Style Aligned Image Generation via Shared Attention
🏠project - CoDi: Conditional Diffusion Distillation for Higher-Fidelity and Faster Image Generation
- Instruct-Imagen: Image Generation with Multi-modal Instruction
- InstanceDiffusion: Instance-level Control for Image Generation
⭐code
🏠project - DemoFusion: Democratising High-Resolution Image Generation With No $$$
⭐code
🏠project - ViewDiff: 3D-Consistent Image Generation with Text-to-Image Models
⭐code
🏠project
⭐code - When StyleGAN Meets Stable Diffusion:a W+ Adapter for Personalized Image Generation
⭐code
🏠project - Correcting Diffusion Generation through Resampling
⭐code - Arbitrary-Scale Image Generation and Upsampling using Latent Diffusion Model and Implicit Neural Decoder
- Condition-Aware Neural Network for Controlled Image Generation
- A Unified and Interpretable Emotion Representation and Expression Generation
⭐code - Rethinking FID: Towards a Better Evaluation Metric for Image Generation
- 主题驱动的图像生成
- 文本-图像
- Zero-Painter: Training-Free Layout Control for Text-to-Image Synthesis
- Learning Multi-Dimensional Human Preference for Text-to-Image Generation
- Customization Assistant for Text-to-Image Generation
- TokenCompose: Text-to-Image Diffusion with Token-level Supervision
- FreeControl: Training-Free Spatial Control of Any Text-to-Image Diffusion Model with Any Condition
- Personalized Residuals for Concept-Driven Text-to-Image Generation
🏠project - Rich Human Feedback for Text-to-Image Generation
- MarkovGen: Structured Prediction for Efficient Text-to-Image Generation
- Intriguing Properties of Diffusion Models: An Empirical Study of the Natural Attack Capability in Text-to-Image Generative Models
- Customization Assistant for Text-to-image Generation
- SwiftBrush: One-Step Text-to-Image Diffusion Model with Variational Score Distillation
- JeDi: Joint-Image Diffusion Models for Finetuning-Free Personalized Text-to-Image Generation
- MIGC: Multi-Instance Generation Controller for Text-to-Image Synthesis
🏠project - Check, Locate, Rectify: A Training-Free Layout Calibration System for Text-to-Image Generation
🏠project - Predicated Diffusion: Predicate Logic-Based Attention Guidance for Text-to-Image Diffusion Models
- DreamMatcher: Appearance Matching Self-Attention for Semantically-Consistent Text-to-Image Personalization
🏠project - UFOGen: You Forward Once Large Scale Text-to-Image Generation via Diffusion GANs
- Prompting Hard or Hardly Prompting: Prompt Inversion for Text-to-Image Diffusion Models
- Countering Personalized Text-to-Image Generation with Influence Watermarks
- Ranni: Taming Text-to-Image Diffusion for Accurate Instruction Following
🏠project - Tailored Visions: Enhancing Text-to-Image Generation with Personalized Prompt Rewriting
⭐code - InstantBooth: Personalized Text-to-Image Generation without Test-Time Finetuning
- FakeInversion: Learning to Detect Images from Unseen Text-to-Image Models by Inverting Stable Diffusion
- Learning Disentangled Identifiers for Action-Customized Text-to-Image Generation
🏠project - LeftRefill: Filling Right Canvas based on Left Reference through Generalized Text-to-Image Diffusion Model
⭐code
🏠project - HyperDreamBooth: HyperNetworks for Fast Personalization of Text-to-Image Models
🏠project - PIA: Your Personalized Image Animator via Plug-and-Play Modules in Text-to-Image Models
🏠project - On the Scalability of Diffusion-based Text-to-Image Generation
- Self-Discovering Interpretable Diffusion Latent Directions for Responsible Text-to-Image Generation
🏠project - EmoGen: Emotional Image Content Generation with Text-to-Image Diffusion Models
⭐code - Grounded Text-to-Image Synthesis with Attention Refocusing
🏠project - OpenBias: Open-set Bias Detection in Text-to-Image Generative Models
⭐code - Prompt-Free Diffusion: Taking “Text” out of Text-to-Image Diffusion Models
⭐code - CONFORM: Contrast is All You Need for High-Fidelity Text-to-Image Diffusion Models
- InitNO: Boosting Text-to-Image Diffusion Models via Initial Noise Optimization
⭐code - Concept Weaver: Enabling Multi-Concept Fusion in Text-to-Image Models
- Cross Initialization for Face Personalization of Text-to-Image Models文本到图像Cross Initialization for Personalized Text-to-Image Generation
- CosmicMan: A Text-to-Image Foundation Model for Humans
⭐code - Dynamic Prompt Optimizing for Text-to-Image Generation
⭐code - WOUAF: Weight Modulation for User Attribution and Fingerprinting in Text-to-Image Diffusion Models
- Attention Calibration for Disentangled Text-to-Image Personalization
- RealCustom: Narrowing Real Text Word for Real-Time Open-Domain Text-to-Image Customization
⭐code - InteractDiffusion: Interaction Control in Text-to-Image Diffusion Models
⭐code
🏠project - Learning Continuous 3D Words for Text-to-Image Generation
⭐code
🏠project - NoiseCollage: A Layout-Aware Text-to-Image Diffusion Model Based on Noise Cropping and Merging
⭐code - HanDiffuser: Text-to-Image Generation With Realistic Hand Appearances
🏠project - Discriminative Probing and Tuning for Text-to-Image Generation
⭐code
🏠project - Selectively Informative Description can Reduce Undesired Embedding Entanglements in Text-to-Image Personalization
- ECLIPSE: A Resource-Efficient Text-to-Image Prior for Image Generations
⭐code
🏠project - FlashEval: Towards Fast and Accurate Evaluation of Text-to-image Diffusion Generative Models
⭐code - MetaCloak: Preventing Unauthorized Subject-driven Text-to-image Diffusion-based Synthesis via Meta-learning
- 主题-图像
- High-fidelity Person-centric Subject-to-Image Synthesis
⭐code - [High Fidelity Person-centric Subject-to-Image Synthesis]
- High-fidelity Person-centric Subject-to-Image Synthesis
- 图像合成
- 视频合成/生成
- 视频生成
- InstructVideo: Instructing Video Diffusion Models with Human Feedback
⭐code
🏠project - Make Pixels Dance: High-Dynamic Video Generation
- GenTron: Diffusion Transformers for Image and Video Generation
- Panacea: Panoramic and Controllable Video Generation for Autonomous Driving
- Generative Rendering: Controllable 4D-Guided Video Generation with 2D Diffusion Models
🏠project - DiffPerformer: Iterative Learning of Consistent Latent Guidance for Diffusion-based Human Video Generation
- VideoBooth: Diffusion-based Video Generation with Image Prompts
🏠project - Hierarchical Patch Diffusion Models for High-Resolution Video Generation
⭐code - On the Content Bias in Fréchet Video Distance
⭐code - 360DVD: Controllable Panorama Video Generation with 360-Degree Video Diffusion Model
- SimDA: Simple Diffusion Adapter for Efficient Video Generation
⭐code
🏠project - GenHowTo: Learning to Generate Actions and State Transformations from Instructional Videos视频生成
- FRESCO: Spatial-Temporal Correspondence for Zero-Shot Video Translation
🏠project
⭐code - Vlogger: Make Your Dream A Vlog
⭐code
🏠project - LAMP: Learn A Motion Pattern for Few-Shot Video Generation
🏠project - EvalCrafter: Benchmarking and Evaluating Large Video Generation Models
🏠project - Co-Speech Gesture Video Generation via Motion-Decoupled Diffusion Model
⭐code - BIVDiff: A Training-free Framework for General-Purpose Video Synthesis via Bridging Image and Video Diffusion Models
⭐code
🏠project视频合成 - DreamVideo: Composing Your Dream Videos with Customized Subject and Motion
🏠project - PEEKABOO: Interactive Video Generation via Masked-Diffusion
🏠project
- InstructVideo: Instructing Video Diffusion Models with Human Feedback
- 文本-视频
- Grid Diffusion Models for Text-to-Video Generation
- Breathing Life Into Sketches Using Text-to-Video Priors
⭐code
🏠project - Dysen-VDM: Empowering Dynamics-aware Text-to-Video Diffusion with LLMs
- TI2V-Zero: Zero-Shot Image Conditioning for Text-to-Video Diffusion Models
🏠project - Animate Anyone: Consistent and Controllable Image-to-Video Synthesis for Character Animation
- Hierarchical Spatio-temporal Decoupling for Text-to-Video Generation
🏠project - A Recipe for Scaling up Text-to-Video Generation with Text-free Videos
🏠project - TRIP: Temporal Residual Learning with Image Noise Prior for Image-to-Video Diffusion Models
⭐code - Snap Video: Scaled Spatiotemporal Transformers for Text-to-Video Synthesis
🏠project - VMC: Video Motion Customization using Temporal Attention Adaption for Text-to-Video Diffusion Models
🏠project - MicroCinema: A Divide-and-Conquer Approach for Text-to-Video Generation
- 图像-视频
- 视频-视频
- 视频生成
- 纹理生成/合成
- 文本-纹理合成
- 纹理合成
- 文本-3D
- DIRECT-3D: Learning Direct Text-to-3D Generation on Massive Noisy 3D Data
⭐code
🏠project - PI3D: Efficient Text-to-3D Generation with Pseudo-Image Diffusion
- Text-to-3D using Gaussian Splatting
⭐code
🏠project - DreamPropeller: Supercharge Text-to-3D Generation with Parallel Sampling
- RichDreamer: A Generalizable Normal-Depth Diffusion Model for Detail Richness in Text-to-3D
- Consistent3D: Towards Consistent High-Fidelity Text-to-3D Generation with Deterministic Sampling Prior
- Sculpt3D: Multi-View Consistent Text-to-3D Generation with Sparse 3D Prior
⭐code - LucidDreamer: Towards High-Fidelity Text-to-3D Generation via Interval Score Matching
⭐code - Direct2.5: Diverse Text-to-3D Generation via Multi-view 2.5D Diffusion
🏠project - Sherpa3D: Boosting High-Fidelity Text-to-3D Generation via Coarse 3D Prior
🏠project文本到 3D - Taming Mode Collapse in Score Distillation for Text-to-3D Generation
🏠project - Text-to-3D Generation with Bidirectional Diffusion using both 2D and 3D priors
🏠project - DreamControl: Control-Based Text-to-3D Generation with 3D Self-Prior
⭐code - VP3D: Unleashing 2D Visual Prompt for Text-to-3D Generation
⭐code - GPT-4V(ision) is a Human-Aligned Evaluator for Text-to-3D Generation
⭐code
🏠project - Enhancing 3D Fidelity of Text-to-3D using Cross-View Correspondences
- DiffusionGAN3D: Boosting Text-guided 3D Generation and Domain Adaptation by Combining 3D GANs and Diffusion Priors
🏠project - HyperSDFusion: Bridging Hierarchical Structures in Language and Geometry for Enhanced 3D Text2Shape Generation
- Diffusion Handles: Enabling 3D Edits for Diffusion Models by Lifting Activations to 3D
🏠project - HarmonyView: Harmonizing Consistency and Diversity in One-Image-to-3D
🏠project
- DIRECT-3D: Learning Direct Text-to-3D Generation on Massive Noisy 3D Data
- 图像-3D
- 文本-4D
- 3D生成
- DreamComposer: Controllable 3D Object Generation via Multi-View Conditions
🏠project - XCube (X3): Large-Scale 3D Generative Modeling using Sparse Voxel Hierarchies
🏠project - CAD: Photorealistic 3D Generation via Adversarial Distillation
⭐code
🏠project - Paint3D: Paint Anything 3D with Lighting-Less Texture Diffusion Models
⭐code3D 内容 - Interactive3D: Create What You Want by Interactive 3D Generation
- DreamComposer: Controllable 3D Object Generation via Multi-View Conditions
- 语义场景生成
- 场景补全
- Unleashing Network Potentials for Semantic Scene Completion
- Not All Voxels Are Equal: Hardness-Aware Semantic Scene Completion with Self-Distillation
⭐code - Symphonize 3D Semantic Scene Completion with Contextual Instance Queries
⭐code3D 语义 - PaSCo: Urban 3D Panoptic Scene Completion with Uncertainty Awareness
- Scaling Diffusion Models to Real-World 3D LiDAR Scene Completion
- 图像-图像翻译
- 图像检测
- 图像编辑
- Focus on Your Instruction: Fine-grained and Multi-instruction Image Editing by Attention Modulation
⭐code - Emu Edit: Precise Image Editing via Recognition and Generation Tasks
- An Edit Friendly DDPM Noise Space: Inversion and Manipulations
- Distraction is All You Need: Memory-Efficient Image Immunization against Diffusion-Based Image Editing
- DiffEditor: Boosting Accuracy and Flexibility on Diffusion-based Image Editing
⭐code - DiffMorpher: Unleashing the Capability of Diffusion Models for Image Morphing
⭐code
🏠project - UniHuman: A Unified Model For Editing Human Images in the Wild
⭐code - Contrastive Denoising Score for Text-guided Latent Diffusion Image Editing
🏠project - Inversion-Free Image Editing with Language-Guided Diffusion Models
⭐code
🏠project - TiNO-Edit: Timestep and Noise Optimization for Robust Diffusion-Based Image Editing
⭐code - Edit One for All: Interactive Batch Image Editing
⭐code
🏠project - SCEdit: Efficient and Controllable Image Diffusion Generation via Skip Connection Editing
🏠project - On Exact Inversion of DPM-Solvers
⭐code
🏠project - Doubly Abductive Counterfactual Inference for Text-based Image Editing
⭐code基于文本的图像编辑 - Towards Understanding Cross and Self-Attention in Stable Diffusion for Text-Guided Image Editing
- ZONE: Zero-Shot Instruction-Guided Local Editing
- HIVE: Harnessing Human Feedback for Instructional Visual Editing
- FreeDrag: Feature Dragging for Reliable Point-based Image Editing
- The Devil is in the Details: StyleFeatureEditor for Detail-Rich StyleGAN Inversion and High Quality Image Editing
- DragDiffusion: Harnessing Diffusion Models for Interactive Point-based Image Editing
⭐code - Text-Driven Image Editing via Learnable Regions
⭐code
🏠project - LEDITS++: Limitless Image Editing using Text-to-Image Models
⭐code
🏠project - SmartEdit: Exploring Complex Instruction-based Image Editing with Multimodal Large Language Models
⭐code
🏠project - Person in Place: Generating Associative Skeleton-Guidance Maps for Human-Object Interaction Image Editing
⭐code
🏠project - PAIR-Diffusion: A Comprehensive Multimodal Object-Level Image Editor
⭐code
🏠project - Referring Image Editing: Object-level Image Editing via Referring Expressions
- Prompt Augmentation for Self-supervised Text-guided Image Manipulation
- Named Entity Driven Zero-Shot Image Manipulation
- Focus on Your Instruction: Fine-grained and Multi-instruction Image Editing by Attention Modulation
- 布局生成
- Constrained Layout Generation with Factor Graphs
- SmartMask: Context Aware High-Fidelity Mask Generation for Fine-grained Object Insertion and Layout Control
- MaskPLAN: Masked Generative Layout Planning from Partial Input
⭐code - Retrieval-Augmented Layout Transformer for Content-Aware Layout Generation
⭐code
🏠project - Visual Layout Composer: Image-Vector Dual Diffusion Model for Design Layout Generation
🏠project
- 手写数学表达式
- NeRF-to-NeRF
- GenN2N: Generative NeRF2NeRF TranslationNeRF-to-NeRF
- 生成伪装图像
- 场景生成
- 交互式编辑
- 视频编辑
- CCEdit: Creative and Controllable Video Editing via Diffusion Models
🏠project
📺video - MaskINT: Video Editing via Interpolative Non-autoregressive Masked Transformers
- RAVE: Randomized Noise Shuffling for Fast and Consistent Video Editing with Diffusion Models
⭐code
🏠project - A Video is Worth 256 Bases: Spatial-Temporal Expectation-Maximization Inversion for Zero-Shot Video Editing
🏠project - Video-P2P: Video Editing with Cross-attention Control
🏠project - VidToMe: Video Token Merging for Zero-Shot Video Editing
🏠project - Video Interpolation with Diffusion Models
⭐code - MotionEditor: Editing Video Motion via Content-Aware Diffusion
- CAMEL: CAusal Motion Enhancement Tailored for Lifting Text-driven Video Editing
⭐code - DynVideo-E: Harnessing Dynamic NeRF for Large-Scale Motion- and View-Change Human-Centric Video Editing
🏠project
- CCEdit: Creative and Controllable Video Editing via Diffusion Models
- 漫画生成
- 文本驱动 3D 风格化
- Image Warping
- 图像重建
- 图像拼接
- 姿势引导的人体图像合成
- 文本引导的人体图像合成
- 文本图像对齐
- 基于文本的图像色调调整
- 图像矢量化
- 文本-矢量
- 矢量字体
- 矢量图形合成
- 二维码生成
- 背景替换