kwanyong
's Collections
Diffusion model
updated
Your Student is Better Than Expected: Adaptive Teacher-Student
Collaboration for Text-Conditional Diffusion Models
Paper
•
2312.10835
•
Published
•
6
LIME: Localized Image Editing via Attention Regularization in Diffusion
Models
Paper
•
2312.09256
•
Published
•
8
PromptBench: A Unified Library for Evaluation of Large Language Models
Paper
•
2312.07910
•
Published
•
15
Prompt Expansion for Adaptive Text-to-Image Generation
Paper
•
2312.16720
•
Published
•
5
A Recipe for Scaling up Text-to-Video Generation with Text-free Videos
Paper
•
2312.15770
•
Published
•
12
DreamTuner: Single Image is Enough for Subject-Driven Generation
Paper
•
2312.13691
•
Published
•
26
AnyV2V: A Plug-and-Play Framework For Any Video-to-Video Editing Tasks
Paper
•
2403.14468
•
Published
•
23
HyperLLaVA: Dynamic Visual and Language Expert Tuning for Multimodal
Large Language Models
Paper
•
2403.13447
•
Published
•
18
AnimateDiff-Lightning: Cross-Model Diffusion Distillation
Paper
•
2403.12706
•
Published
•
17
FouriScale: A Frequency Perspective on Training-Free High-Resolution
Image Synthesis
Paper
•
2403.12963
•
Published
•
7
LightIt: Illumination Modeling and Control for Diffusion Models
Paper
•
2403.10615
•
Published
•
16
DepthFM: Fast Monocular Depth Estimation with Flow Matching
Paper
•
2403.13788
•
Published
•
17
Video Editing via Factorized Diffusion Distillation
Paper
•
2403.09334
•
Published
•
21
ELLA: Equip Diffusion Models with LLM for Enhanced Semantic Alignment
Paper
•
2403.05135
•
Published
•
42
AtomoVideo: High Fidelity Image-to-Video Generation
Paper
•
2403.01800
•
Published
•
20
VideoElevator: Elevating Video Generation Quality with Versatile
Text-to-Image Diffusion Models
Paper
•
2403.05438
•
Published
•
18
DiffuseKronA: A Parameter Efficient Fine-tuning Method for Personalized
Diffusion Model
Paper
•
2402.17412
•
Published
•
21
Playground v2.5: Three Insights towards Enhancing Aesthetic Quality in
Text-to-Image Generation
Paper
•
2402.17245
•
Published
•
10
RealCompo: Dynamic Equilibrium between Realism and Compositionality
Improves Text-to-Image Diffusion Models
Paper
•
2402.12908
•
Published
•
9
Seeing and Hearing: Open-domain Visual-Audio Generation with Diffusion
Latent Aligners
Paper
•
2402.17723
•
Published
•
16
Getting it Right: Improving Spatial Consistency in Text-to-Image Models
Paper
•
2404.01197
•
Published
•
30
StoryDiffusion: Consistent Self-Attention for Long-Range Image and Video
Generation
Paper
•
2405.01434
•
Published
•
52
Paint by Inpaint: Learning to Add Image Objects by Removing Them First
Paper
•
2404.18212
•
Published
•
27
Customizing Text-to-Image Models with a Single Image Pair
Paper
•
2405.01536
•
Published
•
18