Upamanyu098's picture
End of training
ef4d689 verified
|
raw
history blame
12.2 kB



Diffusers

πŸ€— DiffusersλŠ” 이미지, μ˜€λ””μ˜€, 심지어 λΆ„μžμ˜ 3D ꡬ쑰λ₯Ό μƒμ„±ν•˜κΈ° μœ„ν•œ μ΅œμ²¨λ‹¨ 사전 ν›ˆλ ¨λœ diffusion λͺ¨λΈμ„ μœ„ν•œ λΌμ΄λΈŒλŸ¬λ¦¬μž…λ‹ˆλ‹€. κ°„λ‹¨ν•œ μΆ”λ‘  μ†”λ£¨μ…˜μ„ μ°Ύκ³  μžˆλ“ , 자체 diffusion λͺ¨λΈμ„ ν›ˆλ ¨ν•˜κ³  μ‹Άλ“ , πŸ€— DiffusersλŠ” 두 가지 λͺ¨λ‘λ₯Ό μ§€μ›ν•˜λŠ” λͺ¨λ“ˆμ‹ νˆ΄λ°•μŠ€μž…λ‹ˆλ‹€. 저희 λΌμ΄λΈŒλŸ¬λ¦¬λŠ” μ„±λŠ₯보닀 μ‚¬μš©μ„±, κ°„νŽΈν•¨λ³΄λ‹€ λ‹¨μˆœν•¨, 그리고 좔상화보닀 μ‚¬μš©μž 지정 κ°€λŠ₯성에 쀑점을 두고 μ„€κ³„λ˜μ—ˆμŠ΅λ‹ˆλ‹€.

이 λΌμ΄λΈŒλŸ¬λ¦¬μ—λŠ” μ„Έ 가지 μ£Όμš” ꡬ성 μš”μ†Œκ°€ μžˆμŠ΅λ‹ˆλ‹€:

  • λͺ‡ μ€„μ˜ μ½”λ“œλ§ŒμœΌλ‘œ μΆ”λ‘ ν•  수 μžˆλŠ” μ΅œμ²¨λ‹¨ diffusion νŒŒμ΄ν”„λΌμΈ.
  • 생성 속도와 ν’ˆμ§ˆ κ°„μ˜ κ· ν˜•μ„ λ§žμΆ”κΈ° μœ„ν•΄ μƒν˜Έκ΅ν™˜μ μœΌλ‘œ μ‚¬μš©ν•  수 μžˆλŠ” λ…Έμ΄μ¦ˆ μŠ€μΌ€μ€„λŸ¬.
  • λΉŒλ”© λΈ”λ‘μœΌλ‘œ μ‚¬μš©ν•  수 있고 μŠ€μΌ€μ€„λŸ¬μ™€ κ²°ν•©ν•˜μ—¬ 자체적인 end-to-end diffusion μ‹œμŠ€ν…œμ„ λ§Œλ“€ 수 μžˆλŠ” 사전 ν•™μŠ΅λœ λͺ¨λΈ.
Tutorials

결과물을 μƒμ„±ν•˜κ³ , λ‚˜λ§Œμ˜ diffusion μ‹œμŠ€ν…œμ„ κ΅¬μΆ•ν•˜κ³ , ν™•μ‚° λͺ¨λΈμ„ ν›ˆλ ¨ν•˜λŠ” 데 ν•„μš”ν•œ κΈ°λ³Έ κΈ°μˆ μ„ λ°°μ›Œλ³΄μ„Έμš”. πŸ€— Diffusersλ₯Ό 처음 μ‚¬μš©ν•˜λŠ” 경우 μ—¬κΈ°μ—μ„œ μ‹œμž‘ν•˜λŠ” 것이 μ’‹μŠ΅λ‹ˆλ‹€!

How-to guides

νŒŒμ΄ν”„λΌμΈ, λͺ¨λΈ, μŠ€μΌ€μ€„λŸ¬λ₯Ό λ‘œλ“œν•˜λŠ” 데 도움이 λ˜λŠ” μ‹€μš©μ μΈ κ°€μ΄λ“œμž…λ‹ˆλ‹€. λ˜ν•œ νŠΉμ • μž‘μ—…μ— νŒŒμ΄ν”„λΌμΈμ„ μ‚¬μš©ν•˜κ³ , 좜λ ₯ 생성 방식을 μ œμ–΄ν•˜κ³ , μΆ”λ‘  속도에 맞게 μ΅œμ ν™”ν•˜κ³ , λ‹€μ–‘ν•œ ν•™μŠ΅ 기법을 μ‚¬μš©ν•˜λŠ” 방법도 배울 수 μžˆμŠ΅λ‹ˆλ‹€.

Conceptual guides

λΌμ΄λΈŒλŸ¬λ¦¬κ°€ μ™œ 이런 λ°©μ‹μœΌλ‘œ μ„€κ³„λ˜μ—ˆλŠ”μ§€ μ΄ν•΄ν•˜κ³ , 라이브러리 μ΄μš©μ— λŒ€ν•œ 윀리적 κ°€μ΄λ“œλΌμΈκ³Ό μ•ˆμ „ κ΅¬ν˜„μ— λŒ€ν•΄ μžμ„Ένžˆ μ•Œμ•„λ³΄μ„Έμš”.

Reference

πŸ€— Diffusers 클래슀 및 λ©”μ„œλ“œμ˜ μž‘λ™ 방식에 λŒ€ν•œ 기술 μ„€λͺ….

Supported pipelines

Pipeline Paper/Repository Tasks
alt_diffusion AltCLIP: Altering the Language Encoder in CLIP for Extended Language Capabilities Image-to-Image Text-Guided Generation
audio_diffusion Audio Diffusion Unconditional Audio Generation
controlnet Adding Conditional Control to Text-to-Image Diffusion Models Image-to-Image Text-Guided Generation
cycle_diffusion Unifying Diffusion Models' Latent Space, with Applications to CycleDiffusion and Guidance Image-to-Image Text-Guided Generation
dance_diffusion Dance Diffusion Unconditional Audio Generation
ddpm Denoising Diffusion Probabilistic Models Unconditional Image Generation
ddim Denoising Diffusion Implicit Models Unconditional Image Generation
if IF Image Generation
if_img2img IF Image-to-Image Generation
if_inpainting IF Image-to-Image Generation
latent_diffusion High-Resolution Image Synthesis with Latent Diffusion Models Text-to-Image Generation
latent_diffusion High-Resolution Image Synthesis with Latent Diffusion Models Super Resolution Image-to-Image
latent_diffusion_uncond High-Resolution Image Synthesis with Latent Diffusion Models Unconditional Image Generation
paint_by_example Paint by Example: Exemplar-based Image Editing with Diffusion Models Image-Guided Image Inpainting
pndm Pseudo Numerical Methods for Diffusion Models on Manifolds Unconditional Image Generation
score_sde_ve Score-Based Generative Modeling through Stochastic Differential Equations Unconditional Image Generation
score_sde_vp Score-Based Generative Modeling through Stochastic Differential Equations Unconditional Image Generation
semantic_stable_diffusion Semantic Guidance Text-Guided Generation
stable_diffusion_text2img Stable Diffusion Text-to-Image Generation
stable_diffusion_img2img Stable Diffusion Image-to-Image Text-Guided Generation
stable_diffusion_inpaint Stable Diffusion Text-Guided Image Inpainting
stable_diffusion_panorama MultiDiffusion Text-to-Panorama Generation
stable_diffusion_pix2pix InstructPix2Pix: Learning to Follow Image Editing Instructions Text-Guided Image Editing
stable_diffusion_pix2pix_zero Zero-shot Image-to-Image Translation Text-Guided Image Editing
stable_diffusion_attend_and_excite Attend-and-Excite: Attention-Based Semantic Guidance for Text-to-Image Diffusion Models Text-to-Image Generation
stable_diffusion_self_attention_guidance Improving Sample Quality of Diffusion Models Using Self-Attention Guidance Text-to-Image Generation Unconditional Image Generation
stable_diffusion_image_variation Stable Diffusion Image Variations Image-to-Image Generation
stable_diffusion_latent_upscale Stable Diffusion Latent Upscaler Text-Guided Super Resolution Image-to-Image
stable_diffusion_model_editing Editing Implicit Assumptions in Text-to-Image Diffusion Models Text-to-Image Model Editing
stable_diffusion_2 Stable Diffusion 2 Text-to-Image Generation
stable_diffusion_2 Stable Diffusion 2 Text-Guided Image Inpainting
stable_diffusion_2 Depth-Conditional Stable Diffusion Depth-to-Image Generation
stable_diffusion_2 Stable Diffusion 2 Text-Guided Super Resolution Image-to-Image
stable_diffusion_safe Safe Stable Diffusion Text-Guided Generation
stable_unclip Stable unCLIP Text-to-Image Generation
stable_unclip Stable unCLIP Image-to-Image Text-Guided Generation
stochastic_karras_ve Elucidating the Design Space of Diffusion-Based Generative Models Unconditional Image Generation
text_to_video_sd Modelscope's Text-to-video-synthesis Model in Open Domain Text-to-Video Generation
unclip Hierarchical Text-Conditional Image Generation with CLIP Latents(implementation by kakaobrain) Text-to-Image Generation
versatile_diffusion Versatile Diffusion: Text, Images and Variations All in One Diffusion Model Text-to-Image Generation
versatile_diffusion Versatile Diffusion: Text, Images and Variations All in One Diffusion Model Image Variations Generation
versatile_diffusion Versatile Diffusion: Text, Images and Variations All in One Diffusion Model Dual Image and Text Guided Generation
vq_diffusion Vector Quantized Diffusion Model for Text-to-Image Synthesis Text-to-Image Generation