|
23 | 23 | title: Accelerate inference of text-to-image diffusion models
|
24 | 24 | title: Tutorials
|
25 | 25 | - sections:
|
| 26 | + - local: using-diffusers/loading |
| 27 | + title: Load pipelines |
| 28 | + - local: using-diffusers/custom_pipeline_overview |
| 29 | + title: Load community pipelines and components |
| 30 | + - local: using-diffusers/schedulers |
| 31 | + title: Load schedulers and models |
| 32 | + - local: using-diffusers/using_safetensors |
| 33 | + title: Load safetensors |
| 34 | + - local: using-diffusers/other-formats |
| 35 | + title: Load different Stable Diffusion formats |
| 36 | + - local: using-diffusers/loading_adapters |
| 37 | + title: Load adapters |
| 38 | + - local: using-diffusers/push_to_hub |
| 39 | + title: Push files to the Hub |
| 40 | + title: Load pipelines and adapters |
| 41 | +- sections: |
| 42 | + - local: using-diffusers/unconditional_image_generation |
| 43 | + title: Unconditional image generation |
| 44 | + - local: using-diffusers/conditional_image_generation |
| 45 | + title: Text-to-image |
| 46 | + - local: using-diffusers/img2img |
| 47 | + title: Image-to-image |
| 48 | + - local: using-diffusers/inpaint |
| 49 | + title: Inpainting |
| 50 | + - local: using-diffusers/text-img2vid |
| 51 | + title: Text or image-to-video |
| 52 | + - local: using-diffusers/depth2img |
| 53 | + title: Depth-to-image |
| 54 | + title: Generative tasks |
| 55 | +- sections: |
| 56 | + - local: using-diffusers/overview_techniques |
| 57 | + title: Overview |
| 58 | + - local: training/distributed_inference |
| 59 | + title: Distributed inference with multiple GPUs |
| 60 | + - local: using-diffusers/merge_loras |
| 61 | + title: Merge LoRAs |
| 62 | + - local: using-diffusers/callback |
| 63 | + title: Pipeline callbacks |
| 64 | + - local: using-diffusers/reusing_seeds |
| 65 | + title: Reproducible pipelines |
| 66 | + - local: using-diffusers/image_quality |
| 67 | + title: Controlling image quality |
| 68 | + - local: using-diffusers/weighted_prompts |
| 69 | + title: Prompt techniques |
| 70 | + title: Inference techniques |
| 71 | +- sections: |
| 72 | + - local: using-diffusers/sdxl |
| 73 | + title: Stable Diffusion XL |
| 74 | + - local: using-diffusers/sdxl_turbo |
| 75 | + title: SDXL Turbo |
| 76 | + - local: using-diffusers/kandinsky |
| 77 | + title: Kandinsky |
| 78 | + - local: using-diffusers/ip_adapter |
| 79 | + title: IP-Adapter |
| 80 | + - local: using-diffusers/controlnet |
| 81 | + title: ControlNet |
| 82 | + - local: using-diffusers/t2i_adapter |
| 83 | + title: T2I-Adapter |
| 84 | + - local: using-diffusers/textual_inversion_inference |
| 85 | + title: Textual inversion |
| 86 | + - local: using-diffusers/shap-e |
| 87 | + title: Shap-E |
| 88 | + - local: using-diffusers/diffedit |
| 89 | + title: DiffEdit |
| 90 | + - local: using-diffusers/custom_pipeline_examples |
| 91 | + title: Community pipelines |
| 92 | + - local: using-diffusers/contribute_pipeline |
| 93 | + title: Contribute a community pipeline |
| 94 | + - local: using-diffusers/inference_with_lcm_lora |
| 95 | + title: Latent Consistency Model-LoRA |
| 96 | + - local: using-diffusers/inference_with_lcm |
| 97 | + title: Latent Consistency Model |
| 98 | + - local: using-diffusers/inference_with_tcd_lora |
| 99 | + title: Trajectory Consistency Distillation-LoRA |
| 100 | + - local: using-diffusers/svd |
| 101 | + title: Stable Video Diffusion |
| 102 | + title: Specific pipeline examples |
| 103 | +- sections: |
| 104 | + - local: training/overview |
| 105 | + title: Overview |
| 106 | + - local: training/create_dataset |
| 107 | + title: Create a dataset for training |
| 108 | + - local: training/adapt_a_model |
| 109 | + title: Adapt a model to a new task |
26 | 110 | - sections:
|
27 |
| - - local: using-diffusers/loading |
28 |
| - title: Load pipelines |
29 |
| - - local: using-diffusers/custom_pipeline_overview |
30 |
| - title: Load community pipelines and components |
31 |
| - - local: using-diffusers/schedulers |
32 |
| - title: Load schedulers and models |
33 |
| - - local: using-diffusers/using_safetensors |
34 |
| - title: Load safetensors |
35 |
| - - local: using-diffusers/other-formats |
36 |
| - title: Load different Stable Diffusion formats |
37 |
| - - local: using-diffusers/loading_adapters |
38 |
| - title: Load adapters |
39 |
| - - local: using-diffusers/push_to_hub |
40 |
| - title: Push files to the Hub |
41 |
| - title: Loading & Hub |
42 |
| - - sections: |
43 |
| - - local: using-diffusers/pipeline_overview |
44 |
| - title: Overview |
45 |
| - - local: using-diffusers/unconditional_image_generation |
| 111 | + - local: training/unconditional_training |
46 | 112 | title: Unconditional image generation
|
47 |
| - - local: using-diffusers/conditional_image_generation |
| 113 | + - local: training/text2image |
48 | 114 | title: Text-to-image
|
49 |
| - - local: using-diffusers/img2img |
50 |
| - title: Image-to-image |
51 |
| - - local: using-diffusers/inpaint |
52 |
| - title: Inpainting |
53 |
| - - local: using-diffusers/text-img2vid |
54 |
| - title: Text or image-to-video |
55 |
| - - local: using-diffusers/depth2img |
56 |
| - title: Depth-to-image |
57 |
| - title: Tasks |
58 |
| - - sections: |
59 |
| - - local: using-diffusers/textual_inversion_inference |
60 |
| - title: Textual inversion |
61 |
| - - local: using-diffusers/ip_adapter |
62 |
| - title: IP-Adapter |
63 |
| - - local: using-diffusers/merge_loras |
64 |
| - title: Merge LoRAs |
65 |
| - - local: training/distributed_inference |
66 |
| - title: Distributed inference with multiple GPUs |
67 |
| - - local: using-diffusers/reusing_seeds |
68 |
| - title: Improve image quality with deterministic generation |
69 |
| - - local: using-diffusers/control_brightness |
70 |
| - title: Control image brightness |
71 |
| - - local: using-diffusers/weighted_prompts |
72 |
| - title: Prompt techniques |
73 |
| - - local: using-diffusers/freeu |
74 |
| - title: Improve generation quality with FreeU |
75 |
| - title: Techniques |
76 |
| - - sections: |
77 |
| - - local: using-diffusers/pipeline_overview |
78 |
| - title: Overview |
79 |
| - - local: using-diffusers/sdxl |
| 115 | + - local: training/sdxl |
80 | 116 | title: Stable Diffusion XL
|
81 |
| - - local: using-diffusers/sdxl_turbo |
82 |
| - title: SDXL Turbo |
83 |
| - - local: using-diffusers/kandinsky |
84 |
| - title: Kandinsky |
85 |
| - - local: using-diffusers/controlnet |
| 117 | + - local: training/kandinsky |
| 118 | + title: Kandinsky 2.2 |
| 119 | + - local: training/wuerstchen |
| 120 | + title: Wuerstchen |
| 121 | + - local: training/controlnet |
86 | 122 | title: ControlNet
|
87 |
| - - local: using-diffusers/t2i_adapter |
88 |
| - title: T2I-Adapter |
89 |
| - - local: using-diffusers/shap-e |
90 |
| - title: Shap-E |
91 |
| - - local: using-diffusers/diffedit |
92 |
| - title: DiffEdit |
93 |
| - - local: using-diffusers/distilled_sd |
94 |
| - title: Distilled Stable Diffusion inference |
95 |
| - - local: using-diffusers/callback |
96 |
| - title: Pipeline callbacks |
97 |
| - - local: using-diffusers/reproducibility |
98 |
| - title: Create reproducible pipelines |
99 |
| - - local: using-diffusers/custom_pipeline_examples |
100 |
| - title: Community pipelines |
101 |
| - - local: using-diffusers/contribute_pipeline |
102 |
| - title: Contribute a community pipeline |
103 |
| - - local: using-diffusers/inference_with_lcm_lora |
104 |
| - title: Latent Consistency Model-LoRA |
105 |
| - - local: using-diffusers/inference_with_lcm |
106 |
| - title: Latent Consistency Model |
107 |
| - - local: using-diffusers/inference_with_tcd_lora |
108 |
| - title: Trajectory Consistency Distillation-LoRA |
109 |
| - - local: using-diffusers/svd |
110 |
| - title: Stable Video Diffusion |
111 |
| - title: Specific pipeline examples |
112 |
| - - sections: |
113 |
| - - local: training/overview |
114 |
| - title: Overview |
115 |
| - - local: training/create_dataset |
116 |
| - title: Create a dataset for training |
117 |
| - - local: training/adapt_a_model |
118 |
| - title: Adapt a model to a new task |
119 |
| - - sections: |
120 |
| - - local: training/unconditional_training |
121 |
| - title: Unconditional image generation |
122 |
| - - local: training/text2image |
123 |
| - title: Text-to-image |
124 |
| - - local: training/sdxl |
125 |
| - title: Stable Diffusion XL |
126 |
| - - local: training/kandinsky |
127 |
| - title: Kandinsky 2.2 |
128 |
| - - local: training/wuerstchen |
129 |
| - title: Wuerstchen |
130 |
| - - local: training/controlnet |
131 |
| - title: ControlNet |
132 |
| - - local: training/t2i_adapters |
133 |
| - title: T2I-Adapters |
134 |
| - - local: training/instructpix2pix |
135 |
| - title: InstructPix2Pix |
136 |
| - title: Models |
137 |
| - - sections: |
138 |
| - - local: training/text_inversion |
139 |
| - title: Textual Inversion |
140 |
| - - local: training/dreambooth |
141 |
| - title: DreamBooth |
142 |
| - - local: training/lora |
143 |
| - title: LoRA |
144 |
| - - local: training/custom_diffusion |
145 |
| - title: Custom Diffusion |
146 |
| - - local: training/lcm_distill |
147 |
| - title: Latent Consistency Distillation |
148 |
| - - local: training/ddpo |
149 |
| - title: Reinforcement learning training with DDPO |
150 |
| - title: Methods |
151 |
| - title: Training |
| 123 | + - local: training/t2i_adapters |
| 124 | + title: T2I-Adapters |
| 125 | + - local: training/instructpix2pix |
| 126 | + title: InstructPix2Pix |
| 127 | + title: Models |
| 128 | + isExpanded: false |
152 | 129 | - sections:
|
153 |
| - - local: using-diffusers/other-modalities |
154 |
| - title: Other Modalities |
155 |
| - title: Taking Diffusers Beyond Images |
156 |
| - title: Using Diffusers |
| 130 | + - local: training/text_inversion |
| 131 | + title: Textual Inversion |
| 132 | + - local: training/dreambooth |
| 133 | + title: DreamBooth |
| 134 | + - local: training/lora |
| 135 | + title: LoRA |
| 136 | + - local: training/custom_diffusion |
| 137 | + title: Custom Diffusion |
| 138 | + - local: training/lcm_distill |
| 139 | + title: Latent Consistency Distillation |
| 140 | + - local: training/ddpo |
| 141 | + title: Reinforcement learning training with DDPO |
| 142 | + title: Methods |
| 143 | + isExpanded: false |
| 144 | + title: Training |
157 | 145 | - sections:
|
158 |
| - - local: optimization/opt_overview |
159 |
| - title: Overview |
160 |
| - - sections: |
161 |
| - - local: optimization/fp16 |
162 |
| - title: Speed up inference |
163 |
| - - local: optimization/memory |
164 |
| - title: Reduce memory usage |
165 |
| - - local: optimization/torch2.0 |
166 |
| - title: PyTorch 2.0 |
167 |
| - - local: optimization/xformers |
168 |
| - title: xFormers |
169 |
| - - local: optimization/tome |
170 |
| - title: Token merging |
171 |
| - - local: optimization/deepcache |
172 |
| - title: DeepCache |
173 |
| - - local: optimization/tgate |
174 |
| - title: TGATE |
175 |
| - title: General optimizations |
| 146 | + - local: optimization/fp16 |
| 147 | + title: Speed up inference |
| 148 | + - local: using-diffusers/distilled_sd |
| 149 | + title: Distilled Stable Diffusion inference |
| 150 | + - local: optimization/memory |
| 151 | + title: Reduce memory usage |
| 152 | + - local: optimization/torch2.0 |
| 153 | + title: PyTorch 2.0 |
| 154 | + - local: optimization/xformers |
| 155 | + title: xFormers |
| 156 | + - local: optimization/tome |
| 157 | + title: Token merging |
| 158 | + - local: optimization/deepcache |
| 159 | + title: DeepCache |
| 160 | + - local: optimization/tgate |
| 161 | + title: TGATE |
176 | 162 | - sections:
|
177 | 163 | - local: using-diffusers/stable_diffusion_jax_how_to
|
178 | 164 | title: JAX/Flax
|
|
182 | 168 | title: OpenVINO
|
183 | 169 | - local: optimization/coreml
|
184 | 170 | title: Core ML
|
185 |
| - title: Optimized model types |
| 171 | + title: Optimized model formats |
186 | 172 | - sections:
|
187 | 173 | - local: optimization/mps
|
188 | 174 | title: Metal Performance Shaders (MPS)
|
189 | 175 | - local: optimization/habana
|
190 | 176 | title: Habana Gaudi
|
191 | 177 | title: Optimized hardware
|
192 |
| - title: Optimization |
| 178 | + title: Accelerate inference and reduce memory |
193 | 179 | - sections:
|
194 | 180 | - local: conceptual/philosophy
|
195 | 181 | title: Philosophy
|
|
211 | 197 | - local: api/outputs
|
212 | 198 | title: Outputs
|
213 | 199 | title: Main Classes
|
| 200 | + isExpanded: false |
214 | 201 | - sections:
|
215 | 202 | - local: api/loaders/ip_adapter
|
216 | 203 | title: IP-Adapter
|
|
225 | 212 | - local: api/loaders/peft
|
226 | 213 | title: PEFT
|
227 | 214 | title: Loaders
|
| 215 | + isExpanded: false |
228 | 216 | - sections:
|
229 | 217 | - local: api/models/overview
|
230 | 218 | title: Overview
|
|
259 | 247 | - local: api/models/controlnet
|
260 | 248 | title: ControlNet
|
261 | 249 | title: Models
|
| 250 | + isExpanded: false |
262 | 251 | - sections:
|
263 | 252 | - local: api/pipelines/overview
|
264 | 253 | title: Overview
|
|
383 | 372 | - local: api/pipelines/wuerstchen
|
384 | 373 | title: Wuerstchen
|
385 | 374 | title: Pipelines
|
| 375 | + isExpanded: false |
386 | 376 | - sections:
|
387 | 377 | - local: api/schedulers/overview
|
388 | 378 | title: Overview
|
|
443 | 433 | - local: api/schedulers/vq_diffusion
|
444 | 434 | title: VQDiffusionScheduler
|
445 | 435 | title: Schedulers
|
| 436 | + isExpanded: false |
446 | 437 | - sections:
|
447 | 438 | - local: api/internal_classes_overview
|
448 | 439 | title: Overview
|
|
457 | 448 | - local: api/image_processor
|
458 | 449 | title: VAE Image Processor
|
459 | 450 | title: Internal classes
|
| 451 | + isExpanded: false |
460 | 452 | title: API
|
0 commit comments