Skip to content

Commit 14b3f9d

Browse files
author
yiyixuxu
committed
Merge remote-tracking branch 'origin/main' into kandinsky
2 parents 6c95524 + 36f43ea commit 14b3f9d

File tree

115 files changed

+6342
-902
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

115 files changed

+6342
-902
lines changed

.github/workflows/build_documentation.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ on:
55
branches:
66
- main
77
- doc-builder*
8-
- v*-release
8+
- v*-patch
99

1010
jobs:
1111
build:

docs/source/en/_toctree.yml

Lines changed: 7 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -60,6 +60,8 @@
6060
- sections:
6161
- local: training/overview
6262
title: Overview
63+
- local: training/create_dataset
64+
title: Create a dataset for training
6365
- local: training/unconditional_training
6466
title: Unconditional image generation
6567
- local: training/text_inversion
@@ -171,7 +173,7 @@
171173
- local: api/pipelines/semantic_stable_diffusion
172174
title: Semantic Guidance
173175
- local: api/pipelines/spectrogram_diffusion
174-
title: "Spectrogram Diffusion"
176+
title: Spectrogram Diffusion
175177
- sections:
176178
- local: api/pipelines/stable_diffusion/overview
177179
title: Overview
@@ -203,6 +205,8 @@
203205
title: Text-to-Image Generation with ControlNet Conditioning
204206
- local: api/pipelines/stable_diffusion/model_editing
205207
title: Text-to-Image Model Editing
208+
- local: api/pipelines/stable_diffusion/diffedit
209+
title: DiffEdit
206210
title: Stable Diffusion
207211
- local: api/pipelines/stable_diffusion_2
208212
title: Stable Diffusion 2
@@ -238,6 +242,8 @@
238242
title: DPM Discrete Scheduler
239243
- local: api/schedulers/dpm_discrete_ancestral
240244
title: DPM Discrete Scheduler with ancestral sampling
245+
- local: api/schedulers/dpm_sde
246+
title: DPMSolverSDEScheduler
241247
- local: api/schedulers/euler_ancestral
242248
title: Euler Ancestral Scheduler
243249
- local: api/schedulers/euler

docs/source/en/api/logging.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -61,7 +61,7 @@ verbose to the most verbose), those levels (with their corresponding int values
6161
critical errors.
6262
- `diffusers.logging.ERROR` (int value, 40): only report errors.
6363
- `diffusers.logging.WARNING` or `diffusers.logging.WARN` (int value, 30): only reports error and
64-
warnings. This the default level used by the library.
64+
warnings. This is the default level used by the library.
6565
- `diffusers.logging.INFO` (int value, 20): reports error, warnings and basic information.
6666
- `diffusers.logging.DEBUG` (int value, 10): report all information.
6767

docs/source/en/api/pipelines/if.mdx

Lines changed: 16 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@ Our work underscores the potential of larger UNet architectures in the first sta
2929

3030
Before you can use IF, you need to accept its usage conditions. To do so:
3131
1. Make sure to have a [Hugging Face account](https://huggingface.co/join) and be logged in
32-
2. Accept the license on the model card of [DeepFloyd/IF-I-IF-v1.0](https://huggingface.co/DeepFloyd/IF-I-IF-v1.0). Accepting the license on the stage I model card will auto accept for the other IF models.
32+
2. Accept the license on the model card of [DeepFloyd/IF-I-XL-v1.0](https://huggingface.co/DeepFloyd/IF-I-XL-v1.0). Accepting the license on the stage I model card will auto accept for the other IF models.
3333
3. Make sure to login locally. Install `huggingface_hub`
3434
```sh
3535
pip install huggingface_hub --upgrade
@@ -62,7 +62,7 @@ The following sections give more in-detail examples of how to use IF. Specifical
6262

6363
**Available checkpoints**
6464
- *Stage-1*
65-
- [DeepFloyd/IF-I-IF-v1.0](https://huggingface.co/DeepFloyd/IF-I-IF-v1.0)
65+
- [DeepFloyd/IF-I-XL-v1.0](https://huggingface.co/DeepFloyd/IF-I-XL-v1.0)
6666
- [DeepFloyd/IF-I-L-v1.0](https://huggingface.co/DeepFloyd/IF-I-L-v1.0)
6767
- [DeepFloyd/IF-I-M-v1.0](https://huggingface.co/DeepFloyd/IF-I-M-v1.0)
6868

@@ -90,7 +90,7 @@ from diffusers.utils import pt_to_pil
9090
import torch
9191

9292
# stage 1
93-
stage_1 = DiffusionPipeline.from_pretrained("DeepFloyd/IF-I-IF-v1.0", variant="fp16", torch_dtype=torch.float16)
93+
stage_1 = DiffusionPipeline.from_pretrained("DeepFloyd/IF-I-XL-v1.0", variant="fp16", torch_dtype=torch.float16)
9494
stage_1.enable_model_cpu_offload()
9595

9696
# stage 2
@@ -162,7 +162,7 @@ original_image = Image.open(BytesIO(response.content)).convert("RGB")
162162
original_image = original_image.resize((768, 512))
163163

164164
# stage 1
165-
stage_1 = IFImg2ImgPipeline.from_pretrained("DeepFloyd/IF-I-IF-v1.0", variant="fp16", torch_dtype=torch.float16)
165+
stage_1 = IFImg2ImgPipeline.from_pretrained("DeepFloyd/IF-I-XL-v1.0", variant="fp16", torch_dtype=torch.float16)
166166
stage_1.enable_model_cpu_offload()
167167

168168
# stage 2
@@ -244,7 +244,7 @@ mask_image = Image.open(BytesIO(response.content))
244244
mask_image = mask_image
245245

246246
# stage 1
247-
stage_1 = IFInpaintingPipeline.from_pretrained("DeepFloyd/IF-I-IF-v1.0", variant="fp16", torch_dtype=torch.float16)
247+
stage_1 = IFInpaintingPipeline.from_pretrained("DeepFloyd/IF-I-XL-v1.0", variant="fp16", torch_dtype=torch.float16)
248248
stage_1.enable_model_cpu_offload()
249249

250250
# stage 2
@@ -305,7 +305,7 @@ In addition to being loaded with `from_pretrained`, Pipelines can also be loaded
305305
```python
306306
from diffusers import IFPipeline, IFSuperResolutionPipeline
307307

308-
pipe_1 = IFPipeline.from_pretrained("DeepFloyd/IF-I-IF-v1.0")
308+
pipe_1 = IFPipeline.from_pretrained("DeepFloyd/IF-I-XL-v1.0")
309309
pipe_2 = IFSuperResolutionPipeline.from_pretrained("DeepFloyd/IF-II-L-v1.0")
310310

311311

@@ -326,7 +326,7 @@ pipe_2 = IFInpaintingSuperResolutionPipeline(**pipe_2.components)
326326
The simplest optimization to run IF faster is to move all model components to the GPU.
327327

328328
```py
329-
pipe = DiffusionPipeline.from_pretrained("DeepFloyd/IF-I-IF-v1.0", variant="fp16", torch_dtype=torch.float16)
329+
pipe = DiffusionPipeline.from_pretrained("DeepFloyd/IF-I-XL-v1.0", variant="fp16", torch_dtype=torch.float16)
330330
pipe.to("cuda")
331331
```
332332

@@ -352,7 +352,7 @@ the input image which also determines how many steps to run in the denoising pro
352352
A smaller number will vary the image less but run faster.
353353

354354
```py
355-
pipe = IFImg2ImgPipeline.from_pretrained("DeepFloyd/IF-I-IF-v1.0", variant="fp16", torch_dtype=torch.float16)
355+
pipe = IFImg2ImgPipeline.from_pretrained("DeepFloyd/IF-I-XL-v1.0", variant="fp16", torch_dtype=torch.float16)
356356
pipe.to("cuda")
357357

358358
image = pipe(image=image, prompt="<prompt>", strength=0.3).images
@@ -364,7 +364,7 @@ with IF and it might not give expected results.
364364
```py
365365
import torch
366366

367-
pipe = DiffusionPipeline.from_pretrained("DeepFloyd/IF-I-IF-v1.0", variant="fp16", torch_dtype=torch.float16)
367+
pipe = DiffusionPipeline.from_pretrained("DeepFloyd/IF-I-XL-v1.0", variant="fp16", torch_dtype=torch.float16)
368368
pipe.to("cuda")
369369

370370
pipe.text_encoder = torch.compile(pipe.text_encoder)
@@ -378,14 +378,14 @@ When optimizing for GPU memory, we can use the standard diffusers cpu offloading
378378
Either the model based CPU offloading,
379379

380380
```py
381-
pipe = DiffusionPipeline.from_pretrained("DeepFloyd/IF-I-IF-v1.0", variant="fp16", torch_dtype=torch.float16)
381+
pipe = DiffusionPipeline.from_pretrained("DeepFloyd/IF-I-XL-v1.0", variant="fp16", torch_dtype=torch.float16)
382382
pipe.enable_model_cpu_offload()
383383
```
384384

385385
or the more aggressive layer based CPU offloading.
386386

387387
```py
388-
pipe = DiffusionPipeline.from_pretrained("DeepFloyd/IF-I-IF-v1.0", variant="fp16", torch_dtype=torch.float16)
388+
pipe = DiffusionPipeline.from_pretrained("DeepFloyd/IF-I-XL-v1.0", variant="fp16", torch_dtype=torch.float16)
389389
pipe.enable_sequential_cpu_offload()
390390
```
391391

@@ -395,13 +395,13 @@ Additionally, T5 can be loaded in 8bit precision
395395
from transformers import T5EncoderModel
396396

397397
text_encoder = T5EncoderModel.from_pretrained(
398-
"DeepFloyd/IF-I-IF-v1.0", subfolder="text_encoder", device_map="auto", load_in_8bit=True, variant="8bit"
398+
"DeepFloyd/IF-I-XL-v1.0", subfolder="text_encoder", device_map="auto", load_in_8bit=True, variant="8bit"
399399
)
400400

401401
from diffusers import DiffusionPipeline
402402

403403
pipe = DiffusionPipeline.from_pretrained(
404-
"DeepFloyd/IF-I-IF-v1.0",
404+
"DeepFloyd/IF-I-XL-v1.0",
405405
text_encoder=text_encoder, # pass the previously instantiated 8bit text encoder
406406
unet=None,
407407
device_map="auto",
@@ -422,13 +422,13 @@ from transformers import T5EncoderModel
422422
from diffusers.utils import pt_to_pil
423423

424424
text_encoder = T5EncoderModel.from_pretrained(
425-
"DeepFloyd/IF-I-IF-v1.0", subfolder="text_encoder", device_map="auto", load_in_8bit=True, variant="8bit"
425+
"DeepFloyd/IF-I-XL-v1.0", subfolder="text_encoder", device_map="auto", load_in_8bit=True, variant="8bit"
426426
)
427427

428428
# text to image
429429

430430
pipe = DiffusionPipeline.from_pretrained(
431-
"DeepFloyd/IF-I-IF-v1.0",
431+
"DeepFloyd/IF-I-XL-v1.0",
432432
text_encoder=text_encoder, # pass the previously instantiated 8bit text encoder
433433
unet=None,
434434
device_map="auto",
@@ -444,7 +444,7 @@ gc.collect()
444444
torch.cuda.empty_cache()
445445

446446
pipe = IFPipeline.from_pretrained(
447-
"DeepFloyd/IF-I-IF-v1.0", text_encoder=None, variant="fp16", torch_dtype=torch.float16, device_map="auto"
447+
"DeepFloyd/IF-I-XL-v1.0", text_encoder=None, variant="fp16", torch_dtype=torch.float16, device_map="auto"
448448
)
449449

450450
generator = torch.Generator().manual_seed(0)

docs/source/en/api/pipelines/repaint.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -60,7 +60,7 @@ pipe = pipe.to("cuda")
6060

6161
generator = torch.Generator(device="cuda").manual_seed(0)
6262
output = pipe(
63-
original_image=original_image,
63+
image=original_image,
6464
mask_image=mask_image,
6565
num_inference_steps=250,
6666
eta=0.0,

0 commit comments

Comments
 (0)