diff --git a/docs/features/IMG2IMG.md b/docs/features/IMG2IMG.md
index 988caad74b..b396efa91b 100644
--- a/docs/features/IMG2IMG.md
+++ b/docs/features/IMG2IMG.md
@@ -18,13 +18,13 @@ tree on a hill with a river, nature photograph, national geographic -I./test-pic
This will take the original image shown here:
![drawing of a fireplace](../assets/img2img/fire-drawing.png)
@@ -99,13 +99,13 @@ Say I want SD to draw a fire based on this hand-drawn image:
Let's only do 10 steps, to make it easier to see what's happening. If strength is `0.7`, this is what the internal steps the algorithm has to take will look like:
-![](../assets/img2img/000032.steps.gravity.png)
+![gravity32](../assets/img2img/000032.steps.gravity.png)
With strength `0.4`, the steps look more like this:
-![](../assets/img2img/000030.steps.gravity.png)
+![gravity30](../assets/img2img/000030.steps.gravity.png)
Notice how much more fuzzy the starting image is for strength `0.7` compared to `0.4`, and notice also how much longer the sequence is with `0.7`:
@@ -139,7 +139,7 @@ invoke> "fire" -s50 -W384 -H384 -S1592514025 -I /tmp/fire-drawing.png -f 0.4
```
-![](../assets/img2img/000035.1592514025.png)
+![000035.1592514025](../assets/img2img/000035.1592514025.png)
and here is strength `0.7` (note step count `30`, which is roughly `20 รท 0.7` to make sure SD does `20` steps from my image):
@@ -149,28 +149,28 @@ invoke> "fire" -s30 -W384 -H384 -S1592514025 -I /tmp/fire-drawing.png -f 0.7
```
-![](../assets/img2img/000046.1592514025.png)
+![000046.1592514025](../assets/img2img/000046.1592514025.png)
In both cases the image is nice and clean and "finished", but because at strength `0.7` Stable Diffusion has been give so much more freedom to improve on my badly-drawn flames, they've come out looking much better. You can really see the difference when looking at the latent steps. There's more noise on the first image with strength `0.7`:
-![](../assets/img2img/000046.steps.gravity.png)
+![gravity46](../assets/img2img/000046.steps.gravity.png)
than there is for strength `0.4`:
-![](../assets/img2img/000035.steps.gravity.png)
+![gravity35](../assets/img2img/000035.steps.gravity.png)
and that extra noise gives the algorithm more choices when it is evaluating how to denoise any particular pixel in the image.
Unfortunately, it seems that `img2img` is very sensitive to the step count. Here's strength `0.7` with a step count of `29` (SD did 19 steps from my image):
-![](../assets/img2img/000045.1592514025.png)
+![gravity45](../assets/img2img/000045.1592514025.png)
By comparing the latents we can sort of see that something got interpreted differently enough on the third or fourth step to lead to a rather different interpretation of the flames.
-![](../assets/img2img/000046.steps.gravity.png)
-![](../assets/img2img/000045.steps.gravity.png)
+![gravity46](../assets/img2img/000046.steps.gravity.png)
+![gravity45](../assets/img2img/000045.steps.gravity.png)
This is the result of a difference in the de-noising "schedule" - basically the noise has to be cleaned by a certain degree each step or the model won't "converge" on the image properly (see [stable diffusion blog](https://huggingface.co/blog/stable_diffusion) for more about that). A different step count means a different schedule, which means things get interpreted slightly differently at every step.