diff --git a/README.md b/README.md index 063ecfb..87517a9 100644 --- a/README.md +++ b/README.md @@ -39,17 +39,18 @@ LoRA Models: [CivitAI](https://civitai.com/models/) Place downloaded models in the `stable-diffusion-webui/models`. +Some technical background on the latent space of an image model: +![](images/sd-latent-space.jpg) + ## Start Generating ### Image Dimensions (Resolution) -As different models are trained on different image resolutions, it is best to use the training image resolution for generations. For `SD1.5` use 512x512 and for `SDXL1.0` use 1024x1024. You can slightly vary one of the dimensions without significant issues. +As different models are trained on different image resolutions, it is best to use the training image resolution for generations. For SD1.5 use 512x512 and for SDXL1.0 use 1024x1024. You can slightly vary one of the dimensions without significant issues. ### text2img generation -`text2img` can be thought of as generating visual content based on textual descriptions. Popular models include [DALL-E](https://openai.com/dall-e-2), [Midjourney](https://www.midjourney.com/home), and [Stable Diffusion](https://stability.ai/blog/stable-diffusion-public-release). - -![](images/sd-latent-space.jpg) +`text2img` can be thought of as generating visual content based on textual descriptions. Popular models include [DALL-E](https://openai.com/dall-e-2), [Midjourney](https://www.midjourney.com/home), and [Stable Diffusion](https://stability.ai/blog/stable-diffusion-public-release). Stable Diffusion (such as SD1.5 and SDXL1.0) are open and gives us the more control on the image generation process. ### img2img generation @@ -59,5 +60,4 @@ As different models are trained on different image resolutions, it is best to us ![](images/cn-sd.png) - Install extension for A1111: [`sd-webui-controlnet`](https://github.com/Mikubill/sd-webui-controlnet) \ No newline at end of file