sw-yx / prompt-eng
- среда, 19 октября 2022 г. в 00:36:51
notes for prompt engineering
notes for prompt engineering
Table of Contents
The more advanced GPT3 reads have been split out to https://github.com/sw-yx/prompt-eng/blob/main/GPT.md
misc
stable diffusion specific notes
Required reading:
pip install imaginairy
. "just works" on Linux and macOS(M1) (and maybe windows). Memory efficiency improvements, prompt-based editing, face enhancement, upscaling, tiled images, img2img, prompt matrices, prompt variables, BLIP image captions, comes with dockerfile/colab. Has unit tests.Main Stable Diffusion repo: https://github.com/CompVis/stable-diffusion
Name/Link | Stars | Description |
---|---|---|
AUTOMATIC1111 | 9700 | The most well known fork. features: https://github.com/AUTOMATIC1111/stable-diffusion-webui#features launch announcement https://www.reddit.com/r/StableDiffusion/comments/x28a76/stable_diffusion_web_ui/. M1 mac instructions https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Installation-on-Apple-Silicon |
Disco Diffusion | 5600 | A frankensteinian amalgamation of notebooks, models and techniques for the generation of AI Art and Animations. |
sd-webui (formerly hlky fork) | 5100 | A fully-integrated and easy way to work with Stable Diffusion right from a browser window. Long list of UI and SD features (incl textual inversion, alternative samplers, prompt matrix): https://github.com/sd-webui/stable-diffusion-webui#project-features |
InvokeAI (formerly lstein fork) | 3400 | This version of Stable Diffusion features a slick WebGUI, an interactive command-line script that combines text2img and img2img functionality in a "dream bot" style interface, and multiple features and other enhancements. It runs on Windows, Mac and Linux machines, with GPU cards with as little as 4 GB of RAM. |
XavierXiao/Dreambooth-Stable-Diffusion | 2400 | Implementation of Dreambooth (https://arxiv.org/abs/2208.12242) with Stable Diffusion. Dockerized: https://github.com/smy20011/dreambooth-docker |
Basujindal: Optimized Stable Diffusion | 2100 | This repo is a modified version of the Stable Diffusion repo, optimized to use less VRAM than the original by sacrificing inference speed. img2img and txt2img and inpainting under 2.4GB VRAM |
stablediffusion-infinity | 1900 | Outpainting with Stable Diffusion on an infinite canvas. This project mainly works as a proof of concept. |
Waifu Diffusion (huggingface, replicate) | 1100 | stable diffusion finetuned on weeb stuff. "A model trained on danbooru (anime/manga drawing site with also lewds and nsfw on it) over 56k images.Produces FAR BETTER results if you're interested in getting manga and anime stuff out of stable diffusion." |
AbdBarho/stable-diffusion-webui-docker | 929 | Easy Docker setup for Stable Diffusion with both Automatic1111 and hlky UI included. HOWEVER - no mac support yet AbdBarho/stable-diffusion-webui-docker#35 |
fast-stable-diffusion | 753 | +25-50% speed increase + memory efficient + DreamBooth |
imaginAIry |
639 | Pythonic generation of stable diffusion images with just pip install imaginairy . "just works" on Linux and macOS(M1) (and maybe windows). Memory efficiency improvements, prompt-based editing, face enhancement, upscaling, tiled images, img2img, prompt matrices, prompt variables, BLIP image captions, comes with dockerfile/colab. Has unit tests. |
neonsecret/stable-diffusion | 546 | This repo is a modified version of the Stable Diffusion repo, optimized to use less VRAM than the original by sacrificing inference speed. Also I invented the sliced atttention technique, which allows to push the model's abilities even further. It works by automatically determining the slice size from your vram and image size and then allocating it one by one accordingly. You can practically generate any image size, it just depends on the generation speed you are willing to sacrifice. |
Deforum Stable Diffusion | 347 | Animating prompts with stable diffusion. replicate demo: https://replicate.com/deforum/deforum_stable_diffusion |
Doggettx/stable-diffusion | 137 | Allows to use resolutions that require up to 64x more VRAM than possible on the default CompVis build. |
Dormant projects, for historical/research interest:
environment-mac.yaml
from https://github.com/fragmede/stable-diffusion/blob/mps_consistent_seed/environment-mac.yamlUI's that dont come with their own SD distro, just shelling out to one
UI Name/Link | Stars | Self-Description |
---|---|---|
ahrm/UnstableFusion | 815 | UnstableFusion is a desktop frontend for Stable Diffusion which combines image generation, inpainting, img2img and other image editing operation into a seamless workflow. https://www.youtube.com/watch?v=XLOhizAnSfQ&t=1s |
breadthe/sd-buddy | 165 | Companion desktop app for the self-hosted M1 Mac version of Stable Diffusion, with Svelte and Tauri |
leszekhanusz/diffusion-ui | 65 | This is a web interface frontend for the generation of images using diffusion models. The goal is to provide an interface to online and offline backends doing image generation and inpainting like Stable Diffusion. |
GenerationQ | 21 | GenerationQ (for "image generation queue") is a cross-platform desktop application (screens below) designed to provide a general purpose GUI for generating images via text2img and img2img models. Its primary target is Stable Diffusion but since there is such a variety of forked programs with their own particularities, the UI for configuring image generation tasks is designed to be generic enough to accommodate just about any script (even non-SD models). |
See https://github.com/sw-yx/prompt-eng/blob/main/PROMPTS.md for more details and notes
DallE banned so SD https://twitter.com/almost_digital/status/1556216820788609025?s=20&t=GCU5prherJvKebRrv9urdw