UI tool to help you generate art (and experiment) with multimodal (text, image) AI models (stable diffusion)
Project description
Peacasso
Peacasso [Beta] is a UI tool to help you generate art (and experiment) with multimodal (text, image) AI models (stable diffusion).
Why Use Peacasso?
Because you deserve a nice UI and great workflow that makes exploring stable diffusion models fun! But seriously, here are a few things that make Peacasson interesting:
- Easy installation. Instead of cobbling together command line scripts, Peacasso provides a
pip install
flow and a UI that supports a set of curated default operations. - UI with good defaults. The current implementation of Peacasso provides a UI for basic operations - text and image based prompting, remixing generated images as prompts, model parameter selection. Also covers the little things .. like light and dark mode.
- Python API. While the UI is the focus here, there is an underlying python api which will bake in experimentation features (e.g. saving intermediate images in the sampling loop, exploring model explanations etc. . see roadmap below).
Clearly, Peacasso (UI) might not be for those interested in low level code.
Requirements and Installation
-
Step 1: HuggingFace Access
Access to the diffusion model weights requires a HuggingFace model account and access token. Please create an account at huggingface.co, get an access token and agree to the model terms here. Next, create a
HF_API_TOKEN
environment variable containing your token.export HF_API_TOKEN=your_token
. Note that the first time you run peacasso, the weights for the SD model are cached locally on your machine. -
Step 2: Verify Environment - Pythong 3.7+ and CUDA Setup and verify that your python environment is
python 3.7
or higher (preferably, use Conda). Also verify that you have CUDA installed correctly (torch.cuda.is_available()
is true) and your GPU has about 7GB of VRAM memory.
Once requirements are met, run the following command to install the library:
pip install peacasso
Usage - UI and Python API
You can use the library from the ui by running the following command:
peacasso ui --port=8080
Then navigate to http://localhost:8080/ in your browser.
You can also use the python api by running the following command:
import os
from dotenv import load_dotenv
from peacasso.generator import ImageGenerator
from peacasso.datamodel import GeneratorConfig
token = os.environ.get("HF_API_TOKEN")
gen = ImageGenerator(token=token)
prompt = "A sea lion wandering the streets of post apocalyptic London"
prompt_config = GeneratorConfig(
prompt=prompt,
num_images=3,
width=512,
height=512,
guidance_scale=7.5,
num_inference_steps=50,
mode="prompt", # prompt, image
return_intermediates=True, # return intermediate images in the generate dict response
)
result = gen.generate(prompt_config)
for i, image in enumerate(result["images"]):
image.save(f"image_{i}.png")
Features and Road Map
- Command line interface
- UI Features. Query models with multiple parametrs
- Text prompting
- Image based prompting
- Editor (for inpainting and outpainting possibly)
- Latent space exploration
- Experimentation tools
- Save intermediate images in the sampling loop
- Prompt recommendation tools
- Model explanations
- Curation/sharing experiment results
Acknowledgement
This work builds on the stable diffusion model and code is adapted from the HuggingFace implementation. Please note the - CreativeML Open RAIL-M license associated with the stable diffusion model.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Hashes for Peacasso-0.0.10a0-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | cb5f1027862e4d2042506ffac742ad5c4b138df7f9be2c16ab7ea533f10232fb |
|
MD5 | 146f1104a6ddd088069c0463e95d03a9 |
|
BLAKE2b-256 | 7a5df11ea46fcaed9eed91d7612e977184758d87380cd64bb88f4ef9871060f7 |