• Stars
    star
    511
  • Rank 86,473 (Top 2 %)
  • Language
    Jupyter Notebook
  • License
    MIT License
  • Created almost 2 years ago
  • Updated 4 months ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

A prompting enhancement library for transformers-type text embedding systems

Compel

A text prompt weighting and blending library for transformers-type text embedding systems, by @damian0815.

With a flexible and intuitive syntax, you can re-weight different parts of a prompt string and thus re-weight the different parts of the embedding tensor produced from the string.

Tested and developed against Hugging Face's StableDiffusionPipeline but it should work with any diffusers-based system that uses an Tokenizer and a Text Encoder of some kind.

Adapted from the InvokeAI prompting code (also by @damian0815). For now, the syntax is fully documented here.

Note that cross-attention control .swap() is currently ignored by Compel, but you can use it by calling build_conditioning_tensor_for_prompt_object() yourself, and implementing cross-attention control in your diffusion loop.

Installation

pip install compel

Demo

see compel-demo.ipynb

Open In Colab

Quickstart

with Hugging Face diffusers >=0.12:

from diffusers import StableDiffusionPipeline
from compel import Compel

pipeline = StableDiffusionPipeline.from_pretrained("runwayml/stable-diffusion-v1-5")
compel = Compel(tokenizer=pipeline.tokenizer, text_encoder=pipeline.text_encoder)

# upweight "ball"
prompt = "a cat playing with a ball++ in the forest"
conditioning = compel.build_conditioning_tensor(prompt)
# or: conditioning = compel([prompt])

# generate image
images = pipeline(prompt_embeds=conditioning, num_inference_steps=20).images
images[0].save("image.jpg")

For batched input, use the call interface to compel:

from diffusers import StableDiffusionPipeline
from compel import Compel

pipeline = StableDiffusionPipeline.from_pretrained("runwayml/stable-diffusion-v1-5")
compel = Compel(tokenizer=pipeline.tokenizer, text_encoder=pipeline.text_encoder)

prompts = ["a cat playing with a ball++ in the forest", "a dog playing with a ball in the forest"]
prompt_embeds = compel(prompts)
images = pipeline(prompt_embeds=prompt_embeds).images

images[0].save("image0.jpg")
images[1].save("image1.jpg")

Textual Inversion support

If you want to have access to 🤗diffusers textual inversions, instantiate a DiffusersTextualInversionManager and pass it on Compel init:

pipeline = StableDiffusionPipeline.from_pretrained("runwayml/stable-diffusion-v1-5")
textual_inversion_manager = DiffusersTextualInversionManager(pipeline)
compel = Compel(tokenizer=pipeline.tokenizer, text_encoder=pipeline.text_encoder, 
    textual_inversion_manager=textual_inversion_manager)

Memory usage/VRAM leaks

If you are using Compel heavily and repeatedly, you may run into PyTorch memory issues. To alleviate this, according to @kshieh1:

After image generation, you should explictly de-reference the tensor object (i.e., prompt_embeds = None) and call gc.collect()

See #24 for more details. Thanks @kshieh1 !

Changelog

1.2.1 - actually apply .and() weights

1.2.0 - Concatenate embeddings using .and()

For Stable Diffusion 2.1 I've been experimenting with a new feature: concatenated embeddings. What I noticed, for example, is that for more complex prompts image generation quality becomes wildly better when the prompt is broken into multiple parts and fed to OpenCLIP separately.

TL;DR: you can now experiment with breaking up your prompts into segments, which for SD2.1 appears to improve the generated image. The syntax is ("prompt part 1", "prompt part 2").and(). You can have more than one part, and you can also weight them, eg ("a man eating an apple", "sitting on the roof of a car", "high quality, trending on artstation, 8K UHD").and(1, 0.5, 0.5) which will assign weight 1 to man eating an apple and 0.5 to sitting on the roof of a car and high quality, trending on artstation, 8K UHD.

Here's a nonsense example from the InvokeAI discord #garbage-bin channel, created by gogurt enjoyer's incredible nightmare prompt generator:

a moist sloppy pindlesackboy sloppy hamblin' bogomadong, Clem Fandango is pissed-off, Wario's Woods in background, making a noise like ga-woink-a

Plugging this straight into SD2.1 we get this, which is really not a good image:

However, if the prompt is broken up into chunks and fed into OpenCLIP separately as four separate prompts, and then concatenated:

a moist sloppy pindlesackboy sloppy hamblin' bogomadong

Clem Fandango is pissed-off

Wario's Woods in background

making a noise like ga-woink-a

then output image with the same seed is so much better:

In the new .and() syntax you would prompt this as follows:

("a moist sloppy pindlesackboy sloppy hamblin' bogomadong", "Clem Fandango is pissed-off", "Wario's Woods in background", "making a noise like ga-woink-a").and()

The effect can be more or less subtle. Here for example is

A dream of a distant galaxy, by Caspar David Friedrich, matte painting, trending on artstation, HQ

And the same split into two parts:

A dream of a distant galaxy, by Caspar David Friedrich, matte painting

trending on artstation, HQ

The Compel prompt for this is:

("A dream of a distant galaxy, by Caspar David Friedrich, matte painting", "trending on artstation, HQ").and()

1.1.6 - misc small fixes

  • add DiffusersTextualInversionManager (thanks @pdoane)
  • fix batch embedding generation with truncated/non-truncated prompt lengths (#18, thanks @abassino)
  • add note about memory leakage (ref #24, thanks @kshieh1)
  • fix incorrect parsing when commas are not followed by whitespace (#34, thanks @moono)

1.1.5 - fix for compel turning numbers into floats for text inside parentheses

1.1.4 - fixes for #23 (sequential offload) and InvokeAI issue #3442 (allow hyphens in LoRA names)

1.1.3 - enable fetching the penultimate CLIP hidden layer (aka "clip skip")

To use, pass use_penultimate_clip_layer=True when initializing your Compel instance. Note that there's no need to pass this flag for SD2.0/SD2.1 because diffusers already throws away the last hidden layer when loading the SD2.0+ text encoder.

1.1.2 - fix for #21 (crash when parsing long prompts with truncation enabled if there is weighted fragments beyond the truncation boundary)

1.1.1 - fix for #22 (issues parsing . characters inside parentheses)

1.1.0 - support for parsing withLora/useLora on parse_prompt_string().

  • Compel.parse_prompt_string() now returns a Conjunction
  • any appearances of withLora(name[, weight]) or useLora(name[, weight]) anywhere in the prompt string will be parsed to LoraWeight instances, and returned on the outermost Conjunction returned by parse_prompt_string().

1.0.5 - fix incorrect parsing when passing invalid (auto1111) syntax that has a float

also fix test case for default swap parameters

1.0.4 - fix embeddings for empty swap target (eg cat.swap("")) when truncation is disabled

1.0.3 - better defaults for .swap (#8)

1.0.2 - fix padding for non-truncated batched embeddings (#9)

1.0.1 - fix for InvokeAI's --free_gpu_mem option

1.0.0 - new downweighting algorithm

Downweighting now works by applying an attention mask to remove the downweighted tokens, rather than literally removing them from the sequence. This behaviour is the default, but the old behaviour can be re-enabled by passing downweight_mode=DownweightMode.REMOVE on init of the Compel instance.

Formerly, downweighting a token worked by both multiplying the weighting of the token's embedding, and doing an inverse-weighted blend with a copy of the token sequence that had the downweighted tokens removed. The intuition is that as weight approaches zero, the tokens being downweighted should be actually removed from the sequence. However, removing the tokens resulted in the positioning of all downstream tokens becoming messed up. The blend ended up blending a lot more than just the tokens in question.

As of v1.0.0, taking advice from @keturn and @bonlime (#7) the procedure is by default different. Downweighting still involves a blend but what is blended is a version of the token sequence with the downweighted tokens masked out, rather than removed. This correctly preserves positioning embeddings of the other tokens.

Also a bugfix: fix black images on weight 0 (invoke-ai/InvokeAI#2832)

0.1.10 - add support for prompts longer than the model's max token length.

To enable, initialize Compel with truncate_long_prompts=False (default is True). Prompts that are longer than the model's max_token_length will be chunked and padded out to an integer multiple of max_token_length.

Note that even if you don't use a negative prompt, you'll need to build a conditioning tensor for a negative prompt of at least "", and use compel.pad_conditioning_tensors_to_same_length(), otherwise the you'll get an error about mismatched conditioning tensor lengths:

compel = Compel(..., truncate_long_prompts=False)
prompt = "a cat playing with a ball++ in the forest, amazing, exquisite, stunning, masterpiece, skilled, powerful, incredible, amazing, trending on gregstation, greg, greggy, greggs greggson, greggy mcgregface, ..." # very long prompt
conditioning = compel.build_conditioning_tensor(prompt)
negative_prompt = "" # it's necessary to create an empty prompt - it can also be very long, if you want
negative_conditioning = compel.build_conditioning_tensor(negative_prompt)
[conditioning, negative_conditioning] = compel.pad_conditioning_tensors_to_same_length([conditioning, negative_conditioning])

0.1.9 - broken

0.1.8 - downgrade Python min version to 3.7

0.1.7 - InvokeAI compatibility

More Repositories

1

skellington

Implementation of Le, Hodgins 2016 paper "Real-time Skeletal Skinning with Optimized Centers of Rotation"
C++
44
star
2

finetune-clip-huggingface

Finetuning CLIP on a small image/text dataset using huggingface libs
Python
38
star
3

ARKitPointCloud

Basic point cloud / anchor planes rendering with ARKit
Objective-C
24
star
4

insight3d

C++
19
star
5

ofxPd_damians

Pure Data addon for openFrameworks
C
12
star
6

candide

An interface for the Candide-3 face template
C++
12
star
7

ofxProfile

A thread-safe profiler. Stores and manages execution count, total execution time and average time per execution of each labelled section, separated by thread.
C++
12
star
8

grate

Stable diffusion multi-model image matrix generator based on 🧨diffusers
Python
11
star
9

dlib

Damian's pd patches
Pure Data
11
star
10

VegaFEM

Unofficial mirror of Vega-FEM (http://run.usc.edu/vega/) with a fix or two
C++
10
star
11

glutes

GLUT|ES port to unix, based on http://unrealvoodoo.org/git/glutes.git/
C
4
star
12

damian_public_of_apps

damian's public apps folder for of
C++
3
star
13

wind

enivronmental installation
Pure Data
2
star
14

googleSpreadsheetToXml

2
star
15

Metamart-fan-controller

C++
2
star
16

Luciolinae

code parts for the Luciolinae project
C
2
star
17

ortlos-ros

C#
2
star
18

The-Artvertiser

The Artvertiser based on openFrameworks
C++
2
star
19

Synopsis-Inspector

Synopsis Inspector allows for realtime browsing, inspecting, comparison and examination of metadata analyzed by "Synopsis Analyzer". Search, sort, filter based on content of your analyzed movies.
C++
1
star
20

damians-dataset-curation-tools

Tools for curating Stable Diffusion training datasets
Jupyter Notebook
1
star
21

TreeCollectionViewLayoutTest

Testing a tree collection view layout
Objective-C
1
star
22

Loco

animation controller based on cartwheel-3d et al
C++
1
star
23

addacs-module

code to support a beagleboard-based DSP module for the ADDAC system
C
1
star
24

dsLab

damianstewart.com/lab
JavaScript
1
star
25

Multicam-activity

C++
1
star
26

kll

rhythm visualization prototyping
C++
1
star
27

TheArtvertiserCombined

C++
1
star