InvokeAI/invokeai/backend/ip_adapter
2023-09-14 23:06:57 -04:00
..
__init__.py chore: flake8 cleanup 2023-09-05 12:07:12 +12:00
attention_processor.py Pass IP-Adapter conditioning via cross_attention_kwargs instead of concatenating to the text embedding. This avoids interference with other features that manipulate the text embedding (e.g. long prompts). 2023-09-08 11:47:36 -04:00
ip_adapter.py Fix python static checks. 2023-09-14 16:48:47 -04:00
README.md Lookup IP-Adapter linked image encoder from disk instead of storing in model config metadata. 2023-09-14 23:06:57 -04:00
resampler.py Remove need for the image_encoder param in IPAdapter.initialize(). 2023-09-14 14:14:35 -04:00

IP-Adapter Model Formats

The official IP-Adapter models are released here: h94/IP-Adapter

This official model repo does not integrate well with InvokeAI's current approach to model management, so we have defined a new file structure for IP-Adapter models. The InvokeAI format is described below.

CLIP Vision Models

CLIP Vision models are organized in `diffusers`` format. The expected directory structure is:

ip_adapter_sd_image_encoder/
├── config.json
└── model.safetensors

IP-Adapter Models

IP-Adapter models are stored in a directory containing two files

  • image_encoder.txt: A text file containing the model identifier for the CLIP Vision encoder that is intended to be used with this IP-Adapter model.
  • ip_adapter.bin: The IP-Adapter weights.

Sample directory structure:

ip_adapter_sd15/
├── image_encoder.txt
└── ip_adapter.bin

Why save the weights in a .safetensors file?

The weights in ip_adapter.bin are stored in a nested dict, which is not supported by safetensors. This could be solved by splitting ip_adapter.bin into multiple files, but for now we have decided to maintain consistency with the checkpoint structure used in the official h94/IP-Adapter repo.

InvokeAI Hosted IP-Adapters

Image Encoders:

IP-Adapters: