mirror of
https://github.com/invoke-ai/InvokeAI
synced 2024-08-30 20:32:17 +00:00
34 lines
988 B
Plaintext
34 lines
988 B
Plaintext
|
albumentations==0.4.3
|
||
|
einops==0.3.0
|
||
|
huggingface-hub==0.8.1
|
||
|
imageio-ffmpeg==0.4.2
|
||
|
imageio==2.9.0
|
||
|
kornia==0.6.0
|
||
|
# pip will resolve the version which matches torch
|
||
|
numpy
|
||
|
omegaconf==2.1.1
|
||
|
opencv-python==4.6.0.66
|
||
|
pillow==9.2.0
|
||
|
pip>=22
|
||
|
pudb==2019.2
|
||
|
pytorch-lightning==1.4.2
|
||
|
streamlit==1.12.0
|
||
|
# "CompVis/taming-transformers" doesn't work
|
||
|
# ldm\models\autoencoder.py", line 6, in <module>
|
||
|
# from taming.modules.vqvae.quantize import VectorQuantizer2 as VectorQuantizer
|
||
|
# ModuleNotFoundError
|
||
|
taming-transformers-rom1504==0.0.6
|
||
|
test-tube>=0.7.5
|
||
|
torch-fidelity==0.3.0
|
||
|
torchmetrics==0.6.0
|
||
|
transformers==4.19.2
|
||
|
git+https://github.com/openai/CLIP.git@main#egg=clip
|
||
|
git+https://github.com/lstein/k-diffusion.git@master#egg=k-diffusion
|
||
|
git+https://github.com/lstein/GFPGAN@fix-dark-cast-images#egg=gfpgan
|
||
|
# No CUDA in PyPi builds
|
||
|
--extra-index-url https://download.pytorch.org/whl/cu113 --trusted-host https://download.pytorch.org
|
||
|
torch==1.11.0
|
||
|
# Same as numpy - let pip do its thing
|
||
|
torchvision
|
||
|
-e .
|