mirror of
https://github.com/invoke-ai/InvokeAI
synced 2024-08-30 20:32:17 +00:00
Change pad_to_multiple_of to be 8 for all cases. Add comment about it's temporary status
This commit is contained in:
parent
b9f607be56
commit
6001d3d71d
@ -166,14 +166,14 @@ class ModelPatcher:
|
|||||||
init_tokens_count = None
|
init_tokens_count = None
|
||||||
new_tokens_added = None
|
new_tokens_added = None
|
||||||
|
|
||||||
# This is required since Transformers 4.32
|
# TODO: This is required since Transformers 4.32 see
|
||||||
# see https://github.com/huggingface/transformers/pull/25088
|
# https://github.com/huggingface/transformers/pull/25088
|
||||||
# More information: https://docs.nvidia.com/deeplearning/performance/dl-performance-
|
# More information by NVIDIA:
|
||||||
# matrix-multiplication/index.html#requirements-tc
|
# https://docs.nvidia.com/deeplearning/performance/dl-performance-matrix-multiplication/index.html#requirements-tc
|
||||||
if "A100" in torch.cuda.get_device_name():
|
# This value might need to be changed in the future and take the GPUs model into account as there seem
|
||||||
pad_to_multiple_of = 64
|
# to be ideal values for different GPUS. This value is temporary!
|
||||||
else:
|
# For references to the current discussion please see https://github.com/invoke-ai/InvokeAI/pull/4817
|
||||||
pad_to_multiple_of = 8
|
pad_to_multiple_of = 8
|
||||||
|
|
||||||
try:
|
try:
|
||||||
# HACK: The CLIPTokenizer API does not include a way to remove tokens after calling add_tokens(...). As a
|
# HACK: The CLIPTokenizer API does not include a way to remove tokens after calling add_tokens(...). As a
|
||||||
|
Loading…
Reference in New Issue
Block a user