2024-04-15 13:12:49 +00:00
|
|
|
from typing import Dict, Literal, Optional, Union
|
2023-02-18 00:29:03 +00:00
|
|
|
|
2022-08-31 04:33:23 +00:00
|
|
|
import torch
|
2024-04-15 13:12:49 +00:00
|
|
|
from deprecated import deprecated
|
2023-08-18 15:13:28 +00:00
|
|
|
|
2024-04-15 13:12:49 +00:00
|
|
|
from invokeai.app.services.config.config_default import get_config
|
2022-08-31 04:33:23 +00:00
|
|
|
|
2024-04-15 13:12:49 +00:00
|
|
|
# legacy APIs
|
|
|
|
TorchPrecisionNames = Literal["float32", "float16", "bfloat16"]
|
2023-02-18 00:29:03 +00:00
|
|
|
CPU_DEVICE = torch.device("cpu")
|
2023-03-03 05:02:15 +00:00
|
|
|
CUDA_DEVICE = torch.device("cuda")
|
|
|
|
MPS_DEVICE = torch.device("mps")
|
2023-03-03 06:02:00 +00:00
|
|
|
|
2023-07-27 14:54:01 +00:00
|
|
|
|
2024-04-15 13:12:49 +00:00
|
|
|
@deprecated("Use TorchDevice.choose_torch_dtype() instead.") # type: ignore
|
|
|
|
def choose_precision(device: torch.device) -> TorchPrecisionNames:
|
|
|
|
"""Return the string representation of the recommended torch device."""
|
|
|
|
torch_dtype = TorchDevice.choose_torch_dtype(device)
|
|
|
|
return PRECISION_TO_NAME[torch_dtype]
|
|
|
|
|
|
|
|
|
|
|
|
@deprecated("Use TorchDevice.choose_torch_device() instead.") # type: ignore
|
2023-02-18 00:29:03 +00:00
|
|
|
def choose_torch_device() -> torch.device:
|
2024-04-15 13:12:49 +00:00
|
|
|
"""Return the torch.device to use for accelerated inference."""
|
|
|
|
return TorchDevice.choose_torch_device()
|
|
|
|
|
|
|
|
|
|
|
|
@deprecated("Use TorchDevice.choose_torch_dtype() instead.") # type: ignore
|
|
|
|
def torch_dtype(device: torch.device) -> torch.dtype:
|
|
|
|
"""Return the torch precision for the recommended torch device."""
|
|
|
|
return TorchDevice.choose_torch_dtype(device)
|
|
|
|
|
|
|
|
|
|
|
|
NAME_TO_PRECISION: Dict[TorchPrecisionNames, torch.dtype] = {
|
|
|
|
"float32": torch.float32,
|
|
|
|
"float16": torch.float16,
|
|
|
|
"bfloat16": torch.bfloat16,
|
|
|
|
}
|
|
|
|
PRECISION_TO_NAME: Dict[torch.dtype, TorchPrecisionNames] = {v: k for k, v in NAME_TO_PRECISION.items()}
|
|
|
|
|
|
|
|
|
|
|
|
class TorchDevice:
|
|
|
|
"""Abstraction layer for torch devices."""
|
|
|
|
|
2024-06-27 09:15:23 +00:00
|
|
|
CPU_DEVICE = torch.device("cpu")
|
|
|
|
CUDA_DEVICE = torch.device("cuda")
|
|
|
|
MPS_DEVICE = torch.device("mps")
|
|
|
|
|
2024-04-15 13:12:49 +00:00
|
|
|
@classmethod
|
|
|
|
def choose_torch_device(cls) -> torch.device:
|
|
|
|
"""Return the torch.device to use for accelerated inference."""
|
|
|
|
app_config = get_config()
|
|
|
|
if app_config.device != "auto":
|
|
|
|
device = torch.device(app_config.device)
|
|
|
|
elif torch.cuda.is_available():
|
|
|
|
device = CUDA_DEVICE
|
|
|
|
elif torch.backends.mps.is_available():
|
|
|
|
device = MPS_DEVICE
|
2024-04-07 04:28:29 +00:00
|
|
|
else:
|
2024-04-15 13:12:49 +00:00
|
|
|
device = CPU_DEVICE
|
|
|
|
return cls.normalize(device)
|
|
|
|
|
|
|
|
@classmethod
|
|
|
|
def choose_torch_dtype(cls, device: Optional[torch.device] = None) -> torch.dtype:
|
|
|
|
"""Return the precision to use for accelerated inference."""
|
|
|
|
device = device or cls.choose_torch_device()
|
|
|
|
config = get_config()
|
|
|
|
if device.type == "cuda" and torch.cuda.is_available():
|
|
|
|
device_name = torch.cuda.get_device_name(device)
|
|
|
|
if "GeForce GTX 1660" in device_name or "GeForce GTX 1650" in device_name:
|
|
|
|
# These GPUs have limited support for float16
|
|
|
|
return cls._to_dtype("float32")
|
|
|
|
elif config.precision == "auto":
|
|
|
|
# Default to float16 for CUDA devices
|
|
|
|
return cls._to_dtype("float16")
|
|
|
|
else:
|
|
|
|
# Use the user-defined precision
|
|
|
|
return cls._to_dtype(config.precision)
|
|
|
|
|
|
|
|
elif device.type == "mps" and torch.backends.mps.is_available():
|
|
|
|
if config.precision == "auto":
|
|
|
|
# Default to float16 for MPS devices
|
|
|
|
return cls._to_dtype("float16")
|
|
|
|
else:
|
|
|
|
# Use the user-defined precision
|
|
|
|
return cls._to_dtype(config.precision)
|
|
|
|
# CPU / safe fallback
|
|
|
|
return cls._to_dtype("float32")
|
|
|
|
|
|
|
|
@classmethod
|
|
|
|
def get_torch_device_name(cls) -> str:
|
|
|
|
"""Return the device name for the current torch device."""
|
|
|
|
device = cls.choose_torch_device()
|
|
|
|
return torch.cuda.get_device_name(device) if device.type == "cuda" else device.type.upper()
|
|
|
|
|
|
|
|
@classmethod
|
|
|
|
def normalize(cls, device: Union[str, torch.device]) -> torch.device:
|
|
|
|
"""Add the device index to CUDA devices."""
|
|
|
|
device = torch.device(device)
|
|
|
|
if device.index is None and device.type == "cuda" and torch.cuda.is_available():
|
2023-02-18 00:29:03 +00:00
|
|
|
device = torch.device(device.type, torch.cuda.current_device())
|
2024-04-15 13:12:49 +00:00
|
|
|
return device
|
|
|
|
|
|
|
|
@classmethod
|
|
|
|
def empty_cache(cls) -> None:
|
|
|
|
"""Clear the GPU device cache."""
|
|
|
|
if torch.backends.mps.is_available():
|
|
|
|
torch.mps.empty_cache()
|
|
|
|
if torch.cuda.is_available():
|
|
|
|
torch.cuda.empty_cache()
|
|
|
|
|
|
|
|
@classmethod
|
|
|
|
def _to_dtype(cls, precision_name: TorchPrecisionNames) -> torch.dtype:
|
|
|
|
return NAME_TO_PRECISION[precision_name]
|