mirror of
https://github.com/invoke-ai/InvokeAI
synced 2024-08-30 20:32:17 +00:00
remove references to vram_cache in tests
This commit is contained in:
parent
eaa2c68693
commit
a1dcab9c38
@ -80,7 +80,6 @@ class ModelManagerService(ModelManagerServiceBase):
|
|||||||
|
|
||||||
ram_cache = ModelCache(
|
ram_cache = ModelCache(
|
||||||
max_cache_size=app_config.ram,
|
max_cache_size=app_config.ram,
|
||||||
max_vram_cache_size=app_config.vram,
|
|
||||||
logger=logger,
|
logger=logger,
|
||||||
execution_devices=execution_devices,
|
execution_devices=execution_devices,
|
||||||
)
|
)
|
||||||
|
@ -109,7 +109,6 @@ def mm2_loader(mm2_app_config: InvokeAIAppConfig) -> ModelLoadServiceBase:
|
|||||||
ram_cache = ModelCache(
|
ram_cache = ModelCache(
|
||||||
logger=InvokeAILogger.get_logger(),
|
logger=InvokeAILogger.get_logger(),
|
||||||
max_cache_size=mm2_app_config.ram,
|
max_cache_size=mm2_app_config.ram,
|
||||||
max_vram_cache_size=mm2_app_config.vram,
|
|
||||||
)
|
)
|
||||||
convert_cache = ModelConvertCache(mm2_app_config.convert_cache_path)
|
convert_cache = ModelConvertCache(mm2_app_config.convert_cache_path)
|
||||||
return ModelLoadService(
|
return ModelLoadService(
|
||||||
|
@ -92,7 +92,6 @@ def test_migrate_v3_config_from_file(tmp_path: Path, patch_rootdir: None):
|
|||||||
assert config.host == "192.168.1.1"
|
assert config.host == "192.168.1.1"
|
||||||
assert config.port == 8080
|
assert config.port == 8080
|
||||||
assert config.ram == 100
|
assert config.ram == 100
|
||||||
assert config.vram == 50
|
|
||||||
assert config.legacy_models_yaml_path == Path("/custom/models.yaml")
|
assert config.legacy_models_yaml_path == Path("/custom/models.yaml")
|
||||||
# This should be stripped out
|
# This should be stripped out
|
||||||
assert not hasattr(config, "esrgan")
|
assert not hasattr(config, "esrgan")
|
||||||
|
Loading…
Reference in New Issue
Block a user