mirror of
https://github.com/invoke-ai/InvokeAI
synced 2024-08-30 20:32:17 +00:00
e93f4d632d
* introduce new abstraction layer for GPU devices * add unit test for device abstraction * fix ruff * convert TorchDeviceSelect into a stateless class * move logic to select context-specific execution device into context API * add mock hardware environments to pytest * remove dangling mocker fixture * fix unit test for running on non-CUDA systems * remove unimplemented get_execution_device() call * remove autocast precision * Multiple changes: 1. Remove TorchDeviceSelect.get_execution_device(), as well as calls to context.models.get_execution_device(). 2. Rename TorchDeviceSelect to TorchDevice 3. Added back the legacy public API defined in `invocation_api`, including choose_precision(). 4. Added a config file migration script to accommodate removal of precision=autocast. * add deprecation warnings to choose_torch_device() and choose_precision() * fix test crash * remove app_config argument from choose_torch_device() and choose_torch_dtype() --------- Co-authored-by: Lincoln Stein <lstein@gmail.com>
67 lines
2.7 KiB
Python
67 lines
2.7 KiB
Python
# Code from the original DWPose Implementation: https://github.com/IDEA-Research/DWPose
|
|
# Modified pathing to suit Invoke
|
|
|
|
|
|
import numpy as np
|
|
import onnxruntime as ort
|
|
|
|
from invokeai.app.services.config.config_default import get_config
|
|
from invokeai.app.util.download_with_progress import download_with_progress_bar
|
|
from invokeai.backend.util.devices import TorchDevice
|
|
|
|
from .onnxdet import inference_detector
|
|
from .onnxpose import inference_pose
|
|
|
|
DWPOSE_MODELS = {
|
|
"yolox_l.onnx": {
|
|
"local": "any/annotators/dwpose/yolox_l.onnx",
|
|
"url": "https://huggingface.co/yzd-v/DWPose/resolve/main/yolox_l.onnx?download=true",
|
|
},
|
|
"dw-ll_ucoco_384.onnx": {
|
|
"local": "any/annotators/dwpose/dw-ll_ucoco_384.onnx",
|
|
"url": "https://huggingface.co/yzd-v/DWPose/resolve/main/dw-ll_ucoco_384.onnx?download=true",
|
|
},
|
|
}
|
|
|
|
config = get_config()
|
|
|
|
|
|
class Wholebody:
|
|
def __init__(self):
|
|
device = TorchDevice.choose_torch_device()
|
|
|
|
providers = ["CUDAExecutionProvider"] if device.type == "cuda" else ["CPUExecutionProvider"]
|
|
|
|
DET_MODEL_PATH = config.models_path / DWPOSE_MODELS["yolox_l.onnx"]["local"]
|
|
download_with_progress_bar("yolox_l.onnx", DWPOSE_MODELS["yolox_l.onnx"]["url"], DET_MODEL_PATH)
|
|
|
|
POSE_MODEL_PATH = config.models_path / DWPOSE_MODELS["dw-ll_ucoco_384.onnx"]["local"]
|
|
download_with_progress_bar(
|
|
"dw-ll_ucoco_384.onnx", DWPOSE_MODELS["dw-ll_ucoco_384.onnx"]["url"], POSE_MODEL_PATH
|
|
)
|
|
|
|
onnx_det = DET_MODEL_PATH
|
|
onnx_pose = POSE_MODEL_PATH
|
|
|
|
self.session_det = ort.InferenceSession(path_or_bytes=onnx_det, providers=providers)
|
|
self.session_pose = ort.InferenceSession(path_or_bytes=onnx_pose, providers=providers)
|
|
|
|
def __call__(self, oriImg):
|
|
det_result = inference_detector(self.session_det, oriImg)
|
|
keypoints, scores = inference_pose(self.session_pose, det_result, oriImg)
|
|
|
|
keypoints_info = np.concatenate((keypoints, scores[..., None]), axis=-1)
|
|
# compute neck joint
|
|
neck = np.mean(keypoints_info[:, [5, 6]], axis=1)
|
|
# neck score when visualizing pred
|
|
neck[:, 2:4] = np.logical_and(keypoints_info[:, 5, 2:4] > 0.3, keypoints_info[:, 6, 2:4] > 0.3).astype(int)
|
|
new_keypoints_info = np.insert(keypoints_info, 17, neck, axis=1)
|
|
mmpose_idx = [17, 6, 8, 10, 7, 9, 12, 14, 16, 13, 15, 2, 1, 4, 3]
|
|
openpose_idx = [1, 2, 3, 4, 6, 7, 8, 9, 10, 12, 13, 14, 15, 16, 17]
|
|
new_keypoints_info[:, openpose_idx] = new_keypoints_info[:, mmpose_idx]
|
|
keypoints_info = new_keypoints_info
|
|
|
|
keypoints, scores = keypoints_info[..., :2], keypoints_info[..., 2]
|
|
|
|
return keypoints, scores
|