diff --git a/speed.py b/speed.py index 4dc4d10..c4ed93e 100644 --- a/speed.py +++ b/speed.py @@ -6,10 +6,13 @@ Benchmark the inference speed of each module in LivePortrait. TODO: heavy GPT style, need to refactor """ -import yaml import torch +torch._dynamo.config.suppress_errors = True # Suppress errors and fall back to eager execution + +import yaml import time import numpy as np + from src.utils.helper import load_model, concat_feat from src.config.inference_config import InferenceConfig @@ -47,9 +50,6 @@ def load_and_compile_models(cfg, model_config): """ Load and compile models for inference """ - import torch._dynamo - torch._dynamo.config.suppress_errors = True # Suppress errors and fall back to eager execution - appearance_feature_extractor = load_model(cfg.checkpoint_F, model_config, cfg.device_id, 'appearance_feature_extractor') motion_extractor = load_model(cfg.checkpoint_M, model_config, cfg.device_id, 'motion_extractor') warping_module = load_model(cfg.checkpoint_W, model_config, cfg.device_id, 'warping_module')