From fc6c66a4a4265d0cc83e6e275b6c496224365b90 Mon Sep 17 00:00:00 2001 From: Glenn Jocher Date: Sun, 24 Mar 2024 05:56:37 +0100 Subject: [PATCH] Fix OpenVINO missing batch-size bug (#9264) --- ultralytics/nn/autobackend.py | 5 +---- 1 file changed, 1 insertion(+), 4 deletions(-) diff --git a/ultralytics/nn/autobackend.py b/ultralytics/nn/autobackend.py index 19514c65..89057be5 100644 --- a/ultralytics/nn/autobackend.py +++ b/ultralytics/nn/autobackend.py @@ -202,13 +202,10 @@ class AutoBackend(nn.Module): ov_model = core.read_model(model=str(w), weights=w.with_suffix(".bin")) if ov_model.get_parameters()[0].get_layout().empty: ov_model.get_parameters()[0].set_layout(ov.Layout("NCHW")) - batch_dim = ov.get_batch(ov_model) - if batch_dim.is_static: - batch_size = batch_dim.get_length() # OpenVINO inference modes are 'LATENCY', 'THROUGHPUT' (not recommended), or 'CUMULATIVE_THROUGHPUT' inference_mode = "CUMULATIVE_THROUGHPUT" if batch > 1 else "LATENCY" - LOGGER.info(f"Using OpenVINO {inference_mode} mode for batch-size={batch_size} inference...") + LOGGER.info(f"Using OpenVINO {inference_mode} mode for batch={batch} inference...") ov_compiled_model = core.compile_model( ov_model, device_name="AUTO", # AUTO selects best available device, do not modify