From 0c4e97443b7884e219d73d51d31314844c215a74 Mon Sep 17 00:00:00 2001 From: Sulthan Suresh Fazeela <52656038+sulthansf@users.noreply.github.com> Date: Wed, 22 Nov 2023 16:30:30 +0100 Subject: [PATCH] Fix FastSAM text prompt results (#6510) --- ultralytics/models/fastsam/prompt.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/ultralytics/models/fastsam/prompt.py b/ultralytics/models/fastsam/prompt.py index 72ea5ef8..0f43441a 100644 --- a/ultralytics/models/fastsam/prompt.py +++ b/ultralytics/models/fastsam/prompt.py @@ -342,7 +342,7 @@ class FastSAMPrompt: max_idx = scores.argsort() max_idx = max_idx[-1] max_idx += sum(np.array(filter_id) <= int(max_idx)) - self.results[0].masks.data = torch.tensor(np.array([ann['segmentation'] for ann in annotations])) + self.results[0].masks.data = torch.tensor(np.array([annotations[max_idx]['segmentation']])) return self.results def everything_prompt(self):