From 472a02712dab65ef8ef0ff876d6ecdfb6dd50af2 Mon Sep 17 00:00:00 2001 From: Yuta Hayashibe Date: Sat, 7 Jan 2023 01:18:41 +0900 Subject: [PATCH] Updated messages --- whispering/transcriber.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/whispering/transcriber.py b/whispering/transcriber.py index 532c857..3df0c3f 100644 --- a/whispering/transcriber.py +++ b/whispering/transcriber.py @@ -30,9 +30,9 @@ class WhisperStreamingTranscriber: self.dtype = torch.float16 if fp16 else torch.float32 if self.model.device == torch.device("cpu"): if torch.cuda.is_available(): - logger.warning("Performing inference on CPU when CUDA is available") + logger.info("Performing inference on CPU though CUDA is available") if self.dtype == torch.float16: - logger.warning("FP16 is not supported on CPU; using FP32 instead") + logger.info("Using FP32 because FP16 is not supported on CPU") self.dtype = torch.float32 if self.dtype == torch.float32: