#!/usr/bin/env python3 import argparse import asyncio import queue from logging import DEBUG, INFO, basicConfig, getLogger from typing import Optional, Union import sounddevice as sd import torch from whisper import available_models from whisper.audio import N_FRAMES, SAMPLE_RATE from whisper.tokenizer import LANGUAGES, TO_LANGUAGE_CODE from whispering.schema import WhisperConfig from whispering.serve import serve_with_websocket from whispering.transcriber import WhisperStreamingTranscriber from whispering.websocket_client import run_websocket_client logger = getLogger(__name__) def transcribe_from_mic( *, wsp: WhisperStreamingTranscriber, sd_device: Optional[Union[int, str]], num_block: int, ) -> None: q = queue.Queue() def sd_callback(indata, frames, time, status): if status: logger.warning(status) q.put(indata.ravel()) logger.info("Ready to transcribe") with sd.InputStream( samplerate=SAMPLE_RATE, blocksize=N_FRAMES * num_block, device=sd_device, dtype="float32", channels=1, callback=sd_callback, ): idx: int = 0 while True: logger.debug(f"Segment #: {idx}, The rest of queue: {q.qsize()}") segment = q.get() for chunk in wsp.transcribe(segment=segment): print(f"{chunk.start:.2f}->{chunk.end:.2f}\t{chunk.text}") idx += 1 def get_opts() -> argparse.Namespace: parser = argparse.ArgumentParser() parser.add_argument( "--language", type=str, default=None, choices=sorted(LANGUAGES.keys()) + sorted([k.title() for k in TO_LANGUAGE_CODE.keys()]), ) parser.add_argument( "--model", type=str, choices=available_models(), ) parser.add_argument( "--device", default="cuda" if torch.cuda.is_available() else "cpu", help="device to use for PyTorch inference", ) parser.add_argument( "--beam_size", "-b", type=int, default=5, ) parser.add_argument( "--num_block", "-n", type=int, default=160, help="Number of operation unit", ) parser.add_argument( "--temperature", "-t", type=float, action="append", default=[], ) parser.add_argument( "--mic", ) parser.add_argument( "--debug", action="store_true", ) parser.add_argument( "--host", default="0.0.0.0", help="host of websocker server", ) parser.add_argument( "--port", type=int, help="Port number of websocker server", ) parser.add_argument( "--allow-padding", action="store_true", ) parser.add_argument( "--mode", choices=["client"], ) return parser.parse_args() def get_wshiper(*, opts): config = WhisperConfig( model_name=opts.model, language=opts.language, device=opts.device, beam_size=opts.beam_size, temperatures=opts.temperature, allow_padding=opts.allow_padding, ) logger.debug(f"WhisperConfig: {config}") wsp = WhisperStreamingTranscriber(config=config) return wsp def main() -> None: opts = get_opts() basicConfig( level=DEBUG if opts.debug else INFO, format="[%(asctime)s] %(module)s.%(funcName)s:%(lineno)d %(levelname)s -> %(message)s", ) if opts.beam_size <= 0: opts.beam_size = None if len(opts.temperature) == 0: opts.temperature = [0.0, 0.2, 0.4, 0.6, 0.8, 1.0] opts.temperature = sorted(set(opts.temperature)) try: opts.mic = int(opts.mic) except Exception: pass if opts.host is not None and opts.port is not None: if opts.mode == "client": assert opts.language is None assert opts.model is None asyncio.run( run_websocket_client( opts=opts, ) ) else: assert opts.language is not None assert opts.model is not None wsp = get_wshiper(opts=opts) asyncio.run( serve_with_websocket( wsp=wsp, host=opts.host, port=opts.port, ) ) else: assert opts.language is not None assert opts.model is not None wsp = get_wshiper(opts=opts) transcribe_from_mic( wsp=wsp, sd_device=opts.mic, num_block=opts.num_block, ) if __name__ == "__main__": main()