# /// script # requires-python = ">=3.11" # dependencies = [ # "vllm", # "torch", # "huggingface-hub", # "soundfile", # "librosa", # "numpy", # ] # # [[tool.uv.index]] # url = "https://wheels.vllm.ai/nightly/cu129" # # [tool.uv] # prerelease = "allow" # /// """ Transcribe audio files from a directory using Cohere Transcribe via vLLM. Designed to work with HF Buckets mounted as volumes via `hf jobs uv run -v ...`. Long audio files (>30s) are automatically chunked with overlap and reassembled. All chunks from all files are batched through vLLM in a single call for maximum throughput. Input: Output: /input/episode1.mp3 -> /output/episode1.txt /input/sub/clip.wav -> /output/sub/clip.txt Examples: # Local test (requires CUDA GPU) uv run transcribe.py ./test-audio ./test-output --language en # HF Jobs with bucket volumes hf jobs uv run --flavor l4x1 \\ -s HF_TOKEN \\ -v hf://buckets/user/audio-input:/input:ro \\ -v hf://buckets/user/transcripts:/output \\ transcribe.py /input /output --language en Model: CohereLabs/cohere-transcribe-03-2026 (2B, Apache 2.0) - 14 languages: en, de, fr, it, es, pt, el, nl, pl, ar, vi, zh, ja, ko - Up to 3x faster than other dedicated ASR models in same size range """ import argparse import json import logging import sys import time from pathlib import Path import numpy as np import torch logging.basicConfig( level=logging.INFO, format="%(asctime)s %(levelname)s %(message)s" ) logger = logging.getLogger(__name__) MODEL = "CohereLabs/cohere-transcribe-03-2026" AUDIO_EXTENSIONS = {".mp3", ".wav", ".flac", ".ogg", ".m4a", ".wma", ".aac", ".opus"} SUPPORTED_LANGUAGES = { "en", "de", "fr", "it", "es", "pt", "el", "nl", "pl", "ar", "vi", "zh", "ja", "ko", } # Chunking config — matches model's internal chunking behavior CHUNK_DURATION_S = 30 # seconds per chunk OVERLAP_S = 2 # overlap between chunks to avoid cutting mid-word SAMPLE_RATE = 16000 def check_cuda_availability(): if not torch.cuda.is_available(): logger.error("CUDA is not available. This script requires a GPU.") sys.exit(1) logger.info(f"CUDA available. GPU: {torch.cuda.get_device_name(0)}") def discover_audio_files(input_dir: Path) -> list[Path]: """Walk input_dir recursively, returning sorted list of audio files.""" files = [] for path in sorted(input_dir.rglob("*")): if path.is_file() and path.suffix.lower() in AUDIO_EXTENSIONS: files.append(path) return files def load_audio(file_path: Path) -> tuple[np.ndarray, int]: """Load audio file resampled to 16kHz mono.""" import librosa audio, sr = librosa.load(str(file_path), sr=SAMPLE_RATE, mono=True) return audio, sr def chunk_audio(audio: np.ndarray, sr: int) -> list[np.ndarray]: """Split audio into overlapping chunks. Short audio returned as single chunk.""" chunk_samples = CHUNK_DURATION_S * sr overlap_samples = OVERLAP_S * sr step = chunk_samples - overlap_samples if len(audio) <= chunk_samples: return [audio] chunks = [] start = 0 while start < len(audio): end = min(start + chunk_samples, len(audio)) chunks.append(audio[start:end]) if end >= len(audio): break start += step return chunks def build_prompt(language: str) -> str: """Build the CohereASR prompt with language tokens.""" return ( f"<|startofcontext|><|startoftranscript|>" f"<|emo:undefined|><|{language}|><|{language}|><|pnc|><|noitn|>" f"<|notimestamp|><|nodiarize|>" ) def main(): parser = argparse.ArgumentParser( description="Transcribe audio files using Cohere Transcribe via vLLM.", formatter_class=argparse.RawDescriptionHelpFormatter, epilog=""" Languages: en, de, fr, it, es, pt, el, nl, pl, ar, vi, zh, ja, ko Examples: uv run transcribe.py ./audio ./output --language en uv run transcribe.py /input /output --language en --max-files 3 HF Jobs with bucket volumes: hf jobs uv run --flavor l4x1 -s HF_TOKEN \\ -v hf://buckets/user/audio-bucket:/input:ro \\ -v hf://buckets/user/transcripts:/output \\ transcribe.py /input /output --language en """, ) parser.add_argument("input_dir", help="Directory containing audio files") parser.add_argument("output_dir", help="Directory to write transcript text files") parser.add_argument( "--language", required=True, choices=sorted(SUPPORTED_LANGUAGES), help="Language code (required, model does not auto-detect)", ) parser.add_argument( "--max-files", type=int, default=None, help="Limit number of files to process (for testing)", ) parser.add_argument( "--gpu-memory-utilization", type=float, default=0.8, help="GPU memory utilization (default: 0.8)", ) parser.add_argument( "--verbose", action="store_true", help="Print resolved package versions", ) args = parser.parse_args() check_cuda_availability() input_dir = Path(args.input_dir) output_dir = Path(args.output_dir) if not input_dir.is_dir(): logger.error(f"Input directory does not exist: {input_dir}") sys.exit(1) output_dir.mkdir(parents=True, exist_ok=True) # Discover audio files logger.info(f"Scanning {input_dir} for audio files...") files = discover_audio_files(input_dir) if not files: logger.error(f"No audio files found in {input_dir}") logger.error(f"Supported extensions: {', '.join(sorted(AUDIO_EXTENSIONS))}") sys.exit(1) if args.max_files: files = files[: args.max_files] logger.info(f"Found {len(files)} audio file(s)") # Load and chunk audio files logger.info("Loading and chunking audio files...") file_audio = [] # (file_path, full_audio, sr) all_chunks = [] # (file_index, chunk_array) — flat list for batching for i, f in enumerate(files): audio, sr = load_audio(f) file_audio.append((f, audio, sr)) chunks = chunk_audio(audio, sr) duration = len(audio) / sr logger.info(f" {f.name}: {duration:.0f}s -> {len(chunks)} chunk(s)") for chunk in chunks: all_chunks.append((i, chunk)) total_chunks = len(all_chunks) logger.info(f"Total chunks to transcribe: {total_chunks}") # Init vLLM logger.info(f"Initializing vLLM with {MODEL}...") from vllm import LLM, SamplingParams llm = LLM( model=MODEL, trust_remote_code=True, limit_mm_per_prompt={"audio": 1}, gpu_memory_utilization=args.gpu_memory_utilization, ) sampling_params = SamplingParams( temperature=0.0, max_tokens=1024, ) # Build inputs — one prompt per chunk prompt = build_prompt(args.language) inputs = [ { "prompt": prompt, "multi_modal_data": {"audio": [(chunk, SAMPLE_RATE)]}, } for _, chunk in all_chunks ] # Transcribe all chunks in one batch logger.info(f"Transcribing {total_chunks} chunk(s) across {len(files)} file(s)...") start_time = time.time() outputs = llm.generate(inputs, sampling_params) elapsed = time.time() - start_time # Reassemble chunks per file file_texts: dict[int, list[str]] = {} for (file_idx, _), output in zip(all_chunks, outputs): text = output.outputs[0].text.strip() file_texts.setdefault(file_idx, []).append(text) # Write outputs total_audio_duration = 0.0 results = [] for i, (file_path, audio, sr) in enumerate(file_audio): chunks_text = file_texts.get(i, []) full_text = " ".join(chunks_text) rel = file_path.relative_to(input_dir) txt_path = output_dir / rel.with_suffix(".txt") txt_path.parent.mkdir(parents=True, exist_ok=True) txt_path.write_text(full_text, encoding="utf-8") duration = len(audio) / sr total_audio_duration += duration results.append({ "file": str(rel), "duration_s": round(duration, 1), "chunks": len(chunks_text), "transcript_length": len(full_text), "word_count": len(full_text.split()), }) logger.info( f" {rel} -> {txt_path.name} " f"({len(full_text.split())} words, {len(chunks_text)} chunks, " f"{duration:.0f}s audio)" ) # Write summary summary_path = output_dir / "summary.jsonl" with open(summary_path, "w", encoding="utf-8") as f: for r in results: f.write(json.dumps(r) + "\n") # Report elapsed_str = f"{elapsed / 60:.1f} min" if elapsed > 60 else f"{elapsed:.1f}s" rtfx = total_audio_duration / elapsed if elapsed > 0 else 0 logger.info("=" * 50) logger.info(f"Done! Transcribed {len(files)} file(s) in {elapsed_str}") logger.info(f" Output: {output_dir}") logger.info(f" Audio: {total_audio_duration / 60:.1f} min total") logger.info(f" RTFx: {rtfx:.1f}x realtime") logger.info(f" Total chunks: {total_chunks}") logger.info(f" Summary: {summary_path}") if args.verbose: import importlib.metadata logger.info("--- Package versions ---") for pkg in ["vllm", "transformers", "torch", "librosa", "soundfile"]: try: logger.info(f" {pkg}=={importlib.metadata.version(pkg)}") except importlib.metadata.PackageNotFoundError: logger.info(f" {pkg}: not installed") if __name__ == "__main__": if len(sys.argv) == 1: print("=" * 60) print("Audio Transcription with Cohere Transcribe (vLLM)") print("=" * 60) print("\nTranscribe audio files from a directory -> text files.") print("Long audio automatically chunked with overlap.") print("Designed for HF Buckets mounted as volumes.") print() print("Usage:") print(" uv run transcribe.py INPUT_DIR OUTPUT_DIR --language en") print() print("Examples:") print(" uv run transcribe.py ./audio ./output --language en") print(" uv run transcribe.py ./audio ./output --language en --max-files 3") print() print("HF Jobs with bucket volumes:") print(" hf jobs uv run --flavor l4x1 -s HF_TOKEN \\") print(" -v hf://buckets/user/audio-input:/input:ro \\") print(" -v hf://buckets/user/transcripts:/output \\") print(" transcribe.py /input /output --language en") print() print("For full help: uv run transcribe.py --help") sys.exit(0) main()