.. _ten-minutes: NeMo Speech Inference in 5 Minutes =================================== This guide gives you a quick, hands-on tour of NeMo's core speech capabilities. By the end, you'll have transcribed audio, synthesized speech, identified speakers, and used a speech language model — all in about 50 lines of code. .. note:: Make sure you have :doc:`installed NeMo ` before starting. 1. Transcribe Speech (ASR) -------------------------- Automatic Speech Recognition converts audio to text. NeMo's Parakeet model sits at the top of the `HuggingFace OpenASR Leaderboard `_. **Basic transcription** — 3 lines of code: .. code-block:: python import nemo.collections.asr as nemo_asr asr_model = nemo_asr.models.ASRModel.from_pretrained("nvidia/parakeet-tdt-0.6b-v2") transcript = asr_model.transcribe(["audio.wav"])[0].text print(transcript) **With timestamps** — know *when* each word was spoken: .. code-block:: python hypotheses = asr_model.transcribe(["audio.wav"], timestamps=True) for stamp in hypotheses[0].timestamp['word']: print(f"{stamp['start']}s - {stamp['end']}s : {stamp['word']}") **From the command line**: .. code-block:: bash python examples/asr/transcribe_speech.py \ pretrained_name="nvidia/parakeet-tdt-0.6b-v2" \ audio_dir=./my_audio_files/ 2. Synthesize Speech (TTS) -------------------------- Text-to-Speech generates natural audio from text. NeMo's **Magpie TTS** is a multilingual, codec-based model that supports multiple speakers and languages: .. code-block:: python from nemo.collections.tts.models import MagpieTTSModel import soundfile as sf # Load model (multilingual 357M, from Hugging Face) model = MagpieTTSModel.from_pretrained("nvidia/magpie_tts_multilingual_357m") model.eval() # Generate speech audio, audio_len = model.do_tts( transcript="Hello! Welcome to NeMo speech AI.", language="en", ) # Save to file sf.write("output.wav", audio[0].cpu().numpy(), 22050) print("Speech saved to output.wav") 3. Identify Speakers (Diarization) ---------------------------------- Speaker diarization answers "who spoke when?" in multi-speaker audio. .. code-block:: python from nemo.collections.asr.models import SortformerEncLabelModel diar_model = SortformerEncLabelModel.from_pretrained("nvidia/diar_streaming_sortformer_4spk-v2") diar_model.eval() segments = diar_model.diarize(audio=["meeting.wav"], batch_size=1) for seg in segments[0]: print(seg) # (begin_seconds, end_seconds, speaker_index) 4. Speech Language Models (SpeechLM2) ------------------------------------- SpeechLM2 augments large language models with speech understanding. Canary-Qwen combines an ASR encoder with a Qwen LLM: .. code-block:: python from nemo.collections.speechlm2.models import SALM model = SALM.from_pretrained('nvidia/canary-qwen-2.5b') answer_ids = model.generate( prompts=[[{ "role": "user", "content": f"Transcribe the following: {model.audio_locator_tag}", "audio": ["speech.wav"], }]], max_new_tokens=128, ) print(model.tokenizer.ids_to_text(answer_ids[0].cpu())) What's Next? ------------ Now that you've seen the basics, dive deeper: - :doc:`key_concepts` — Understand the speech AI fundamentals behind these models - :doc:`choosing_a_model` — Find the best model for your specific use case - :doc:`../asr/intro` — Full ASR documentation - :doc:`../tts/intro` — Full TTS documentation - :doc:`../asr/speaker_diarization/intro` — Speaker diarization and recognition - :doc:`../starthere/tutorials` — Tutorial notebooks