CDLI SLAM-ASR Luganda Atypical Speech Encoder-LoRA Checkpoint (Step 5000)

Encoder-LoRA atypical-speech adaptation checkpoint for SLAM-ASR on the CDLI Luganda atypical speech dataset. The base Whisper encoder and Sunflower-14B decoder are kept frozen while low-rank adapters are applied to the encoder q_proj and v_proj modules.

What this repository contains

This Hub repository stores a partial SLAM-ASR checkpoint for use with the SLAM-LLM codebase. It is not a standalone transformers checkpoint.

  • Checkpoint type: encoder_lora_projector
  • Architecture: Whisper encoder (Sunbird/asr-whisper-large-v3-salt) + linear projector + Sunflower-14B decoder; encoder LoRA on q_proj/v_proj; LLM frozen.
  • Base encoder: Sunbird/asr-whisper-large-v3-salt
  • Base LLM: Sunbird/Sunflower-14B
  • Exported files: model.pt

Training / evaluation context

  • Dataset: cdli/ugandan_luganda_nonstandard_speech_v1.0
  • Evaluation split: test
  • Training speakers: 36
  • Validation speakers: 5
  • Speaker overlap: No speaker overlap between train and validation/test

Reported metrics

  • Normalized WER (JiWER scorer): 60.84%
  • Normalized CER (JiWER scorer): 24.12%
  • Atypical overall normalized WER: 61.00%
  • Atypical overall normalized CER: 24.14%
  • Atypical averaged utterance WER: 55.83%
  • Atypical averaged utterance CER: 20.02%

Decode settings used for the reported metrics

Test decode used MAX_NEW_TOKENS=200, NUM_BEAMS=4, REPETITION_PENALTY=2.0, NO_REPEAT_NGRAM_SIZE=2, USE_ENCODER_PEFT=true, ENCODER_TARGET_MODULES=[q_proj,v_proj].

Additional results notes

Test subgroup breakdown: Mild 50.75% WER, Moderate 53.58%, Severe 64.20%. By disorder: Dysarthria 50.68%, Stuttering 55.57%, Articulation Disorders 55.95%, Voice disorder 70.17%. Average hyp/ref ratio was 97.14%, indicating stable decoding without catastrophic looping.

Loading notes

Load through SLAM-LLM; this repository stores a partial SLAM-ASR checkpoint, not a standalone Transformers model.

Typical decode flow in this project uses:

  • examples/asr_luganda/scripts/decode_luganda_sunflower.sh
  • USE_ENCODER_PEFT=true for encoder-LoRA checkpoints
  • matching LoRA target modules at decode time

Caveats

  • This repository stores SLAM-ASR training artifacts intended for research use.
  • The checkpoint must be used with the matching SLAM-LLM model code and base components.
  • Results can be sensitive to decode settings and evaluation protocol.
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for KasuleTrevor/cdli-slam-asr-luganda-atypical-encoder-lora-step5000

Finetuned
Qwen/Qwen3-14B
Adapter
(9)
this model

Dataset used to train KasuleTrevor/cdli-slam-asr-luganda-atypical-encoder-lora-step5000