If you want toggle for thinking for gguf's at LmStudio

#3
by tech77 - opened

For nohurry gguf's:

Go to your folder -> C:\Users\YOURNAME.lmstudio\hub\models\google\gemma-4-26B-A4B-it-heretic-GUFF (if you dont have last folders make them). Make file model.yaml
Open notepad and copy:

model.yaml is an open standard for defining cross-platform, composable AI models

Learn more at https://modelyaml.org

model: google/gemma-4-26B-A4B-it-heretic-GUFF
base:

  • key: nohurry/gemma-4-26B-A4B-it-heretic-GUFF
    sources:
    • type: huggingface
      user: nohurry
      repo: gemma-4-26B-A4B-it-heretic-GUFF

config:
operation:
fields:
- key: llm.prediction.temperature
value: 1.0
- key: llm.prediction.topPSampling
value:
checked: true
value: 0.95
- key: llm.prediction.topKSampling
value: 64
- key: llm.prediction.reasoning.parsing
value:
enabled: true
startString: "<|channel>thought"
endString: "<channel|>"
customFields:

  • key: enableThinking
    displayName: Enable Thinking
    description: Controls whether the model will think before replying
    type: boolean
    defaultValue: true
    effects:
    • type: setJinjaVariable
      variable: enable_thinking

metadataOverrides:
domain: llm
architectures:
- gemma4
compatibilityTypes:
- gguf
paramsStrings:
- 26B
minMemoryUsageBytes: 17000000000
contextLengths:
- 262144
vision: true
reasoning: true
trainedForToolUse: true

lms

I unfortunately couldn't get this to work, though it is probably a formatting issue. The formatting changes some of the symbols and indentations in your post, but using 'Quote' from the triple dots in the top right of your post can give your original text in a reply, and then removing the '> ' quotes at the start of each line can give the clean text.
Also added the slash after C:\Users\YOURNAME\ which might have gotten stripped.

Google suggested putting the yaml file in the gguf file location, though that caused LMStudio to not list the model at all.

edit: It might be because I didn't have the mmproj weights for vision. Grabbing those now and maybe it will prevent a silent failure which was happening when trying to add the model to the list.

edit 2: I had to rename the mmproj file to mmproj-gemma-4-26B-A4B-it-heretic.bf16.gguf to get it be detected by LM Studio and auto associated with gemma-4-26b-a4b-it-heretic.q4_k_m.gguf, and now Vision is automatically enabled.

Someone else make step by step tutorial at Reddit - https://www.reddit.com/r/LocalLLaMA/comments/1sc9s1x/tutorial_how_to_toggle_onoff_the_thinking_mode/
(but you need only 1 file - model.yaml)

Thanks! Once I added the mmproj vision weights with the correct name, and put the yaml file in the original location you said, it all works. :)

edit: Well the think button is there and lit up, though the model didn't actually seem to do a reasoning step on the first attempt. Will keep investigating. (Update: Should mention it works fine)

Sign up or log in to comment