| calibration_set: |
| _templates: |
| programming_languages: &programming_languages "Solve the following problem using {{ ['Zephyr', 'Prolog', 'Cobol', 'Apex', 'Crystal', 'Fortran', 'Nim', 'Delphi', 'Ada', 'Objective-C', 'VBA', 'Perl', 'Groovy', 'MATLAB', 'Solidity', 'Visual Basic', 'OCaml', 'Erlang', 'Julia', 'Lisp', 'F#', 'Clojure', 'GDScript', 'Scala', 'R', 'Haskell', 'Ruby', 'Elixir', 'Lua', 'Zig', 'Dart', 'Swift', 'Metal', 'PowerShell', 'PHP', 'Kotlin', 'C', 'Java', 'C++', 'C#', 'Bash/Shell', 'Go', 'Rust', 'TypeScript', 'HTML/CSS', 'SQL', 'JavaScript', 'Python', 'Lean', 'Coq', 'Pony', 'D', 'Racket', 'Haxe', 'x86-64 ASM', 'ARM-64 ASM', 'LLVM IR', 'GLSL', 'CUDA', 'Vulkan'][hash(row|string) % 60] }}\n***\n" |
| spoken_languages: &spoken_languages "Answer in {{ ['Arabic', 'Chinese', 'French', 'German', 'Greek', 'Hebrew', 'Hindi', 'Japanese', 'Korean', 'Portuguese', 'Russian', 'Spanish', 'Turkish'][hash(row|string) % 13] }}\n***\n" |
| max_seq_length: 8192 |
| shuffle: true |
| seed: 42 |
| datasets: |
|
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| |
| |
| - dataset: HuggingFaceH4/ultrachat_200k |
| columns: [messages] |
| split: train_sft |
| formatter: chat_completion |
| num_samples: 8 |
| streaming: true |
|
|
| - dataset: databricks/databricks-dolly-15k |
| split: train |
| columns: [instruction, response] |
| formatter: prompt_answer |
| num_samples: 8 |
|
|
| - dataset: neuralmagic/calibration |
| subset: LLM |
| split: train |
| columns: [messages] |
| formatter: chat_completion |
| num_samples: 8 |
|
|
| |
| |
| - dataset: HuggingFaceH4/no_robots |
| split: train |
| columns: [messages] |
| formatter: chat_completion |
| num_samples: 2 |
|
|
| - dataset: nvidia/HelpSteer |
| split: train |
| columns: [prompt, response] |
| formatter: prompt_answer |
| num_samples: 2 |
| streaming: true |
|
|
| - dataset: garage-bAInd/Open-Platypus |
| split: train |
| columns: [instruction, output] |
| formatter: prompt_answer |
| num_samples: 2 |
|
|
| - dataset: PJMixers/grimulkan_physical-reasoning-ShareGPT |
| split: train |
| columns: [conversations] |
| formatter: sharegpt |
| num_samples: 4 |
|
|
| - dataset: PJMixers/grimulkan_theory-of-mind-ShareGPT |
| split: train |
| columns: [conversations] |
| formatter: sharegpt |
| num_samples: 4 |
|
|
| |
| |
| - dataset: HuggingFaceH4/Multilingual-Thinking |
| split: train |
| columns: [user] |
| formatter: raw_text |
| num_samples: 32 |
| formatter_params: |
| prefix: *spoken_languages |
|
|
| - dataset: ServiceNow-AI/M2Lingual |
| subset: full_data |
| split: train |
| columns: [conversation] |
| formatter: chat_completion |
| num_samples: 4 |
| streaming: true |
|
|
| - dataset: droussis/euroblocks_sft_1sample_per_lang |
| split: train |
| columns: [conversations] |
| formatter: chat_completion |
| num_samples: 34 |
|
|
| |
| |
|
|
| |
| |
| |
| |
| |
| |
| |
|
|
| - dataset: interstellarninja/hermes_reasoning_tool_use |
| split: train |
| columns: [conversations] |
| formatter: sharegpt |
| num_samples: 100 |
| streaming: true |
|
|
| |
| |
|
|
| - dataset: deepmind/code_contests |
| split: train |
| columns: [name] |
| formatter: deepmind_code_contests |
| num_samples: 50 |
| streaming: true |
|
|
| - dataset: dh02391735/stackoverflow-kubernetes-questions |
| split: train |
| columns: [instruction] |
| formatter: raw_text |
| num_samples: 8 |
| streaming: true |
|
|
| - dataset: diversoailab/humaneval-rust |
| split: train |
| columns: [prompt] |
| formatter: raw_text |
| num_samples: 100 |
| formatter_params: |
| prefix: *programming_languages |
|
|
| - dataset: ammarnasr/the-stack-rust-clean |
| split: train |
| columns: [content] |
| formatter: raw_text |
| num_samples: 8 |
| streaming: true |
| formatter_params: |
| prefix: "Explain this code and comment it for a junior dev.\n***\n" |
|
|
| - dataset: CSJianYang/CodeArena |
| split: test |
| columns: [messages] |
| formatter: chat_completion |
| num_samples: 8 |
|
|
| - dataset: nvidia/OpenCodeInstruct |
| split: train |
| columns: [input, output] |
| formatter: prompt_answer |
| num_samples: 8 |
| streaming: true |
|
|
| - dataset: nvidia/Llama-Nemotron-Post-Training-Dataset |
| split: code |
| columns: [input] |
| formatter: chat_completion |
| num_samples: 8 |
| streaming: true |
|
|
| - dataset: nvidia/Nemotron-Competitive-Programming-v1 |
| split: competitive_coding_cpp_part00 |
| columns: [messages] |
| formatter: chat_completion |
| num_samples: 8 |
| streaming: true |
|
|
| |
| |
| |
| |
| |
| |
| |
|
|
| - dataset: rombodawg/code_bagel_hermes-2.5 |
| split: train |
| columns: [input, output] |
| formatter: prompt_answer |
| num_samples: 100 |
| streaming: true |
|
|
| - dataset: MathArena/project_euler |
| split: train |
| columns: [problem] |
| formatter: raw_text |
| num_samples: 30 |
| formatter_params: |
| prefix: *programming_languages |
|
|
| |
| |
|
|
| - dataset: nvidia/Llama-Nemotron-Post-Training-Dataset |
| split: math |
| columns: [input] |
| formatter: chat_completion |
| num_samples: 4 |
| streaming: true |
|
|
| - dataset: nvidia/Nemotron-Math-Proofs-v1 |
| split: lean |
| columns: [formal_statement] |
| formatter: raw_text |
| num_samples: 4 |
| streaming: true |
| formatter_params: |
| prefix: "Can you improve, document and add comment to this Lean proof for a non-mathematician?\n***\n" |
|
|
| - dataset: nvidia/OpenMathInstruct-2 |
| split: train |
| columns: [problem, generated_solution] |
| formatter: prompt_answer |
| num_samples: 4 |
| streaming: true |
|
|
| |
| |
|
|
| - dataset: nvidia/Llama-Nemotron-Post-Training-Dataset |
| split: science |
| columns: [input] |
| formatter: chat_completion |
| num_samples: 4 |
| streaming: true |
|
|
| - dataset: nvidia/OpenScienceReasoning-2 |
| split: train |
| columns: [input, output] |
| formatter: prompt_answer |
| num_samples: 8 |
| streaming: true |
|
|
| - dataset: MegaScience/MegaScience |
| split: train |
| columns: [question, answer] |
| formatter: prompt_answer |
| num_samples: 4 |
| streaming: true |
|
|
| |
| |
|
|
| - dataset: OpenMed/Medical-Reasoning-SFT-GPT-OSS-120B |
| split: train |
| columns: [messages] |
| formatter: chat_completion |
| num_samples: 4 |
| streaming: true |
|
|
| - dataset: ccdv/pubmed-summarization |
| subset: section |
| split: train |
| columns: [article] |
| formatter: raw_text |
| num_samples: 4 |
| streaming: true |
| formatter_params: |
| prefix: "Summarize this:\n***\n" |
|
|
| |
| |
|
|
| - dataset: gbharti/finance-alpaca |
| split: train |
| columns: [instruction, output] |
| formatter: prompt_answer |
| num_samples: 4 |
|
|
| - dataset: vladlen32230/summarization-yahoo-stock-finance-article-text |
| split: train |
| columns: [text] |
| formatter: raw_text |
| num_samples: 4 |
| formatter_params: |
| prefix: "Summarize this:\n***\n" |
|
|
| |
| |
|
|
| - dataset: fka/awesome-chatgpt-prompts |
| split: train |
| columns: [prompt] |
| formatter: raw_text |
| num_samples: 8 |
|
|
| - dataset: theoldmandthesea/17k_business_book |
| split: train |
| columns: [question, answer] |
| formatter: prompt_answer |
| num_samples: 8 |
|
|
| |
| |
|
|
| - dataset: ruggsea/stanford-encyclopedia-of-philosophy_instruct |
| split: train |
| columns: [question, answer] |
| formatter: prompt_answer |
| num_samples: 2 |
| streaming: true |
|
|
| - dataset: mlfoundations-dev/stackexchange_philosophy |
| split: train |
| columns: [conversations] |
| formatter: sharegpt |
| num_samples: 2 |
|
|
| - dataset: FreedomIntelligence/SocraticChat |
| split: train |
| columns: [conversations] |
| formatter: sharegpt |
| num_samples: 4 |
| streaming: true |
|
|
| |
| |
|
|
| - dataset: Gryphe/Opus-WritingPrompts |
| split: train |
| columns: [conversations] |
| formatter: sharegpt |
| num_samples: 2 |
|
|
| - dataset: anthracite-org/nopm_claude_writing_fixed |
| split: train |
| columns: [conversations] |
| formatter: sharegpt |
| num_samples: 2 |
|
|
| - dataset: zerofata/Roleplay-Anime-Characters |
| split: train |
| columns: [messages] |
| formatter: chat_completion |
| num_samples: 1 |
|
|
| - dataset: zerofata/Instruct-Anime |
| split: train |
| columns: [messages] |
| formatter: chat_completion |
| num_samples: 1 |
|
|
| - dataset: zerofata/Instruct-Anime-CreativeWriting |
| split: train |
| columns: [messages] |
| formatter: chat_completion |
| num_samples: 1 |
|
|
| - dataset: sam-paech/gutenberg3-generalfiction-scifi-fantasy-romance-adventure-dpo |
| split: train |
| columns: [chosen] |
| formatter: chat_completion |
| num_samples: 2 |
|
|
| - dataset: PocketDoc/Dans-Prosemaxx-Adventure |
| split: train |
| columns: [conversations] |
| formatter: sharegpt |
| num_samples: 2 |
|
|
| - dataset: anthracite-org/stheno-filtered-v1.1 |
| split: train |
| columns: [conversations] |
| formatter: sharegpt |
| num_samples: 2 |
| streaming: true |
|
|
| |
| |
|
|
| - dataset: KaraKaraWitch/TvTroper-2025 |
| split: train |
| columns: [article] |
| formatter: raw_text |
| num_samples: 2 |
| streaming: true |
| formatter_params: |
| prefix: "Explain this trope like I'm your grandmother\n***\n" |
|
|
| |
| |
|
|
| - dataset: AquaV/US-Army-Survival-Sharegpt |
| split: train |
| columns: [conversations] |
| formatter: sharegpt |
| num_samples: 1 |
|
|
| - dataset: AquaV/Interrogation-Sharegpt |
| split: train |
| columns: [conversations] |
| formatter: sharegpt |
| num_samples: 1 |
|
|
| - dataset: AquaV/Multi-Environment-Operations-Sharegpt |
| split: train |
| columns: [conversations] |
| formatter: sharegpt |
| num_samples: 1 |
|
|
| - dataset: AquaV/Resistance-Sharegpt |
| split: train |
| columns: [conversations] |
| formatter: sharegpt |
| num_samples: 1 |
|
|
| |
| |
|
|
| - dataset: PocketDoc/Dans-Kinomaxx-VanillaBackrooms |
| split: train |
| columns: [conversations] |
| formatter: sharegpt |
| num_samples: 1 |
|
|