gpt-oss-20b-Ja-Fin-CPT

A Japanese financial domain model built through continued pre-training of gpt-oss-20b on a curated Japanese financial corpus.

Model Overview

A domain-adapted base model for Japanese finance, intended for further fine-tuning on specific financial tasks.

  • Base Model: gpt-oss-20b
  • Training Stage: Continued Pre-Training (CPT)
  • Domain: Japanese Finance
  • Language: Japanese, English

Training

Continued Pre-Training

Trained on a Japanese financial corpus constructed from Common Crawl and other public sources, with domain classification and quality filtering.

Training Infrastructure:

  • Hardware: AWS p5en.48xlarge (NVIDIA H200 Tensor Core GPU x 8)

Intended Use

Primarily intended as a foundation for further fine-tuning. For a reasoning model with instruction-following capabilities, see gpt-oss-20b-Ja-Fin-Thinking.

Primary Use Cases

  • Base model for financial domain SFT
  • Feature extraction for financial text
  • Further pre-training on proprietary financial data

Limitations

  • Not instruction-tuned: This model has not undergone supervised fine-tuning and may not follow instructions well
  • Domain specificity: Optimized for Japanese financial domain; performance on other domains may vary
  • Language coverage: Primarily Japanese and English

License

This model is released under the Apache 2.0 license.

Privacy Notice

For details on how personal information is handled, please see the Privacy Notice (日本語).

Citation

@inproceedings{okochiDomainSpecificLLM2026,
  author    = {大河内 悠磨 and Sim, Fabio Milentiansen and 岡田 智靖},
  title     = {ドメイン特化LLMの推論能力向上を目的とした合成指示データセットの構築と金融ドメインにおける評価},
  booktitle = {言語処理学会第32回年次大会 (NLP2026) },
  year      = {2026},
  month     = mar,
  address   = {Utsunomiya, Tochigi, Japan},
  publisher = {言語処理学会},
  note      = {Paper ID: C7-2},
  url       = {https://www.anlp.jp/proceedings/annual_meeting/2026/pdf_dir/C7-2.pdf}
}
@misc{okochi2026constructingsyntheticinstructiondatasets,
  title     = {Constructing Synthetic Instruction Datasets for Improving Reasoning in Domain-Specific LLMs: A Case Study in the Japanese Financial Domain},
  author    = {Yuma Okochi and Fabio Milentiansen Sim and Tomoyasu Okada},
  year      = {2026},
  eprint    = {2603.01353},
  archivePrefix = {arXiv},
  primaryClass  = {cs.LG},
  url       = {https://arxiv.org/abs/2603.01353}
}

Acknowledgments

This model was developed with the support of the "GENIAC (Generative AI Accelerator Challenge)" project, implemented by the Ministry of Economy, Trade and Industry (METI) and the New Energy and Industrial Technology Development Organization (NEDO), with the aim of strengthening Japan's development capabilities in generative AI.

Downloads last month
12
Safetensors
Model size
21B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for nri-ai/gpt-oss-20b-Ja-Fin-CPT

Finetuned
(499)
this model
Finetunes
1 model

Collection including nri-ai/gpt-oss-20b-Ja-Fin-CPT

Paper for nri-ai/gpt-oss-20b-Ja-Fin-CPT