Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism
Paper • 1909.08053 • Published • 5
1.3 billion parameter Megatron version of FinBERT.
Results have been measured on 10% of the datasets
| Model | Accuracy |
|---|---|
| Megatron-finbert-1.3B-cased | 82.64% |
| FinBERT-cased | 79.18% |
| Model | Accuracy |
|---|---|
| Megatron-finbert-1.3B-cased | 82.64% |
| FinBERT-cased | 79.18% |
More documentation coming soon...