Multilingual GPT2 Models
Collection
8 items • Updated • 1
This model is a fine-tuned version of CausalNLP/gpt2-hf_multilingual-20 on an unknown dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
| Training Loss | Epoch | Step | Validation Loss |
|---|---|---|---|
| 3.2904 | 0.0876 | 500 | 3.3086 |
| 3.2581 | 0.1752 | 1000 | 3.2938 |
| 3.2796 | 0.2628 | 1500 | 3.2979 |
| 3.2409 | 0.3504 | 2000 | 3.2964 |
| 3.2632 | 0.4380 | 2500 | 3.2832 |
| 3.2446 | 0.5256 | 3000 | 3.2724 |
| 3.2425 | 0.6132 | 3500 | 3.2638 |
| 3.2372 | 0.7008 | 4000 | 3.2551 |
| 3.2201 | 0.7884 | 4500 | 3.2476 |
| 3.2328 | 0.8760 | 5000 | 3.2399 |
| 3.2056 | 0.9636 | 5500 | 3.2325 |
| 3.1739 | 1.0512 | 6000 | 3.2286 |
| 3.1632 | 1.1388 | 6500 | 3.2228 |
| 3.1633 | 1.2264 | 7000 | 3.2166 |
| 3.1675 | 1.3140 | 7500 | 3.2105 |
| 3.1314 | 1.4016 | 8000 | 3.2037 |
| 3.1463 | 1.4892 | 8500 | 3.1977 |
| 3.1327 | 1.5768 | 9000 | 3.1925 |
| 3.125 | 1.6644 | 9500 | 3.1866 |
| 3.1211 | 1.7520 | 10000 | 3.1822 |
| 3.1329 | 1.8396 | 10500 | 3.1781 |
| 3.1597 | 1.9272 | 11000 | 3.1746 |
| 3.1055 | 2.0147 | 11500 | 3.1727 |
| 3.0945 | 2.1023 | 12000 | 3.1712 |
| 3.1067 | 2.1899 | 12500 | 3.1698 |
| 3.077 | 2.2775 | 13000 | 3.1686 |
| 3.0829 | 2.3651 | 13500 | 3.1675 |
| 3.0893 | 2.4527 | 14000 | 3.1666 |
| 3.1241 | 2.5403 | 14500 | 3.1660 |
| 3.0701 | 2.6279 | 15000 | 3.1658 |
| 3.0995 | 2.7155 | 15500 | 3.1657 |
| 3.0706 | 2.8032 | 16000 | 3.1656 |
| 3.1012 | 2.8908 | 16500 | 3.1656 |
| 3.115 | 2.9784 | 17000 | 3.1656 |
Base model
CausalNLP/gpt2-hf_multilingual-20