| --- |
| language: |
| - ko |
| base_model: |
| - EleutherAI/polyglot-ko-1.3b |
| pipeline_tag: text-generation |
| tags: |
| - ingredient |
| - chatbot |
| - review |
| - usage |
| license: apache-2.0 |
| --- |
| |
| ## 🐥 Base Model |
|
|
| **EleutherAI/polyglot-ko-1.3b** |
|
|
| This model is based on the Korean version of Polyglot-1.3B, an open-source language model released by EleutherAI. |
| It is pre-trained on a large-scale Korean corpus and designed for general-purpose Korean language understanding and generation tasks. |
|
|
| --- |
|
|
| ## Training Procedure |
|
|
| ### Training Hyperparameters |
|
|
| The following hyperparameters were used during training: |
|
|
| - `output_dir`: `./qlora_model_eleutherai` |
| - `per_device_train_batch_size`: `2` |
| - `gradient_accumulation_steps`: `4` |
| - `total_batch_size`: `8 (2 x 4)` |
| - `learning_rate`: `2e-5` |
| - `num_train_epochs`: `2` |
| - `fp16`: `True` |
| - `logging_dir`: `./logs` |
| - `logging_steps`: `5` |
| - `save_steps`: `100` |
| - `save_total_limit`: `1` |
| - `load_best_model_at_end`: `True` |
| - `metric_for_best_model`: `loss` |