DreamWideWebWorld
Collection
Web Agent Foundation Model Series • 2 items • Updated
This model is a fine-tuned version of Meta-Llama-3-8B-Instruct on the web_policy_sft dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
| Training Loss | Epoch | Step | Validation Loss |
|---|---|---|---|
| 1.0437 | 0.1878 | 50 | 1.0606 |
| 0.5727 | 0.3756 | 100 | 0.5909 |
| 0.5279 | 0.5634 | 150 | 0.4970 |
| 0.4961 | 0.7512 | 200 | 0.4700 |
| 0.4797 | 0.9390 | 250 | 0.4632 |
Base model
meta-llama/Meta-Llama-3-8B-Instruct