tempgraphrag-grpo / sft /checkpoint-490 /trainer_state.json
Guen's picture
sync 2026-04-24T07:52:39+00:00
68a4af1 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.0,
"eval_steps": 500,
"global_step": 490,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.04094165813715456,
"grad_norm": 7.529000759124756,
"learning_rate": 9.000000000000001e-07,
"loss": 6.639524841308594,
"step": 10
},
{
"epoch": 0.08188331627430911,
"grad_norm": 7.1427388191223145,
"learning_rate": 1.9000000000000002e-06,
"loss": 6.582770538330078,
"step": 20
},
{
"epoch": 0.12282497441146366,
"grad_norm": 6.251121997833252,
"learning_rate": 2.9e-06,
"loss": 6.114261245727539,
"step": 30
},
{
"epoch": 0.16376663254861823,
"grad_norm": 7.014133453369141,
"learning_rate": 3.900000000000001e-06,
"loss": 6.333962631225586,
"step": 40
},
{
"epoch": 0.2047082906857728,
"grad_norm": 7.072969913482666,
"learning_rate": 4.9000000000000005e-06,
"loss": 6.248543930053711,
"step": 50
},
{
"epoch": 0.24564994882292732,
"grad_norm": 5.674424171447754,
"learning_rate": 5.9e-06,
"loss": 5.6220745086669925,
"step": 60
},
{
"epoch": 0.2865916069600819,
"grad_norm": 5.3739800453186035,
"learning_rate": 6.9e-06,
"loss": 5.0251930236816404,
"step": 70
},
{
"epoch": 0.32753326509723646,
"grad_norm": 6.372293472290039,
"learning_rate": 7.9e-06,
"loss": 5.033017349243164,
"step": 80
},
{
"epoch": 0.368474923234391,
"grad_norm": 7.22106409072876,
"learning_rate": 8.900000000000001e-06,
"loss": 3.514456939697266,
"step": 90
},
{
"epoch": 0.4094165813715456,
"grad_norm": 4.990958213806152,
"learning_rate": 9.9e-06,
"loss": 2.256962776184082,
"step": 100
},
{
"epoch": 0.4503582395087001,
"grad_norm": 3.2661993503570557,
"learning_rate": 9.76923076923077e-06,
"loss": 2.0482555389404298,
"step": 110
},
{
"epoch": 0.49129989764585463,
"grad_norm": 2.997323513031006,
"learning_rate": 9.512820512820514e-06,
"loss": 1.6077747344970703,
"step": 120
},
{
"epoch": 0.5322415557830092,
"grad_norm": 2.422741413116455,
"learning_rate": 9.256410256410257e-06,
"loss": 1.538028621673584,
"step": 130
},
{
"epoch": 0.5731832139201638,
"grad_norm": 2.4594290256500244,
"learning_rate": 9e-06,
"loss": 1.5614049911499024,
"step": 140
},
{
"epoch": 0.6141248720573184,
"grad_norm": 2.7256577014923096,
"learning_rate": 8.743589743589743e-06,
"loss": 1.7180768966674804,
"step": 150
},
{
"epoch": 0.6550665301944729,
"grad_norm": 2.6614902019500732,
"learning_rate": 8.487179487179488e-06,
"loss": 1.6412044525146485,
"step": 160
},
{
"epoch": 0.6960081883316275,
"grad_norm": 2.1517934799194336,
"learning_rate": 8.230769230769232e-06,
"loss": 1.729467010498047,
"step": 170
},
{
"epoch": 0.736949846468782,
"grad_norm": 2.4588229656219482,
"learning_rate": 7.974358974358975e-06,
"loss": 1.305363941192627,
"step": 180
},
{
"epoch": 0.7778915046059366,
"grad_norm": 2.5282061100006104,
"learning_rate": 7.717948717948718e-06,
"loss": 1.4986873626708985,
"step": 190
},
{
"epoch": 0.8188331627430911,
"grad_norm": 3.699396848678589,
"learning_rate": 7.461538461538462e-06,
"loss": 1.5059691429138184,
"step": 200
},
{
"epoch": 0.8597748208802457,
"grad_norm": 2.2632241249084473,
"learning_rate": 7.205128205128206e-06,
"loss": 1.4249659538269044,
"step": 210
},
{
"epoch": 0.9007164790174002,
"grad_norm": 1.6525421142578125,
"learning_rate": 6.948717948717949e-06,
"loss": 1.2684237480163574,
"step": 220
},
{
"epoch": 0.9416581371545547,
"grad_norm": 4.144842624664307,
"learning_rate": 6.692307692307692e-06,
"loss": 1.3331901550292968,
"step": 230
},
{
"epoch": 0.9825997952917093,
"grad_norm": 4.287148475646973,
"learning_rate": 6.435897435897437e-06,
"loss": 1.3455743789672852,
"step": 240
},
{
"epoch": 1.0204708290685773,
"grad_norm": 1.7868481874465942,
"learning_rate": 6.17948717948718e-06,
"loss": 1.2180957794189453,
"step": 250
},
{
"epoch": 1.0614124872057318,
"grad_norm": 3.4123318195343018,
"learning_rate": 5.923076923076924e-06,
"loss": 1.0844655990600587,
"step": 260
},
{
"epoch": 1.1023541453428864,
"grad_norm": 2.9107143878936768,
"learning_rate": 5.666666666666667e-06,
"loss": 1.2759632110595702,
"step": 270
},
{
"epoch": 1.143295803480041,
"grad_norm": 2.166071653366089,
"learning_rate": 5.41025641025641e-06,
"loss": 1.2902444839477538,
"step": 280
},
{
"epoch": 1.1842374616171956,
"grad_norm": 3.7286741733551025,
"learning_rate": 5.1538461538461534e-06,
"loss": 1.1132170677185058,
"step": 290
},
{
"epoch": 1.22517911975435,
"grad_norm": 3.644495964050293,
"learning_rate": 4.8974358974358975e-06,
"loss": 1.3006702423095704,
"step": 300
},
{
"epoch": 1.2661207778915047,
"grad_norm": 2.9774715900421143,
"learning_rate": 4.641025641025642e-06,
"loss": 1.3277738571166993,
"step": 310
},
{
"epoch": 1.3070624360286591,
"grad_norm": 3.5428693294525146,
"learning_rate": 4.384615384615385e-06,
"loss": 1.2050466537475586,
"step": 320
},
{
"epoch": 1.3480040941658138,
"grad_norm": 2.9129176139831543,
"learning_rate": 4.128205128205128e-06,
"loss": 1.0705522537231444,
"step": 330
},
{
"epoch": 1.3889457523029682,
"grad_norm": 3.6323633193969727,
"learning_rate": 3.871794871794872e-06,
"loss": 1.1811490058898926,
"step": 340
},
{
"epoch": 1.429887410440123,
"grad_norm": 2.344156265258789,
"learning_rate": 3.6153846153846156e-06,
"loss": 1.2702527046203613,
"step": 350
},
{
"epoch": 1.4708290685772774,
"grad_norm": 2.93733286857605,
"learning_rate": 3.358974358974359e-06,
"loss": 1.2678390502929688,
"step": 360
},
{
"epoch": 1.511770726714432,
"grad_norm": 3.3147552013397217,
"learning_rate": 3.102564102564103e-06,
"loss": 0.975819206237793,
"step": 370
},
{
"epoch": 1.5527123848515865,
"grad_norm": 3.5217061042785645,
"learning_rate": 2.846153846153846e-06,
"loss": 1.5488268852233886,
"step": 380
},
{
"epoch": 1.593654042988741,
"grad_norm": 2.5279595851898193,
"learning_rate": 2.5897435897435903e-06,
"loss": 1.1411288261413575,
"step": 390
},
{
"epoch": 1.6345957011258956,
"grad_norm": 2.848078489303589,
"learning_rate": 2.3333333333333336e-06,
"loss": 1.1716268539428711,
"step": 400
},
{
"epoch": 1.6755373592630503,
"grad_norm": 2.8351571559906006,
"learning_rate": 2.0769230769230773e-06,
"loss": 1.139915370941162,
"step": 410
},
{
"epoch": 1.7164790174002047,
"grad_norm": 5.1406707763671875,
"learning_rate": 1.8205128205128205e-06,
"loss": 1.016810417175293,
"step": 420
},
{
"epoch": 1.7574206755373591,
"grad_norm": 2.3416874408721924,
"learning_rate": 1.5641025641025642e-06,
"loss": 1.0270769119262695,
"step": 430
},
{
"epoch": 1.7983623336745138,
"grad_norm": 2.4181466102600098,
"learning_rate": 1.307692307692308e-06,
"loss": 1.2266251564025878,
"step": 440
},
{
"epoch": 1.8393039918116685,
"grad_norm": 2.907052755355835,
"learning_rate": 1.0512820512820514e-06,
"loss": 1.1883393287658692,
"step": 450
},
{
"epoch": 1.880245649948823,
"grad_norm": 2.765097141265869,
"learning_rate": 7.948717948717949e-07,
"loss": 1.2259196281433105,
"step": 460
},
{
"epoch": 1.9211873080859774,
"grad_norm": 3.7463090419769287,
"learning_rate": 5.384615384615386e-07,
"loss": 1.016909694671631,
"step": 470
},
{
"epoch": 1.962128966223132,
"grad_norm": 3.9370245933532715,
"learning_rate": 2.820512820512821e-07,
"loss": 1.1455986976623536,
"step": 480
},
{
"epoch": 2.0,
"grad_norm": 4.772556304931641,
"learning_rate": 2.5641025641025643e-08,
"loss": 1.0747014999389648,
"step": 490
}
],
"logging_steps": 10,
"max_steps": 490,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 200,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.4297687441485824e+16,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}