Zip-1 / training_history.json
vdmbrsv's picture
Upload model
f30f9b2 verified
[
{
"epoch": 1,
"eval_results": {
"sts_spearman": 0.8289338745472172,
"sts_pearson": 0.8098611426918698,
"retrieval_recall_at_1": 0.428,
"retrieval_recall_at_5": 0.746,
"retrieval_recall_at_10": 0.85,
"nli_accuracy": 0.5,
"nli_similarity": 0.7193673849105835,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8657987713813782,
"composite_score": 0.7549336039402753
},
"losses": {
"sts": {
"total": 0.09194075802098149,
"distill": 0.07957400041429893,
"task": 0.2431019875018493,
"count": 23
},
"retrieval": {
"total": 0.268772747288359,
"distill": 0.0272021861945061,
"task": 1.241009996292439,
"count": 47
},
"nli": {
"total": 0.3314921881290192,
"distill": 0.01526121862549731,
"task": 2.3350987434387207,
"count": 47
},
"paraphrase": {
"total": 0.060653433203697205,
"distill": 0.008428953401744366,
"task": 0.8303535401821136,
"count": 10
}
},
"distill_weight": 0.3
},
{
"epoch": 2,
"eval_results": {
"sts_spearman": 0.8275112970530258,
"sts_pearson": 0.8086337285455611,
"retrieval_recall_at_1": 0.452,
"retrieval_recall_at_5": 0.78,
"retrieval_recall_at_10": 0.876,
"nli_accuracy": 0.5,
"nli_similarity": 0.7263807654380798,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8608207106590271,
"composite_score": 0.7644223151931796
},
"losses": {
"sts": {
"total": 0.05613501726285271,
"distill": 0.00838314470551584,
"task": 0.19135420996209848,
"count": 23
},
"retrieval": {
"total": 0.21614952480539362,
"distill": 0.006375962016271784,
"task": 1.019319458210722,
"count": 47
},
"nli": {
"total": 0.30814098867964235,
"distill": 0.004959808186964786,
"task": 2.188524269043131,
"count": 47
},
"paraphrase": {
"total": 0.051446066424250605,
"distill": 0.0038678635377436877,
"task": 0.7177851617336273,
"count": 10
}
},
"distill_weight": 0.2994
},
{
"epoch": 3,
"eval_results": {
"sts_spearman": 0.827182891132439,
"sts_pearson": 0.809022237364101,
"retrieval_recall_at_1": 0.47,
"retrieval_recall_at_5": 0.8,
"retrieval_recall_at_10": 0.9,
"nli_accuracy": 0.5,
"nli_similarity": 0.7274820804595947,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8593040704727173,
"composite_score": 0.7702581122328862
},
"losses": {
"sts": {
"total": 0.05474994104841481,
"distill": 0.0038318179709755855,
"task": 0.191118774206742,
"count": 23
},
"retrieval": {
"total": 0.1865766501807152,
"distill": 0.0033635680979870737,
"task": 0.8821620687525323,
"count": 47
},
"nli": {
"total": 0.2887607974574921,
"distill": 0.002587009018207801,
"task": 2.0535353625074344,
"count": 47
},
"paraphrase": {
"total": 0.04753798432648182,
"distill": 0.0023388142231851815,
"task": 0.6679855525493622,
"count": 10
}
},
"distill_weight": 0.2988
},
{
"epoch": 4,
"eval_results": {
"sts_spearman": 0.8273207309309314,
"sts_pearson": 0.8086748676612024,
"retrieval_recall_at_1": 0.496,
"retrieval_recall_at_5": 0.822,
"retrieval_recall_at_10": 0.914,
"nli_accuracy": 0.5,
"nli_similarity": 0.7310156226158142,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8595395684242249,
"composite_score": 0.7769270321321324
},
"losses": {
"sts": {
"total": 0.05435633513590564,
"distill": 0.0022464350635266824,
"task": 0.19124554292015408,
"count": 23
},
"retrieval": {
"total": 0.16951107297171938,
"distill": 0.0021174795668017356,
"task": 0.8021261704728958,
"count": 47
},
"nli": {
"total": 0.280746913019647,
"distill": 0.0015788413361309374,
"task": 1.9968374790029322,
"count": 47
},
"paraphrase": {
"total": 0.044596540927886966,
"distill": 0.0015972736524417996,
"task": 0.628672468662262,
"count": 10
}
},
"distill_weight": 0.29819999999999997
},
{
"epoch": 5,
"eval_results": {
"sts_spearman": 0.8252767207778248,
"sts_pearson": 0.8062386165350521,
"retrieval_recall_at_1": 0.51,
"retrieval_recall_at_5": 0.828,
"retrieval_recall_at_10": 0.922,
"nli_accuracy": 0.5,
"nli_similarity": 0.7306257486343384,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8570234775543213,
"composite_score": 0.7777050270555791
},
"losses": {
"sts": {
"total": 0.05338770879999451,
"distill": 0.0014984115867106163,
"task": 0.1884317449901415,
"count": 23
},
"retrieval": {
"total": 0.15480095338314137,
"distill": 0.0014994340983437414,
"task": 0.7325110093076178,
"count": 47
},
"nli": {
"total": 0.27232939100011866,
"distill": 0.0010938003134140944,
"task": 1.936246319020048,
"count": 47
},
"paraphrase": {
"total": 0.04112741462886334,
"distill": 0.0012359182350337506,
"task": 0.5802905261516571,
"count": 10
}
},
"distill_weight": 0.2976
},
{
"epoch": 6,
"eval_results": {
"sts_spearman": 0.8248752970132536,
"sts_pearson": 0.805502215995378,
"retrieval_recall_at_1": 0.524,
"retrieval_recall_at_5": 0.83,
"retrieval_recall_at_10": 0.922,
"nli_accuracy": 0.5,
"nli_similarity": 0.731809675693512,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8566122651100159,
"composite_score": 0.7781043151732935
},
"losses": {
"sts": {
"total": 0.05317179696715396,
"distill": 0.0011427759790145185,
"task": 0.1878819122262623,
"count": 23
},
"retrieval": {
"total": 0.14514635503292084,
"distill": 0.001184225406874209,
"task": 0.6865559283723223,
"count": 47
},
"nli": {
"total": 0.2656283996840741,
"distill": 0.0008388215606596242,
"task": 1.8874770808727184,
"count": 47
},
"paraphrase": {
"total": 0.038727284595370295,
"distill": 0.0010345598682761192,
"task": 0.546515229344368,
"count": 10
}
},
"distill_weight": 0.297
},
{
"epoch": 7,
"eval_results": {
"sts_spearman": 0.82402492098156,
"sts_pearson": 0.8039468293688954,
"retrieval_recall_at_1": 0.538,
"retrieval_recall_at_5": 0.842,
"retrieval_recall_at_10": 0.926,
"nli_accuracy": 0.5,
"nli_similarity": 0.7345654368400574,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8579738140106201,
"composite_score": 0.7812791271574466
},
"losses": {
"sts": {
"total": 0.052247073501348495,
"distill": 0.0009337503781906613,
"task": 0.18465858309165292,
"count": 23
},
"retrieval": {
"total": 0.13885294297274123,
"distill": 0.0010064611387280548,
"task": 0.6564081304884971,
"count": 47
},
"nli": {
"total": 0.2596762865147692,
"distill": 0.0006988341003616756,
"task": 1.8438684280882491,
"count": 47
},
"paraphrase": {
"total": 0.03767538573592901,
"distill": 0.0009226226538885385,
"task": 0.5315793305635452,
"count": 10
}
},
"distill_weight": 0.2964
},
{
"epoch": 8,
"eval_results": {
"sts_spearman": 0.823563554115123,
"sts_pearson": 0.8030950418925904,
"retrieval_recall_at_1": 0.558,
"retrieval_recall_at_5": 0.85,
"retrieval_recall_at_10": 0.936,
"nli_accuracy": 0.5,
"nli_similarity": 0.7343195080757141,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8582324385643005,
"composite_score": 0.7834484437242282
},
"losses": {
"sts": {
"total": 0.05065638410008472,
"distill": 0.0008207539668428185,
"task": 0.1789747496014056,
"count": 23
},
"retrieval": {
"total": 0.13189882991161753,
"distill": 0.0008978993072115044,
"task": 0.6230863763930949,
"count": 47
},
"nli": {
"total": 0.2552652054644646,
"distill": 0.0006161105479708218,
"task": 1.8111542945212507,
"count": 47
},
"paraphrase": {
"total": 0.03772428296506405,
"distill": 0.0008593377890065313,
"task": 0.5320944607257843,
"count": 10
}
},
"distill_weight": 0.2958
},
{
"epoch": 9,
"eval_results": {
"sts_spearman": 0.8224264234371386,
"sts_pearson": 0.801554516870732,
"retrieval_recall_at_1": 0.57,
"retrieval_recall_at_5": 0.86,
"retrieval_recall_at_10": 0.94,
"nli_accuracy": 0.5,
"nli_similarity": 0.733967125415802,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8588850498199463,
"composite_score": 0.785879878385236
},
"losses": {
"sts": {
"total": 0.049955147601988006,
"distill": 0.0007517727663861992,
"task": 0.17640900611877441,
"count": 23
},
"retrieval": {
"total": 0.12419923918044314,
"distill": 0.0008363095508452426,
"task": 0.5862294866683635,
"count": 47
},
"nli": {
"total": 0.2503235790323704,
"distill": 0.0005653487615029704,
"task": 1.7746644527354138,
"count": 47
},
"paraphrase": {
"total": 0.035123253054916856,
"distill": 0.0008159744320437312,
"task": 0.49492592811584474,
"count": 10
}
},
"distill_weight": 0.2952
},
{
"epoch": 10,
"eval_results": {
"sts_spearman": 0.8213443657508973,
"sts_pearson": 0.8000685657254155,
"retrieval_recall_at_1": 0.574,
"retrieval_recall_at_5": 0.876,
"retrieval_recall_at_10": 0.942,
"nli_accuracy": 0.5,
"nli_similarity": 0.7336843609809875,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8580488562583923,
"composite_score": 0.7901388495421153
},
"losses": {
"sts": {
"total": 0.050120412493529526,
"distill": 0.0007037564241529807,
"task": 0.1768963829330776,
"count": 23
},
"retrieval": {
"total": 0.11885773881952813,
"distill": 0.0007918142426283436,
"task": 0.5605541461325706,
"count": 47
},
"nli": {
"total": 0.24667784539943047,
"distill": 0.0005330673104370052,
"task": 1.7473829781755488,
"count": 47
},
"paraphrase": {
"total": 0.03359868098050356,
"distill": 0.0007893668138422072,
"task": 0.47301009893417356,
"count": 10
}
},
"distill_weight": 0.29460000000000003
},
{
"epoch": 11,
"eval_results": {
"sts_spearman": 0.8200398862029752,
"sts_pearson": 0.7984632150530062,
"retrieval_recall_at_1": 0.574,
"retrieval_recall_at_5": 0.88,
"retrieval_recall_at_10": 0.948,
"nli_accuracy": 0.5,
"nli_similarity": 0.7349022030830383,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8587563037872314,
"composite_score": 0.7906866097681543
},
"losses": {
"sts": {
"total": 0.049540048222179,
"distill": 0.0006752971321870775,
"task": 0.17472205732179724,
"count": 23
},
"retrieval": {
"total": 0.1152741235304386,
"distill": 0.0007609698123873231,
"task": 0.543203037469945,
"count": 47
},
"nli": {
"total": 0.2454570157730833,
"distill": 0.000510894448218986,
"task": 1.7373002945108618,
"count": 47
},
"paraphrase": {
"total": 0.032351000048220155,
"distill": 0.0007691650826018304,
"task": 0.4550263941287994,
"count": 10
}
},
"distill_weight": 0.29400000000000004
},
{
"epoch": 12,
"eval_results": {
"sts_spearman": 0.8193103945963425,
"sts_pearson": 0.7974055277518916,
"retrieval_recall_at_1": 0.594,
"retrieval_recall_at_5": 0.884,
"retrieval_recall_at_10": 0.948,
"nli_accuracy": 0.5,
"nli_similarity": 0.7340207695960999,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8586063981056213,
"composite_score": 0.7915218639648379
},
"losses": {
"sts": {
"total": 0.048518418131963066,
"distill": 0.0006525755262650225,
"task": 0.1709841172332349,
"count": 23
},
"retrieval": {
"total": 0.11415320951887901,
"distill": 0.0007357201898133659,
"task": 0.5374910330518763,
"count": 47
},
"nli": {
"total": 0.24104384951134947,
"distill": 0.0004927666555654179,
"task": 1.704636736119047,
"count": 47
},
"paraphrase": {
"total": 0.031325037218630315,
"distill": 0.0007606724102515727,
"task": 0.44016211330890653,
"count": 10
}
},
"distill_weight": 0.2934
},
{
"epoch": 13,
"eval_results": {
"sts_spearman": 0.8188842839819943,
"sts_pearson": 0.7964862034231038,
"retrieval_recall_at_1": 0.596,
"retrieval_recall_at_5": 0.896,
"retrieval_recall_at_10": 0.948,
"nli_accuracy": 0.5,
"nli_similarity": 0.7331050634384155,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8598613142967224,
"composite_score": 0.7949088086576639
},
"losses": {
"sts": {
"total": 0.047971061390379204,
"distill": 0.000634684164137782,
"task": 0.1689240109661351,
"count": 23
},
"retrieval": {
"total": 0.109825795159695,
"distill": 0.000719507896103599,
"task": 0.5166625177606623,
"count": 47
},
"nli": {
"total": 0.23657965850322804,
"distill": 0.00048042560786385326,
"task": 1.6716556980254802,
"count": 47
},
"paraphrase": {
"total": 0.030759319849312305,
"distill": 0.0007484613277483731,
"task": 0.43184628784656526,
"count": 10
}
},
"distill_weight": 0.2928
},
{
"epoch": 14,
"eval_results": {
"sts_spearman": 0.8170877062297013,
"sts_pearson": 0.794295936858975,
"retrieval_recall_at_1": 0.596,
"retrieval_recall_at_5": 0.896,
"retrieval_recall_at_10": 0.948,
"nli_accuracy": 0.5,
"nli_similarity": 0.7348426580429077,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8593454957008362,
"composite_score": 0.7940105197815173
},
"losses": {
"sts": {
"total": 0.04733733427913293,
"distill": 0.0006224965605803806,
"task": 0.16655636870342752,
"count": 23
},
"retrieval": {
"total": 0.10522193921373245,
"distill": 0.0007066282906886586,
"task": 0.49456279772393247,
"count": 47
},
"nli": {
"total": 0.23459308768840545,
"distill": 0.000471937776037908,
"task": 1.6562247961125476,
"count": 47
},
"paraphrase": {
"total": 0.02917077410966158,
"distill": 0.0007398281595669687,
"task": 0.40907877683639526,
"count": 10
}
},
"distill_weight": 0.2922
},
{
"epoch": 15,
"eval_results": {
"sts_spearman": 0.8157736839705673,
"sts_pearson": 0.7925540172374534,
"retrieval_recall_at_1": 0.606,
"retrieval_recall_at_5": 0.894,
"retrieval_recall_at_10": 0.952,
"nli_accuracy": 0.5,
"nli_similarity": 0.7331643104553223,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8580483794212341,
"composite_score": 0.7927535086519504
},
"losses": {
"sts": {
"total": 0.04638928920030594,
"distill": 0.0006136388072501059,
"task": 0.16308000878147458,
"count": 23
},
"retrieval": {
"total": 0.10471683328456068,
"distill": 0.000696288561696147,
"task": 0.49178333104924954,
"count": 47
},
"nli": {
"total": 0.23013398304898688,
"distill": 0.00046382398540253174,
"task": 1.6233676443708704,
"count": 47
},
"paraphrase": {
"total": 0.028761257790029048,
"distill": 0.0007339846633840352,
"task": 0.4029817461967468,
"count": 10
}
},
"distill_weight": 0.2916
},
{
"epoch": 16,
"eval_results": {
"sts_spearman": 0.8157617853858435,
"sts_pearson": 0.7923932840541673,
"retrieval_recall_at_1": 0.612,
"retrieval_recall_at_5": 0.898,
"retrieval_recall_at_10": 0.954,
"nli_accuracy": 0.5,
"nli_similarity": 0.7322084903717041,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8596447706222534,
"composite_score": 0.7939475593595884
},
"losses": {
"sts": {
"total": 0.04621984566683355,
"distill": 0.0006054732543618783,
"task": 0.1623542036699212,
"count": 23
},
"retrieval": {
"total": 0.10093737583845219,
"distill": 0.0006863788065024989,
"task": 0.473613737111396,
"count": 47
},
"nli": {
"total": 0.22668360205406837,
"distill": 0.00045654376067141904,
"task": 1.597678039936309,
"count": 47
},
"paraphrase": {
"total": 0.02780891638249159,
"distill": 0.0007290024077519774,
"task": 0.3892352104187012,
"count": 10
}
},
"distill_weight": 0.29100000000000004
},
{
"epoch": 17,
"eval_results": {
"sts_spearman": 0.8138228214142917,
"sts_pearson": 0.7899716079426576,
"retrieval_recall_at_1": 0.61,
"retrieval_recall_at_5": 0.9,
"retrieval_recall_at_10": 0.952,
"nli_accuracy": 0.5,
"nli_similarity": 0.7340738773345947,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8595521450042725,
"composite_score": 0.7935780773738126
},
"losses": {
"sts": {
"total": 0.04601614598346793,
"distill": 0.0006001048604957759,
"task": 0.1615060425322989,
"count": 23
},
"retrieval": {
"total": 0.09758770988976702,
"distill": 0.0006810528612596557,
"task": 0.4574874652192948,
"count": 47
},
"nli": {
"total": 0.22612765399699516,
"distill": 0.00045324116202190203,
"task": 1.5924184956449143,
"count": 47
},
"paraphrase": {
"total": 0.02633261661976576,
"distill": 0.0007273759925737977,
"task": 0.3681142464280128,
"count": 10
}
},
"distill_weight": 0.2904
},
{
"epoch": 18,
"eval_results": {
"sts_spearman": 0.8137941541347218,
"sts_pearson": 0.7900114938692636,
"retrieval_recall_at_1": 0.618,
"retrieval_recall_at_5": 0.902,
"retrieval_recall_at_10": 0.954,
"nli_accuracy": 0.5,
"nli_similarity": 0.7320045232772827,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8595419526100159,
"composite_score": 0.7941637437340276
},
"losses": {
"sts": {
"total": 0.04508868280960166,
"distill": 0.0005957435021865303,
"task": 0.15811052076194598,
"count": 23
},
"retrieval": {
"total": 0.09446902795040861,
"distill": 0.0006735043134540319,
"task": 0.4424755630340982,
"count": 47
},
"nli": {
"total": 0.2214733213186264,
"distill": 0.00044834198977580255,
"task": 1.5583173239484747,
"count": 47
},
"paraphrase": {
"total": 0.026047103106975555,
"distill": 0.0007218799262773245,
"task": 0.36381163746118544,
"count": 10
}
},
"distill_weight": 0.2898
},
{
"epoch": 19,
"eval_results": {
"sts_spearman": 0.8131324437136556,
"sts_pearson": 0.7888827667888834,
"retrieval_recall_at_1": 0.622,
"retrieval_recall_at_5": 0.906,
"retrieval_recall_at_10": 0.954,
"nli_accuracy": 0.5,
"nli_similarity": 0.7326275110244751,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8593710064888,
"composite_score": 0.7950328885234945
},
"losses": {
"sts": {
"total": 0.044820788761843804,
"distill": 0.0005908587026288328,
"task": 0.1570410553527915,
"count": 23
},
"retrieval": {
"total": 0.0947175715514954,
"distill": 0.0006690982015843087,
"task": 0.44327550555797335,
"count": 47
},
"nli": {
"total": 0.22196958103078476,
"distill": 0.00044533125981886654,
"task": 1.5605007993414046,
"count": 47
},
"paraphrase": {
"total": 0.02539514433592558,
"distill": 0.000717665534466505,
"task": 0.35435559451580045,
"count": 10
}
},
"distill_weight": 0.2892
},
{
"epoch": 20,
"eval_results": {
"sts_spearman": 0.8125557315718881,
"sts_pearson": 0.7883415732948076,
"retrieval_recall_at_1": 0.622,
"retrieval_recall_at_5": 0.906,
"retrieval_recall_at_10": 0.956,
"nli_accuracy": 0.5,
"nli_similarity": 0.7300553321838379,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8589670062065125,
"composite_score": 0.7947445324526108
},
"losses": {
"sts": {
"total": 0.04461472562473753,
"distill": 0.0005882442093697255,
"task": 0.15618835778340048,
"count": 23
},
"retrieval": {
"total": 0.09131274277225454,
"distill": 0.0006642562456111959,
"task": 0.4269564062991041,
"count": 47
},
"nli": {
"total": 0.21616570017439254,
"distill": 0.0004426293698030504,
"task": 1.518400065442349,
"count": 47
},
"paraphrase": {
"total": 0.024884730763733386,
"distill": 0.0007182472327258438,
"task": 0.3468856424093246,
"count": 10
}
},
"distill_weight": 0.2886
},
{
"epoch": 21,
"eval_results": {
"sts_spearman": 0.8124046923217468,
"sts_pearson": 0.7879715870379221,
"retrieval_recall_at_1": 0.626,
"retrieval_recall_at_5": 0.908,
"retrieval_recall_at_10": 0.954,
"nli_accuracy": 0.5,
"nli_similarity": 0.7299230694770813,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8608139157295227,
"composite_score": 0.7952690128275401
},
"losses": {
"sts": {
"total": 0.044788526128167694,
"distill": 0.0005869252971656945,
"task": 0.15666956746059915,
"count": 23
},
"retrieval": {
"total": 0.09129251508002585,
"distill": 0.0006613465486728447,
"task": 0.42650772028781,
"count": 47
},
"nli": {
"total": 0.21487902199968378,
"distill": 0.0004392804477631332,
"task": 1.5080934889773105,
"count": 47
},
"paraphrase": {
"total": 0.02408289248123765,
"distill": 0.0007130690268240869,
"task": 0.3353585585951805,
"count": 10
}
},
"distill_weight": 0.28800000000000003
},
{
"epoch": 22,
"eval_results": {
"sts_spearman": 0.8108396163396415,
"sts_pearson": 0.7859428479306386,
"retrieval_recall_at_1": 0.626,
"retrieval_recall_at_5": 0.91,
"retrieval_recall_at_10": 0.956,
"nli_accuracy": 0.5,
"nli_similarity": 0.7304312586784363,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8598558306694031,
"composite_score": 0.7950864748364874
},
"losses": {
"sts": {
"total": 0.04427984421667845,
"distill": 0.0005824335992498243,
"task": 0.15475882071515787,
"count": 23
},
"retrieval": {
"total": 0.08749476669633642,
"distill": 0.0006578407458406179,
"task": 0.40839041484163163,
"count": 47
},
"nli": {
"total": 0.21218161253218956,
"distill": 0.0004366559572171103,
"task": 1.487904317835544,
"count": 47
},
"paraphrase": {
"total": 0.02473886413499713,
"distill": 0.0007142506423406303,
"task": 0.34428276121616364,
"count": 10
}
},
"distill_weight": 0.2874
},
{
"epoch": 23,
"eval_results": {
"sts_spearman": 0.8101206106743032,
"sts_pearson": 0.7852421682051318,
"retrieval_recall_at_1": 0.626,
"retrieval_recall_at_5": 0.91,
"retrieval_recall_at_10": 0.956,
"nli_accuracy": 0.5,
"nli_similarity": 0.7298892140388489,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8599591255187988,
"composite_score": 0.7947269720038184
},
"losses": {
"sts": {
"total": 0.043497207696023193,
"distill": 0.0005803209623200414,
"task": 0.15188857848229614,
"count": 23
},
"retrieval": {
"total": 0.08651064019253914,
"distill": 0.0006545566213119062,
"task": 0.40345349781056666,
"count": 47
},
"nli": {
"total": 0.21238021647676508,
"distill": 0.00043464487014794126,
"task": 1.4880508184432983,
"count": 47
},
"paraphrase": {
"total": 0.022220892272889613,
"distill": 0.0007110251928679645,
"task": 0.30870682895183565,
"count": 10
}
},
"distill_weight": 0.2868
},
{
"epoch": 24,
"eval_results": {
"sts_spearman": 0.8090374451655163,
"sts_pearson": 0.783875236477,
"retrieval_recall_at_1": 0.64,
"retrieval_recall_at_5": 0.91,
"retrieval_recall_at_10": 0.956,
"nli_accuracy": 0.5,
"nli_similarity": 0.7307232618331909,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8604580163955688,
"composite_score": 0.7941853892494248
},
"losses": {
"sts": {
"total": 0.04361056018134822,
"distill": 0.0005793894398147645,
"task": 0.15216006111839545,
"count": 23
},
"retrieval": {
"total": 0.08273381509996475,
"distill": 0.0006529860733512868,
"task": 0.3854811432513785,
"count": 47
},
"nli": {
"total": 0.2092740849611607,
"distill": 0.00043218842759589724,
"task": 1.4650490512239172,
"count": 47
},
"paraphrase": {
"total": 0.02230049455538392,
"distill": 0.0007111194601748139,
"task": 0.30956813097000124,
"count": 10
}
},
"distill_weight": 0.2862
},
{
"epoch": 25,
"eval_results": {
"sts_spearman": 0.8088290083594672,
"sts_pearson": 0.7836111268009822,
"retrieval_recall_at_1": 0.642,
"retrieval_recall_at_5": 0.912,
"retrieval_recall_at_10": 0.958,
"nli_accuracy": 0.5,
"nli_similarity": 0.7291015386581421,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8604010939598083,
"composite_score": 0.7946811708464003
},
"losses": {
"sts": {
"total": 0.042825429821791855,
"distill": 0.0005763423598735877,
"task": 0.14928901389889096,
"count": 23
},
"retrieval": {
"total": 0.08428762876924048,
"distill": 0.0006495943922113548,
"task": 0.3924136973441915,
"count": 47
},
"nli": {
"total": 0.20725905958642352,
"distill": 0.0004320865307508552,
"task": 1.4497177093587024,
"count": 47
},
"paraphrase": {
"total": 0.021814299654215573,
"distill": 0.0007105717668309808,
"task": 0.30251066088676454,
"count": 10
}
},
"distill_weight": 0.2856
},
{
"epoch": 26,
"eval_results": {
"sts_spearman": 0.8086444673068466,
"sts_pearson": 0.7835424818834511,
"retrieval_recall_at_1": 0.65,
"retrieval_recall_at_5": 0.914,
"retrieval_recall_at_10": 0.96,
"nli_accuracy": 0.5,
"nli_similarity": 0.7279387712478638,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8617146015167236,
"composite_score": 0.79518890032009
},
"losses": {
"sts": {
"total": 0.04305816587546597,
"distill": 0.0005753715294818192,
"task": 0.14997966134029886,
"count": 23
},
"retrieval": {
"total": 0.08245786469667515,
"distill": 0.0006471946550671883,
"task": 0.3835590494439957,
"count": 47
},
"nli": {
"total": 0.20547382248208879,
"distill": 0.00042902108806958225,
"task": 1.4360247748963377,
"count": 47
},
"paraphrase": {
"total": 0.02213638899847865,
"distill": 0.0007105463359039277,
"task": 0.3067675843834877,
"count": 10
}
},
"distill_weight": 0.28500000000000003
},
{
"epoch": 27,
"eval_results": {
"sts_spearman": 0.8075945769025906,
"sts_pearson": 0.7817250896683322,
"retrieval_recall_at_1": 0.648,
"retrieval_recall_at_5": 0.916,
"retrieval_recall_at_10": 0.964,
"nli_accuracy": 0.5,
"nli_similarity": 0.7290590405464172,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8608205318450928,
"composite_score": 0.795263955117962
},
"losses": {
"sts": {
"total": 0.041979301558888474,
"distill": 0.0005742409064070038,
"task": 0.1460871498869813,
"count": 23
},
"retrieval": {
"total": 0.07854667179127957,
"distill": 0.0006465072064918089,
"task": 0.3650214504054252,
"count": 47
},
"nli": {
"total": 0.20393373991580718,
"distill": 0.00042719431060485227,
"task": 1.4240653768498848,
"count": 47
},
"paraphrase": {
"total": 0.02039573285728693,
"distill": 0.0007079597911797464,
"task": 0.28220218420028687,
"count": 10
}
},
"distill_weight": 0.2844
},
{
"epoch": 28,
"eval_results": {
"sts_spearman": 0.8072111574266839,
"sts_pearson": 0.7814816033631548,
"retrieval_recall_at_1": 0.656,
"retrieval_recall_at_5": 0.916,
"retrieval_recall_at_10": 0.962,
"nli_accuracy": 0.5,
"nli_similarity": 0.7277703881263733,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8611229062080383,
"composite_score": 0.7950722453800086
},
"losses": {
"sts": {
"total": 0.04250466094716736,
"distill": 0.0005732332761196986,
"task": 0.14780080836752188,
"count": 23
},
"retrieval": {
"total": 0.07828945341579457,
"distill": 0.0006427260170234962,
"task": 0.363525298681665,
"count": 47
},
"nli": {
"total": 0.20262091527593898,
"distill": 0.0004252295209184051,
"task": 1.4137128566173798,
"count": 47
},
"paraphrase": {
"total": 0.021627166587859393,
"distill": 0.000706591084599495,
"task": 0.29917111396789553,
"count": 10
}
},
"distill_weight": 0.2838
},
{
"epoch": 29,
"eval_results": {
"sts_spearman": 0.8061150812196531,
"sts_pearson": 0.7799263057126831,
"retrieval_recall_at_1": 0.654,
"retrieval_recall_at_5": 0.918,
"retrieval_recall_at_10": 0.964,
"nli_accuracy": 0.5,
"nli_similarity": 0.7274843454360962,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8605352640151978,
"composite_score": 0.7951242072764932
},
"losses": {
"sts": {
"total": 0.04178553026007569,
"distill": 0.0005705209104749172,
"task": 0.14517284346663434,
"count": 23
},
"retrieval": {
"total": 0.07806510272178244,
"distill": 0.0006415697936701806,
"task": 0.3621810408348733,
"count": 47
},
"nli": {
"total": 0.20029219636257659,
"distill": 0.00042426467754263825,
"task": 1.3962893257749842,
"count": 47
},
"paraphrase": {
"total": 0.020195775851607322,
"distill": 0.000706263561733067,
"task": 0.2789587274193764,
"count": 10
}
},
"distill_weight": 0.2832
},
{
"epoch": 30,
"eval_results": {
"sts_spearman": 0.805952102459493,
"sts_pearson": 0.7796381825168456,
"retrieval_recall_at_1": 0.658,
"retrieval_recall_at_5": 0.92,
"retrieval_recall_at_10": 0.966,
"nli_accuracy": 0.5,
"nli_similarity": 0.7271153926849365,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8611077666282654,
"composite_score": 0.7956427178964132
},
"losses": {
"sts": {
"total": 0.041891162486180016,
"distill": 0.0005706739384154587,
"task": 0.14542057948267978,
"count": 23
},
"retrieval": {
"total": 0.07668351699063118,
"distill": 0.0006396458459463208,
"task": 0.35546302129613594,
"count": 47
},
"nli": {
"total": 0.19826114177703857,
"distill": 0.00042267177598253686,
"task": 1.3809708052493157,
"count": 47
},
"paraphrase": {
"total": 0.019489498622715474,
"distill": 0.0007044955214951188,
"task": 0.268893338739872,
"count": 10
}
},
"distill_weight": 0.2826
},
{
"epoch": 31,
"eval_results": {
"sts_spearman": 0.8054395024616142,
"sts_pearson": 0.7790870358568797,
"retrieval_recall_at_1": 0.662,
"retrieval_recall_at_5": 0.92,
"retrieval_recall_at_10": 0.968,
"nli_accuracy": 0.5,
"nli_similarity": 0.7270572781562805,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8614012002944946,
"composite_score": 0.7953864178974739
},
"losses": {
"sts": {
"total": 0.04180093708893527,
"distill": 0.000569236811513648,
"task": 0.14498750833065613,
"count": 23
},
"retrieval": {
"total": 0.07472586061092133,
"distill": 0.0006379124688658308,
"task": 0.346081572644254,
"count": 47
},
"nli": {
"total": 0.19744586469011105,
"distill": 0.0004216810487745766,
"task": 1.3741430931902947,
"count": 47
},
"paraphrase": {
"total": 0.02020758679136634,
"distill": 0.0007039231946691871,
"task": 0.27867799401283266,
"count": 10
}
},
"distill_weight": 0.28200000000000003
},
{
"epoch": 32,
"eval_results": {
"sts_spearman": 0.8039734694214193,
"sts_pearson": 0.7774542134736094,
"retrieval_recall_at_1": 0.658,
"retrieval_recall_at_5": 0.922,
"retrieval_recall_at_10": 0.972,
"nli_accuracy": 0.5,
"nli_similarity": 0.7258086204528809,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8613040447235107,
"composite_score": 0.7952534013773764
},
"losses": {
"sts": {
"total": 0.04136973813824032,
"distill": 0.000568263144131102,
"task": 0.14336845570284387,
"count": 23
},
"retrieval": {
"total": 0.07349338850125353,
"distill": 0.000635797770921775,
"task": 0.3400801315586618,
"count": 47
},
"nli": {
"total": 0.195538673629152,
"distill": 0.00042041142513242333,
"task": 1.3597298099639568,
"count": 47
},
"paraphrase": {
"total": 0.01802230104804039,
"distill": 0.0007017374911811203,
"task": 0.24804942756891252,
"count": 10
}
},
"distill_weight": 0.2814
},
{
"epoch": 33,
"eval_results": {
"sts_spearman": 0.8034526879232323,
"sts_pearson": 0.7767872504658838,
"retrieval_recall_at_1": 0.664,
"retrieval_recall_at_5": 0.92,
"retrieval_recall_at_10": 0.972,
"nli_accuracy": 0.5,
"nli_similarity": 0.7267512083053589,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8609178066253662,
"composite_score": 0.7943930106282828
},
"losses": {
"sts": {
"total": 0.04124606368334397,
"distill": 0.0005672681145370007,
"task": 0.14282110203867374,
"count": 23
},
"retrieval": {
"total": 0.07269588295132556,
"distill": 0.0006344794367558937,
"task": 0.3361036536541391,
"count": 47
},
"nli": {
"total": 0.19442298532800473,
"distill": 0.0004187757124569505,
"task": 1.3508439393753702,
"count": 47
},
"paraphrase": {
"total": 0.01792764011770487,
"distill": 0.0007010513567365706,
"task": 0.246534825861454,
"count": 10
}
},
"distill_weight": 0.2808
},
{
"epoch": 34,
"eval_results": {
"sts_spearman": 0.8029147605057396,
"sts_pearson": 0.7758699525760386,
"retrieval_recall_at_1": 0.656,
"retrieval_recall_at_5": 0.924,
"retrieval_recall_at_10": 0.976,
"nli_accuracy": 0.5,
"nli_similarity": 0.7263163924217224,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.861183762550354,
"composite_score": 0.7953240469195365
},
"losses": {
"sts": {
"total": 0.04087469354271889,
"distill": 0.0005659387705078268,
"task": 0.14141468995291254,
"count": 23
},
"retrieval": {
"total": 0.07193009151106185,
"distill": 0.0006315128511174562,
"task": 0.3322827714554807,
"count": 47
},
"nli": {
"total": 0.19458080923303644,
"distill": 0.0004175955335550169,
"task": 1.3508182814780703,
"count": 47
},
"paraphrase": {
"total": 0.018750363681465387,
"distill": 0.0007004878658335656,
"task": 0.2577672630548477,
"count": 10
}
},
"distill_weight": 0.2802
},
{
"epoch": 35,
"eval_results": {
"sts_spearman": 0.8029120487600959,
"sts_pearson": 0.7759091259908452,
"retrieval_recall_at_1": 0.652,
"retrieval_recall_at_5": 0.924,
"retrieval_recall_at_10": 0.976,
"nli_accuracy": 0.5,
"nli_similarity": 0.7263759970664978,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8618313670158386,
"composite_score": 0.7953226910467146
},
"losses": {
"sts": {
"total": 0.040802220611468605,
"distill": 0.0005649596442589941,
"task": 0.1410475363549979,
"count": 23
},
"retrieval": {
"total": 0.07252203149998442,
"distill": 0.0006298948205670619,
"task": 0.33474880393515244,
"count": 47
},
"nli": {
"total": 0.19197505680804558,
"distill": 0.0004152855129932311,
"task": 1.3316139946592616,
"count": 47
},
"paraphrase": {
"total": 0.01693342700600624,
"distill": 0.0006995417177677154,
"task": 0.23234085589647294,
"count": 10
}
},
"distill_weight": 0.2796
},
{
"epoch": 36,
"eval_results": {
"sts_spearman": 0.802407902504706,
"sts_pearson": 0.7756569556734015,
"retrieval_recall_at_1": 0.658,
"retrieval_recall_at_5": 0.924,
"retrieval_recall_at_10": 0.974,
"nli_accuracy": 0.5,
"nli_similarity": 0.7243511080741882,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8609747290611267,
"composite_score": 0.7950706179190197
},
"losses": {
"sts": {
"total": 0.040637732200000595,
"distill": 0.0005651830908154016,
"task": 0.14036077325758728,
"count": 23
},
"retrieval": {
"total": 0.07090983135586089,
"distill": 0.0006288531424596588,
"task": 0.32701979894587335,
"count": 47
},
"nli": {
"total": 0.1879324754501911,
"distill": 0.0004152300212770066,
"task": 1.3024731620829155,
"count": 47
},
"paraphrase": {
"total": 0.01749492483213544,
"distill": 0.0006997698335908353,
"task": 0.23994021117687225,
"count": 10
}
},
"distill_weight": 0.279
},
{
"epoch": 37,
"eval_results": {
"sts_spearman": 0.8014589498925283,
"sts_pearson": 0.7745302218047226,
"retrieval_recall_at_1": 0.662,
"retrieval_recall_at_5": 0.926,
"retrieval_recall_at_10": 0.974,
"nli_accuracy": 0.5,
"nli_similarity": 0.7247843742370605,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8621292114257812,
"composite_score": 0.7951961416129308
},
"losses": {
"sts": {
"total": 0.0404453170688256,
"distill": 0.0005647512937329062,
"task": 0.13957903113054193,
"count": 23
},
"retrieval": {
"total": 0.07041878134329269,
"distill": 0.000626888732980699,
"task": 0.3244838112212242,
"count": 47
},
"nli": {
"total": 0.18824834677767247,
"distill": 0.00041336335856071175,
"task": 1.3035842015388164,
"count": 47
},
"paraphrase": {
"total": 0.017420530039817094,
"distill": 0.0006980633072089404,
"task": 0.23872213810682297,
"count": 10
}
},
"distill_weight": 0.2784
},
{
"epoch": 38,
"eval_results": {
"sts_spearman": 0.8015481972850798,
"sts_pearson": 0.7748134816243155,
"retrieval_recall_at_1": 0.662,
"retrieval_recall_at_5": 0.926,
"retrieval_recall_at_10": 0.974,
"nli_accuracy": 0.5,
"nli_similarity": 0.7238280177116394,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8617278337478638,
"composite_score": 0.7952407653092066
},
"losses": {
"sts": {
"total": 0.040147524246055145,
"distill": 0.000563749264034888,
"task": 0.13843435340601465,
"count": 23
},
"retrieval": {
"total": 0.06983992354349887,
"distill": 0.0006249398723779682,
"task": 0.3215467447930194,
"count": 47
},
"nli": {
"total": 0.18727405749736947,
"distill": 0.0004120995005731411,
"task": 1.2957600558057745,
"count": 47
},
"paraphrase": {
"total": 0.016873036976903677,
"distill": 0.0006976060452871024,
"task": 0.23095046430826188,
"count": 10
}
},
"distill_weight": 0.2778
},
{
"epoch": 39,
"eval_results": {
"sts_spearman": 0.8010951080044263,
"sts_pearson": 0.7744044551086896,
"retrieval_recall_at_1": 0.66,
"retrieval_recall_at_5": 0.926,
"retrieval_recall_at_10": 0.976,
"nli_accuracy": 0.5,
"nli_similarity": 0.7240303158760071,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8622973561286926,
"composite_score": 0.7950142206688798
},
"losses": {
"sts": {
"total": 0.040290426786827004,
"distill": 0.0005619644627744412,
"task": 0.1388165892466255,
"count": 23
},
"retrieval": {
"total": 0.06733930118857546,
"distill": 0.0006235166393378948,
"task": 0.3097512569833309,
"count": 47
},
"nli": {
"total": 0.18707014081325937,
"distill": 0.00041061582445028297,
"task": 1.29327839993416,
"count": 47
},
"paraphrase": {
"total": 0.017511206772178413,
"distill": 0.0006984172330703586,
"task": 0.23959056437015533,
"count": 10
}
},
"distill_weight": 0.2772
},
{
"epoch": 40,
"eval_results": {
"sts_spearman": 0.8008200971384819,
"sts_pearson": 0.774130829271817,
"retrieval_recall_at_1": 0.666,
"retrieval_recall_at_5": 0.932,
"retrieval_recall_at_10": 0.974,
"nli_accuracy": 0.5,
"nli_similarity": 0.7223304510116577,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.862885594367981,
"composite_score": 0.7966767152359077
},
"losses": {
"sts": {
"total": 0.039851526122378265,
"distill": 0.0005616477422375718,
"task": 0.1371861229772153,
"count": 23
},
"retrieval": {
"total": 0.0681393450879036,
"distill": 0.0006218473711843661,
"task": 0.3131846848954546,
"count": 47
},
"nli": {
"total": 0.18474397950984062,
"distill": 0.00041012915824738114,
"task": 1.2761304099509057,
"count": 47
},
"paraphrase": {
"total": 0.01582443034276366,
"distill": 0.0006965516076888889,
"task": 0.2160874292254448,
"count": 10
}
},
"distill_weight": 0.2766
},
{
"epoch": 41,
"eval_results": {
"sts_spearman": 0.8004397143136981,
"sts_pearson": 0.7737034113801735,
"retrieval_recall_at_1": 0.662,
"retrieval_recall_at_5": 0.93,
"retrieval_recall_at_10": 0.976,
"nli_accuracy": 0.5,
"nli_similarity": 0.7216349840164185,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8629198670387268,
"composite_score": 0.7958865238235158
},
"losses": {
"sts": {
"total": 0.040168939077335854,
"distill": 0.0005606161886016312,
"task": 0.1381706041486367,
"count": 23
},
"retrieval": {
"total": 0.06701801083189377,
"distill": 0.0006201347891003527,
"task": 0.3077663652440335,
"count": 47
},
"nli": {
"total": 0.18171446183894543,
"distill": 0.00040893261246581345,
"task": 1.254154608604756,
"count": 47
},
"paraphrase": {
"total": 0.016577236633747817,
"distill": 0.0006964895350392908,
"task": 0.2263122245669365,
"count": 10
}
},
"distill_weight": 0.276
},
{
"epoch": 42,
"eval_results": {
"sts_spearman": 0.799541139671189,
"sts_pearson": 0.7726619557363898,
"retrieval_recall_at_1": 0.664,
"retrieval_recall_at_5": 0.932,
"retrieval_recall_at_10": 0.976,
"nli_accuracy": 0.5,
"nli_similarity": 0.7218517065048218,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.862092912197113,
"composite_score": 0.7960372365022612
},
"losses": {
"sts": {
"total": 0.039756916949282524,
"distill": 0.0005605145656179799,
"task": 0.13663589921982391,
"count": 23
},
"retrieval": {
"total": 0.06787618566700752,
"distill": 0.0006185190620871776,
"task": 0.3114630852607971,
"count": 47
},
"nli": {
"total": 0.1818689741986863,
"distill": 0.0004076707292041008,
"task": 1.254186414657755,
"count": 47
},
"paraphrase": {
"total": 0.015548030100762843,
"distill": 0.0006947330140974373,
"task": 0.21193347945809365,
"count": 10
}
},
"distill_weight": 0.2754
},
{
"epoch": 43,
"eval_results": {
"sts_spearman": 0.798959430373989,
"sts_pearson": 0.7717844926047425,
"retrieval_recall_at_1": 0.658,
"retrieval_recall_at_5": 0.934,
"retrieval_recall_at_10": 0.976,
"nli_accuracy": 0.5,
"nli_similarity": 0.7230818867683411,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8633172512054443,
"composite_score": 0.7963463818536611
},
"losses": {
"sts": {
"total": 0.03980643334596053,
"distill": 0.0005611540931884361,
"task": 0.13669410898633624,
"count": 23
},
"retrieval": {
"total": 0.0670052819905129,
"distill": 0.0006171439581134535,
"task": 0.3072057971928982,
"count": 47
},
"nli": {
"total": 0.181170218168421,
"distill": 0.0004060288384665755,
"task": 1.2483358712906534,
"count": 47
},
"paraphrase": {
"total": 0.016034624353051186,
"distill": 0.0006926223228219897,
"task": 0.21848167926073075,
"count": 10
}
},
"distill_weight": 0.2748
},
{
"epoch": 44,
"eval_results": {
"sts_spearman": 0.7992766546129851,
"sts_pearson": 0.7723631373159269,
"retrieval_recall_at_1": 0.668,
"retrieval_recall_at_5": 0.936,
"retrieval_recall_at_10": 0.976,
"nli_accuracy": 0.5,
"nli_similarity": 0.7214555740356445,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8633242845535278,
"composite_score": 0.7971049939731593
},
"losses": {
"sts": {
"total": 0.03956834745147954,
"distill": 0.0005598463517937647,
"task": 0.13576342230257782,
"count": 23
},
"retrieval": {
"total": 0.0675140564587522,
"distill": 0.0006155266111934597,
"task": 0.30929218010699494,
"count": 47
},
"nli": {
"total": 0.18059916984527669,
"distill": 0.00040597010755594426,
"task": 1.2433717808824905,
"count": 47
},
"paraphrase": {
"total": 0.016109541151672603,
"distill": 0.000693465251242742,
"task": 0.21933580189943314,
"count": 10
}
},
"distill_weight": 0.2742
},
{
"epoch": 45,
"eval_results": {
"sts_spearman": 0.7987902718942615,
"sts_pearson": 0.7716241234848017,
"retrieval_recall_at_1": 0.67,
"retrieval_recall_at_5": 0.934,
"retrieval_recall_at_10": 0.976,
"nli_accuracy": 0.5,
"nli_similarity": 0.7223896384239197,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8632831573486328,
"composite_score": 0.7962618026137975
},
"losses": {
"sts": {
"total": 0.03940606570762137,
"distill": 0.0005587598197567074,
"task": 0.13509494876084122,
"count": 23
},
"retrieval": {
"total": 0.06665098445212587,
"distill": 0.0006136801812124062,
"task": 0.3050802118600683,
"count": 47
},
"nli": {
"total": 0.18052284007376812,
"distill": 0.00040472789662592906,
"task": 1.2418234031251136,
"count": 47
},
"paraphrase": {
"total": 0.015592311229556798,
"distill": 0.0006917931721545755,
"task": 0.21204619854688644,
"count": 10
}
},
"distill_weight": 0.2736
},
{
"epoch": 46,
"eval_results": {
"sts_spearman": 0.7980845899126008,
"sts_pearson": 0.7708174356684444,
"retrieval_recall_at_1": 0.662,
"retrieval_recall_at_5": 0.934,
"retrieval_recall_at_10": 0.974,
"nli_accuracy": 0.5,
"nli_similarity": 0.7227649688720703,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8638308644294739,
"composite_score": 0.7959089616229671
},
"losses": {
"sts": {
"total": 0.03914242918076723,
"distill": 0.0005580810273228133,
"task": 0.13407865684965384,
"count": 23
},
"retrieval": {
"total": 0.06371399538314089,
"distill": 0.0006122274517497801,
"task": 0.2913656944924213,
"count": 47
},
"nli": {
"total": 0.1788206826499168,
"distill": 0.00040350851176821804,
"task": 1.2290957481303113,
"count": 47
},
"paraphrase": {
"total": 0.016117287054657935,
"distill": 0.0006897407933138311,
"task": 0.2191057413816452,
"count": 10
}
},
"distill_weight": 0.273
},
{
"epoch": 47,
"eval_results": {
"sts_spearman": 0.7986050030831314,
"sts_pearson": 0.7713344721815415,
"retrieval_recall_at_1": 0.668,
"retrieval_recall_at_5": 0.936,
"retrieval_recall_at_10": 0.974,
"nli_accuracy": 0.5,
"nli_similarity": 0.7211058139801025,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8641474843025208,
"composite_score": 0.7967691682082324
},
"losses": {
"sts": {
"total": 0.03924534291676853,
"distill": 0.0005579624380713896,
"task": 0.1343229581480441,
"count": 23
},
"retrieval": {
"total": 0.06390803307294846,
"distill": 0.0006106501616081817,
"task": 0.29201801152939494,
"count": 47
},
"nli": {
"total": 0.17669346231095334,
"distill": 0.0004022118878213966,
"task": 1.2134682447352307,
"count": 47
},
"paraphrase": {
"total": 0.014580465480685234,
"distill": 0.0006913242454174906,
"task": 0.1978030323982239,
"count": 10
}
},
"distill_weight": 0.2724
},
{
"epoch": 48,
"eval_results": {
"sts_spearman": 0.7978328086000539,
"sts_pearson": 0.7705122749494036,
"retrieval_recall_at_1": 0.67,
"retrieval_recall_at_5": 0.936,
"retrieval_recall_at_10": 0.976,
"nli_accuracy": 0.5,
"nli_similarity": 0.7213853001594543,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8641984462738037,
"composite_score": 0.7963830709666937
},
"losses": {
"sts": {
"total": 0.038881031796336174,
"distill": 0.0005576102867843988,
"task": 0.13296303638945456,
"count": 23
},
"retrieval": {
"total": 0.06267972861198669,
"distill": 0.0006095131591675764,
"task": 0.2861579453691523,
"count": 47
},
"nli": {
"total": 0.17544851499669095,
"distill": 0.0004012782276518881,
"task": 1.2039236971672544,
"count": 47
},
"paraphrase": {
"total": 0.015775612369179726,
"distill": 0.0006886643706820906,
"task": 0.2140680193901062,
"count": 10
}
},
"distill_weight": 0.2718
},
{
"epoch": 49,
"eval_results": {
"sts_spearman": 0.7973172210556293,
"sts_pearson": 0.7701713894839444,
"retrieval_recall_at_1": 0.67,
"retrieval_recall_at_5": 0.936,
"retrieval_recall_at_10": 0.976,
"nli_accuracy": 0.5,
"nli_similarity": 0.7203776240348816,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8631545305252075,
"composite_score": 0.7961252771944813
},
"losses": {
"sts": {
"total": 0.03882810230488363,
"distill": 0.0005557498307493718,
"task": 0.13267488712849823,
"count": 23
},
"retrieval": {
"total": 0.0626571145146451,
"distill": 0.000607952565955434,
"task": 0.28582252847387435,
"count": 47
},
"nli": {
"total": 0.1763826126747943,
"distill": 0.0004007951009701541,
"task": 1.2093435627348879,
"count": 47
},
"paraphrase": {
"total": 0.015177728608250618,
"distill": 0.0006882219109684229,
"task": 0.20569542646408082,
"count": 10
}
},
"distill_weight": 0.2712
},
{
"epoch": 50,
"eval_results": {
"sts_spearman": 0.7968594218073127,
"sts_pearson": 0.7695483662129042,
"retrieval_recall_at_1": 0.67,
"retrieval_recall_at_5": 0.936,
"retrieval_recall_at_10": 0.974,
"nli_accuracy": 0.5,
"nli_similarity": 0.7204389572143555,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8636235594749451,
"composite_score": 0.795896377570323
},
"losses": {
"sts": {
"total": 0.03918402480042499,
"distill": 0.0005560298273137406,
"task": 0.13378654711920282,
"count": 23
},
"retrieval": {
"total": 0.06285070390143294,
"distill": 0.000605922420211929,
"task": 0.2864762779245985,
"count": 47
},
"nli": {
"total": 0.1742387483728693,
"distill": 0.00039892817941553733,
"task": 1.1936577951654475,
"count": 47
},
"paraphrase": {
"total": 0.014562566205859185,
"distill": 0.0006869392120279372,
"task": 0.197102826833725,
"count": 10
}
},
"distill_weight": 0.2706
},
{
"epoch": 51,
"eval_results": {
"sts_spearman": 0.7961018874767011,
"sts_pearson": 0.7685896198859151,
"retrieval_recall_at_1": 0.672,
"retrieval_recall_at_5": 0.938,
"retrieval_recall_at_10": 0.976,
"nli_accuracy": 0.5,
"nli_similarity": 0.7212554216384888,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8639085292816162,
"composite_score": 0.7961176104050172
},
"losses": {
"sts": {
"total": 0.038832955305343086,
"distill": 0.000555693042849231,
"task": 0.13247574898211853,
"count": 23
},
"retrieval": {
"total": 0.061906919121108156,
"distill": 0.0006047165407700107,
"task": 0.2819344585246228,
"count": 47
},
"nli": {
"total": 0.17475084611710082,
"distill": 0.0003985753827827408,
"task": 1.1961865387064345,
"count": 47
},
"paraphrase": {
"total": 0.014561035577207804,
"distill": 0.0006853513652458787,
"task": 0.19693138003349303,
"count": 10
}
},
"distill_weight": 0.27
},
{
"epoch": 52,
"eval_results": {
"sts_spearman": 0.7957712004157503,
"sts_pearson": 0.7682444812474918,
"retrieval_recall_at_1": 0.668,
"retrieval_recall_at_5": 0.938,
"retrieval_recall_at_10": 0.976,
"nli_accuracy": 0.5,
"nli_similarity": 0.7211222052574158,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8639959096908569,
"composite_score": 0.7959522668745418
},
"losses": {
"sts": {
"total": 0.038104739849982056,
"distill": 0.0005552242105097874,
"task": 0.12987668682699618,
"count": 23
},
"retrieval": {
"total": 0.06224994496145147,
"distill": 0.0006037014598482625,
"task": 0.2832708587037756,
"count": 47
},
"nli": {
"total": 0.17524569560872746,
"distill": 0.00039716928756419333,
"task": 1.198594996269713,
"count": 47
},
"paraphrase": {
"total": 0.01391328014433384,
"distill": 0.0006847139564342796,
"task": 0.18791155964136125,
"count": 10
}
},
"distill_weight": 0.2694
},
{
"epoch": 53,
"eval_results": {
"sts_spearman": 0.7948627602870467,
"sts_pearson": 0.7672766924578068,
"retrieval_recall_at_1": 0.674,
"retrieval_recall_at_5": 0.936,
"retrieval_recall_at_10": 0.974,
"nli_accuracy": 0.5,
"nli_similarity": 0.7208473682403564,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8635191321372986,
"composite_score": 0.79489804681019
},
"losses": {
"sts": {
"total": 0.03807506864161595,
"distill": 0.0005540588349306389,
"task": 0.12967087879129077,
"count": 23
},
"retrieval": {
"total": 0.06309370720323096,
"distill": 0.0006028104070673122,
"task": 0.2868876425509757,
"count": 47
},
"nli": {
"total": 0.17431603213574023,
"distill": 0.00039660765918249146,
"task": 1.1912570329422647,
"count": 47
},
"paraphrase": {
"total": 0.013855330273509025,
"distill": 0.0006855725019704551,
"task": 0.18696729764342307,
"count": 10
}
},
"distill_weight": 0.2688
},
{
"epoch": 54,
"eval_results": {
"sts_spearman": 0.7948670645604796,
"sts_pearson": 0.7673114563945611,
"retrieval_recall_at_1": 0.672,
"retrieval_recall_at_5": 0.936,
"retrieval_recall_at_10": 0.976,
"nli_accuracy": 0.5,
"nli_similarity": 0.7204586267471313,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8637556433677673,
"composite_score": 0.7949001989469064
},
"losses": {
"sts": {
"total": 0.03859076422193776,
"distill": 0.000553890953913493,
"task": 0.1313275915125142,
"count": 23
},
"retrieval": {
"total": 0.06243176108345072,
"distill": 0.000601062821125255,
"task": 0.2836410492024523,
"count": 47
},
"nli": {
"total": 0.17191652573169547,
"distill": 0.0003955346752612039,
"task": 1.173889383356622,
"count": 47
},
"paraphrase": {
"total": 0.014081115927547216,
"distill": 0.0006832460989244282,
"task": 0.1899135023355484,
"count": 10
}
},
"distill_weight": 0.2682
},
{
"epoch": 55,
"eval_results": {
"sts_spearman": 0.7940184198467072,
"sts_pearson": 0.7664062690593959,
"retrieval_recall_at_1": 0.67,
"retrieval_recall_at_5": 0.936,
"retrieval_recall_at_10": 0.974,
"nli_accuracy": 0.5,
"nli_similarity": 0.7194271087646484,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8639047145843506,
"composite_score": 0.7944758765900203
},
"losses": {
"sts": {
"total": 0.03801367480469787,
"distill": 0.0005527152021860946,
"task": 0.12925235538379007,
"count": 23
},
"retrieval": {
"total": 0.06092823788206628,
"distill": 0.0005999853217756336,
"task": 0.276568721900595,
"count": 47
},
"nli": {
"total": 0.1702281376148792,
"distill": 0.0003946685254514376,
"task": 1.161404493007254,
"count": 47
},
"paraphrase": {
"total": 0.012735517043620349,
"distill": 0.0006815990433096886,
"task": 0.17139707654714584,
"count": 10
}
},
"distill_weight": 0.2676
},
{
"epoch": 56,
"eval_results": {
"sts_spearman": 0.7940615711220378,
"sts_pearson": 0.7665661594709636,
"retrieval_recall_at_1": 0.672,
"retrieval_recall_at_5": 0.94,
"retrieval_recall_at_10": 0.976,
"nli_accuracy": 0.5,
"nli_similarity": 0.7197932004928589,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8638368248939514,
"composite_score": 0.7956974522276856
},
"losses": {
"sts": {
"total": 0.03830795994271403,
"distill": 0.0005517558893188834,
"task": 0.13015226015578146,
"count": 23
},
"retrieval": {
"total": 0.060107594633356055,
"distill": 0.0005995195837037519,
"task": 0.27261265953804587,
"count": 47
},
"nli": {
"total": 0.1709233709472291,
"distill": 0.0003937529740567775,
"task": 1.165199505521896,
"count": 47
},
"paraphrase": {
"total": 0.014191682077944278,
"distill": 0.0006824985903222114,
"task": 0.19112490341067315,
"count": 10
}
},
"distill_weight": 0.267
},
{
"epoch": 57,
"eval_results": {
"sts_spearman": 0.7937036029034699,
"sts_pearson": 0.7662776909565029,
"retrieval_recall_at_1": 0.672,
"retrieval_recall_at_5": 0.942,
"retrieval_recall_at_10": 0.978,
"nli_accuracy": 0.5,
"nli_similarity": 0.7196366190910339,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.864041268825531,
"composite_score": 0.7961184681184016
},
"losses": {
"sts": {
"total": 0.038414100913897804,
"distill": 0.0005509398837128411,
"task": 0.1304093819597493,
"count": 23
},
"retrieval": {
"total": 0.057978192938769116,
"distill": 0.0005975660850650611,
"task": 0.26271810398456896,
"count": 47
},
"nli": {
"total": 0.17013170490873622,
"distill": 0.00039284772232194687,
"task": 1.1588538897798417,
"count": 47
},
"paraphrase": {
"total": 0.014643667824566364,
"distill": 0.0006814012362156063,
"task": 0.19713934063911437,
"count": 10
}
},
"distill_weight": 0.26639999999999997
},
{
"epoch": 58,
"eval_results": {
"sts_spearman": 0.7928266663552261,
"sts_pearson": 0.765244563485406,
"retrieval_recall_at_1": 0.678,
"retrieval_recall_at_5": 0.94,
"retrieval_recall_at_10": 0.976,
"nli_accuracy": 0.5,
"nli_similarity": 0.7203112840652466,
"paraphrase_accuracy": 0.5,
"paraphrase_f1": 0.6666666666666666,
"paraphrase_similarity": 0.8644272685050964,
"composite_score": 0.7950799998442797
},
"losses": {
"sts": {
"total": 0.03789730612998423,
"distill": 0.0005501627937242713,
"task": 0.1285449188688527,
"count": 23
},
"retrieval": {
"total": 0.06108237723720834,
"distill": 0.0005970121710561216,
"task": 0.27659898869534755,
"count": 47
},
"nli": {
"total": 0.1709375616083754,
"distill": 0.0003915692932230044,
"task": 1.1633987756485635,
"count": 47
},
"paraphrase": {
"total": 0.013374552130699158,
"distill": 0.0006797625974286348,
"task": 0.17970404177904128,
"count": 10
}
},
"distill_weight": 0.26580000000000004
}
]