| [ |
| { |
| "epoch": 1, |
| "eval_results": { |
| "sts_spearman": 0.8289338745472172, |
| "sts_pearson": 0.8098611426918698, |
| "retrieval_recall_at_1": 0.428, |
| "retrieval_recall_at_5": 0.746, |
| "retrieval_recall_at_10": 0.85, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7193673849105835, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8657987713813782, |
| "composite_score": 0.7549336039402753 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.09194075802098149, |
| "distill": 0.07957400041429893, |
| "task": 0.2431019875018493, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.268772747288359, |
| "distill": 0.0272021861945061, |
| "task": 1.241009996292439, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.3314921881290192, |
| "distill": 0.01526121862549731, |
| "task": 2.3350987434387207, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.060653433203697205, |
| "distill": 0.008428953401744366, |
| "task": 0.8303535401821136, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.3 |
| }, |
| { |
| "epoch": 2, |
| "eval_results": { |
| "sts_spearman": 0.8275112970530258, |
| "sts_pearson": 0.8086337285455611, |
| "retrieval_recall_at_1": 0.452, |
| "retrieval_recall_at_5": 0.78, |
| "retrieval_recall_at_10": 0.876, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7263807654380798, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8608207106590271, |
| "composite_score": 0.7644223151931796 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.05613501726285271, |
| "distill": 0.00838314470551584, |
| "task": 0.19135420996209848, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.21614952480539362, |
| "distill": 0.006375962016271784, |
| "task": 1.019319458210722, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.30814098867964235, |
| "distill": 0.004959808186964786, |
| "task": 2.188524269043131, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.051446066424250605, |
| "distill": 0.0038678635377436877, |
| "task": 0.7177851617336273, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2994 |
| }, |
| { |
| "epoch": 3, |
| "eval_results": { |
| "sts_spearman": 0.827182891132439, |
| "sts_pearson": 0.809022237364101, |
| "retrieval_recall_at_1": 0.47, |
| "retrieval_recall_at_5": 0.8, |
| "retrieval_recall_at_10": 0.9, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7274820804595947, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8593040704727173, |
| "composite_score": 0.7702581122328862 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.05474994104841481, |
| "distill": 0.0038318179709755855, |
| "task": 0.191118774206742, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.1865766501807152, |
| "distill": 0.0033635680979870737, |
| "task": 0.8821620687525323, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.2887607974574921, |
| "distill": 0.002587009018207801, |
| "task": 2.0535353625074344, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.04753798432648182, |
| "distill": 0.0023388142231851815, |
| "task": 0.6679855525493622, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2988 |
| }, |
| { |
| "epoch": 4, |
| "eval_results": { |
| "sts_spearman": 0.8273207309309314, |
| "sts_pearson": 0.8086748676612024, |
| "retrieval_recall_at_1": 0.496, |
| "retrieval_recall_at_5": 0.822, |
| "retrieval_recall_at_10": 0.914, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7310156226158142, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8595395684242249, |
| "composite_score": 0.7769270321321324 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.05435633513590564, |
| "distill": 0.0022464350635266824, |
| "task": 0.19124554292015408, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.16951107297171938, |
| "distill": 0.0021174795668017356, |
| "task": 0.8021261704728958, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.280746913019647, |
| "distill": 0.0015788413361309374, |
| "task": 1.9968374790029322, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.044596540927886966, |
| "distill": 0.0015972736524417996, |
| "task": 0.628672468662262, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.29819999999999997 |
| }, |
| { |
| "epoch": 5, |
| "eval_results": { |
| "sts_spearman": 0.8252767207778248, |
| "sts_pearson": 0.8062386165350521, |
| "retrieval_recall_at_1": 0.51, |
| "retrieval_recall_at_5": 0.828, |
| "retrieval_recall_at_10": 0.922, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7306257486343384, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8570234775543213, |
| "composite_score": 0.7777050270555791 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.05338770879999451, |
| "distill": 0.0014984115867106163, |
| "task": 0.1884317449901415, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.15480095338314137, |
| "distill": 0.0014994340983437414, |
| "task": 0.7325110093076178, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.27232939100011866, |
| "distill": 0.0010938003134140944, |
| "task": 1.936246319020048, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.04112741462886334, |
| "distill": 0.0012359182350337506, |
| "task": 0.5802905261516571, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2976 |
| }, |
| { |
| "epoch": 6, |
| "eval_results": { |
| "sts_spearman": 0.8248752970132536, |
| "sts_pearson": 0.805502215995378, |
| "retrieval_recall_at_1": 0.524, |
| "retrieval_recall_at_5": 0.83, |
| "retrieval_recall_at_10": 0.922, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.731809675693512, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8566122651100159, |
| "composite_score": 0.7781043151732935 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.05317179696715396, |
| "distill": 0.0011427759790145185, |
| "task": 0.1878819122262623, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.14514635503292084, |
| "distill": 0.001184225406874209, |
| "task": 0.6865559283723223, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.2656283996840741, |
| "distill": 0.0008388215606596242, |
| "task": 1.8874770808727184, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.038727284595370295, |
| "distill": 0.0010345598682761192, |
| "task": 0.546515229344368, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.297 |
| }, |
| { |
| "epoch": 7, |
| "eval_results": { |
| "sts_spearman": 0.82402492098156, |
| "sts_pearson": 0.8039468293688954, |
| "retrieval_recall_at_1": 0.538, |
| "retrieval_recall_at_5": 0.842, |
| "retrieval_recall_at_10": 0.926, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7345654368400574, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8579738140106201, |
| "composite_score": 0.7812791271574466 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.052247073501348495, |
| "distill": 0.0009337503781906613, |
| "task": 0.18465858309165292, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.13885294297274123, |
| "distill": 0.0010064611387280548, |
| "task": 0.6564081304884971, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.2596762865147692, |
| "distill": 0.0006988341003616756, |
| "task": 1.8438684280882491, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.03767538573592901, |
| "distill": 0.0009226226538885385, |
| "task": 0.5315793305635452, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2964 |
| }, |
| { |
| "epoch": 8, |
| "eval_results": { |
| "sts_spearman": 0.823563554115123, |
| "sts_pearson": 0.8030950418925904, |
| "retrieval_recall_at_1": 0.558, |
| "retrieval_recall_at_5": 0.85, |
| "retrieval_recall_at_10": 0.936, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7343195080757141, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8582324385643005, |
| "composite_score": 0.7834484437242282 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.05065638410008472, |
| "distill": 0.0008207539668428185, |
| "task": 0.1789747496014056, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.13189882991161753, |
| "distill": 0.0008978993072115044, |
| "task": 0.6230863763930949, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.2552652054644646, |
| "distill": 0.0006161105479708218, |
| "task": 1.8111542945212507, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.03772428296506405, |
| "distill": 0.0008593377890065313, |
| "task": 0.5320944607257843, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2958 |
| }, |
| { |
| "epoch": 9, |
| "eval_results": { |
| "sts_spearman": 0.8224264234371386, |
| "sts_pearson": 0.801554516870732, |
| "retrieval_recall_at_1": 0.57, |
| "retrieval_recall_at_5": 0.86, |
| "retrieval_recall_at_10": 0.94, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.733967125415802, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8588850498199463, |
| "composite_score": 0.785879878385236 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.049955147601988006, |
| "distill": 0.0007517727663861992, |
| "task": 0.17640900611877441, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.12419923918044314, |
| "distill": 0.0008363095508452426, |
| "task": 0.5862294866683635, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.2503235790323704, |
| "distill": 0.0005653487615029704, |
| "task": 1.7746644527354138, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.035123253054916856, |
| "distill": 0.0008159744320437312, |
| "task": 0.49492592811584474, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2952 |
| }, |
| { |
| "epoch": 10, |
| "eval_results": { |
| "sts_spearman": 0.8213443657508973, |
| "sts_pearson": 0.8000685657254155, |
| "retrieval_recall_at_1": 0.574, |
| "retrieval_recall_at_5": 0.876, |
| "retrieval_recall_at_10": 0.942, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7336843609809875, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8580488562583923, |
| "composite_score": 0.7901388495421153 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.050120412493529526, |
| "distill": 0.0007037564241529807, |
| "task": 0.1768963829330776, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.11885773881952813, |
| "distill": 0.0007918142426283436, |
| "task": 0.5605541461325706, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.24667784539943047, |
| "distill": 0.0005330673104370052, |
| "task": 1.7473829781755488, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.03359868098050356, |
| "distill": 0.0007893668138422072, |
| "task": 0.47301009893417356, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.29460000000000003 |
| }, |
| { |
| "epoch": 11, |
| "eval_results": { |
| "sts_spearman": 0.8200398862029752, |
| "sts_pearson": 0.7984632150530062, |
| "retrieval_recall_at_1": 0.574, |
| "retrieval_recall_at_5": 0.88, |
| "retrieval_recall_at_10": 0.948, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7349022030830383, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8587563037872314, |
| "composite_score": 0.7906866097681543 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.049540048222179, |
| "distill": 0.0006752971321870775, |
| "task": 0.17472205732179724, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.1152741235304386, |
| "distill": 0.0007609698123873231, |
| "task": 0.543203037469945, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.2454570157730833, |
| "distill": 0.000510894448218986, |
| "task": 1.7373002945108618, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.032351000048220155, |
| "distill": 0.0007691650826018304, |
| "task": 0.4550263941287994, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.29400000000000004 |
| }, |
| { |
| "epoch": 12, |
| "eval_results": { |
| "sts_spearman": 0.8193103945963425, |
| "sts_pearson": 0.7974055277518916, |
| "retrieval_recall_at_1": 0.594, |
| "retrieval_recall_at_5": 0.884, |
| "retrieval_recall_at_10": 0.948, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7340207695960999, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8586063981056213, |
| "composite_score": 0.7915218639648379 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.048518418131963066, |
| "distill": 0.0006525755262650225, |
| "task": 0.1709841172332349, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.11415320951887901, |
| "distill": 0.0007357201898133659, |
| "task": 0.5374910330518763, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.24104384951134947, |
| "distill": 0.0004927666555654179, |
| "task": 1.704636736119047, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.031325037218630315, |
| "distill": 0.0007606724102515727, |
| "task": 0.44016211330890653, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2934 |
| }, |
| { |
| "epoch": 13, |
| "eval_results": { |
| "sts_spearman": 0.8188842839819943, |
| "sts_pearson": 0.7964862034231038, |
| "retrieval_recall_at_1": 0.596, |
| "retrieval_recall_at_5": 0.896, |
| "retrieval_recall_at_10": 0.948, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7331050634384155, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8598613142967224, |
| "composite_score": 0.7949088086576639 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.047971061390379204, |
| "distill": 0.000634684164137782, |
| "task": 0.1689240109661351, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.109825795159695, |
| "distill": 0.000719507896103599, |
| "task": 0.5166625177606623, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.23657965850322804, |
| "distill": 0.00048042560786385326, |
| "task": 1.6716556980254802, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.030759319849312305, |
| "distill": 0.0007484613277483731, |
| "task": 0.43184628784656526, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2928 |
| }, |
| { |
| "epoch": 14, |
| "eval_results": { |
| "sts_spearman": 0.8170877062297013, |
| "sts_pearson": 0.794295936858975, |
| "retrieval_recall_at_1": 0.596, |
| "retrieval_recall_at_5": 0.896, |
| "retrieval_recall_at_10": 0.948, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7348426580429077, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8593454957008362, |
| "composite_score": 0.7940105197815173 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.04733733427913293, |
| "distill": 0.0006224965605803806, |
| "task": 0.16655636870342752, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.10522193921373245, |
| "distill": 0.0007066282906886586, |
| "task": 0.49456279772393247, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.23459308768840545, |
| "distill": 0.000471937776037908, |
| "task": 1.6562247961125476, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.02917077410966158, |
| "distill": 0.0007398281595669687, |
| "task": 0.40907877683639526, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2922 |
| }, |
| { |
| "epoch": 15, |
| "eval_results": { |
| "sts_spearman": 0.8157736839705673, |
| "sts_pearson": 0.7925540172374534, |
| "retrieval_recall_at_1": 0.606, |
| "retrieval_recall_at_5": 0.894, |
| "retrieval_recall_at_10": 0.952, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7331643104553223, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8580483794212341, |
| "composite_score": 0.7927535086519504 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.04638928920030594, |
| "distill": 0.0006136388072501059, |
| "task": 0.16308000878147458, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.10471683328456068, |
| "distill": 0.000696288561696147, |
| "task": 0.49178333104924954, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.23013398304898688, |
| "distill": 0.00046382398540253174, |
| "task": 1.6233676443708704, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.028761257790029048, |
| "distill": 0.0007339846633840352, |
| "task": 0.4029817461967468, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2916 |
| }, |
| { |
| "epoch": 16, |
| "eval_results": { |
| "sts_spearman": 0.8157617853858435, |
| "sts_pearson": 0.7923932840541673, |
| "retrieval_recall_at_1": 0.612, |
| "retrieval_recall_at_5": 0.898, |
| "retrieval_recall_at_10": 0.954, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7322084903717041, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8596447706222534, |
| "composite_score": 0.7939475593595884 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.04621984566683355, |
| "distill": 0.0006054732543618783, |
| "task": 0.1623542036699212, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.10093737583845219, |
| "distill": 0.0006863788065024989, |
| "task": 0.473613737111396, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.22668360205406837, |
| "distill": 0.00045654376067141904, |
| "task": 1.597678039936309, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.02780891638249159, |
| "distill": 0.0007290024077519774, |
| "task": 0.3892352104187012, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.29100000000000004 |
| }, |
| { |
| "epoch": 17, |
| "eval_results": { |
| "sts_spearman": 0.8138228214142917, |
| "sts_pearson": 0.7899716079426576, |
| "retrieval_recall_at_1": 0.61, |
| "retrieval_recall_at_5": 0.9, |
| "retrieval_recall_at_10": 0.952, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7340738773345947, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8595521450042725, |
| "composite_score": 0.7935780773738126 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.04601614598346793, |
| "distill": 0.0006001048604957759, |
| "task": 0.1615060425322989, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.09758770988976702, |
| "distill": 0.0006810528612596557, |
| "task": 0.4574874652192948, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.22612765399699516, |
| "distill": 0.00045324116202190203, |
| "task": 1.5924184956449143, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.02633261661976576, |
| "distill": 0.0007273759925737977, |
| "task": 0.3681142464280128, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2904 |
| }, |
| { |
| "epoch": 18, |
| "eval_results": { |
| "sts_spearman": 0.8137941541347218, |
| "sts_pearson": 0.7900114938692636, |
| "retrieval_recall_at_1": 0.618, |
| "retrieval_recall_at_5": 0.902, |
| "retrieval_recall_at_10": 0.954, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7320045232772827, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8595419526100159, |
| "composite_score": 0.7941637437340276 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.04508868280960166, |
| "distill": 0.0005957435021865303, |
| "task": 0.15811052076194598, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.09446902795040861, |
| "distill": 0.0006735043134540319, |
| "task": 0.4424755630340982, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.2214733213186264, |
| "distill": 0.00044834198977580255, |
| "task": 1.5583173239484747, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.026047103106975555, |
| "distill": 0.0007218799262773245, |
| "task": 0.36381163746118544, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2898 |
| }, |
| { |
| "epoch": 19, |
| "eval_results": { |
| "sts_spearman": 0.8131324437136556, |
| "sts_pearson": 0.7888827667888834, |
| "retrieval_recall_at_1": 0.622, |
| "retrieval_recall_at_5": 0.906, |
| "retrieval_recall_at_10": 0.954, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7326275110244751, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8593710064888, |
| "composite_score": 0.7950328885234945 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.044820788761843804, |
| "distill": 0.0005908587026288328, |
| "task": 0.1570410553527915, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.0947175715514954, |
| "distill": 0.0006690982015843087, |
| "task": 0.44327550555797335, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.22196958103078476, |
| "distill": 0.00044533125981886654, |
| "task": 1.5605007993414046, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.02539514433592558, |
| "distill": 0.000717665534466505, |
| "task": 0.35435559451580045, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2892 |
| }, |
| { |
| "epoch": 20, |
| "eval_results": { |
| "sts_spearman": 0.8125557315718881, |
| "sts_pearson": 0.7883415732948076, |
| "retrieval_recall_at_1": 0.622, |
| "retrieval_recall_at_5": 0.906, |
| "retrieval_recall_at_10": 0.956, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7300553321838379, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8589670062065125, |
| "composite_score": 0.7947445324526108 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.04461472562473753, |
| "distill": 0.0005882442093697255, |
| "task": 0.15618835778340048, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.09131274277225454, |
| "distill": 0.0006642562456111959, |
| "task": 0.4269564062991041, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.21616570017439254, |
| "distill": 0.0004426293698030504, |
| "task": 1.518400065442349, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.024884730763733386, |
| "distill": 0.0007182472327258438, |
| "task": 0.3468856424093246, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2886 |
| }, |
| { |
| "epoch": 21, |
| "eval_results": { |
| "sts_spearman": 0.8124046923217468, |
| "sts_pearson": 0.7879715870379221, |
| "retrieval_recall_at_1": 0.626, |
| "retrieval_recall_at_5": 0.908, |
| "retrieval_recall_at_10": 0.954, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7299230694770813, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8608139157295227, |
| "composite_score": 0.7952690128275401 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.044788526128167694, |
| "distill": 0.0005869252971656945, |
| "task": 0.15666956746059915, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.09129251508002585, |
| "distill": 0.0006613465486728447, |
| "task": 0.42650772028781, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.21487902199968378, |
| "distill": 0.0004392804477631332, |
| "task": 1.5080934889773105, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.02408289248123765, |
| "distill": 0.0007130690268240869, |
| "task": 0.3353585585951805, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.28800000000000003 |
| }, |
| { |
| "epoch": 22, |
| "eval_results": { |
| "sts_spearman": 0.8108396163396415, |
| "sts_pearson": 0.7859428479306386, |
| "retrieval_recall_at_1": 0.626, |
| "retrieval_recall_at_5": 0.91, |
| "retrieval_recall_at_10": 0.956, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7304312586784363, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8598558306694031, |
| "composite_score": 0.7950864748364874 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.04427984421667845, |
| "distill": 0.0005824335992498243, |
| "task": 0.15475882071515787, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.08749476669633642, |
| "distill": 0.0006578407458406179, |
| "task": 0.40839041484163163, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.21218161253218956, |
| "distill": 0.0004366559572171103, |
| "task": 1.487904317835544, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.02473886413499713, |
| "distill": 0.0007142506423406303, |
| "task": 0.34428276121616364, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2874 |
| }, |
| { |
| "epoch": 23, |
| "eval_results": { |
| "sts_spearman": 0.8101206106743032, |
| "sts_pearson": 0.7852421682051318, |
| "retrieval_recall_at_1": 0.626, |
| "retrieval_recall_at_5": 0.91, |
| "retrieval_recall_at_10": 0.956, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7298892140388489, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8599591255187988, |
| "composite_score": 0.7947269720038184 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.043497207696023193, |
| "distill": 0.0005803209623200414, |
| "task": 0.15188857848229614, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.08651064019253914, |
| "distill": 0.0006545566213119062, |
| "task": 0.40345349781056666, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.21238021647676508, |
| "distill": 0.00043464487014794126, |
| "task": 1.4880508184432983, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.022220892272889613, |
| "distill": 0.0007110251928679645, |
| "task": 0.30870682895183565, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2868 |
| }, |
| { |
| "epoch": 24, |
| "eval_results": { |
| "sts_spearman": 0.8090374451655163, |
| "sts_pearson": 0.783875236477, |
| "retrieval_recall_at_1": 0.64, |
| "retrieval_recall_at_5": 0.91, |
| "retrieval_recall_at_10": 0.956, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7307232618331909, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8604580163955688, |
| "composite_score": 0.7941853892494248 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.04361056018134822, |
| "distill": 0.0005793894398147645, |
| "task": 0.15216006111839545, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.08273381509996475, |
| "distill": 0.0006529860733512868, |
| "task": 0.3854811432513785, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.2092740849611607, |
| "distill": 0.00043218842759589724, |
| "task": 1.4650490512239172, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.02230049455538392, |
| "distill": 0.0007111194601748139, |
| "task": 0.30956813097000124, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2862 |
| }, |
| { |
| "epoch": 25, |
| "eval_results": { |
| "sts_spearman": 0.8088290083594672, |
| "sts_pearson": 0.7836111268009822, |
| "retrieval_recall_at_1": 0.642, |
| "retrieval_recall_at_5": 0.912, |
| "retrieval_recall_at_10": 0.958, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7291015386581421, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8604010939598083, |
| "composite_score": 0.7946811708464003 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.042825429821791855, |
| "distill": 0.0005763423598735877, |
| "task": 0.14928901389889096, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.08428762876924048, |
| "distill": 0.0006495943922113548, |
| "task": 0.3924136973441915, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.20725905958642352, |
| "distill": 0.0004320865307508552, |
| "task": 1.4497177093587024, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.021814299654215573, |
| "distill": 0.0007105717668309808, |
| "task": 0.30251066088676454, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2856 |
| }, |
| { |
| "epoch": 26, |
| "eval_results": { |
| "sts_spearman": 0.8086444673068466, |
| "sts_pearson": 0.7835424818834511, |
| "retrieval_recall_at_1": 0.65, |
| "retrieval_recall_at_5": 0.914, |
| "retrieval_recall_at_10": 0.96, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7279387712478638, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8617146015167236, |
| "composite_score": 0.79518890032009 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.04305816587546597, |
| "distill": 0.0005753715294818192, |
| "task": 0.14997966134029886, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.08245786469667515, |
| "distill": 0.0006471946550671883, |
| "task": 0.3835590494439957, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.20547382248208879, |
| "distill": 0.00042902108806958225, |
| "task": 1.4360247748963377, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.02213638899847865, |
| "distill": 0.0007105463359039277, |
| "task": 0.3067675843834877, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.28500000000000003 |
| }, |
| { |
| "epoch": 27, |
| "eval_results": { |
| "sts_spearman": 0.8075945769025906, |
| "sts_pearson": 0.7817250896683322, |
| "retrieval_recall_at_1": 0.648, |
| "retrieval_recall_at_5": 0.916, |
| "retrieval_recall_at_10": 0.964, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7290590405464172, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8608205318450928, |
| "composite_score": 0.795263955117962 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.041979301558888474, |
| "distill": 0.0005742409064070038, |
| "task": 0.1460871498869813, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.07854667179127957, |
| "distill": 0.0006465072064918089, |
| "task": 0.3650214504054252, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.20393373991580718, |
| "distill": 0.00042719431060485227, |
| "task": 1.4240653768498848, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.02039573285728693, |
| "distill": 0.0007079597911797464, |
| "task": 0.28220218420028687, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2844 |
| }, |
| { |
| "epoch": 28, |
| "eval_results": { |
| "sts_spearman": 0.8072111574266839, |
| "sts_pearson": 0.7814816033631548, |
| "retrieval_recall_at_1": 0.656, |
| "retrieval_recall_at_5": 0.916, |
| "retrieval_recall_at_10": 0.962, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7277703881263733, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8611229062080383, |
| "composite_score": 0.7950722453800086 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.04250466094716736, |
| "distill": 0.0005732332761196986, |
| "task": 0.14780080836752188, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.07828945341579457, |
| "distill": 0.0006427260170234962, |
| "task": 0.363525298681665, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.20262091527593898, |
| "distill": 0.0004252295209184051, |
| "task": 1.4137128566173798, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.021627166587859393, |
| "distill": 0.000706591084599495, |
| "task": 0.29917111396789553, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2838 |
| }, |
| { |
| "epoch": 29, |
| "eval_results": { |
| "sts_spearman": 0.8061150812196531, |
| "sts_pearson": 0.7799263057126831, |
| "retrieval_recall_at_1": 0.654, |
| "retrieval_recall_at_5": 0.918, |
| "retrieval_recall_at_10": 0.964, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7274843454360962, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8605352640151978, |
| "composite_score": 0.7951242072764932 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.04178553026007569, |
| "distill": 0.0005705209104749172, |
| "task": 0.14517284346663434, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.07806510272178244, |
| "distill": 0.0006415697936701806, |
| "task": 0.3621810408348733, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.20029219636257659, |
| "distill": 0.00042426467754263825, |
| "task": 1.3962893257749842, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.020195775851607322, |
| "distill": 0.000706263561733067, |
| "task": 0.2789587274193764, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2832 |
| }, |
| { |
| "epoch": 30, |
| "eval_results": { |
| "sts_spearman": 0.805952102459493, |
| "sts_pearson": 0.7796381825168456, |
| "retrieval_recall_at_1": 0.658, |
| "retrieval_recall_at_5": 0.92, |
| "retrieval_recall_at_10": 0.966, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7271153926849365, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8611077666282654, |
| "composite_score": 0.7956427178964132 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.041891162486180016, |
| "distill": 0.0005706739384154587, |
| "task": 0.14542057948267978, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.07668351699063118, |
| "distill": 0.0006396458459463208, |
| "task": 0.35546302129613594, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.19826114177703857, |
| "distill": 0.00042267177598253686, |
| "task": 1.3809708052493157, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.019489498622715474, |
| "distill": 0.0007044955214951188, |
| "task": 0.268893338739872, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2826 |
| }, |
| { |
| "epoch": 31, |
| "eval_results": { |
| "sts_spearman": 0.8054395024616142, |
| "sts_pearson": 0.7790870358568797, |
| "retrieval_recall_at_1": 0.662, |
| "retrieval_recall_at_5": 0.92, |
| "retrieval_recall_at_10": 0.968, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7270572781562805, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8614012002944946, |
| "composite_score": 0.7953864178974739 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.04180093708893527, |
| "distill": 0.000569236811513648, |
| "task": 0.14498750833065613, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.07472586061092133, |
| "distill": 0.0006379124688658308, |
| "task": 0.346081572644254, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.19744586469011105, |
| "distill": 0.0004216810487745766, |
| "task": 1.3741430931902947, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.02020758679136634, |
| "distill": 0.0007039231946691871, |
| "task": 0.27867799401283266, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.28200000000000003 |
| }, |
| { |
| "epoch": 32, |
| "eval_results": { |
| "sts_spearman": 0.8039734694214193, |
| "sts_pearson": 0.7774542134736094, |
| "retrieval_recall_at_1": 0.658, |
| "retrieval_recall_at_5": 0.922, |
| "retrieval_recall_at_10": 0.972, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7258086204528809, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8613040447235107, |
| "composite_score": 0.7952534013773764 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.04136973813824032, |
| "distill": 0.000568263144131102, |
| "task": 0.14336845570284387, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.07349338850125353, |
| "distill": 0.000635797770921775, |
| "task": 0.3400801315586618, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.195538673629152, |
| "distill": 0.00042041142513242333, |
| "task": 1.3597298099639568, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.01802230104804039, |
| "distill": 0.0007017374911811203, |
| "task": 0.24804942756891252, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2814 |
| }, |
| { |
| "epoch": 33, |
| "eval_results": { |
| "sts_spearman": 0.8034526879232323, |
| "sts_pearson": 0.7767872504658838, |
| "retrieval_recall_at_1": 0.664, |
| "retrieval_recall_at_5": 0.92, |
| "retrieval_recall_at_10": 0.972, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7267512083053589, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8609178066253662, |
| "composite_score": 0.7943930106282828 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.04124606368334397, |
| "distill": 0.0005672681145370007, |
| "task": 0.14282110203867374, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.07269588295132556, |
| "distill": 0.0006344794367558937, |
| "task": 0.3361036536541391, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.19442298532800473, |
| "distill": 0.0004187757124569505, |
| "task": 1.3508439393753702, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.01792764011770487, |
| "distill": 0.0007010513567365706, |
| "task": 0.246534825861454, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2808 |
| }, |
| { |
| "epoch": 34, |
| "eval_results": { |
| "sts_spearman": 0.8029147605057396, |
| "sts_pearson": 0.7758699525760386, |
| "retrieval_recall_at_1": 0.656, |
| "retrieval_recall_at_5": 0.924, |
| "retrieval_recall_at_10": 0.976, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7263163924217224, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.861183762550354, |
| "composite_score": 0.7953240469195365 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.04087469354271889, |
| "distill": 0.0005659387705078268, |
| "task": 0.14141468995291254, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.07193009151106185, |
| "distill": 0.0006315128511174562, |
| "task": 0.3322827714554807, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.19458080923303644, |
| "distill": 0.0004175955335550169, |
| "task": 1.3508182814780703, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.018750363681465387, |
| "distill": 0.0007004878658335656, |
| "task": 0.2577672630548477, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2802 |
| }, |
| { |
| "epoch": 35, |
| "eval_results": { |
| "sts_spearman": 0.8029120487600959, |
| "sts_pearson": 0.7759091259908452, |
| "retrieval_recall_at_1": 0.652, |
| "retrieval_recall_at_5": 0.924, |
| "retrieval_recall_at_10": 0.976, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7263759970664978, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8618313670158386, |
| "composite_score": 0.7953226910467146 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.040802220611468605, |
| "distill": 0.0005649596442589941, |
| "task": 0.1410475363549979, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.07252203149998442, |
| "distill": 0.0006298948205670619, |
| "task": 0.33474880393515244, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.19197505680804558, |
| "distill": 0.0004152855129932311, |
| "task": 1.3316139946592616, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.01693342700600624, |
| "distill": 0.0006995417177677154, |
| "task": 0.23234085589647294, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2796 |
| }, |
| { |
| "epoch": 36, |
| "eval_results": { |
| "sts_spearman": 0.802407902504706, |
| "sts_pearson": 0.7756569556734015, |
| "retrieval_recall_at_1": 0.658, |
| "retrieval_recall_at_5": 0.924, |
| "retrieval_recall_at_10": 0.974, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7243511080741882, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8609747290611267, |
| "composite_score": 0.7950706179190197 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.040637732200000595, |
| "distill": 0.0005651830908154016, |
| "task": 0.14036077325758728, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.07090983135586089, |
| "distill": 0.0006288531424596588, |
| "task": 0.32701979894587335, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.1879324754501911, |
| "distill": 0.0004152300212770066, |
| "task": 1.3024731620829155, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.01749492483213544, |
| "distill": 0.0006997698335908353, |
| "task": 0.23994021117687225, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.279 |
| }, |
| { |
| "epoch": 37, |
| "eval_results": { |
| "sts_spearman": 0.8014589498925283, |
| "sts_pearson": 0.7745302218047226, |
| "retrieval_recall_at_1": 0.662, |
| "retrieval_recall_at_5": 0.926, |
| "retrieval_recall_at_10": 0.974, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7247843742370605, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8621292114257812, |
| "composite_score": 0.7951961416129308 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.0404453170688256, |
| "distill": 0.0005647512937329062, |
| "task": 0.13957903113054193, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.07041878134329269, |
| "distill": 0.000626888732980699, |
| "task": 0.3244838112212242, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.18824834677767247, |
| "distill": 0.00041336335856071175, |
| "task": 1.3035842015388164, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.017420530039817094, |
| "distill": 0.0006980633072089404, |
| "task": 0.23872213810682297, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2784 |
| }, |
| { |
| "epoch": 38, |
| "eval_results": { |
| "sts_spearman": 0.8015481972850798, |
| "sts_pearson": 0.7748134816243155, |
| "retrieval_recall_at_1": 0.662, |
| "retrieval_recall_at_5": 0.926, |
| "retrieval_recall_at_10": 0.974, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7238280177116394, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8617278337478638, |
| "composite_score": 0.7952407653092066 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.040147524246055145, |
| "distill": 0.000563749264034888, |
| "task": 0.13843435340601465, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.06983992354349887, |
| "distill": 0.0006249398723779682, |
| "task": 0.3215467447930194, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.18727405749736947, |
| "distill": 0.0004120995005731411, |
| "task": 1.2957600558057745, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.016873036976903677, |
| "distill": 0.0006976060452871024, |
| "task": 0.23095046430826188, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2778 |
| }, |
| { |
| "epoch": 39, |
| "eval_results": { |
| "sts_spearman": 0.8010951080044263, |
| "sts_pearson": 0.7744044551086896, |
| "retrieval_recall_at_1": 0.66, |
| "retrieval_recall_at_5": 0.926, |
| "retrieval_recall_at_10": 0.976, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7240303158760071, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8622973561286926, |
| "composite_score": 0.7950142206688798 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.040290426786827004, |
| "distill": 0.0005619644627744412, |
| "task": 0.1388165892466255, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.06733930118857546, |
| "distill": 0.0006235166393378948, |
| "task": 0.3097512569833309, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.18707014081325937, |
| "distill": 0.00041061582445028297, |
| "task": 1.29327839993416, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.017511206772178413, |
| "distill": 0.0006984172330703586, |
| "task": 0.23959056437015533, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2772 |
| }, |
| { |
| "epoch": 40, |
| "eval_results": { |
| "sts_spearman": 0.8008200971384819, |
| "sts_pearson": 0.774130829271817, |
| "retrieval_recall_at_1": 0.666, |
| "retrieval_recall_at_5": 0.932, |
| "retrieval_recall_at_10": 0.974, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7223304510116577, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.862885594367981, |
| "composite_score": 0.7966767152359077 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.039851526122378265, |
| "distill": 0.0005616477422375718, |
| "task": 0.1371861229772153, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.0681393450879036, |
| "distill": 0.0006218473711843661, |
| "task": 0.3131846848954546, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.18474397950984062, |
| "distill": 0.00041012915824738114, |
| "task": 1.2761304099509057, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.01582443034276366, |
| "distill": 0.0006965516076888889, |
| "task": 0.2160874292254448, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2766 |
| }, |
| { |
| "epoch": 41, |
| "eval_results": { |
| "sts_spearman": 0.8004397143136981, |
| "sts_pearson": 0.7737034113801735, |
| "retrieval_recall_at_1": 0.662, |
| "retrieval_recall_at_5": 0.93, |
| "retrieval_recall_at_10": 0.976, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7216349840164185, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8629198670387268, |
| "composite_score": 0.7958865238235158 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.040168939077335854, |
| "distill": 0.0005606161886016312, |
| "task": 0.1381706041486367, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.06701801083189377, |
| "distill": 0.0006201347891003527, |
| "task": 0.3077663652440335, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.18171446183894543, |
| "distill": 0.00040893261246581345, |
| "task": 1.254154608604756, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.016577236633747817, |
| "distill": 0.0006964895350392908, |
| "task": 0.2263122245669365, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.276 |
| }, |
| { |
| "epoch": 42, |
| "eval_results": { |
| "sts_spearman": 0.799541139671189, |
| "sts_pearson": 0.7726619557363898, |
| "retrieval_recall_at_1": 0.664, |
| "retrieval_recall_at_5": 0.932, |
| "retrieval_recall_at_10": 0.976, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7218517065048218, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.862092912197113, |
| "composite_score": 0.7960372365022612 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.039756916949282524, |
| "distill": 0.0005605145656179799, |
| "task": 0.13663589921982391, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.06787618566700752, |
| "distill": 0.0006185190620871776, |
| "task": 0.3114630852607971, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.1818689741986863, |
| "distill": 0.0004076707292041008, |
| "task": 1.254186414657755, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.015548030100762843, |
| "distill": 0.0006947330140974373, |
| "task": 0.21193347945809365, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2754 |
| }, |
| { |
| "epoch": 43, |
| "eval_results": { |
| "sts_spearman": 0.798959430373989, |
| "sts_pearson": 0.7717844926047425, |
| "retrieval_recall_at_1": 0.658, |
| "retrieval_recall_at_5": 0.934, |
| "retrieval_recall_at_10": 0.976, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7230818867683411, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8633172512054443, |
| "composite_score": 0.7963463818536611 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.03980643334596053, |
| "distill": 0.0005611540931884361, |
| "task": 0.13669410898633624, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.0670052819905129, |
| "distill": 0.0006171439581134535, |
| "task": 0.3072057971928982, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.181170218168421, |
| "distill": 0.0004060288384665755, |
| "task": 1.2483358712906534, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.016034624353051186, |
| "distill": 0.0006926223228219897, |
| "task": 0.21848167926073075, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2748 |
| }, |
| { |
| "epoch": 44, |
| "eval_results": { |
| "sts_spearman": 0.7992766546129851, |
| "sts_pearson": 0.7723631373159269, |
| "retrieval_recall_at_1": 0.668, |
| "retrieval_recall_at_5": 0.936, |
| "retrieval_recall_at_10": 0.976, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7214555740356445, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8633242845535278, |
| "composite_score": 0.7971049939731593 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.03956834745147954, |
| "distill": 0.0005598463517937647, |
| "task": 0.13576342230257782, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.0675140564587522, |
| "distill": 0.0006155266111934597, |
| "task": 0.30929218010699494, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.18059916984527669, |
| "distill": 0.00040597010755594426, |
| "task": 1.2433717808824905, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.016109541151672603, |
| "distill": 0.000693465251242742, |
| "task": 0.21933580189943314, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2742 |
| }, |
| { |
| "epoch": 45, |
| "eval_results": { |
| "sts_spearman": 0.7987902718942615, |
| "sts_pearson": 0.7716241234848017, |
| "retrieval_recall_at_1": 0.67, |
| "retrieval_recall_at_5": 0.934, |
| "retrieval_recall_at_10": 0.976, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7223896384239197, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8632831573486328, |
| "composite_score": 0.7962618026137975 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.03940606570762137, |
| "distill": 0.0005587598197567074, |
| "task": 0.13509494876084122, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.06665098445212587, |
| "distill": 0.0006136801812124062, |
| "task": 0.3050802118600683, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.18052284007376812, |
| "distill": 0.00040472789662592906, |
| "task": 1.2418234031251136, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.015592311229556798, |
| "distill": 0.0006917931721545755, |
| "task": 0.21204619854688644, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2736 |
| }, |
| { |
| "epoch": 46, |
| "eval_results": { |
| "sts_spearman": 0.7980845899126008, |
| "sts_pearson": 0.7708174356684444, |
| "retrieval_recall_at_1": 0.662, |
| "retrieval_recall_at_5": 0.934, |
| "retrieval_recall_at_10": 0.974, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7227649688720703, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8638308644294739, |
| "composite_score": 0.7959089616229671 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.03914242918076723, |
| "distill": 0.0005580810273228133, |
| "task": 0.13407865684965384, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.06371399538314089, |
| "distill": 0.0006122274517497801, |
| "task": 0.2913656944924213, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.1788206826499168, |
| "distill": 0.00040350851176821804, |
| "task": 1.2290957481303113, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.016117287054657935, |
| "distill": 0.0006897407933138311, |
| "task": 0.2191057413816452, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.273 |
| }, |
| { |
| "epoch": 47, |
| "eval_results": { |
| "sts_spearman": 0.7986050030831314, |
| "sts_pearson": 0.7713344721815415, |
| "retrieval_recall_at_1": 0.668, |
| "retrieval_recall_at_5": 0.936, |
| "retrieval_recall_at_10": 0.974, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7211058139801025, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8641474843025208, |
| "composite_score": 0.7967691682082324 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.03924534291676853, |
| "distill": 0.0005579624380713896, |
| "task": 0.1343229581480441, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.06390803307294846, |
| "distill": 0.0006106501616081817, |
| "task": 0.29201801152939494, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.17669346231095334, |
| "distill": 0.0004022118878213966, |
| "task": 1.2134682447352307, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.014580465480685234, |
| "distill": 0.0006913242454174906, |
| "task": 0.1978030323982239, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2724 |
| }, |
| { |
| "epoch": 48, |
| "eval_results": { |
| "sts_spearman": 0.7978328086000539, |
| "sts_pearson": 0.7705122749494036, |
| "retrieval_recall_at_1": 0.67, |
| "retrieval_recall_at_5": 0.936, |
| "retrieval_recall_at_10": 0.976, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7213853001594543, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8641984462738037, |
| "composite_score": 0.7963830709666937 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.038881031796336174, |
| "distill": 0.0005576102867843988, |
| "task": 0.13296303638945456, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.06267972861198669, |
| "distill": 0.0006095131591675764, |
| "task": 0.2861579453691523, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.17544851499669095, |
| "distill": 0.0004012782276518881, |
| "task": 1.2039236971672544, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.015775612369179726, |
| "distill": 0.0006886643706820906, |
| "task": 0.2140680193901062, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2718 |
| }, |
| { |
| "epoch": 49, |
| "eval_results": { |
| "sts_spearman": 0.7973172210556293, |
| "sts_pearson": 0.7701713894839444, |
| "retrieval_recall_at_1": 0.67, |
| "retrieval_recall_at_5": 0.936, |
| "retrieval_recall_at_10": 0.976, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7203776240348816, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8631545305252075, |
| "composite_score": 0.7961252771944813 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.03882810230488363, |
| "distill": 0.0005557498307493718, |
| "task": 0.13267488712849823, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.0626571145146451, |
| "distill": 0.000607952565955434, |
| "task": 0.28582252847387435, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.1763826126747943, |
| "distill": 0.0004007951009701541, |
| "task": 1.2093435627348879, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.015177728608250618, |
| "distill": 0.0006882219109684229, |
| "task": 0.20569542646408082, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2712 |
| }, |
| { |
| "epoch": 50, |
| "eval_results": { |
| "sts_spearman": 0.7968594218073127, |
| "sts_pearson": 0.7695483662129042, |
| "retrieval_recall_at_1": 0.67, |
| "retrieval_recall_at_5": 0.936, |
| "retrieval_recall_at_10": 0.974, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7204389572143555, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8636235594749451, |
| "composite_score": 0.795896377570323 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.03918402480042499, |
| "distill": 0.0005560298273137406, |
| "task": 0.13378654711920282, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.06285070390143294, |
| "distill": 0.000605922420211929, |
| "task": 0.2864762779245985, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.1742387483728693, |
| "distill": 0.00039892817941553733, |
| "task": 1.1936577951654475, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.014562566205859185, |
| "distill": 0.0006869392120279372, |
| "task": 0.197102826833725, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2706 |
| }, |
| { |
| "epoch": 51, |
| "eval_results": { |
| "sts_spearman": 0.7961018874767011, |
| "sts_pearson": 0.7685896198859151, |
| "retrieval_recall_at_1": 0.672, |
| "retrieval_recall_at_5": 0.938, |
| "retrieval_recall_at_10": 0.976, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7212554216384888, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8639085292816162, |
| "composite_score": 0.7961176104050172 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.038832955305343086, |
| "distill": 0.000555693042849231, |
| "task": 0.13247574898211853, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.061906919121108156, |
| "distill": 0.0006047165407700107, |
| "task": 0.2819344585246228, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.17475084611710082, |
| "distill": 0.0003985753827827408, |
| "task": 1.1961865387064345, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.014561035577207804, |
| "distill": 0.0006853513652458787, |
| "task": 0.19693138003349303, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.27 |
| }, |
| { |
| "epoch": 52, |
| "eval_results": { |
| "sts_spearman": 0.7957712004157503, |
| "sts_pearson": 0.7682444812474918, |
| "retrieval_recall_at_1": 0.668, |
| "retrieval_recall_at_5": 0.938, |
| "retrieval_recall_at_10": 0.976, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7211222052574158, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8639959096908569, |
| "composite_score": 0.7959522668745418 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.038104739849982056, |
| "distill": 0.0005552242105097874, |
| "task": 0.12987668682699618, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.06224994496145147, |
| "distill": 0.0006037014598482625, |
| "task": 0.2832708587037756, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.17524569560872746, |
| "distill": 0.00039716928756419333, |
| "task": 1.198594996269713, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.01391328014433384, |
| "distill": 0.0006847139564342796, |
| "task": 0.18791155964136125, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2694 |
| }, |
| { |
| "epoch": 53, |
| "eval_results": { |
| "sts_spearman": 0.7948627602870467, |
| "sts_pearson": 0.7672766924578068, |
| "retrieval_recall_at_1": 0.674, |
| "retrieval_recall_at_5": 0.936, |
| "retrieval_recall_at_10": 0.974, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7208473682403564, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8635191321372986, |
| "composite_score": 0.79489804681019 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.03807506864161595, |
| "distill": 0.0005540588349306389, |
| "task": 0.12967087879129077, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.06309370720323096, |
| "distill": 0.0006028104070673122, |
| "task": 0.2868876425509757, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.17431603213574023, |
| "distill": 0.00039660765918249146, |
| "task": 1.1912570329422647, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.013855330273509025, |
| "distill": 0.0006855725019704551, |
| "task": 0.18696729764342307, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2688 |
| }, |
| { |
| "epoch": 54, |
| "eval_results": { |
| "sts_spearman": 0.7948670645604796, |
| "sts_pearson": 0.7673114563945611, |
| "retrieval_recall_at_1": 0.672, |
| "retrieval_recall_at_5": 0.936, |
| "retrieval_recall_at_10": 0.976, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7204586267471313, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8637556433677673, |
| "composite_score": 0.7949001989469064 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.03859076422193776, |
| "distill": 0.000553890953913493, |
| "task": 0.1313275915125142, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.06243176108345072, |
| "distill": 0.000601062821125255, |
| "task": 0.2836410492024523, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.17191652573169547, |
| "distill": 0.0003955346752612039, |
| "task": 1.173889383356622, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.014081115927547216, |
| "distill": 0.0006832460989244282, |
| "task": 0.1899135023355484, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2682 |
| }, |
| { |
| "epoch": 55, |
| "eval_results": { |
| "sts_spearman": 0.7940184198467072, |
| "sts_pearson": 0.7664062690593959, |
| "retrieval_recall_at_1": 0.67, |
| "retrieval_recall_at_5": 0.936, |
| "retrieval_recall_at_10": 0.974, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7194271087646484, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8639047145843506, |
| "composite_score": 0.7944758765900203 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.03801367480469787, |
| "distill": 0.0005527152021860946, |
| "task": 0.12925235538379007, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.06092823788206628, |
| "distill": 0.0005999853217756336, |
| "task": 0.276568721900595, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.1702281376148792, |
| "distill": 0.0003946685254514376, |
| "task": 1.161404493007254, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.012735517043620349, |
| "distill": 0.0006815990433096886, |
| "task": 0.17139707654714584, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.2676 |
| }, |
| { |
| "epoch": 56, |
| "eval_results": { |
| "sts_spearman": 0.7940615711220378, |
| "sts_pearson": 0.7665661594709636, |
| "retrieval_recall_at_1": 0.672, |
| "retrieval_recall_at_5": 0.94, |
| "retrieval_recall_at_10": 0.976, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7197932004928589, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8638368248939514, |
| "composite_score": 0.7956974522276856 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.03830795994271403, |
| "distill": 0.0005517558893188834, |
| "task": 0.13015226015578146, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.060107594633356055, |
| "distill": 0.0005995195837037519, |
| "task": 0.27261265953804587, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.1709233709472291, |
| "distill": 0.0003937529740567775, |
| "task": 1.165199505521896, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.014191682077944278, |
| "distill": 0.0006824985903222114, |
| "task": 0.19112490341067315, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.267 |
| }, |
| { |
| "epoch": 57, |
| "eval_results": { |
| "sts_spearman": 0.7937036029034699, |
| "sts_pearson": 0.7662776909565029, |
| "retrieval_recall_at_1": 0.672, |
| "retrieval_recall_at_5": 0.942, |
| "retrieval_recall_at_10": 0.978, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7196366190910339, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.864041268825531, |
| "composite_score": 0.7961184681184016 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.038414100913897804, |
| "distill": 0.0005509398837128411, |
| "task": 0.1304093819597493, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.057978192938769116, |
| "distill": 0.0005975660850650611, |
| "task": 0.26271810398456896, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.17013170490873622, |
| "distill": 0.00039284772232194687, |
| "task": 1.1588538897798417, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.014643667824566364, |
| "distill": 0.0006814012362156063, |
| "task": 0.19713934063911437, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.26639999999999997 |
| }, |
| { |
| "epoch": 58, |
| "eval_results": { |
| "sts_spearman": 0.7928266663552261, |
| "sts_pearson": 0.765244563485406, |
| "retrieval_recall_at_1": 0.678, |
| "retrieval_recall_at_5": 0.94, |
| "retrieval_recall_at_10": 0.976, |
| "nli_accuracy": 0.5, |
| "nli_similarity": 0.7203112840652466, |
| "paraphrase_accuracy": 0.5, |
| "paraphrase_f1": 0.6666666666666666, |
| "paraphrase_similarity": 0.8644272685050964, |
| "composite_score": 0.7950799998442797 |
| }, |
| "losses": { |
| "sts": { |
| "total": 0.03789730612998423, |
| "distill": 0.0005501627937242713, |
| "task": 0.1285449188688527, |
| "count": 23 |
| }, |
| "retrieval": { |
| "total": 0.06108237723720834, |
| "distill": 0.0005970121710561216, |
| "task": 0.27659898869534755, |
| "count": 47 |
| }, |
| "nli": { |
| "total": 0.1709375616083754, |
| "distill": 0.0003915692932230044, |
| "task": 1.1633987756485635, |
| "count": 47 |
| }, |
| "paraphrase": { |
| "total": 0.013374552130699158, |
| "distill": 0.0006797625974286348, |
| "task": 0.17970404177904128, |
| "count": 10 |
| } |
| }, |
| "distill_weight": 0.26580000000000004 |
| } |
| ] |