| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.6454566185356987, | |
| "eval_steps": 500, | |
| "global_step": 72000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.000896467525744026, | |
| "grad_norm": 0.5214600563049316, | |
| "learning_rate": 0.0004999990104957794, | |
| "loss": 4.1615, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.001792935051488052, | |
| "grad_norm": 0.5064386129379272, | |
| "learning_rate": 0.0004999960380279981, | |
| "loss": 3.6511, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.002689402577232078, | |
| "grad_norm": 0.5299622416496277, | |
| "learning_rate": 0.0004999910826192423, | |
| "loss": 3.537, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.003585870102976104, | |
| "grad_norm": 0.49445027112960815, | |
| "learning_rate": 0.0004999841443088176, | |
| "loss": 3.4766, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.00448233762872013, | |
| "grad_norm": 0.5226134061813354, | |
| "learning_rate": 0.0004999752231517578, | |
| "loss": 3.4303, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.005378805154464156, | |
| "grad_norm": 0.4533003568649292, | |
| "learning_rate": 0.0004999643192188245, | |
| "loss": 3.387, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.0062752726802081825, | |
| "grad_norm": 0.4646398723125458, | |
| "learning_rate": 0.0004999514325965069, | |
| "loss": 3.3508, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.007171740205952208, | |
| "grad_norm": 0.47208401560783386, | |
| "learning_rate": 0.0004999365633870197, | |
| "loss": 3.3181, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.008068207731696234, | |
| "grad_norm": 0.45212841033935547, | |
| "learning_rate": 0.0004999197117083043, | |
| "loss": 3.2928, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.00896467525744026, | |
| "grad_norm": 0.4529814124107361, | |
| "learning_rate": 0.0004999008776940261, | |
| "loss": 3.2734, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.009861142783184286, | |
| "grad_norm": 0.4323786795139313, | |
| "learning_rate": 0.0004998800614935745, | |
| "loss": 3.2607, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.010757610308928313, | |
| "grad_norm": 0.45383089780807495, | |
| "learning_rate": 0.0004998572632720614, | |
| "loss": 3.2526, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.011654077834672339, | |
| "grad_norm": 0.4236689805984497, | |
| "learning_rate": 0.0004998324832103195, | |
| "loss": 3.216, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.012550545360416365, | |
| "grad_norm": 0.3684336841106415, | |
| "learning_rate": 0.0004998057215049018, | |
| "loss": 3.2024, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.013447012886160391, | |
| "grad_norm": 0.39953121542930603, | |
| "learning_rate": 0.000499776978368079, | |
| "loss": 3.1853, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.014343480411904416, | |
| "grad_norm": 0.3904390335083008, | |
| "learning_rate": 0.0004997462540278386, | |
| "loss": 3.1799, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.015239947937648442, | |
| "grad_norm": 0.3936985433101654, | |
| "learning_rate": 0.0004997135487278825, | |
| "loss": 3.1613, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.016136415463392468, | |
| "grad_norm": 0.36848315596580505, | |
| "learning_rate": 0.0004996788627276258, | |
| "loss": 3.1438, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.017032882989136494, | |
| "grad_norm": 0.3514796495437622, | |
| "learning_rate": 0.0004996421963021939, | |
| "loss": 3.1358, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.01792935051488052, | |
| "grad_norm": 0.3697972893714905, | |
| "learning_rate": 0.0004996035497424208, | |
| "loss": 3.1413, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.018825818040624547, | |
| "grad_norm": 0.3728257715702057, | |
| "learning_rate": 0.0004995629233548467, | |
| "loss": 3.1187, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.019722285566368573, | |
| "grad_norm": 0.3539767861366272, | |
| "learning_rate": 0.0004995203174617158, | |
| "loss": 3.1038, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.0206187530921126, | |
| "grad_norm": 0.3395736813545227, | |
| "learning_rate": 0.0004994757324009732, | |
| "loss": 3.0942, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.021515220617856625, | |
| "grad_norm": 0.36709773540496826, | |
| "learning_rate": 0.0004994291685262627, | |
| "loss": 3.0961, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.02241168814360065, | |
| "grad_norm": 0.32454609870910645, | |
| "learning_rate": 0.0004993806262069239, | |
| "loss": 3.0847, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.023308155669344677, | |
| "grad_norm": 0.33812129497528076, | |
| "learning_rate": 0.000499330105827989, | |
| "loss": 3.0767, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.024204623195088704, | |
| "grad_norm": 0.33684587478637695, | |
| "learning_rate": 0.0004992776077901801, | |
| "loss": 3.0641, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.02510109072083273, | |
| "grad_norm": 0.3293500542640686, | |
| "learning_rate": 0.0004992231325099059, | |
| "loss": 3.0715, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.025997558246576756, | |
| "grad_norm": 0.3323623836040497, | |
| "learning_rate": 0.0004991666804192582, | |
| "loss": 3.055, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.026894025772320782, | |
| "grad_norm": 0.3230840861797333, | |
| "learning_rate": 0.0004991082519660089, | |
| "loss": 3.0366, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.02779049329806481, | |
| "grad_norm": 0.31436941027641296, | |
| "learning_rate": 0.0004990478476136059, | |
| "loss": 3.035, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.02868696082380883, | |
| "grad_norm": 0.32258903980255127, | |
| "learning_rate": 0.0004989854678411699, | |
| "loss": 3.0321, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.029583428349552857, | |
| "grad_norm": 0.3148656487464905, | |
| "learning_rate": 0.0004989211131434904, | |
| "loss": 3.0146, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.030479895875296883, | |
| "grad_norm": 0.31875699758529663, | |
| "learning_rate": 0.0004988547840310217, | |
| "loss": 3.0214, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.03137636340104091, | |
| "grad_norm": 0.3068808913230896, | |
| "learning_rate": 0.0004987864810298791, | |
| "loss": 3.0135, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.032272830926784936, | |
| "grad_norm": 0.3052406311035156, | |
| "learning_rate": 0.0004987162046818343, | |
| "loss": 3.0042, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.033169298452528966, | |
| "grad_norm": 0.3155302405357361, | |
| "learning_rate": 0.0004986439555443118, | |
| "loss": 3.008, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.03406576597827299, | |
| "grad_norm": 0.31643468141555786, | |
| "learning_rate": 0.0004985697341903834, | |
| "loss": 2.9999, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.03496223350401702, | |
| "grad_norm": 0.29758599400520325, | |
| "learning_rate": 0.0004984935412087651, | |
| "loss": 2.9957, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.03585870102976104, | |
| "grad_norm": 0.30581027269363403, | |
| "learning_rate": 0.0004984153772038109, | |
| "loss": 2.9873, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.03675516855550507, | |
| "grad_norm": 0.30977991223335266, | |
| "learning_rate": 0.0004983352427955092, | |
| "loss": 2.9864, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.03765163608124909, | |
| "grad_norm": 0.30028942227363586, | |
| "learning_rate": 0.0004982531386194774, | |
| "loss": 2.9801, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.038548103606993116, | |
| "grad_norm": 0.2785912752151489, | |
| "learning_rate": 0.0004981690653269564, | |
| "loss": 2.9707, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.039444571132737145, | |
| "grad_norm": 0.30985763669013977, | |
| "learning_rate": 0.0004980830235848066, | |
| "loss": 2.9673, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.04034103865848117, | |
| "grad_norm": 0.2833133339881897, | |
| "learning_rate": 0.0004979950140755015, | |
| "loss": 2.9683, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.0412375061842252, | |
| "grad_norm": 0.27662980556488037, | |
| "learning_rate": 0.0004979050374971228, | |
| "loss": 2.9525, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.04213397370996922, | |
| "grad_norm": 0.284396231174469, | |
| "learning_rate": 0.0004978130945633548, | |
| "loss": 2.9643, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.04303044123571325, | |
| "grad_norm": 0.2856046259403229, | |
| "learning_rate": 0.0004977191860034787, | |
| "loss": 2.9566, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.04392690876145727, | |
| "grad_norm": 0.27780982851982117, | |
| "learning_rate": 0.0004976233125623668, | |
| "loss": 2.9501, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.0448233762872013, | |
| "grad_norm": 0.282685786485672, | |
| "learning_rate": 0.0004975254750004767, | |
| "loss": 2.943, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.045719843812945325, | |
| "grad_norm": 0.28598085045814514, | |
| "learning_rate": 0.0004974256740938451, | |
| "loss": 2.9507, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.046616311338689355, | |
| "grad_norm": 0.2788136899471283, | |
| "learning_rate": 0.0004973239106340817, | |
| "loss": 2.9437, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.04751277886443338, | |
| "grad_norm": 0.2748227119445801, | |
| "learning_rate": 0.0004972201854283632, | |
| "loss": 2.9439, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 0.04840924639017741, | |
| "grad_norm": 0.27777764201164246, | |
| "learning_rate": 0.0004971144992994265, | |
| "loss": 2.9362, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 0.04930571391592143, | |
| "grad_norm": 0.273762583732605, | |
| "learning_rate": 0.0004970068530855621, | |
| "loss": 2.9336, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.05020218144166546, | |
| "grad_norm": 0.27693507075309753, | |
| "learning_rate": 0.0004968972476406079, | |
| "loss": 2.9285, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 0.05109864896740948, | |
| "grad_norm": 0.26276636123657227, | |
| "learning_rate": 0.0004967856838339424, | |
| "loss": 2.9161, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 0.05199511649315351, | |
| "grad_norm": 0.26518794894218445, | |
| "learning_rate": 0.0004966721625504772, | |
| "loss": 2.9268, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 0.052891584018897535, | |
| "grad_norm": 0.2554951012134552, | |
| "learning_rate": 0.0004965566846906506, | |
| "loss": 2.922, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 0.053788051544641564, | |
| "grad_norm": 0.25945886969566345, | |
| "learning_rate": 0.0004964392511704205, | |
| "loss": 2.9179, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.05468451907038559, | |
| "grad_norm": 0.2560843527317047, | |
| "learning_rate": 0.0004963198629212563, | |
| "loss": 2.9127, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 0.05558098659612962, | |
| "grad_norm": 0.29079943895339966, | |
| "learning_rate": 0.0004961985208901326, | |
| "loss": 2.9097, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 0.05647745412187364, | |
| "grad_norm": 0.2522130608558655, | |
| "learning_rate": 0.0004960752260395208, | |
| "loss": 2.918, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 0.05737392164761766, | |
| "grad_norm": 0.26022017002105713, | |
| "learning_rate": 0.0004959499793473822, | |
| "loss": 2.9088, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 0.05827038917336169, | |
| "grad_norm": 0.25249192118644714, | |
| "learning_rate": 0.0004958227818071597, | |
| "loss": 2.8943, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.059166856699105715, | |
| "grad_norm": 0.24774658679962158, | |
| "learning_rate": 0.0004956936344277703, | |
| "loss": 2.9003, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 0.060063324224849744, | |
| "grad_norm": 0.243002787232399, | |
| "learning_rate": 0.0004955625382335964, | |
| "loss": 2.8918, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 0.06095979175059377, | |
| "grad_norm": 0.2450036108493805, | |
| "learning_rate": 0.0004954294942644788, | |
| "loss": 2.8959, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 0.0618562592763378, | |
| "grad_norm": 0.24282772839069366, | |
| "learning_rate": 0.0004952945035757076, | |
| "loss": 2.8954, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 0.06275272680208183, | |
| "grad_norm": 0.24141670763492584, | |
| "learning_rate": 0.0004951575672380139, | |
| "loss": 2.8933, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.06364919432782584, | |
| "grad_norm": 0.23969806730747223, | |
| "learning_rate": 0.0004950186863375616, | |
| "loss": 2.8874, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 0.06454566185356987, | |
| "grad_norm": 0.24008285999298096, | |
| "learning_rate": 0.0004948778619759389, | |
| "loss": 2.8894, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 0.0654421293793139, | |
| "grad_norm": 0.24920688569545746, | |
| "learning_rate": 0.0004947350952701489, | |
| "loss": 2.8913, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 0.06633859690505793, | |
| "grad_norm": 0.23809005320072174, | |
| "learning_rate": 0.0004945903873526013, | |
| "loss": 2.8832, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 0.06723506443080195, | |
| "grad_norm": 0.2459540218114853, | |
| "learning_rate": 0.0004944437393711034, | |
| "loss": 2.8836, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.06813153195654598, | |
| "grad_norm": 0.2407006472349167, | |
| "learning_rate": 0.0004942951524888509, | |
| "loss": 2.8717, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 0.06902799948229, | |
| "grad_norm": 0.22954203188419342, | |
| "learning_rate": 0.0004941446278844185, | |
| "loss": 2.8766, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 0.06992446700803404, | |
| "grad_norm": 0.2322009801864624, | |
| "learning_rate": 0.0004939921667517505, | |
| "loss": 2.8691, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 0.07082093453377805, | |
| "grad_norm": 0.2393633872270584, | |
| "learning_rate": 0.0004938377703001521, | |
| "loss": 2.874, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 0.07171740205952208, | |
| "grad_norm": 0.23171070218086243, | |
| "learning_rate": 0.0004936814397542788, | |
| "loss": 2.8763, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.07261386958526611, | |
| "grad_norm": 0.23274166882038116, | |
| "learning_rate": 0.0004935231763541272, | |
| "loss": 2.8668, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 0.07351033711101014, | |
| "grad_norm": 0.23799176514148712, | |
| "learning_rate": 0.0004933629813550251, | |
| "loss": 2.8675, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 0.07440680463675416, | |
| "grad_norm": 0.2373514026403427, | |
| "learning_rate": 0.0004932008560276217, | |
| "loss": 2.8648, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 0.07530327216249819, | |
| "grad_norm": 0.22208142280578613, | |
| "learning_rate": 0.0004930368016578769, | |
| "loss": 2.8605, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 0.07619973968824222, | |
| "grad_norm": 0.230060413479805, | |
| "learning_rate": 0.0004928708195470521, | |
| "loss": 2.8626, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.07709620721398623, | |
| "grad_norm": 0.22975729405879974, | |
| "learning_rate": 0.000492702911011699, | |
| "loss": 2.8612, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 0.07799267473973026, | |
| "grad_norm": 0.2305164784193039, | |
| "learning_rate": 0.0004925330773836496, | |
| "loss": 2.8615, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 0.07888914226547429, | |
| "grad_norm": 0.22835880517959595, | |
| "learning_rate": 0.0004923613200100054, | |
| "loss": 2.8563, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 0.07978560979121832, | |
| "grad_norm": 0.23437829315662384, | |
| "learning_rate": 0.000492187640253127, | |
| "loss": 2.8429, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 0.08068207731696234, | |
| "grad_norm": 0.21821853518486023, | |
| "learning_rate": 0.0004920120394906231, | |
| "loss": 2.8556, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.08157854484270637, | |
| "grad_norm": 0.22391639649868011, | |
| "learning_rate": 0.0004918345191153395, | |
| "loss": 2.8439, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 0.0824750123684504, | |
| "grad_norm": 0.22824177145957947, | |
| "learning_rate": 0.0004916550805353483, | |
| "loss": 2.851, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 0.08337147989419443, | |
| "grad_norm": 0.22843202948570251, | |
| "learning_rate": 0.0004914737251739363, | |
| "loss": 2.8493, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 0.08426794741993844, | |
| "grad_norm": 0.22368155419826508, | |
| "learning_rate": 0.0004912904544695944, | |
| "loss": 2.8447, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 0.08516441494568247, | |
| "grad_norm": 0.22078120708465576, | |
| "learning_rate": 0.0004911052698760055, | |
| "loss": 2.8498, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.0860608824714265, | |
| "grad_norm": 0.21971289813518524, | |
| "learning_rate": 0.0004909181728620333, | |
| "loss": 2.846, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 0.08695734999717053, | |
| "grad_norm": 0.217596173286438, | |
| "learning_rate": 0.0004907291649117109, | |
| "loss": 2.8415, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 0.08785381752291455, | |
| "grad_norm": 0.2179151475429535, | |
| "learning_rate": 0.0004905382475242282, | |
| "loss": 2.8454, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 0.08875028504865858, | |
| "grad_norm": 0.2107924073934555, | |
| "learning_rate": 0.0004903454222139212, | |
| "loss": 2.8413, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 0.0896467525744026, | |
| "grad_norm": 0.20650827884674072, | |
| "learning_rate": 0.0004901506905102585, | |
| "loss": 2.825, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.09054322010014663, | |
| "grad_norm": 0.2175520658493042, | |
| "learning_rate": 0.0004899540539578308, | |
| "loss": 2.8297, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 0.09143968762589065, | |
| "grad_norm": 0.21441052854061127, | |
| "learning_rate": 0.0004897555141163376, | |
| "loss": 2.8265, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 0.09233615515163468, | |
| "grad_norm": 0.21620461344718933, | |
| "learning_rate": 0.0004895550725605748, | |
| "loss": 2.8266, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 0.09323262267737871, | |
| "grad_norm": 0.21449413895606995, | |
| "learning_rate": 0.0004893527308804227, | |
| "loss": 2.8308, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 0.09412909020312274, | |
| "grad_norm": 0.2166638970375061, | |
| "learning_rate": 0.0004891484906808332, | |
| "loss": 2.8243, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.09502555772886676, | |
| "grad_norm": 0.2123963087797165, | |
| "learning_rate": 0.0004889423535818172, | |
| "loss": 2.8224, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 0.09592202525461078, | |
| "grad_norm": 0.20958571135997772, | |
| "learning_rate": 0.0004887343212184311, | |
| "loss": 2.8235, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 0.09681849278035481, | |
| "grad_norm": 0.20609615743160248, | |
| "learning_rate": 0.0004885243952407648, | |
| "loss": 2.8304, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 0.09771496030609883, | |
| "grad_norm": 0.20418639481067657, | |
| "learning_rate": 0.000488312577313928, | |
| "loss": 2.8251, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 0.09861142783184286, | |
| "grad_norm": 0.20903444290161133, | |
| "learning_rate": 0.00048809886911803705, | |
| "loss": 2.8244, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.09950789535758689, | |
| "grad_norm": 0.20328710973262787, | |
| "learning_rate": 0.00048788327234820187, | |
| "loss": 2.8212, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 0.10040436288333092, | |
| "grad_norm": 0.21136407554149628, | |
| "learning_rate": 0.0004876657887145121, | |
| "loss": 2.8116, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 0.10130083040907494, | |
| "grad_norm": 0.20518264174461365, | |
| "learning_rate": 0.00048744641994202397, | |
| "loss": 2.8183, | |
| "step": 11300 | |
| }, | |
| { | |
| "epoch": 0.10219729793481896, | |
| "grad_norm": 0.20106618106365204, | |
| "learning_rate": 0.0004872251677707464, | |
| "loss": 2.8147, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 0.103093765460563, | |
| "grad_norm": 0.20347937941551208, | |
| "learning_rate": 0.00048700203395562715, | |
| "loss": 2.8081, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.10399023298630702, | |
| "grad_norm": 0.1995743364095688, | |
| "learning_rate": 0.000486777020266539, | |
| "loss": 2.8204, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 0.10488670051205104, | |
| "grad_norm": 0.20354920625686646, | |
| "learning_rate": 0.00048655012848826574, | |
| "loss": 2.8077, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 0.10578316803779507, | |
| "grad_norm": 0.1998705416917801, | |
| "learning_rate": 0.00048632136042048784, | |
| "loss": 2.8129, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 0.1066796355635391, | |
| "grad_norm": 0.19472312927246094, | |
| "learning_rate": 0.00048609071787776846, | |
| "loss": 2.8141, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 0.10757610308928313, | |
| "grad_norm": 0.1991991549730301, | |
| "learning_rate": 0.0004858582026895387, | |
| "loss": 2.8148, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.10847257061502714, | |
| "grad_norm": 0.20118767023086548, | |
| "learning_rate": 0.0004856238167000834, | |
| "loss": 2.8126, | |
| "step": 12100 | |
| }, | |
| { | |
| "epoch": 0.10936903814077117, | |
| "grad_norm": 0.20079652965068817, | |
| "learning_rate": 0.0004853875617685263, | |
| "loss": 2.8128, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 0.1102655056665152, | |
| "grad_norm": 0.20117126405239105, | |
| "learning_rate": 0.0004851494397688154, | |
| "loss": 2.8065, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 0.11116197319225923, | |
| "grad_norm": 0.19336390495300293, | |
| "learning_rate": 0.00048490945258970833, | |
| "loss": 2.8034, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 0.11205844071800325, | |
| "grad_norm": 0.20018301904201508, | |
| "learning_rate": 0.00048466760213475665, | |
| "loss": 2.8033, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 0.11295490824374728, | |
| "grad_norm": 0.19468124210834503, | |
| "learning_rate": 0.00048442389032229163, | |
| "loss": 2.7959, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 0.11385137576949131, | |
| "grad_norm": 0.1926909238100052, | |
| "learning_rate": 0.00048417831908540836, | |
| "loss": 2.7999, | |
| "step": 12700 | |
| }, | |
| { | |
| "epoch": 0.11474784329523532, | |
| "grad_norm": 0.19469407200813293, | |
| "learning_rate": 0.00048393089037195085, | |
| "loss": 2.8011, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 0.11564431082097935, | |
| "grad_norm": 0.19478073716163635, | |
| "learning_rate": 0.0004836816061444964, | |
| "loss": 2.8136, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 0.11654077834672338, | |
| "grad_norm": 0.19283322989940643, | |
| "learning_rate": 0.00048343046838033975, | |
| "loss": 2.7914, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.11743724587246741, | |
| "grad_norm": 0.19198790192604065, | |
| "learning_rate": 0.0004831774790714781, | |
| "loss": 2.8133, | |
| "step": 13100 | |
| }, | |
| { | |
| "epoch": 0.11833371339821143, | |
| "grad_norm": 0.18972335755825043, | |
| "learning_rate": 0.0004829226402245948, | |
| "loss": 2.7983, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 0.11923018092395546, | |
| "grad_norm": 0.18818192183971405, | |
| "learning_rate": 0.0004826659538610433, | |
| "loss": 2.7911, | |
| "step": 13300 | |
| }, | |
| { | |
| "epoch": 0.12012664844969949, | |
| "grad_norm": 0.19898109138011932, | |
| "learning_rate": 0.0004824074220168315, | |
| "loss": 2.7961, | |
| "step": 13400 | |
| }, | |
| { | |
| "epoch": 0.12102311597544352, | |
| "grad_norm": 0.1899373084306717, | |
| "learning_rate": 0.00048214704674260544, | |
| "loss": 2.7923, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 0.12191958350118753, | |
| "grad_norm": 0.19022135436534882, | |
| "learning_rate": 0.00048188483010363315, | |
| "loss": 2.7933, | |
| "step": 13600 | |
| }, | |
| { | |
| "epoch": 0.12281605102693156, | |
| "grad_norm": 0.19756515324115753, | |
| "learning_rate": 0.000481620774179788, | |
| "loss": 2.7928, | |
| "step": 13700 | |
| }, | |
| { | |
| "epoch": 0.1237125185526756, | |
| "grad_norm": 0.18114596605300903, | |
| "learning_rate": 0.00048135488106553246, | |
| "loss": 2.7896, | |
| "step": 13800 | |
| }, | |
| { | |
| "epoch": 0.12460898607841962, | |
| "grad_norm": 0.19043464958667755, | |
| "learning_rate": 0.0004810871528699013, | |
| "loss": 2.8021, | |
| "step": 13900 | |
| }, | |
| { | |
| "epoch": 0.12550545360416365, | |
| "grad_norm": 0.18425244092941284, | |
| "learning_rate": 0.000480817591716485, | |
| "loss": 2.7863, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.12640192112990767, | |
| "grad_norm": 0.186711385846138, | |
| "learning_rate": 0.00048054619974341293, | |
| "loss": 2.7866, | |
| "step": 14100 | |
| }, | |
| { | |
| "epoch": 0.12729838865565168, | |
| "grad_norm": 0.18656401336193085, | |
| "learning_rate": 0.00048027297910333634, | |
| "loss": 2.7888, | |
| "step": 14200 | |
| }, | |
| { | |
| "epoch": 0.12819485618139573, | |
| "grad_norm": 0.19122888147830963, | |
| "learning_rate": 0.00048000069146883305, | |
| "loss": 2.7899, | |
| "step": 14300 | |
| }, | |
| { | |
| "epoch": 0.12909132370713974, | |
| "grad_norm": 0.19195199012756348, | |
| "learning_rate": 0.000479723838243038, | |
| "loss": 2.7819, | |
| "step": 14400 | |
| }, | |
| { | |
| "epoch": 0.1299877912328838, | |
| "grad_norm": 0.1867845505475998, | |
| "learning_rate": 0.00047944516287311896, | |
| "loss": 2.7853, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 0.1308842587586278, | |
| "grad_norm": 0.18608401715755463, | |
| "learning_rate": 0.0004791646675694982, | |
| "loss": 2.7899, | |
| "step": 14600 | |
| }, | |
| { | |
| "epoch": 0.13178072628437182, | |
| "grad_norm": 0.18311014771461487, | |
| "learning_rate": 0.0004788823545570335, | |
| "loss": 2.7835, | |
| "step": 14700 | |
| }, | |
| { | |
| "epoch": 0.13267719381011586, | |
| "grad_norm": 0.18312698602676392, | |
| "learning_rate": 0.00047859822607500046, | |
| "loss": 2.7814, | |
| "step": 14800 | |
| }, | |
| { | |
| "epoch": 0.13357366133585988, | |
| "grad_norm": 0.1876700073480606, | |
| "learning_rate": 0.00047831228437707495, | |
| "loss": 2.7925, | |
| "step": 14900 | |
| }, | |
| { | |
| "epoch": 0.1344701288616039, | |
| "grad_norm": 0.18215568363666534, | |
| "learning_rate": 0.0004780245317313149, | |
| "loss": 2.7815, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.13536659638734794, | |
| "grad_norm": 0.18465806543827057, | |
| "learning_rate": 0.0004777349704201426, | |
| "loss": 2.7845, | |
| "step": 15100 | |
| }, | |
| { | |
| "epoch": 0.13626306391309195, | |
| "grad_norm": 0.18615123629570007, | |
| "learning_rate": 0.00047744360274032616, | |
| "loss": 2.7788, | |
| "step": 15200 | |
| }, | |
| { | |
| "epoch": 0.13715953143883597, | |
| "grad_norm": 0.18580564856529236, | |
| "learning_rate": 0.0004771504310029621, | |
| "loss": 2.7727, | |
| "step": 15300 | |
| }, | |
| { | |
| "epoch": 0.13805599896458, | |
| "grad_norm": 0.19075210392475128, | |
| "learning_rate": 0.00047685545753345615, | |
| "loss": 2.7777, | |
| "step": 15400 | |
| }, | |
| { | |
| "epoch": 0.13895246649032403, | |
| "grad_norm": 0.17943909764289856, | |
| "learning_rate": 0.00047655868467150534, | |
| "loss": 2.7842, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 0.13984893401606807, | |
| "grad_norm": 0.19626812636852264, | |
| "learning_rate": 0.00047626011477107925, | |
| "loss": 2.7736, | |
| "step": 15600 | |
| }, | |
| { | |
| "epoch": 0.1407454015418121, | |
| "grad_norm": 0.18597416579723358, | |
| "learning_rate": 0.0004759597502004014, | |
| "loss": 2.7789, | |
| "step": 15700 | |
| }, | |
| { | |
| "epoch": 0.1416418690675561, | |
| "grad_norm": 0.18150849640369415, | |
| "learning_rate": 0.0004756575933419304, | |
| "loss": 2.7771, | |
| "step": 15800 | |
| }, | |
| { | |
| "epoch": 0.14253833659330015, | |
| "grad_norm": 0.17691569030284882, | |
| "learning_rate": 0.00047535364659234095, | |
| "loss": 2.7689, | |
| "step": 15900 | |
| }, | |
| { | |
| "epoch": 0.14343480411904416, | |
| "grad_norm": 0.1855591982603073, | |
| "learning_rate": 0.00047504791236250535, | |
| "loss": 2.7755, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 0.14433127164478818, | |
| "grad_norm": 0.18740758299827576, | |
| "learning_rate": 0.00047474039307747354, | |
| "loss": 2.7756, | |
| "step": 16100 | |
| }, | |
| { | |
| "epoch": 0.14522773917053222, | |
| "grad_norm": 0.17985744774341583, | |
| "learning_rate": 0.00047443109117645466, | |
| "loss": 2.774, | |
| "step": 16200 | |
| }, | |
| { | |
| "epoch": 0.14612420669627624, | |
| "grad_norm": 0.17564240097999573, | |
| "learning_rate": 0.0004741200091127973, | |
| "loss": 2.7719, | |
| "step": 16300 | |
| }, | |
| { | |
| "epoch": 0.14702067422202028, | |
| "grad_norm": 0.17614321410655975, | |
| "learning_rate": 0.00047380714935396986, | |
| "loss": 2.7679, | |
| "step": 16400 | |
| }, | |
| { | |
| "epoch": 0.1479171417477643, | |
| "grad_norm": 0.17652449011802673, | |
| "learning_rate": 0.00047349251438154154, | |
| "loss": 2.77, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 0.1488136092735083, | |
| "grad_norm": 0.1796472817659378, | |
| "learning_rate": 0.0004731761066911622, | |
| "loss": 2.7652, | |
| "step": 16600 | |
| }, | |
| { | |
| "epoch": 0.14971007679925236, | |
| "grad_norm": 0.17830465734004974, | |
| "learning_rate": 0.00047285792879254274, | |
| "loss": 2.7684, | |
| "step": 16700 | |
| }, | |
| { | |
| "epoch": 0.15060654432499637, | |
| "grad_norm": 0.17882998287677765, | |
| "learning_rate": 0.000472537983209435, | |
| "loss": 2.7709, | |
| "step": 16800 | |
| }, | |
| { | |
| "epoch": 0.1515030118507404, | |
| "grad_norm": 0.17409980297088623, | |
| "learning_rate": 0.0004722162724796122, | |
| "loss": 2.7692, | |
| "step": 16900 | |
| }, | |
| { | |
| "epoch": 0.15239947937648443, | |
| "grad_norm": 0.1783556491136551, | |
| "learning_rate": 0.00047189279915484816, | |
| "loss": 2.7736, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 0.15329594690222845, | |
| "grad_norm": 0.17761445045471191, | |
| "learning_rate": 0.00047156756580089766, | |
| "loss": 2.7668, | |
| "step": 17100 | |
| }, | |
| { | |
| "epoch": 0.15419241442797246, | |
| "grad_norm": 0.17673403024673462, | |
| "learning_rate": 0.00047124057499747573, | |
| "loss": 2.7712, | |
| "step": 17200 | |
| }, | |
| { | |
| "epoch": 0.1550888819537165, | |
| "grad_norm": 0.1734510213136673, | |
| "learning_rate": 0.00047091182933823737, | |
| "loss": 2.7579, | |
| "step": 17300 | |
| }, | |
| { | |
| "epoch": 0.15598534947946052, | |
| "grad_norm": 0.17702840268611908, | |
| "learning_rate": 0.0004705813314307569, | |
| "loss": 2.7612, | |
| "step": 17400 | |
| }, | |
| { | |
| "epoch": 0.15688181700520457, | |
| "grad_norm": 0.17082242667675018, | |
| "learning_rate": 0.00047024908389650704, | |
| "loss": 2.7597, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 0.15777828453094858, | |
| "grad_norm": 0.18025079369544983, | |
| "learning_rate": 0.00046991508937083875, | |
| "loss": 2.7559, | |
| "step": 17600 | |
| }, | |
| { | |
| "epoch": 0.1586747520566926, | |
| "grad_norm": 0.17427673935890198, | |
| "learning_rate": 0.00046957935050295963, | |
| "loss": 2.7577, | |
| "step": 17700 | |
| }, | |
| { | |
| "epoch": 0.15957121958243664, | |
| "grad_norm": 0.17229335010051727, | |
| "learning_rate": 0.0004692418699559134, | |
| "loss": 2.7559, | |
| "step": 17800 | |
| }, | |
| { | |
| "epoch": 0.16046768710818066, | |
| "grad_norm": 0.17168091237545013, | |
| "learning_rate": 0.0004689026504065585, | |
| "loss": 2.7582, | |
| "step": 17900 | |
| }, | |
| { | |
| "epoch": 0.16136415463392467, | |
| "grad_norm": 0.17163404822349548, | |
| "learning_rate": 0.0004685616945455469, | |
| "loss": 2.7621, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 0.16226062215966872, | |
| "grad_norm": 0.16957524418830872, | |
| "learning_rate": 0.00046821900507730275, | |
| "loss": 2.7586, | |
| "step": 18100 | |
| }, | |
| { | |
| "epoch": 0.16315708968541273, | |
| "grad_norm": 0.165022075176239, | |
| "learning_rate": 0.0004678745847200012, | |
| "loss": 2.7541, | |
| "step": 18200 | |
| }, | |
| { | |
| "epoch": 0.16405355721115678, | |
| "grad_norm": 0.17016685009002686, | |
| "learning_rate": 0.00046752843620554655, | |
| "loss": 2.7522, | |
| "step": 18300 | |
| }, | |
| { | |
| "epoch": 0.1649500247369008, | |
| "grad_norm": 0.16751013696193695, | |
| "learning_rate": 0.00046718056227955043, | |
| "loss": 2.7501, | |
| "step": 18400 | |
| }, | |
| { | |
| "epoch": 0.1658464922626448, | |
| "grad_norm": 0.1747274398803711, | |
| "learning_rate": 0.0004668309657013106, | |
| "loss": 2.7551, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 0.16674295978838885, | |
| "grad_norm": 0.1704045832157135, | |
| "learning_rate": 0.0004664796492437884, | |
| "loss": 2.7522, | |
| "step": 18600 | |
| }, | |
| { | |
| "epoch": 0.16763942731413287, | |
| "grad_norm": 0.16978523135185242, | |
| "learning_rate": 0.0004661266156935873, | |
| "loss": 2.7494, | |
| "step": 18700 | |
| }, | |
| { | |
| "epoch": 0.16853589483987688, | |
| "grad_norm": 0.1745961755514145, | |
| "learning_rate": 0.0004657718678509303, | |
| "loss": 2.7546, | |
| "step": 18800 | |
| }, | |
| { | |
| "epoch": 0.16943236236562093, | |
| "grad_norm": 0.16892167925834656, | |
| "learning_rate": 0.0004654154085296382, | |
| "loss": 2.7597, | |
| "step": 18900 | |
| }, | |
| { | |
| "epoch": 0.17032882989136494, | |
| "grad_norm": 0.17090590298175812, | |
| "learning_rate": 0.000465057240557107, | |
| "loss": 2.7545, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 0.17122529741710896, | |
| "grad_norm": 0.1694989651441574, | |
| "learning_rate": 0.00046469736677428556, | |
| "loss": 2.7525, | |
| "step": 19100 | |
| }, | |
| { | |
| "epoch": 0.172121764942853, | |
| "grad_norm": 0.16777247190475464, | |
| "learning_rate": 0.00046433579003565286, | |
| "loss": 2.7478, | |
| "step": 19200 | |
| }, | |
| { | |
| "epoch": 0.17301823246859702, | |
| "grad_norm": 0.1667390614748001, | |
| "learning_rate": 0.00046397251320919584, | |
| "loss": 2.7606, | |
| "step": 19300 | |
| }, | |
| { | |
| "epoch": 0.17391469999434106, | |
| "grad_norm": 0.16497531533241272, | |
| "learning_rate": 0.00046360753917638604, | |
| "loss": 2.7459, | |
| "step": 19400 | |
| }, | |
| { | |
| "epoch": 0.17481116752008508, | |
| "grad_norm": 0.16512423753738403, | |
| "learning_rate": 0.00046324087083215727, | |
| "loss": 2.748, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 0.1757076350458291, | |
| "grad_norm": 0.1689414232969284, | |
| "learning_rate": 0.0004628725110848823, | |
| "loss": 2.7553, | |
| "step": 19600 | |
| }, | |
| { | |
| "epoch": 0.17660410257157313, | |
| "grad_norm": 0.17020919919013977, | |
| "learning_rate": 0.00046250246285635, | |
| "loss": 2.7498, | |
| "step": 19700 | |
| }, | |
| { | |
| "epoch": 0.17750057009731715, | |
| "grad_norm": 0.16616366803646088, | |
| "learning_rate": 0.00046213072908174213, | |
| "loss": 2.7459, | |
| "step": 19800 | |
| }, | |
| { | |
| "epoch": 0.17839703762306117, | |
| "grad_norm": 0.16394533216953278, | |
| "learning_rate": 0.00046175731270961, | |
| "loss": 2.7494, | |
| "step": 19900 | |
| }, | |
| { | |
| "epoch": 0.1792935051488052, | |
| "grad_norm": 0.1853574514389038, | |
| "learning_rate": 0.0004613822167018508, | |
| "loss": 2.7406, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 0.18018997267454923, | |
| "grad_norm": 0.16652604937553406, | |
| "learning_rate": 0.00046100544403368483, | |
| "loss": 2.7364, | |
| "step": 20100 | |
| }, | |
| { | |
| "epoch": 0.18108644020029327, | |
| "grad_norm": 0.16384656727313995, | |
| "learning_rate": 0.0004606269976936314, | |
| "loss": 2.747, | |
| "step": 20200 | |
| }, | |
| { | |
| "epoch": 0.18198290772603729, | |
| "grad_norm": 0.16769762337207794, | |
| "learning_rate": 0.00046024688068348484, | |
| "loss": 2.7465, | |
| "step": 20300 | |
| }, | |
| { | |
| "epoch": 0.1828793752517813, | |
| "grad_norm": 0.15770003199577332, | |
| "learning_rate": 0.0004598650960182915, | |
| "loss": 2.7459, | |
| "step": 20400 | |
| }, | |
| { | |
| "epoch": 0.18377584277752534, | |
| "grad_norm": 0.16936561465263367, | |
| "learning_rate": 0.00045948164672632514, | |
| "loss": 2.7496, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 0.18467231030326936, | |
| "grad_norm": 0.16903486847877502, | |
| "learning_rate": 0.0004590965358490632, | |
| "loss": 2.7447, | |
| "step": 20600 | |
| }, | |
| { | |
| "epoch": 0.18556877782901338, | |
| "grad_norm": 0.16535791754722595, | |
| "learning_rate": 0.00045870976644116267, | |
| "loss": 2.7375, | |
| "step": 20700 | |
| }, | |
| { | |
| "epoch": 0.18646524535475742, | |
| "grad_norm": 0.16924481093883514, | |
| "learning_rate": 0.0004583213415704358, | |
| "loss": 2.7494, | |
| "step": 20800 | |
| }, | |
| { | |
| "epoch": 0.18736171288050144, | |
| "grad_norm": 0.16382168233394623, | |
| "learning_rate": 0.00045793126431782584, | |
| "loss": 2.7411, | |
| "step": 20900 | |
| }, | |
| { | |
| "epoch": 0.18825818040624548, | |
| "grad_norm": 0.16305121779441833, | |
| "learning_rate": 0.00045753953777738233, | |
| "loss": 2.7465, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 0.1891546479319895, | |
| "grad_norm": 0.1712082326412201, | |
| "learning_rate": 0.00045714616505623703, | |
| "loss": 2.7368, | |
| "step": 21100 | |
| }, | |
| { | |
| "epoch": 0.1900511154577335, | |
| "grad_norm": 0.15486562252044678, | |
| "learning_rate": 0.00045675114927457895, | |
| "loss": 2.7385, | |
| "step": 21200 | |
| }, | |
| { | |
| "epoch": 0.19094758298347755, | |
| "grad_norm": 0.16223442554473877, | |
| "learning_rate": 0.0004563544935656296, | |
| "loss": 2.7404, | |
| "step": 21300 | |
| }, | |
| { | |
| "epoch": 0.19184405050922157, | |
| "grad_norm": 0.16115334630012512, | |
| "learning_rate": 0.0004559562010756185, | |
| "loss": 2.7325, | |
| "step": 21400 | |
| }, | |
| { | |
| "epoch": 0.19274051803496559, | |
| "grad_norm": 0.15933333337306976, | |
| "learning_rate": 0.0004555562749637574, | |
| "loss": 2.7438, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 0.19363698556070963, | |
| "grad_norm": 0.16439130902290344, | |
| "learning_rate": 0.00045515471840221623, | |
| "loss": 2.7353, | |
| "step": 21600 | |
| }, | |
| { | |
| "epoch": 0.19453345308645364, | |
| "grad_norm": 0.1617208570241928, | |
| "learning_rate": 0.0004547515345760973, | |
| "loss": 2.7307, | |
| "step": 21700 | |
| }, | |
| { | |
| "epoch": 0.19542992061219766, | |
| "grad_norm": 0.16066919267177582, | |
| "learning_rate": 0.00045434672668341017, | |
| "loss": 2.7379, | |
| "step": 21800 | |
| }, | |
| { | |
| "epoch": 0.1963263881379417, | |
| "grad_norm": 0.16751715540885925, | |
| "learning_rate": 0.0004539402979350464, | |
| "loss": 2.7363, | |
| "step": 21900 | |
| }, | |
| { | |
| "epoch": 0.19722285566368572, | |
| "grad_norm": 0.16093355417251587, | |
| "learning_rate": 0.00045353225155475384, | |
| "loss": 2.7412, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 0.19811932318942976, | |
| "grad_norm": 0.16404040157794952, | |
| "learning_rate": 0.00045312259077911153, | |
| "loss": 2.7436, | |
| "step": 22100 | |
| }, | |
| { | |
| "epoch": 0.19901579071517378, | |
| "grad_norm": 0.15982191264629364, | |
| "learning_rate": 0.0004527113188575034, | |
| "loss": 2.735, | |
| "step": 22200 | |
| }, | |
| { | |
| "epoch": 0.1999122582409178, | |
| "grad_norm": 0.15941210091114044, | |
| "learning_rate": 0.00045229843905209287, | |
| "loss": 2.7457, | |
| "step": 22300 | |
| }, | |
| { | |
| "epoch": 0.20080872576666184, | |
| "grad_norm": 0.15950894355773926, | |
| "learning_rate": 0.00045188395463779707, | |
| "loss": 2.736, | |
| "step": 22400 | |
| }, | |
| { | |
| "epoch": 0.20170519329240585, | |
| "grad_norm": 0.1595742106437683, | |
| "learning_rate": 0.0004514678689022606, | |
| "loss": 2.7345, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 0.20260166081814987, | |
| "grad_norm": 0.16472984850406647, | |
| "learning_rate": 0.00045105018514582953, | |
| "loss": 2.7316, | |
| "step": 22600 | |
| }, | |
| { | |
| "epoch": 0.2034981283438939, | |
| "grad_norm": 0.1674964427947998, | |
| "learning_rate": 0.0004506309066815254, | |
| "loss": 2.7273, | |
| "step": 22700 | |
| }, | |
| { | |
| "epoch": 0.20439459586963793, | |
| "grad_norm": 0.159923255443573, | |
| "learning_rate": 0.00045021003683501853, | |
| "loss": 2.7275, | |
| "step": 22800 | |
| }, | |
| { | |
| "epoch": 0.20529106339538197, | |
| "grad_norm": 0.16554875671863556, | |
| "learning_rate": 0.0004497875789446023, | |
| "loss": 2.7388, | |
| "step": 22900 | |
| }, | |
| { | |
| "epoch": 0.206187530921126, | |
| "grad_norm": 0.16701550781726837, | |
| "learning_rate": 0.0004493635363611659, | |
| "loss": 2.7303, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 0.20708399844687, | |
| "grad_norm": 0.16078519821166992, | |
| "learning_rate": 0.0004489379124481683, | |
| "loss": 2.7312, | |
| "step": 23100 | |
| }, | |
| { | |
| "epoch": 0.20798046597261405, | |
| "grad_norm": 0.16200828552246094, | |
| "learning_rate": 0.0004485107105816115, | |
| "loss": 2.7358, | |
| "step": 23200 | |
| }, | |
| { | |
| "epoch": 0.20887693349835806, | |
| "grad_norm": 0.15977811813354492, | |
| "learning_rate": 0.00044808193415001344, | |
| "loss": 2.7309, | |
| "step": 23300 | |
| }, | |
| { | |
| "epoch": 0.20977340102410208, | |
| "grad_norm": 0.16767093539237976, | |
| "learning_rate": 0.0004476515865543815, | |
| "loss": 2.7315, | |
| "step": 23400 | |
| }, | |
| { | |
| "epoch": 0.21066986854984612, | |
| "grad_norm": 0.15728136897087097, | |
| "learning_rate": 0.0004472196712081852, | |
| "loss": 2.735, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 0.21156633607559014, | |
| "grad_norm": 0.15546581149101257, | |
| "learning_rate": 0.0004467861915373295, | |
| "loss": 2.7301, | |
| "step": 23600 | |
| }, | |
| { | |
| "epoch": 0.21246280360133415, | |
| "grad_norm": 0.15766388177871704, | |
| "learning_rate": 0.0004463511509801273, | |
| "loss": 2.7248, | |
| "step": 23700 | |
| }, | |
| { | |
| "epoch": 0.2133592711270782, | |
| "grad_norm": 0.15457496047019958, | |
| "learning_rate": 0.00044591455298727213, | |
| "loss": 2.7234, | |
| "step": 23800 | |
| }, | |
| { | |
| "epoch": 0.21425573865282221, | |
| "grad_norm": 0.15532919764518738, | |
| "learning_rate": 0.0004454764010218112, | |
| "loss": 2.7234, | |
| "step": 23900 | |
| }, | |
| { | |
| "epoch": 0.21515220617856626, | |
| "grad_norm": 0.15373288094997406, | |
| "learning_rate": 0.00044503669855911756, | |
| "loss": 2.7254, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 0.21604867370431027, | |
| "grad_norm": 0.15703774988651276, | |
| "learning_rate": 0.00044459544908686236, | |
| "loss": 2.7221, | |
| "step": 24100 | |
| }, | |
| { | |
| "epoch": 0.2169451412300543, | |
| "grad_norm": 0.15091052651405334, | |
| "learning_rate": 0.000444152656104988, | |
| "loss": 2.7288, | |
| "step": 24200 | |
| }, | |
| { | |
| "epoch": 0.21784160875579833, | |
| "grad_norm": 0.15579521656036377, | |
| "learning_rate": 0.0004437083231256794, | |
| "loss": 2.7192, | |
| "step": 24300 | |
| }, | |
| { | |
| "epoch": 0.21873807628154235, | |
| "grad_norm": 0.15548183023929596, | |
| "learning_rate": 0.0004432624536733367, | |
| "loss": 2.735, | |
| "step": 24400 | |
| }, | |
| { | |
| "epoch": 0.21963454380728636, | |
| "grad_norm": 0.15607748925685883, | |
| "learning_rate": 0.00044281505128454713, | |
| "loss": 2.7218, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 0.2205310113330304, | |
| "grad_norm": 0.16912153363227844, | |
| "learning_rate": 0.00044236611950805707, | |
| "loss": 2.7198, | |
| "step": 24600 | |
| }, | |
| { | |
| "epoch": 0.22142747885877442, | |
| "grad_norm": 0.15691158175468445, | |
| "learning_rate": 0.0004419156619047439, | |
| "loss": 2.7233, | |
| "step": 24700 | |
| }, | |
| { | |
| "epoch": 0.22232394638451847, | |
| "grad_norm": 0.1513551026582718, | |
| "learning_rate": 0.0004414636820475875, | |
| "loss": 2.7198, | |
| "step": 24800 | |
| }, | |
| { | |
| "epoch": 0.22322041391026248, | |
| "grad_norm": 0.15347003936767578, | |
| "learning_rate": 0.0004410101835216422, | |
| "loss": 2.7185, | |
| "step": 24900 | |
| }, | |
| { | |
| "epoch": 0.2241168814360065, | |
| "grad_norm": 0.15076322853565216, | |
| "learning_rate": 0.00044055516992400827, | |
| "loss": 2.7232, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 0.22501334896175054, | |
| "grad_norm": 0.1548827737569809, | |
| "learning_rate": 0.00044009864486380335, | |
| "loss": 2.7195, | |
| "step": 25100 | |
| }, | |
| { | |
| "epoch": 0.22590981648749456, | |
| "grad_norm": 0.15956267714500427, | |
| "learning_rate": 0.0004396406119621338, | |
| "loss": 2.7231, | |
| "step": 25200 | |
| }, | |
| { | |
| "epoch": 0.22680628401323857, | |
| "grad_norm": 0.14837653934955597, | |
| "learning_rate": 0.00043918107485206603, | |
| "loss": 2.7211, | |
| "step": 25300 | |
| }, | |
| { | |
| "epoch": 0.22770275153898262, | |
| "grad_norm": 0.15196016430854797, | |
| "learning_rate": 0.0004387200371785977, | |
| "loss": 2.7209, | |
| "step": 25400 | |
| }, | |
| { | |
| "epoch": 0.22859921906472663, | |
| "grad_norm": 0.15298311412334442, | |
| "learning_rate": 0.00043825750259862873, | |
| "loss": 2.7132, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 0.22949568659047065, | |
| "grad_norm": 0.16172164678573608, | |
| "learning_rate": 0.00043779347478093225, | |
| "loss": 2.72, | |
| "step": 25600 | |
| }, | |
| { | |
| "epoch": 0.2303921541162147, | |
| "grad_norm": 0.15316250920295715, | |
| "learning_rate": 0.0004373279574061258, | |
| "loss": 2.7284, | |
| "step": 25700 | |
| }, | |
| { | |
| "epoch": 0.2312886216419587, | |
| "grad_norm": 0.1559763252735138, | |
| "learning_rate": 0.0004368609541666417, | |
| "loss": 2.7287, | |
| "step": 25800 | |
| }, | |
| { | |
| "epoch": 0.23218508916770275, | |
| "grad_norm": 0.15110167860984802, | |
| "learning_rate": 0.0004363924687666982, | |
| "loss": 2.7144, | |
| "step": 25900 | |
| }, | |
| { | |
| "epoch": 0.23308155669344677, | |
| "grad_norm": 0.15237966179847717, | |
| "learning_rate": 0.0004359225049222697, | |
| "loss": 2.7236, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 0.23397802421919078, | |
| "grad_norm": 0.1501353681087494, | |
| "learning_rate": 0.0004354510663610576, | |
| "loss": 2.7132, | |
| "step": 26100 | |
| }, | |
| { | |
| "epoch": 0.23487449174493483, | |
| "grad_norm": 0.15224196016788483, | |
| "learning_rate": 0.00043497815682246044, | |
| "loss": 2.7213, | |
| "step": 26200 | |
| }, | |
| { | |
| "epoch": 0.23577095927067884, | |
| "grad_norm": 0.1475629061460495, | |
| "learning_rate": 0.00043450378005754453, | |
| "loss": 2.7138, | |
| "step": 26300 | |
| }, | |
| { | |
| "epoch": 0.23666742679642286, | |
| "grad_norm": 0.15193389356136322, | |
| "learning_rate": 0.000434027939829014, | |
| "loss": 2.7079, | |
| "step": 26400 | |
| }, | |
| { | |
| "epoch": 0.2375638943221669, | |
| "grad_norm": 0.1544477641582489, | |
| "learning_rate": 0.00043355063991118095, | |
| "loss": 2.7213, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 0.23846036184791092, | |
| "grad_norm": 0.14745964109897614, | |
| "learning_rate": 0.0004330718840899357, | |
| "loss": 2.7175, | |
| "step": 26600 | |
| }, | |
| { | |
| "epoch": 0.23935682937365496, | |
| "grad_norm": 0.1513916403055191, | |
| "learning_rate": 0.00043259167616271644, | |
| "loss": 2.7168, | |
| "step": 26700 | |
| }, | |
| { | |
| "epoch": 0.24025329689939898, | |
| "grad_norm": 0.15179474651813507, | |
| "learning_rate": 0.00043211001993847967, | |
| "loss": 2.7099, | |
| "step": 26800 | |
| }, | |
| { | |
| "epoch": 0.241149764425143, | |
| "grad_norm": 0.1514696627855301, | |
| "learning_rate": 0.00043162691923766917, | |
| "loss": 2.7201, | |
| "step": 26900 | |
| }, | |
| { | |
| "epoch": 0.24204623195088704, | |
| "grad_norm": 0.15021684765815735, | |
| "learning_rate": 0.0004311423778921865, | |
| "loss": 2.7178, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 0.24294269947663105, | |
| "grad_norm": 0.15158309042453766, | |
| "learning_rate": 0.0004306563997453601, | |
| "loss": 2.7076, | |
| "step": 27100 | |
| }, | |
| { | |
| "epoch": 0.24383916700237507, | |
| "grad_norm": 0.14725959300994873, | |
| "learning_rate": 0.00043016898865191487, | |
| "loss": 2.7078, | |
| "step": 27200 | |
| }, | |
| { | |
| "epoch": 0.2447356345281191, | |
| "grad_norm": 0.14671307802200317, | |
| "learning_rate": 0.0004296801484779419, | |
| "loss": 2.7109, | |
| "step": 27300 | |
| }, | |
| { | |
| "epoch": 0.24563210205386313, | |
| "grad_norm": 0.1534958928823471, | |
| "learning_rate": 0.0004291898831008675, | |
| "loss": 2.7154, | |
| "step": 27400 | |
| }, | |
| { | |
| "epoch": 0.24652856957960714, | |
| "grad_norm": 0.14696376025676727, | |
| "learning_rate": 0.00042869819640942245, | |
| "loss": 2.704, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 0.2474250371053512, | |
| "grad_norm": 0.15045034885406494, | |
| "learning_rate": 0.00042820509230361134, | |
| "loss": 2.7145, | |
| "step": 27600 | |
| }, | |
| { | |
| "epoch": 0.2483215046310952, | |
| "grad_norm": 0.1504196673631668, | |
| "learning_rate": 0.0004277105746946814, | |
| "loss": 2.7076, | |
| "step": 27700 | |
| }, | |
| { | |
| "epoch": 0.24921797215683925, | |
| "grad_norm": 0.14724665880203247, | |
| "learning_rate": 0.00042721464750509195, | |
| "loss": 2.7034, | |
| "step": 27800 | |
| }, | |
| { | |
| "epoch": 0.25011443968258323, | |
| "grad_norm": 0.14712798595428467, | |
| "learning_rate": 0.00042671731466848253, | |
| "loss": 2.7139, | |
| "step": 27900 | |
| }, | |
| { | |
| "epoch": 0.2510109072083273, | |
| "grad_norm": 0.15577539801597595, | |
| "learning_rate": 0.0004262185801296422, | |
| "loss": 2.7077, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 0.2519073747340713, | |
| "grad_norm": 0.14991851150989532, | |
| "learning_rate": 0.0004257184478444785, | |
| "loss": 2.7028, | |
| "step": 28100 | |
| }, | |
| { | |
| "epoch": 0.25280384225981534, | |
| "grad_norm": 0.14650234580039978, | |
| "learning_rate": 0.00042521692177998537, | |
| "loss": 2.7069, | |
| "step": 28200 | |
| }, | |
| { | |
| "epoch": 0.25370030978555935, | |
| "grad_norm": 0.1475357860326767, | |
| "learning_rate": 0.0004247140059142123, | |
| "loss": 2.7074, | |
| "step": 28300 | |
| }, | |
| { | |
| "epoch": 0.25459677731130337, | |
| "grad_norm": 0.14877575635910034, | |
| "learning_rate": 0.0004242097042362322, | |
| "loss": 2.6997, | |
| "step": 28400 | |
| }, | |
| { | |
| "epoch": 0.25549324483704744, | |
| "grad_norm": 0.15285931527614594, | |
| "learning_rate": 0.0004237040207461104, | |
| "loss": 2.7016, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 0.25638971236279146, | |
| "grad_norm": 0.1494821012020111, | |
| "learning_rate": 0.00042319695945487257, | |
| "loss": 2.7093, | |
| "step": 28600 | |
| }, | |
| { | |
| "epoch": 0.25728617988853547, | |
| "grad_norm": 0.14891798794269562, | |
| "learning_rate": 0.00042268852438447297, | |
| "loss": 2.7037, | |
| "step": 28700 | |
| }, | |
| { | |
| "epoch": 0.2581826474142795, | |
| "grad_norm": 0.14432930946350098, | |
| "learning_rate": 0.0004221787195677623, | |
| "loss": 2.6977, | |
| "step": 28800 | |
| }, | |
| { | |
| "epoch": 0.2590791149400235, | |
| "grad_norm": 0.1493179351091385, | |
| "learning_rate": 0.0004216675490484561, | |
| "loss": 2.7135, | |
| "step": 28900 | |
| }, | |
| { | |
| "epoch": 0.2599755824657676, | |
| "grad_norm": 0.14740417897701263, | |
| "learning_rate": 0.00042115501688110257, | |
| "loss": 2.7079, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 0.2608720499915116, | |
| "grad_norm": 0.1458692103624344, | |
| "learning_rate": 0.0004206411271310502, | |
| "loss": 2.7029, | |
| "step": 29100 | |
| }, | |
| { | |
| "epoch": 0.2617685175172556, | |
| "grad_norm": 0.14873549342155457, | |
| "learning_rate": 0.00042012588387441586, | |
| "loss": 2.7096, | |
| "step": 29200 | |
| }, | |
| { | |
| "epoch": 0.2626649850429996, | |
| "grad_norm": 0.14170105755329132, | |
| "learning_rate": 0.00041960929119805215, | |
| "loss": 2.7045, | |
| "step": 29300 | |
| }, | |
| { | |
| "epoch": 0.26356145256874364, | |
| "grad_norm": 0.14562779664993286, | |
| "learning_rate": 0.00041909135319951495, | |
| "loss": 2.7044, | |
| "step": 29400 | |
| }, | |
| { | |
| "epoch": 0.26445792009448765, | |
| "grad_norm": 0.14775702357292175, | |
| "learning_rate": 0.00041857207398703154, | |
| "loss": 2.7016, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 0.2653543876202317, | |
| "grad_norm": 0.14575980603694916, | |
| "learning_rate": 0.0004180514576794673, | |
| "loss": 2.7095, | |
| "step": 29600 | |
| }, | |
| { | |
| "epoch": 0.26625085514597574, | |
| "grad_norm": 0.14411023259162903, | |
| "learning_rate": 0.0004175295084062931, | |
| "loss": 2.7031, | |
| "step": 29700 | |
| }, | |
| { | |
| "epoch": 0.26714732267171976, | |
| "grad_norm": 0.15136118233203888, | |
| "learning_rate": 0.0004170062303075531, | |
| "loss": 2.7052, | |
| "step": 29800 | |
| }, | |
| { | |
| "epoch": 0.26804379019746377, | |
| "grad_norm": 0.14870133996009827, | |
| "learning_rate": 0.00041648162753383144, | |
| "loss": 2.7044, | |
| "step": 29900 | |
| }, | |
| { | |
| "epoch": 0.2689402577232078, | |
| "grad_norm": 0.1507934033870697, | |
| "learning_rate": 0.0004159557042462193, | |
| "loss": 2.7113, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 0.26983672524895186, | |
| "grad_norm": 0.14786238968372345, | |
| "learning_rate": 0.0004154284646162822, | |
| "loss": 2.6978, | |
| "step": 30100 | |
| }, | |
| { | |
| "epoch": 0.2707331927746959, | |
| "grad_norm": 0.14871017634868622, | |
| "learning_rate": 0.00041489991282602667, | |
| "loss": 2.6969, | |
| "step": 30200 | |
| }, | |
| { | |
| "epoch": 0.2716296603004399, | |
| "grad_norm": 0.1434573084115982, | |
| "learning_rate": 0.00041437005306786716, | |
| "loss": 2.7015, | |
| "step": 30300 | |
| }, | |
| { | |
| "epoch": 0.2725261278261839, | |
| "grad_norm": 0.14657068252563477, | |
| "learning_rate": 0.0004138388895445928, | |
| "loss": 2.6979, | |
| "step": 30400 | |
| }, | |
| { | |
| "epoch": 0.2734225953519279, | |
| "grad_norm": 0.14329074323177338, | |
| "learning_rate": 0.00041330642646933397, | |
| "loss": 2.7004, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 0.27431906287767194, | |
| "grad_norm": 0.14054018259048462, | |
| "learning_rate": 0.00041277266806552906, | |
| "loss": 2.7023, | |
| "step": 30600 | |
| }, | |
| { | |
| "epoch": 0.275215530403416, | |
| "grad_norm": 0.1455359160900116, | |
| "learning_rate": 0.00041223761856689067, | |
| "loss": 2.6972, | |
| "step": 30700 | |
| }, | |
| { | |
| "epoch": 0.27611199792916, | |
| "grad_norm": 0.14418181777000427, | |
| "learning_rate": 0.0004117012822173725, | |
| "loss": 2.7018, | |
| "step": 30800 | |
| }, | |
| { | |
| "epoch": 0.27700846545490404, | |
| "grad_norm": 0.14956611394882202, | |
| "learning_rate": 0.0004111636632711353, | |
| "loss": 2.7062, | |
| "step": 30900 | |
| }, | |
| { | |
| "epoch": 0.27790493298064806, | |
| "grad_norm": 0.15083587169647217, | |
| "learning_rate": 0.00041062476599251297, | |
| "loss": 2.6939, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 0.27880140050639207, | |
| "grad_norm": 0.14435066282749176, | |
| "learning_rate": 0.00041008459465597947, | |
| "loss": 2.699, | |
| "step": 31100 | |
| }, | |
| { | |
| "epoch": 0.27969786803213614, | |
| "grad_norm": 0.14453698694705963, | |
| "learning_rate": 0.00040954315354611403, | |
| "loss": 2.7017, | |
| "step": 31200 | |
| }, | |
| { | |
| "epoch": 0.28059433555788016, | |
| "grad_norm": 0.14497888088226318, | |
| "learning_rate": 0.0004090004469575679, | |
| "loss": 2.7007, | |
| "step": 31300 | |
| }, | |
| { | |
| "epoch": 0.2814908030836242, | |
| "grad_norm": 0.1438531130552292, | |
| "learning_rate": 0.0004084564791950298, | |
| "loss": 2.7045, | |
| "step": 31400 | |
| }, | |
| { | |
| "epoch": 0.2823872706093682, | |
| "grad_norm": 0.14402426779270172, | |
| "learning_rate": 0.0004079112545731919, | |
| "loss": 2.6936, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 0.2832837381351122, | |
| "grad_norm": 0.14424774050712585, | |
| "learning_rate": 0.00040736477741671576, | |
| "loss": 2.6976, | |
| "step": 31600 | |
| }, | |
| { | |
| "epoch": 0.2841802056608562, | |
| "grad_norm": 0.1417878419160843, | |
| "learning_rate": 0.0004068170520601978, | |
| "loss": 2.6965, | |
| "step": 31700 | |
| }, | |
| { | |
| "epoch": 0.2850766731866003, | |
| "grad_norm": 0.14787960052490234, | |
| "learning_rate": 0.0004062680828481352, | |
| "loss": 2.6921, | |
| "step": 31800 | |
| }, | |
| { | |
| "epoch": 0.2859731407123443, | |
| "grad_norm": 0.14073611795902252, | |
| "learning_rate": 0.00040571787413489104, | |
| "loss": 2.6997, | |
| "step": 31900 | |
| }, | |
| { | |
| "epoch": 0.2868696082380883, | |
| "grad_norm": 0.13744498789310455, | |
| "learning_rate": 0.0004051664302846601, | |
| "loss": 2.6971, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 0.28776607576383234, | |
| "grad_norm": 0.1451827883720398, | |
| "learning_rate": 0.00040461375567143413, | |
| "loss": 2.6972, | |
| "step": 32100 | |
| }, | |
| { | |
| "epoch": 0.28866254328957636, | |
| "grad_norm": 0.14334948360919952, | |
| "learning_rate": 0.0004040598546789672, | |
| "loss": 2.6915, | |
| "step": 32200 | |
| }, | |
| { | |
| "epoch": 0.28955901081532043, | |
| "grad_norm": 0.14681456983089447, | |
| "learning_rate": 0.00040350473170074075, | |
| "loss": 2.6943, | |
| "step": 32300 | |
| }, | |
| { | |
| "epoch": 0.29045547834106444, | |
| "grad_norm": 0.14370225369930267, | |
| "learning_rate": 0.00040294839113992907, | |
| "loss": 2.6919, | |
| "step": 32400 | |
| }, | |
| { | |
| "epoch": 0.29135194586680846, | |
| "grad_norm": 0.14142639935016632, | |
| "learning_rate": 0.000402390837409364, | |
| "loss": 2.6997, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 0.2922484133925525, | |
| "grad_norm": 0.14680539071559906, | |
| "learning_rate": 0.0004018320749315001, | |
| "loss": 2.6939, | |
| "step": 32600 | |
| }, | |
| { | |
| "epoch": 0.2931448809182965, | |
| "grad_norm": 0.14420635998249054, | |
| "learning_rate": 0.00040127210813837975, | |
| "loss": 2.6892, | |
| "step": 32700 | |
| }, | |
| { | |
| "epoch": 0.29404134844404056, | |
| "grad_norm": 0.1437712162733078, | |
| "learning_rate": 0.00040071094147159776, | |
| "loss": 2.695, | |
| "step": 32800 | |
| }, | |
| { | |
| "epoch": 0.2949378159697846, | |
| "grad_norm": 0.14098462462425232, | |
| "learning_rate": 0.00040014857938226614, | |
| "loss": 2.6851, | |
| "step": 32900 | |
| }, | |
| { | |
| "epoch": 0.2958342834955286, | |
| "grad_norm": 0.13981753587722778, | |
| "learning_rate": 0.00039958502633097895, | |
| "loss": 2.6938, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 0.2967307510212726, | |
| "grad_norm": 0.13768906891345978, | |
| "learning_rate": 0.0003990202867877766, | |
| "loss": 2.6834, | |
| "step": 33100 | |
| }, | |
| { | |
| "epoch": 0.2976272185470166, | |
| "grad_norm": 0.13921628892421722, | |
| "learning_rate": 0.000398454365232111, | |
| "loss": 2.6977, | |
| "step": 33200 | |
| }, | |
| { | |
| "epoch": 0.29852368607276064, | |
| "grad_norm": 0.14550244808197021, | |
| "learning_rate": 0.0003978872661528094, | |
| "loss": 2.6897, | |
| "step": 33300 | |
| }, | |
| { | |
| "epoch": 0.2994201535985047, | |
| "grad_norm": 0.13877898454666138, | |
| "learning_rate": 0.00039731899404803905, | |
| "loss": 2.7007, | |
| "step": 33400 | |
| }, | |
| { | |
| "epoch": 0.30031662112424873, | |
| "grad_norm": 0.14092972874641418, | |
| "learning_rate": 0.00039674955342527165, | |
| "loss": 2.6964, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 0.30121308864999274, | |
| "grad_norm": 0.13762860000133514, | |
| "learning_rate": 0.00039617894880124716, | |
| "loss": 2.6908, | |
| "step": 33600 | |
| }, | |
| { | |
| "epoch": 0.30210955617573676, | |
| "grad_norm": 0.14549992978572845, | |
| "learning_rate": 0.00039560718470193866, | |
| "loss": 2.6874, | |
| "step": 33700 | |
| }, | |
| { | |
| "epoch": 0.3030060237014808, | |
| "grad_norm": 0.1463141292333603, | |
| "learning_rate": 0.00039503426566251575, | |
| "loss": 2.6867, | |
| "step": 33800 | |
| }, | |
| { | |
| "epoch": 0.30390249122722485, | |
| "grad_norm": 0.13913968205451965, | |
| "learning_rate": 0.0003944601962273091, | |
| "loss": 2.6949, | |
| "step": 33900 | |
| }, | |
| { | |
| "epoch": 0.30479895875296886, | |
| "grad_norm": 0.14150671660900116, | |
| "learning_rate": 0.000393884980949774, | |
| "loss": 2.6921, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 0.3056954262787129, | |
| "grad_norm": 0.14346466958522797, | |
| "learning_rate": 0.0003933086243924545, | |
| "loss": 2.6935, | |
| "step": 34100 | |
| }, | |
| { | |
| "epoch": 0.3065918938044569, | |
| "grad_norm": 0.14826858043670654, | |
| "learning_rate": 0.00039273113112694736, | |
| "loss": 2.6841, | |
| "step": 34200 | |
| }, | |
| { | |
| "epoch": 0.3074883613302009, | |
| "grad_norm": 0.13878753781318665, | |
| "learning_rate": 0.0003921525057338652, | |
| "loss": 2.6858, | |
| "step": 34300 | |
| }, | |
| { | |
| "epoch": 0.3083848288559449, | |
| "grad_norm": 0.1484604924917221, | |
| "learning_rate": 0.0003915727528028009, | |
| "loss": 2.6928, | |
| "step": 34400 | |
| }, | |
| { | |
| "epoch": 0.309281296381689, | |
| "grad_norm": 0.14599835872650146, | |
| "learning_rate": 0.00039099187693229066, | |
| "loss": 2.6878, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 0.310177763907433, | |
| "grad_norm": 0.14739172160625458, | |
| "learning_rate": 0.0003904098827297777, | |
| "loss": 2.6846, | |
| "step": 34600 | |
| }, | |
| { | |
| "epoch": 0.31107423143317703, | |
| "grad_norm": 0.14112310111522675, | |
| "learning_rate": 0.0003898267748115759, | |
| "loss": 2.6879, | |
| "step": 34700 | |
| }, | |
| { | |
| "epoch": 0.31197069895892104, | |
| "grad_norm": 0.13632678985595703, | |
| "learning_rate": 0.00038924255780283277, | |
| "loss": 2.6899, | |
| "step": 34800 | |
| }, | |
| { | |
| "epoch": 0.31286716648466506, | |
| "grad_norm": 0.14148631691932678, | |
| "learning_rate": 0.0003886572363374933, | |
| "loss": 2.6901, | |
| "step": 34900 | |
| }, | |
| { | |
| "epoch": 0.31376363401040913, | |
| "grad_norm": 0.14057014882564545, | |
| "learning_rate": 0.0003880708150582626, | |
| "loss": 2.6896, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 0.31466010153615315, | |
| "grad_norm": 0.1407003551721573, | |
| "learning_rate": 0.00038748329861656945, | |
| "loss": 2.6833, | |
| "step": 35100 | |
| }, | |
| { | |
| "epoch": 0.31555656906189716, | |
| "grad_norm": 0.1437043398618698, | |
| "learning_rate": 0.00038689469167252957, | |
| "loss": 2.6878, | |
| "step": 35200 | |
| }, | |
| { | |
| "epoch": 0.3164530365876412, | |
| "grad_norm": 0.13795500993728638, | |
| "learning_rate": 0.0003863049988949081, | |
| "loss": 2.6831, | |
| "step": 35300 | |
| }, | |
| { | |
| "epoch": 0.3173495041133852, | |
| "grad_norm": 0.13427403569221497, | |
| "learning_rate": 0.0003857142249610833, | |
| "loss": 2.6855, | |
| "step": 35400 | |
| }, | |
| { | |
| "epoch": 0.31824597163912927, | |
| "grad_norm": 0.14448797702789307, | |
| "learning_rate": 0.0003851223745570085, | |
| "loss": 2.6934, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 0.3191424391648733, | |
| "grad_norm": 0.13982614874839783, | |
| "learning_rate": 0.00038452945237717597, | |
| "loss": 2.6853, | |
| "step": 35600 | |
| }, | |
| { | |
| "epoch": 0.3200389066906173, | |
| "grad_norm": 0.1487363576889038, | |
| "learning_rate": 0.0003839354631245789, | |
| "loss": 2.6812, | |
| "step": 35700 | |
| }, | |
| { | |
| "epoch": 0.3209353742163613, | |
| "grad_norm": 0.1380891501903534, | |
| "learning_rate": 0.00038334041151067447, | |
| "loss": 2.6825, | |
| "step": 35800 | |
| }, | |
| { | |
| "epoch": 0.32183184174210533, | |
| "grad_norm": 0.14058321714401245, | |
| "learning_rate": 0.00038274430225534637, | |
| "loss": 2.6868, | |
| "step": 35900 | |
| }, | |
| { | |
| "epoch": 0.32272830926784934, | |
| "grad_norm": 0.1391235589981079, | |
| "learning_rate": 0.00038214714008686745, | |
| "loss": 2.6914, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 0.3236247767935934, | |
| "grad_norm": 0.13752683997154236, | |
| "learning_rate": 0.0003815489297418621, | |
| "loss": 2.6854, | |
| "step": 36100 | |
| }, | |
| { | |
| "epoch": 0.32452124431933743, | |
| "grad_norm": 0.14746305346488953, | |
| "learning_rate": 0.00038094967596526873, | |
| "loss": 2.6827, | |
| "step": 36200 | |
| }, | |
| { | |
| "epoch": 0.32541771184508145, | |
| "grad_norm": 0.14420130848884583, | |
| "learning_rate": 0.00038034938351030206, | |
| "loss": 2.6876, | |
| "step": 36300 | |
| }, | |
| { | |
| "epoch": 0.32631417937082546, | |
| "grad_norm": 0.1397886425256729, | |
| "learning_rate": 0.0003797480571384157, | |
| "loss": 2.6798, | |
| "step": 36400 | |
| }, | |
| { | |
| "epoch": 0.3272106468965695, | |
| "grad_norm": 0.14020013809204102, | |
| "learning_rate": 0.00037914570161926405, | |
| "loss": 2.6868, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 0.32810711442231355, | |
| "grad_norm": 0.13931307196617126, | |
| "learning_rate": 0.0003785423217306645, | |
| "loss": 2.6836, | |
| "step": 36600 | |
| }, | |
| { | |
| "epoch": 0.32900358194805757, | |
| "grad_norm": 0.14019936323165894, | |
| "learning_rate": 0.00037793792225855973, | |
| "loss": 2.6871, | |
| "step": 36700 | |
| }, | |
| { | |
| "epoch": 0.3299000494738016, | |
| "grad_norm": 0.13788673281669617, | |
| "learning_rate": 0.0003773325079969796, | |
| "loss": 2.6785, | |
| "step": 36800 | |
| }, | |
| { | |
| "epoch": 0.3307965169995456, | |
| "grad_norm": 0.13793878257274628, | |
| "learning_rate": 0.0003767260837480032, | |
| "loss": 2.6785, | |
| "step": 36900 | |
| }, | |
| { | |
| "epoch": 0.3316929845252896, | |
| "grad_norm": 0.13882067799568176, | |
| "learning_rate": 0.0003761186543217209, | |
| "loss": 2.6791, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 0.33258945205103363, | |
| "grad_norm": 0.1409679502248764, | |
| "learning_rate": 0.00037551022453619564, | |
| "loss": 2.6809, | |
| "step": 37100 | |
| }, | |
| { | |
| "epoch": 0.3334859195767777, | |
| "grad_norm": 0.1374359279870987, | |
| "learning_rate": 0.0003749007992174254, | |
| "loss": 2.6836, | |
| "step": 37200 | |
| }, | |
| { | |
| "epoch": 0.3343823871025217, | |
| "grad_norm": 0.14376886188983917, | |
| "learning_rate": 0.00037429038319930453, | |
| "loss": 2.6798, | |
| "step": 37300 | |
| }, | |
| { | |
| "epoch": 0.33527885462826573, | |
| "grad_norm": 0.1401790827512741, | |
| "learning_rate": 0.00037367898132358544, | |
| "loss": 2.6711, | |
| "step": 37400 | |
| }, | |
| { | |
| "epoch": 0.33617532215400975, | |
| "grad_norm": 0.13772746920585632, | |
| "learning_rate": 0.0003730665984398404, | |
| "loss": 2.6768, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 0.33707178967975376, | |
| "grad_norm": 0.14339695870876312, | |
| "learning_rate": 0.0003724532394054228, | |
| "loss": 2.6825, | |
| "step": 37600 | |
| }, | |
| { | |
| "epoch": 0.33796825720549784, | |
| "grad_norm": 0.13825613260269165, | |
| "learning_rate": 0.0003718389090854287, | |
| "loss": 2.6796, | |
| "step": 37700 | |
| }, | |
| { | |
| "epoch": 0.33886472473124185, | |
| "grad_norm": 0.13653145730495453, | |
| "learning_rate": 0.00037122361235265855, | |
| "loss": 2.6715, | |
| "step": 37800 | |
| }, | |
| { | |
| "epoch": 0.33976119225698587, | |
| "grad_norm": 0.13996466994285583, | |
| "learning_rate": 0.00037060735408757795, | |
| "loss": 2.6852, | |
| "step": 37900 | |
| }, | |
| { | |
| "epoch": 0.3406576597827299, | |
| "grad_norm": 0.13708461821079254, | |
| "learning_rate": 0.0003699901391782795, | |
| "loss": 2.6801, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 0.3415541273084739, | |
| "grad_norm": 0.1354062259197235, | |
| "learning_rate": 0.00036937197252044374, | |
| "loss": 2.6748, | |
| "step": 38100 | |
| }, | |
| { | |
| "epoch": 0.3424505948342179, | |
| "grad_norm": 0.13771981000900269, | |
| "learning_rate": 0.00036875285901730035, | |
| "loss": 2.6823, | |
| "step": 38200 | |
| }, | |
| { | |
| "epoch": 0.343347062359962, | |
| "grad_norm": 0.13679315149784088, | |
| "learning_rate": 0.0003681328035795892, | |
| "loss": 2.6733, | |
| "step": 38300 | |
| }, | |
| { | |
| "epoch": 0.344243529885706, | |
| "grad_norm": 0.14163535833358765, | |
| "learning_rate": 0.0003675118111255216, | |
| "loss": 2.6796, | |
| "step": 38400 | |
| }, | |
| { | |
| "epoch": 0.34513999741145, | |
| "grad_norm": 0.13825653493404388, | |
| "learning_rate": 0.00036688988658074124, | |
| "loss": 2.6762, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 0.34603646493719403, | |
| "grad_norm": 0.14156493544578552, | |
| "learning_rate": 0.00036626703487828487, | |
| "loss": 2.6694, | |
| "step": 38600 | |
| }, | |
| { | |
| "epoch": 0.34693293246293805, | |
| "grad_norm": 0.13615700602531433, | |
| "learning_rate": 0.0003656432609585435, | |
| "loss": 2.6779, | |
| "step": 38700 | |
| }, | |
| { | |
| "epoch": 0.3478293999886821, | |
| "grad_norm": 0.13561411201953888, | |
| "learning_rate": 0.0003650185697692229, | |
| "loss": 2.6838, | |
| "step": 38800 | |
| }, | |
| { | |
| "epoch": 0.34872586751442614, | |
| "grad_norm": 0.1366826444864273, | |
| "learning_rate": 0.0003643929662653046, | |
| "loss": 2.6663, | |
| "step": 38900 | |
| }, | |
| { | |
| "epoch": 0.34962233504017015, | |
| "grad_norm": 0.14089661836624146, | |
| "learning_rate": 0.00036376645540900663, | |
| "loss": 2.6708, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 0.35051880256591417, | |
| "grad_norm": 0.13665175437927246, | |
| "learning_rate": 0.000363139042169744, | |
| "loss": 2.6705, | |
| "step": 39100 | |
| }, | |
| { | |
| "epoch": 0.3514152700916582, | |
| "grad_norm": 0.13564077019691467, | |
| "learning_rate": 0.0003625107315240891, | |
| "loss": 2.6698, | |
| "step": 39200 | |
| }, | |
| { | |
| "epoch": 0.35231173761740225, | |
| "grad_norm": 0.14218759536743164, | |
| "learning_rate": 0.0003618815284557326, | |
| "loss": 2.6816, | |
| "step": 39300 | |
| }, | |
| { | |
| "epoch": 0.35320820514314627, | |
| "grad_norm": 0.14430591464042664, | |
| "learning_rate": 0.00036125143795544387, | |
| "loss": 2.6763, | |
| "step": 39400 | |
| }, | |
| { | |
| "epoch": 0.3541046726688903, | |
| "grad_norm": 0.13792584836483002, | |
| "learning_rate": 0.0003606204650210312, | |
| "loss": 2.6674, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 0.3550011401946343, | |
| "grad_norm": 0.136796236038208, | |
| "learning_rate": 0.0003599886146573022, | |
| "loss": 2.6718, | |
| "step": 39600 | |
| }, | |
| { | |
| "epoch": 0.3558976077203783, | |
| "grad_norm": 0.13962402939796448, | |
| "learning_rate": 0.00035935589187602426, | |
| "loss": 2.6776, | |
| "step": 39700 | |
| }, | |
| { | |
| "epoch": 0.35679407524612233, | |
| "grad_norm": 0.1402391791343689, | |
| "learning_rate": 0.0003587223016958845, | |
| "loss": 2.6739, | |
| "step": 39800 | |
| }, | |
| { | |
| "epoch": 0.3576905427718664, | |
| "grad_norm": 0.13909810781478882, | |
| "learning_rate": 0.0003580878491424504, | |
| "loss": 2.6799, | |
| "step": 39900 | |
| }, | |
| { | |
| "epoch": 0.3585870102976104, | |
| "grad_norm": 0.13748426735401154, | |
| "learning_rate": 0.0003574525392481295, | |
| "loss": 2.6765, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 0.35948347782335444, | |
| "grad_norm": 0.132966548204422, | |
| "learning_rate": 0.0003568163770521299, | |
| "loss": 2.6702, | |
| "step": 40100 | |
| }, | |
| { | |
| "epoch": 0.36037994534909845, | |
| "grad_norm": 0.13667432963848114, | |
| "learning_rate": 0.00035617936760041976, | |
| "loss": 2.6645, | |
| "step": 40200 | |
| }, | |
| { | |
| "epoch": 0.36127641287484247, | |
| "grad_norm": 0.14251597225666046, | |
| "learning_rate": 0.00035554151594568767, | |
| "loss": 2.6719, | |
| "step": 40300 | |
| }, | |
| { | |
| "epoch": 0.36217288040058654, | |
| "grad_norm": 0.1350788176059723, | |
| "learning_rate": 0.00035490282714730273, | |
| "loss": 2.6701, | |
| "step": 40400 | |
| }, | |
| { | |
| "epoch": 0.36306934792633055, | |
| "grad_norm": 0.1342398077249527, | |
| "learning_rate": 0.0003542633062712738, | |
| "loss": 2.6653, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 0.36396581545207457, | |
| "grad_norm": 0.13306699693202972, | |
| "learning_rate": 0.00035362295839020996, | |
| "loss": 2.6724, | |
| "step": 40600 | |
| }, | |
| { | |
| "epoch": 0.3648622829778186, | |
| "grad_norm": 0.13337896764278412, | |
| "learning_rate": 0.0003529817885832799, | |
| "loss": 2.6633, | |
| "step": 40700 | |
| }, | |
| { | |
| "epoch": 0.3657587505035626, | |
| "grad_norm": 0.13504943251609802, | |
| "learning_rate": 0.00035233980193617184, | |
| "loss": 2.6678, | |
| "step": 40800 | |
| }, | |
| { | |
| "epoch": 0.3666552180293066, | |
| "grad_norm": 0.1362612545490265, | |
| "learning_rate": 0.0003516970035410527, | |
| "loss": 2.6669, | |
| "step": 40900 | |
| }, | |
| { | |
| "epoch": 0.3675516855550507, | |
| "grad_norm": 0.1384200006723404, | |
| "learning_rate": 0.00035105339849652873, | |
| "loss": 2.6706, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 0.3684481530807947, | |
| "grad_norm": 0.1365375816822052, | |
| "learning_rate": 0.0003504089919076039, | |
| "loss": 2.6647, | |
| "step": 41100 | |
| }, | |
| { | |
| "epoch": 0.3693446206065387, | |
| "grad_norm": 0.13598588109016418, | |
| "learning_rate": 0.00034976378888564014, | |
| "loss": 2.6647, | |
| "step": 41200 | |
| }, | |
| { | |
| "epoch": 0.37024108813228274, | |
| "grad_norm": 0.13357794284820557, | |
| "learning_rate": 0.00034911779454831665, | |
| "loss": 2.6694, | |
| "step": 41300 | |
| }, | |
| { | |
| "epoch": 0.37113755565802675, | |
| "grad_norm": 0.13431541621685028, | |
| "learning_rate": 0.00034847101401958914, | |
| "loss": 2.6644, | |
| "step": 41400 | |
| }, | |
| { | |
| "epoch": 0.3720340231837708, | |
| "grad_norm": 0.13836424052715302, | |
| "learning_rate": 0.0003478234524296494, | |
| "loss": 2.6752, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 0.37293049070951484, | |
| "grad_norm": 0.13321442902088165, | |
| "learning_rate": 0.00034717511491488454, | |
| "loss": 2.6755, | |
| "step": 41600 | |
| }, | |
| { | |
| "epoch": 0.37382695823525885, | |
| "grad_norm": 0.13987848162651062, | |
| "learning_rate": 0.00034652600661783594, | |
| "loss": 2.6666, | |
| "step": 41700 | |
| }, | |
| { | |
| "epoch": 0.37472342576100287, | |
| "grad_norm": 0.1362723857164383, | |
| "learning_rate": 0.00034587613268715917, | |
| "loss": 2.6736, | |
| "step": 41800 | |
| }, | |
| { | |
| "epoch": 0.3756198932867469, | |
| "grad_norm": 0.13275477290153503, | |
| "learning_rate": 0.0003452254982775824, | |
| "loss": 2.6658, | |
| "step": 41900 | |
| }, | |
| { | |
| "epoch": 0.37651636081249096, | |
| "grad_norm": 0.13551685214042664, | |
| "learning_rate": 0.0003445741085498659, | |
| "loss": 2.6666, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 0.377412828338235, | |
| "grad_norm": 0.13056619465351105, | |
| "learning_rate": 0.00034392196867076124, | |
| "loss": 2.6638, | |
| "step": 42100 | |
| }, | |
| { | |
| "epoch": 0.378309295863979, | |
| "grad_norm": 0.1297016590833664, | |
| "learning_rate": 0.0003432690838129698, | |
| "loss": 2.6666, | |
| "step": 42200 | |
| }, | |
| { | |
| "epoch": 0.379205763389723, | |
| "grad_norm": 0.1312173455953598, | |
| "learning_rate": 0.00034261545915510223, | |
| "loss": 2.6665, | |
| "step": 42300 | |
| }, | |
| { | |
| "epoch": 0.380102230915467, | |
| "grad_norm": 0.13310836255550385, | |
| "learning_rate": 0.00034196109988163714, | |
| "loss": 2.6629, | |
| "step": 42400 | |
| }, | |
| { | |
| "epoch": 0.38099869844121104, | |
| "grad_norm": 0.13753747940063477, | |
| "learning_rate": 0.0003413060111828801, | |
| "loss": 2.6682, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 0.3818951659669551, | |
| "grad_norm": 0.13357776403427124, | |
| "learning_rate": 0.00034065019825492237, | |
| "loss": 2.6657, | |
| "step": 42600 | |
| }, | |
| { | |
| "epoch": 0.3827916334926991, | |
| "grad_norm": 0.13426382839679718, | |
| "learning_rate": 0.00033999366629959956, | |
| "loss": 2.6663, | |
| "step": 42700 | |
| }, | |
| { | |
| "epoch": 0.38368810101844314, | |
| "grad_norm": 0.14345508813858032, | |
| "learning_rate": 0.0003393364205244508, | |
| "loss": 2.6652, | |
| "step": 42800 | |
| }, | |
| { | |
| "epoch": 0.38458456854418716, | |
| "grad_norm": 0.1440078616142273, | |
| "learning_rate": 0.00033867846614267695, | |
| "loss": 2.6729, | |
| "step": 42900 | |
| }, | |
| { | |
| "epoch": 0.38548103606993117, | |
| "grad_norm": 0.1365692913532257, | |
| "learning_rate": 0.00033801980837309953, | |
| "loss": 2.6725, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 0.38637750359567524, | |
| "grad_norm": 0.13232523202896118, | |
| "learning_rate": 0.0003373604524401193, | |
| "loss": 2.6655, | |
| "step": 43100 | |
| }, | |
| { | |
| "epoch": 0.38727397112141926, | |
| "grad_norm": 0.13531488180160522, | |
| "learning_rate": 0.0003367004035736747, | |
| "loss": 2.6694, | |
| "step": 43200 | |
| }, | |
| { | |
| "epoch": 0.3881704386471633, | |
| "grad_norm": 0.1343296617269516, | |
| "learning_rate": 0.0003360396670092004, | |
| "loss": 2.6576, | |
| "step": 43300 | |
| }, | |
| { | |
| "epoch": 0.3890669061729073, | |
| "grad_norm": 0.13563938438892365, | |
| "learning_rate": 0.00033537824798758603, | |
| "loss": 2.6682, | |
| "step": 43400 | |
| }, | |
| { | |
| "epoch": 0.3899633736986513, | |
| "grad_norm": 0.13184553384780884, | |
| "learning_rate": 0.00033471615175513416, | |
| "loss": 2.6544, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 0.3908598412243953, | |
| "grad_norm": 0.13147491216659546, | |
| "learning_rate": 0.00033405338356351906, | |
| "loss": 2.659, | |
| "step": 43600 | |
| }, | |
| { | |
| "epoch": 0.3917563087501394, | |
| "grad_norm": 0.128835991024971, | |
| "learning_rate": 0.00033338994866974477, | |
| "loss": 2.668, | |
| "step": 43700 | |
| }, | |
| { | |
| "epoch": 0.3926527762758834, | |
| "grad_norm": 0.13172951340675354, | |
| "learning_rate": 0.00033272585233610363, | |
| "loss": 2.668, | |
| "step": 43800 | |
| }, | |
| { | |
| "epoch": 0.3935492438016274, | |
| "grad_norm": 0.13852158188819885, | |
| "learning_rate": 0.00033206109983013464, | |
| "loss": 2.6618, | |
| "step": 43900 | |
| }, | |
| { | |
| "epoch": 0.39444571132737144, | |
| "grad_norm": 0.13343538343906403, | |
| "learning_rate": 0.00033139569642458104, | |
| "loss": 2.6694, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 0.39534217885311546, | |
| "grad_norm": 0.13819359242916107, | |
| "learning_rate": 0.00033072964739734936, | |
| "loss": 2.6645, | |
| "step": 44100 | |
| }, | |
| { | |
| "epoch": 0.3962386463788595, | |
| "grad_norm": 0.14136740565299988, | |
| "learning_rate": 0.0003300629580314668, | |
| "loss": 2.6596, | |
| "step": 44200 | |
| }, | |
| { | |
| "epoch": 0.39713511390460354, | |
| "grad_norm": 0.13478617370128632, | |
| "learning_rate": 0.00032939563361503995, | |
| "loss": 2.6682, | |
| "step": 44300 | |
| }, | |
| { | |
| "epoch": 0.39803158143034756, | |
| "grad_norm": 0.13910335302352905, | |
| "learning_rate": 0.00032872767944121234, | |
| "loss": 2.6561, | |
| "step": 44400 | |
| }, | |
| { | |
| "epoch": 0.3989280489560916, | |
| "grad_norm": 0.12869343161582947, | |
| "learning_rate": 0.0003280591008081227, | |
| "loss": 2.6698, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 0.3998245164818356, | |
| "grad_norm": 0.13684409856796265, | |
| "learning_rate": 0.00032738990301886306, | |
| "loss": 2.6606, | |
| "step": 44600 | |
| }, | |
| { | |
| "epoch": 0.4007209840075796, | |
| "grad_norm": 0.13658647239208221, | |
| "learning_rate": 0.00032672009138143634, | |
| "loss": 2.6641, | |
| "step": 44700 | |
| }, | |
| { | |
| "epoch": 0.4016174515333237, | |
| "grad_norm": 0.13844211399555206, | |
| "learning_rate": 0.00032604967120871444, | |
| "loss": 2.6548, | |
| "step": 44800 | |
| }, | |
| { | |
| "epoch": 0.4025139190590677, | |
| "grad_norm": 0.12905430793762207, | |
| "learning_rate": 0.0003253786478183963, | |
| "loss": 2.6601, | |
| "step": 44900 | |
| }, | |
| { | |
| "epoch": 0.4034103865848117, | |
| "grad_norm": 0.1291283369064331, | |
| "learning_rate": 0.00032470702653296515, | |
| "loss": 2.659, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 0.4043068541105557, | |
| "grad_norm": 0.1292611062526703, | |
| "learning_rate": 0.0003240348126796471, | |
| "loss": 2.6575, | |
| "step": 45100 | |
| }, | |
| { | |
| "epoch": 0.40520332163629974, | |
| "grad_norm": 0.13455387949943542, | |
| "learning_rate": 0.00032336201159036815, | |
| "loss": 2.6514, | |
| "step": 45200 | |
| }, | |
| { | |
| "epoch": 0.4060997891620438, | |
| "grad_norm": 0.12952381372451782, | |
| "learning_rate": 0.00032268862860171214, | |
| "loss": 2.6559, | |
| "step": 45300 | |
| }, | |
| { | |
| "epoch": 0.4069962566877878, | |
| "grad_norm": 0.13285262882709503, | |
| "learning_rate": 0.00032201466905487885, | |
| "loss": 2.66, | |
| "step": 45400 | |
| }, | |
| { | |
| "epoch": 0.40789272421353184, | |
| "grad_norm": 0.13195694983005524, | |
| "learning_rate": 0.0003213401382956408, | |
| "loss": 2.6638, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 0.40878919173927586, | |
| "grad_norm": 0.12948913872241974, | |
| "learning_rate": 0.0003206650416743017, | |
| "loss": 2.6568, | |
| "step": 45600 | |
| }, | |
| { | |
| "epoch": 0.4096856592650199, | |
| "grad_norm": 0.13518835604190826, | |
| "learning_rate": 0.0003199893845456531, | |
| "loss": 2.6584, | |
| "step": 45700 | |
| }, | |
| { | |
| "epoch": 0.41058212679076395, | |
| "grad_norm": 0.13390909135341644, | |
| "learning_rate": 0.00031931317226893295, | |
| "loss": 2.6477, | |
| "step": 45800 | |
| }, | |
| { | |
| "epoch": 0.41147859431650796, | |
| "grad_norm": 0.13253462314605713, | |
| "learning_rate": 0.00031863641020778247, | |
| "loss": 2.6549, | |
| "step": 45900 | |
| }, | |
| { | |
| "epoch": 0.412375061842252, | |
| "grad_norm": 0.13521717488765717, | |
| "learning_rate": 0.0003179591037302035, | |
| "loss": 2.6527, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 0.413271529367996, | |
| "grad_norm": 0.14094848930835724, | |
| "learning_rate": 0.0003172812582085163, | |
| "loss": 2.6608, | |
| "step": 46100 | |
| }, | |
| { | |
| "epoch": 0.41416799689374, | |
| "grad_norm": 0.1350257843732834, | |
| "learning_rate": 0.0003166028790193166, | |
| "loss": 2.6568, | |
| "step": 46200 | |
| }, | |
| { | |
| "epoch": 0.415064464419484, | |
| "grad_norm": 0.1422998309135437, | |
| "learning_rate": 0.00031592397154343334, | |
| "loss": 2.6567, | |
| "step": 46300 | |
| }, | |
| { | |
| "epoch": 0.4159609319452281, | |
| "grad_norm": 0.12808671593666077, | |
| "learning_rate": 0.0003152445411658856, | |
| "loss": 2.6528, | |
| "step": 46400 | |
| }, | |
| { | |
| "epoch": 0.4168573994709721, | |
| "grad_norm": 0.12911584973335266, | |
| "learning_rate": 0.00031456459327584027, | |
| "loss": 2.6641, | |
| "step": 46500 | |
| }, | |
| { | |
| "epoch": 0.41775386699671613, | |
| "grad_norm": 0.13506324589252472, | |
| "learning_rate": 0.00031388413326656885, | |
| "loss": 2.655, | |
| "step": 46600 | |
| }, | |
| { | |
| "epoch": 0.41865033452246014, | |
| "grad_norm": 0.13051171600818634, | |
| "learning_rate": 0.00031320316653540495, | |
| "loss": 2.6625, | |
| "step": 46700 | |
| }, | |
| { | |
| "epoch": 0.41954680204820416, | |
| "grad_norm": 0.1310121864080429, | |
| "learning_rate": 0.00031252169848370155, | |
| "loss": 2.6587, | |
| "step": 46800 | |
| }, | |
| { | |
| "epoch": 0.42044326957394823, | |
| "grad_norm": 0.13700339198112488, | |
| "learning_rate": 0.00031183973451678806, | |
| "loss": 2.6501, | |
| "step": 46900 | |
| }, | |
| { | |
| "epoch": 0.42133973709969225, | |
| "grad_norm": 0.13302241265773773, | |
| "learning_rate": 0.0003111572800439273, | |
| "loss": 2.6557, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 0.42223620462543626, | |
| "grad_norm": 0.13645824790000916, | |
| "learning_rate": 0.00031047434047827294, | |
| "loss": 2.655, | |
| "step": 47100 | |
| }, | |
| { | |
| "epoch": 0.4231326721511803, | |
| "grad_norm": 0.13405530154705048, | |
| "learning_rate": 0.0003097909212368261, | |
| "loss": 2.6616, | |
| "step": 47200 | |
| }, | |
| { | |
| "epoch": 0.4240291396769243, | |
| "grad_norm": 0.1315753012895584, | |
| "learning_rate": 0.0003091070277403927, | |
| "loss": 2.6516, | |
| "step": 47300 | |
| }, | |
| { | |
| "epoch": 0.4249256072026683, | |
| "grad_norm": 0.13063915073871613, | |
| "learning_rate": 0.0003084226654135406, | |
| "loss": 2.6581, | |
| "step": 47400 | |
| }, | |
| { | |
| "epoch": 0.4258220747284124, | |
| "grad_norm": 0.1367933750152588, | |
| "learning_rate": 0.00030773783968455614, | |
| "loss": 2.6541, | |
| "step": 47500 | |
| }, | |
| { | |
| "epoch": 0.4267185422541564, | |
| "grad_norm": 0.13026690483093262, | |
| "learning_rate": 0.0003070525559854015, | |
| "loss": 2.651, | |
| "step": 47600 | |
| }, | |
| { | |
| "epoch": 0.4276150097799004, | |
| "grad_norm": 0.13831181824207306, | |
| "learning_rate": 0.00030636681975167114, | |
| "loss": 2.6596, | |
| "step": 47700 | |
| }, | |
| { | |
| "epoch": 0.42851147730564443, | |
| "grad_norm": 0.13564659655094147, | |
| "learning_rate": 0.0003056806364225493, | |
| "loss": 2.6502, | |
| "step": 47800 | |
| }, | |
| { | |
| "epoch": 0.42940794483138844, | |
| "grad_norm": 0.13234366476535797, | |
| "learning_rate": 0.00030499401144076636, | |
| "loss": 2.647, | |
| "step": 47900 | |
| }, | |
| { | |
| "epoch": 0.4303044123571325, | |
| "grad_norm": 0.13098286092281342, | |
| "learning_rate": 0.00030430695025255596, | |
| "loss": 2.6519, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 0.43120087988287653, | |
| "grad_norm": 0.13402335345745087, | |
| "learning_rate": 0.00030361945830761146, | |
| "loss": 2.6581, | |
| "step": 48100 | |
| }, | |
| { | |
| "epoch": 0.43209734740862055, | |
| "grad_norm": 0.13898345828056335, | |
| "learning_rate": 0.0003029315410590431, | |
| "loss": 2.652, | |
| "step": 48200 | |
| }, | |
| { | |
| "epoch": 0.43299381493436456, | |
| "grad_norm": 0.1286890059709549, | |
| "learning_rate": 0.00030224320396333456, | |
| "loss": 2.6529, | |
| "step": 48300 | |
| }, | |
| { | |
| "epoch": 0.4338902824601086, | |
| "grad_norm": 0.13320975005626678, | |
| "learning_rate": 0.00030155445248029975, | |
| "loss": 2.6533, | |
| "step": 48400 | |
| }, | |
| { | |
| "epoch": 0.4347867499858526, | |
| "grad_norm": 0.128444641828537, | |
| "learning_rate": 0.00030086529207303935, | |
| "loss": 2.6537, | |
| "step": 48500 | |
| }, | |
| { | |
| "epoch": 0.43568321751159667, | |
| "grad_norm": 0.1326085776090622, | |
| "learning_rate": 0.00030017572820789765, | |
| "loss": 2.6522, | |
| "step": 48600 | |
| }, | |
| { | |
| "epoch": 0.4365796850373407, | |
| "grad_norm": 0.13274775445461273, | |
| "learning_rate": 0.00029948576635441905, | |
| "loss": 2.6506, | |
| "step": 48700 | |
| }, | |
| { | |
| "epoch": 0.4374761525630847, | |
| "grad_norm": 0.13352380692958832, | |
| "learning_rate": 0.0002987954119853048, | |
| "loss": 2.6552, | |
| "step": 48800 | |
| }, | |
| { | |
| "epoch": 0.4383726200888287, | |
| "grad_norm": 0.13431380689144135, | |
| "learning_rate": 0.0002981046705763696, | |
| "loss": 2.6532, | |
| "step": 48900 | |
| }, | |
| { | |
| "epoch": 0.43926908761457273, | |
| "grad_norm": 0.1325884610414505, | |
| "learning_rate": 0.0002974135476064981, | |
| "loss": 2.6586, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 0.4401655551403168, | |
| "grad_norm": 0.13121579587459564, | |
| "learning_rate": 0.0002967220485576013, | |
| "loss": 2.6486, | |
| "step": 49100 | |
| }, | |
| { | |
| "epoch": 0.4410620226660608, | |
| "grad_norm": 0.12927518784999847, | |
| "learning_rate": 0.0002960301789145733, | |
| "loss": 2.6553, | |
| "step": 49200 | |
| }, | |
| { | |
| "epoch": 0.44195849019180483, | |
| "grad_norm": 0.1333087980747223, | |
| "learning_rate": 0.0002953379441652478, | |
| "loss": 2.6525, | |
| "step": 49300 | |
| }, | |
| { | |
| "epoch": 0.44285495771754885, | |
| "grad_norm": 0.13250477612018585, | |
| "learning_rate": 0.0002946453498003543, | |
| "loss": 2.6441, | |
| "step": 49400 | |
| }, | |
| { | |
| "epoch": 0.44375142524329286, | |
| "grad_norm": 0.13053111732006073, | |
| "learning_rate": 0.00029395240131347507, | |
| "loss": 2.6549, | |
| "step": 49500 | |
| }, | |
| { | |
| "epoch": 0.44464789276903693, | |
| "grad_norm": 0.1300216168165207, | |
| "learning_rate": 0.00029325910420100083, | |
| "loss": 2.6509, | |
| "step": 49600 | |
| }, | |
| { | |
| "epoch": 0.44554436029478095, | |
| "grad_norm": 0.13556109368801117, | |
| "learning_rate": 0.00029256546396208766, | |
| "loss": 2.6403, | |
| "step": 49700 | |
| }, | |
| { | |
| "epoch": 0.44644082782052497, | |
| "grad_norm": 0.12819913029670715, | |
| "learning_rate": 0.00029187148609861353, | |
| "loss": 2.6435, | |
| "step": 49800 | |
| }, | |
| { | |
| "epoch": 0.447337295346269, | |
| "grad_norm": 0.1305069774389267, | |
| "learning_rate": 0.0002911771761151342, | |
| "loss": 2.654, | |
| "step": 49900 | |
| }, | |
| { | |
| "epoch": 0.448233762872013, | |
| "grad_norm": 0.1274418979883194, | |
| "learning_rate": 0.0002904825395188397, | |
| "loss": 2.65, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 0.449130230397757, | |
| "grad_norm": 0.13119454681873322, | |
| "learning_rate": 0.0002897875818195111, | |
| "loss": 2.6608, | |
| "step": 50100 | |
| }, | |
| { | |
| "epoch": 0.4500266979235011, | |
| "grad_norm": 0.13181808590888977, | |
| "learning_rate": 0.00028909230852947575, | |
| "loss": 2.6457, | |
| "step": 50200 | |
| }, | |
| { | |
| "epoch": 0.4509231654492451, | |
| "grad_norm": 0.1294698864221573, | |
| "learning_rate": 0.00028839672516356495, | |
| "loss": 2.6537, | |
| "step": 50300 | |
| }, | |
| { | |
| "epoch": 0.4518196329749891, | |
| "grad_norm": 0.13653060793876648, | |
| "learning_rate": 0.00028770083723906904, | |
| "loss": 2.6575, | |
| "step": 50400 | |
| }, | |
| { | |
| "epoch": 0.45271610050073313, | |
| "grad_norm": 0.12975502014160156, | |
| "learning_rate": 0.0002870046502756942, | |
| "loss": 2.645, | |
| "step": 50500 | |
| }, | |
| { | |
| "epoch": 0.45361256802647715, | |
| "grad_norm": 0.13355258107185364, | |
| "learning_rate": 0.0002863081697955187, | |
| "loss": 2.6435, | |
| "step": 50600 | |
| }, | |
| { | |
| "epoch": 0.4545090355522212, | |
| "grad_norm": 0.1305021345615387, | |
| "learning_rate": 0.00028561140132294863, | |
| "loss": 2.6454, | |
| "step": 50700 | |
| }, | |
| { | |
| "epoch": 0.45540550307796523, | |
| "grad_norm": 0.13262712955474854, | |
| "learning_rate": 0.00028491435038467466, | |
| "loss": 2.6534, | |
| "step": 50800 | |
| }, | |
| { | |
| "epoch": 0.45630197060370925, | |
| "grad_norm": 0.13376927375793457, | |
| "learning_rate": 0.00028421702250962786, | |
| "loss": 2.646, | |
| "step": 50900 | |
| }, | |
| { | |
| "epoch": 0.45719843812945327, | |
| "grad_norm": 0.13164885342121124, | |
| "learning_rate": 0.0002835194232289361, | |
| "loss": 2.6567, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 0.4580949056551973, | |
| "grad_norm": 0.13637390732765198, | |
| "learning_rate": 0.0002828215580758798, | |
| "loss": 2.6466, | |
| "step": 51100 | |
| }, | |
| { | |
| "epoch": 0.4589913731809413, | |
| "grad_norm": 0.13366563618183136, | |
| "learning_rate": 0.0002821234325858482, | |
| "loss": 2.6475, | |
| "step": 51200 | |
| }, | |
| { | |
| "epoch": 0.45988784070668537, | |
| "grad_norm": 0.1311609447002411, | |
| "learning_rate": 0.0002814250522962956, | |
| "loss": 2.653, | |
| "step": 51300 | |
| }, | |
| { | |
| "epoch": 0.4607843082324294, | |
| "grad_norm": 0.13002333045005798, | |
| "learning_rate": 0.0002807264227466975, | |
| "loss": 2.6472, | |
| "step": 51400 | |
| }, | |
| { | |
| "epoch": 0.4616807757581734, | |
| "grad_norm": 0.12868809700012207, | |
| "learning_rate": 0.00028002754947850623, | |
| "loss": 2.6347, | |
| "step": 51500 | |
| }, | |
| { | |
| "epoch": 0.4625772432839174, | |
| "grad_norm": 0.1310162991285324, | |
| "learning_rate": 0.00027932843803510755, | |
| "loss": 2.6414, | |
| "step": 51600 | |
| }, | |
| { | |
| "epoch": 0.46347371080966143, | |
| "grad_norm": 0.12595106661319733, | |
| "learning_rate": 0.00027862909396177615, | |
| "loss": 2.6427, | |
| "step": 51700 | |
| }, | |
| { | |
| "epoch": 0.4643701783354055, | |
| "grad_norm": 0.130056232213974, | |
| "learning_rate": 0.000277929522805632, | |
| "loss": 2.6461, | |
| "step": 51800 | |
| }, | |
| { | |
| "epoch": 0.4652666458611495, | |
| "grad_norm": 0.12970997393131256, | |
| "learning_rate": 0.00027722973011559633, | |
| "loss": 2.6483, | |
| "step": 51900 | |
| }, | |
| { | |
| "epoch": 0.46616311338689354, | |
| "grad_norm": 0.13723085820674896, | |
| "learning_rate": 0.00027652972144234745, | |
| "loss": 2.6384, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 0.46705958091263755, | |
| "grad_norm": 0.13320958614349365, | |
| "learning_rate": 0.000275829502338277, | |
| "loss": 2.6573, | |
| "step": 52100 | |
| }, | |
| { | |
| "epoch": 0.46795604843838157, | |
| "grad_norm": 0.1280314177274704, | |
| "learning_rate": 0.00027512907835744547, | |
| "loss": 2.6469, | |
| "step": 52200 | |
| }, | |
| { | |
| "epoch": 0.46885251596412564, | |
| "grad_norm": 0.13650980591773987, | |
| "learning_rate": 0.0002744284550555385, | |
| "loss": 2.6457, | |
| "step": 52300 | |
| }, | |
| { | |
| "epoch": 0.46974898348986965, | |
| "grad_norm": 0.1296728104352951, | |
| "learning_rate": 0.0002737276379898229, | |
| "loss": 2.6386, | |
| "step": 52400 | |
| }, | |
| { | |
| "epoch": 0.47064545101561367, | |
| "grad_norm": 0.13340330123901367, | |
| "learning_rate": 0.0002730266327191023, | |
| "loss": 2.6419, | |
| "step": 52500 | |
| }, | |
| { | |
| "epoch": 0.4715419185413577, | |
| "grad_norm": 0.1351691633462906, | |
| "learning_rate": 0.0002723254448036731, | |
| "loss": 2.6474, | |
| "step": 52600 | |
| }, | |
| { | |
| "epoch": 0.4724383860671017, | |
| "grad_norm": 0.135370671749115, | |
| "learning_rate": 0.00027162407980528037, | |
| "loss": 2.6486, | |
| "step": 52700 | |
| }, | |
| { | |
| "epoch": 0.4733348535928457, | |
| "grad_norm": 0.12964408099651337, | |
| "learning_rate": 0.0002709225432870738, | |
| "loss": 2.6437, | |
| "step": 52800 | |
| }, | |
| { | |
| "epoch": 0.4742313211185898, | |
| "grad_norm": 0.12840570509433746, | |
| "learning_rate": 0.0002702208408135637, | |
| "loss": 2.6409, | |
| "step": 52900 | |
| }, | |
| { | |
| "epoch": 0.4751277886443338, | |
| "grad_norm": 0.13321109116077423, | |
| "learning_rate": 0.00026951897795057653, | |
| "loss": 2.6443, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 0.4760242561700778, | |
| "grad_norm": 0.13473448157310486, | |
| "learning_rate": 0.0002688169602652113, | |
| "loss": 2.6404, | |
| "step": 53100 | |
| }, | |
| { | |
| "epoch": 0.47692072369582184, | |
| "grad_norm": 0.1295260488986969, | |
| "learning_rate": 0.00026811479332579445, | |
| "loss": 2.6359, | |
| "step": 53200 | |
| }, | |
| { | |
| "epoch": 0.47781719122156585, | |
| "grad_norm": 0.13443715870380402, | |
| "learning_rate": 0.0002674124827018368, | |
| "loss": 2.6423, | |
| "step": 53300 | |
| }, | |
| { | |
| "epoch": 0.4787136587473099, | |
| "grad_norm": 0.13394586741924286, | |
| "learning_rate": 0.0002667100339639886, | |
| "loss": 2.6353, | |
| "step": 53400 | |
| }, | |
| { | |
| "epoch": 0.47961012627305394, | |
| "grad_norm": 0.1299637407064438, | |
| "learning_rate": 0.00026600745268399583, | |
| "loss": 2.6376, | |
| "step": 53500 | |
| }, | |
| { | |
| "epoch": 0.48050659379879795, | |
| "grad_norm": 0.12686701118946075, | |
| "learning_rate": 0.0002653047444346556, | |
| "loss": 2.6426, | |
| "step": 53600 | |
| }, | |
| { | |
| "epoch": 0.48140306132454197, | |
| "grad_norm": 0.13109557330608368, | |
| "learning_rate": 0.00026460191478977203, | |
| "loss": 2.6413, | |
| "step": 53700 | |
| }, | |
| { | |
| "epoch": 0.482299528850286, | |
| "grad_norm": 0.12681497633457184, | |
| "learning_rate": 0.00026389896932411234, | |
| "loss": 2.6469, | |
| "step": 53800 | |
| }, | |
| { | |
| "epoch": 0.48319599637603, | |
| "grad_norm": 0.12837442755699158, | |
| "learning_rate": 0.0002631959136133624, | |
| "loss": 2.6444, | |
| "step": 53900 | |
| }, | |
| { | |
| "epoch": 0.4840924639017741, | |
| "grad_norm": 0.13397662341594696, | |
| "learning_rate": 0.0002624927532340825, | |
| "loss": 2.6385, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 0.4849889314275181, | |
| "grad_norm": 0.1271597146987915, | |
| "learning_rate": 0.00026178949376366316, | |
| "loss": 2.637, | |
| "step": 54100 | |
| }, | |
| { | |
| "epoch": 0.4858853989532621, | |
| "grad_norm": 0.12960323691368103, | |
| "learning_rate": 0.00026108614078028077, | |
| "loss": 2.6347, | |
| "step": 54200 | |
| }, | |
| { | |
| "epoch": 0.4867818664790061, | |
| "grad_norm": 0.13047392666339874, | |
| "learning_rate": 0.0002603826998628536, | |
| "loss": 2.6414, | |
| "step": 54300 | |
| }, | |
| { | |
| "epoch": 0.48767833400475014, | |
| "grad_norm": 0.1311124563217163, | |
| "learning_rate": 0.0002596791765909973, | |
| "loss": 2.6422, | |
| "step": 54400 | |
| }, | |
| { | |
| "epoch": 0.4885748015304942, | |
| "grad_norm": 0.13057972490787506, | |
| "learning_rate": 0.00025897557654498086, | |
| "loss": 2.6416, | |
| "step": 54500 | |
| }, | |
| { | |
| "epoch": 0.4894712690562382, | |
| "grad_norm": 0.12886947393417358, | |
| "learning_rate": 0.0002582719053056822, | |
| "loss": 2.6402, | |
| "step": 54600 | |
| }, | |
| { | |
| "epoch": 0.49036773658198224, | |
| "grad_norm": 0.12779085338115692, | |
| "learning_rate": 0.00025756816845454384, | |
| "loss": 2.6349, | |
| "step": 54700 | |
| }, | |
| { | |
| "epoch": 0.49126420410772625, | |
| "grad_norm": 0.12994657456874847, | |
| "learning_rate": 0.0002568643715735288, | |
| "loss": 2.6384, | |
| "step": 54800 | |
| }, | |
| { | |
| "epoch": 0.49216067163347027, | |
| "grad_norm": 0.1322993040084839, | |
| "learning_rate": 0.00025616052024507625, | |
| "loss": 2.6413, | |
| "step": 54900 | |
| }, | |
| { | |
| "epoch": 0.4930571391592143, | |
| "grad_norm": 0.1368289738893509, | |
| "learning_rate": 0.00025545662005205716, | |
| "loss": 2.641, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 0.49395360668495836, | |
| "grad_norm": 0.1309647113084793, | |
| "learning_rate": 0.00025475267657773027, | |
| "loss": 2.6389, | |
| "step": 55100 | |
| }, | |
| { | |
| "epoch": 0.4948500742107024, | |
| "grad_norm": 0.13188649713993073, | |
| "learning_rate": 0.0002540486954056975, | |
| "loss": 2.6445, | |
| "step": 55200 | |
| }, | |
| { | |
| "epoch": 0.4957465417364464, | |
| "grad_norm": 0.13085578382015228, | |
| "learning_rate": 0.0002533446821198597, | |
| "loss": 2.6373, | |
| "step": 55300 | |
| }, | |
| { | |
| "epoch": 0.4966430092621904, | |
| "grad_norm": 0.13119032979011536, | |
| "learning_rate": 0.0002526406423043725, | |
| "loss": 2.6449, | |
| "step": 55400 | |
| }, | |
| { | |
| "epoch": 0.4975394767879344, | |
| "grad_norm": 0.1324346512556076, | |
| "learning_rate": 0.0002519365815436021, | |
| "loss": 2.6332, | |
| "step": 55500 | |
| }, | |
| { | |
| "epoch": 0.4984359443136785, | |
| "grad_norm": 0.13353855907917023, | |
| "learning_rate": 0.0002512325054220807, | |
| "loss": 2.6365, | |
| "step": 55600 | |
| }, | |
| { | |
| "epoch": 0.4993324118394225, | |
| "grad_norm": 0.13728708028793335, | |
| "learning_rate": 0.0002505284195244624, | |
| "loss": 2.6378, | |
| "step": 55700 | |
| }, | |
| { | |
| "epoch": 0.5002288793651665, | |
| "grad_norm": 0.12542614340782166, | |
| "learning_rate": 0.0002498243294354787, | |
| "loss": 2.6436, | |
| "step": 55800 | |
| }, | |
| { | |
| "epoch": 0.5011253468909106, | |
| "grad_norm": 0.12808193266391754, | |
| "learning_rate": 0.0002491202407398945, | |
| "loss": 2.6439, | |
| "step": 55900 | |
| }, | |
| { | |
| "epoch": 0.5020218144166546, | |
| "grad_norm": 0.1336560845375061, | |
| "learning_rate": 0.0002484161590224637, | |
| "loss": 2.6317, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 0.5029182819423986, | |
| "grad_norm": 0.1294037103652954, | |
| "learning_rate": 0.0002477120898678847, | |
| "loss": 2.6397, | |
| "step": 56100 | |
| }, | |
| { | |
| "epoch": 0.5038147494681426, | |
| "grad_norm": 0.12634062767028809, | |
| "learning_rate": 0.0002470080388607563, | |
| "loss": 2.6345, | |
| "step": 56200 | |
| }, | |
| { | |
| "epoch": 0.5047112169938867, | |
| "grad_norm": 0.12850134074687958, | |
| "learning_rate": 0.0002463040115855333, | |
| "loss": 2.6355, | |
| "step": 56300 | |
| }, | |
| { | |
| "epoch": 0.5056076845196307, | |
| "grad_norm": 0.13224093616008759, | |
| "learning_rate": 0.00024560001362648233, | |
| "loss": 2.6333, | |
| "step": 56400 | |
| }, | |
| { | |
| "epoch": 0.5065041520453747, | |
| "grad_norm": 0.129663348197937, | |
| "learning_rate": 0.00024489605056763757, | |
| "loss": 2.6363, | |
| "step": 56500 | |
| }, | |
| { | |
| "epoch": 0.5074006195711187, | |
| "grad_norm": 0.1328994631767273, | |
| "learning_rate": 0.00024419212799275627, | |
| "loss": 2.6307, | |
| "step": 56600 | |
| }, | |
| { | |
| "epoch": 0.5082970870968627, | |
| "grad_norm": 0.12859898805618286, | |
| "learning_rate": 0.0002434882514852745, | |
| "loss": 2.638, | |
| "step": 56700 | |
| }, | |
| { | |
| "epoch": 0.5091935546226067, | |
| "grad_norm": 0.13029265403747559, | |
| "learning_rate": 0.0002427844266282631, | |
| "loss": 2.6369, | |
| "step": 56800 | |
| }, | |
| { | |
| "epoch": 0.5100900221483509, | |
| "grad_norm": 0.1283547580242157, | |
| "learning_rate": 0.00024208065900438309, | |
| "loss": 2.634, | |
| "step": 56900 | |
| }, | |
| { | |
| "epoch": 0.5109864896740949, | |
| "grad_norm": 0.12610678374767303, | |
| "learning_rate": 0.00024137695419584163, | |
| "loss": 2.6389, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 0.5118829571998389, | |
| "grad_norm": 0.12752105295658112, | |
| "learning_rate": 0.0002406733177843475, | |
| "loss": 2.6364, | |
| "step": 57100 | |
| }, | |
| { | |
| "epoch": 0.5127794247255829, | |
| "grad_norm": 0.13300961256027222, | |
| "learning_rate": 0.0002399697553510671, | |
| "loss": 2.6272, | |
| "step": 57200 | |
| }, | |
| { | |
| "epoch": 0.5136758922513269, | |
| "grad_norm": 0.1362060010433197, | |
| "learning_rate": 0.00023926627247657987, | |
| "loss": 2.6327, | |
| "step": 57300 | |
| }, | |
| { | |
| "epoch": 0.5145723597770709, | |
| "grad_norm": 0.12719836831092834, | |
| "learning_rate": 0.00023856287474083437, | |
| "loss": 2.6335, | |
| "step": 57400 | |
| }, | |
| { | |
| "epoch": 0.515468827302815, | |
| "grad_norm": 0.13754625618457794, | |
| "learning_rate": 0.0002378595677231038, | |
| "loss": 2.627, | |
| "step": 57500 | |
| }, | |
| { | |
| "epoch": 0.516365294828559, | |
| "grad_norm": 0.13351771235466003, | |
| "learning_rate": 0.00023715635700194185, | |
| "loss": 2.641, | |
| "step": 57600 | |
| }, | |
| { | |
| "epoch": 0.517261762354303, | |
| "grad_norm": 0.12546195089817047, | |
| "learning_rate": 0.0002364532481551382, | |
| "loss": 2.639, | |
| "step": 57700 | |
| }, | |
| { | |
| "epoch": 0.518158229880047, | |
| "grad_norm": 0.13150018453598022, | |
| "learning_rate": 0.00023575024675967464, | |
| "loss": 2.6328, | |
| "step": 57800 | |
| }, | |
| { | |
| "epoch": 0.519054697405791, | |
| "grad_norm": 0.12908001244068146, | |
| "learning_rate": 0.00023504735839168062, | |
| "loss": 2.6236, | |
| "step": 57900 | |
| }, | |
| { | |
| "epoch": 0.5199511649315351, | |
| "grad_norm": 0.12657436728477478, | |
| "learning_rate": 0.00023434458862638919, | |
| "loss": 2.6363, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 0.5208476324572792, | |
| "grad_norm": 0.13314677774906158, | |
| "learning_rate": 0.0002336419430380924, | |
| "loss": 2.6304, | |
| "step": 58100 | |
| }, | |
| { | |
| "epoch": 0.5217440999830232, | |
| "grad_norm": 0.1295042335987091, | |
| "learning_rate": 0.0002329394272000976, | |
| "loss": 2.6269, | |
| "step": 58200 | |
| }, | |
| { | |
| "epoch": 0.5226405675087672, | |
| "grad_norm": 0.13147129118442535, | |
| "learning_rate": 0.00023223704668468275, | |
| "loss": 2.6355, | |
| "step": 58300 | |
| }, | |
| { | |
| "epoch": 0.5235370350345112, | |
| "grad_norm": 0.13861297070980072, | |
| "learning_rate": 0.0002315348070630526, | |
| "loss": 2.6307, | |
| "step": 58400 | |
| }, | |
| { | |
| "epoch": 0.5244335025602552, | |
| "grad_norm": 0.12782630324363708, | |
| "learning_rate": 0.0002308327139052943, | |
| "loss": 2.6396, | |
| "step": 58500 | |
| }, | |
| { | |
| "epoch": 0.5253299700859992, | |
| "grad_norm": 0.1270517110824585, | |
| "learning_rate": 0.00023013077278033334, | |
| "loss": 2.6343, | |
| "step": 58600 | |
| }, | |
| { | |
| "epoch": 0.5262264376117433, | |
| "grad_norm": 0.12780235707759857, | |
| "learning_rate": 0.00022942898925588912, | |
| "loss": 2.635, | |
| "step": 58700 | |
| }, | |
| { | |
| "epoch": 0.5271229051374873, | |
| "grad_norm": 0.12590976059436798, | |
| "learning_rate": 0.00022872736889843116, | |
| "loss": 2.6274, | |
| "step": 58800 | |
| }, | |
| { | |
| "epoch": 0.5280193726632313, | |
| "grad_norm": 0.1275419443845749, | |
| "learning_rate": 0.0002280259172731346, | |
| "loss": 2.6303, | |
| "step": 58900 | |
| }, | |
| { | |
| "epoch": 0.5289158401889753, | |
| "grad_norm": 0.12726274132728577, | |
| "learning_rate": 0.00022732463994383641, | |
| "loss": 2.6297, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 0.5298123077147194, | |
| "grad_norm": 0.13650214672088623, | |
| "learning_rate": 0.00022662354247299086, | |
| "loss": 2.6233, | |
| "step": 59100 | |
| }, | |
| { | |
| "epoch": 0.5307087752404634, | |
| "grad_norm": 0.12631866335868835, | |
| "learning_rate": 0.00022592263042162577, | |
| "loss": 2.6333, | |
| "step": 59200 | |
| }, | |
| { | |
| "epoch": 0.5316052427662075, | |
| "grad_norm": 0.12630638480186462, | |
| "learning_rate": 0.00022522190934929801, | |
| "loss": 2.6275, | |
| "step": 59300 | |
| }, | |
| { | |
| "epoch": 0.5325017102919515, | |
| "grad_norm": 0.12765736877918243, | |
| "learning_rate": 0.00022452138481404986, | |
| "loss": 2.6306, | |
| "step": 59400 | |
| }, | |
| { | |
| "epoch": 0.5333981778176955, | |
| "grad_norm": 0.12750397622585297, | |
| "learning_rate": 0.00022382106237236457, | |
| "loss": 2.6271, | |
| "step": 59500 | |
| }, | |
| { | |
| "epoch": 0.5342946453434395, | |
| "grad_norm": 0.13162820041179657, | |
| "learning_rate": 0.0002231209475791225, | |
| "loss": 2.6287, | |
| "step": 59600 | |
| }, | |
| { | |
| "epoch": 0.5351911128691835, | |
| "grad_norm": 0.13135115802288055, | |
| "learning_rate": 0.0002224210459875568, | |
| "loss": 2.6322, | |
| "step": 59700 | |
| }, | |
| { | |
| "epoch": 0.5360875803949275, | |
| "grad_norm": 0.12867547571659088, | |
| "learning_rate": 0.00022172136314920968, | |
| "loss": 2.6353, | |
| "step": 59800 | |
| }, | |
| { | |
| "epoch": 0.5369840479206716, | |
| "grad_norm": 0.1288844645023346, | |
| "learning_rate": 0.00022102190461388818, | |
| "loss": 2.6355, | |
| "step": 59900 | |
| }, | |
| { | |
| "epoch": 0.5378805154464156, | |
| "grad_norm": 0.12920548021793365, | |
| "learning_rate": 0.00022032267592962023, | |
| "loss": 2.6274, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 0.5387769829721596, | |
| "grad_norm": 0.1328190267086029, | |
| "learning_rate": 0.0002196236826426104, | |
| "loss": 2.6318, | |
| "step": 60100 | |
| }, | |
| { | |
| "epoch": 0.5396734504979037, | |
| "grad_norm": 0.12695330381393433, | |
| "learning_rate": 0.00021892493029719652, | |
| "loss": 2.6287, | |
| "step": 60200 | |
| }, | |
| { | |
| "epoch": 0.5405699180236477, | |
| "grad_norm": 0.13476692140102386, | |
| "learning_rate": 0.00021822642443580483, | |
| "loss": 2.6335, | |
| "step": 60300 | |
| }, | |
| { | |
| "epoch": 0.5414663855493917, | |
| "grad_norm": 0.13216422498226166, | |
| "learning_rate": 0.00021752817059890673, | |
| "loss": 2.6262, | |
| "step": 60400 | |
| }, | |
| { | |
| "epoch": 0.5423628530751358, | |
| "grad_norm": 0.13027572631835938, | |
| "learning_rate": 0.00021683017432497464, | |
| "loss": 2.6189, | |
| "step": 60500 | |
| }, | |
| { | |
| "epoch": 0.5432593206008798, | |
| "grad_norm": 0.12778045237064362, | |
| "learning_rate": 0.00021613244115043802, | |
| "loss": 2.6295, | |
| "step": 60600 | |
| }, | |
| { | |
| "epoch": 0.5441557881266238, | |
| "grad_norm": 0.13152672350406647, | |
| "learning_rate": 0.00021543497660963924, | |
| "loss": 2.6303, | |
| "step": 60700 | |
| }, | |
| { | |
| "epoch": 0.5450522556523678, | |
| "grad_norm": 0.12792283296585083, | |
| "learning_rate": 0.00021473778623479006, | |
| "loss": 2.6272, | |
| "step": 60800 | |
| }, | |
| { | |
| "epoch": 0.5459487231781118, | |
| "grad_norm": 0.1273530125617981, | |
| "learning_rate": 0.00021404087555592762, | |
| "loss": 2.625, | |
| "step": 60900 | |
| }, | |
| { | |
| "epoch": 0.5468451907038558, | |
| "grad_norm": 0.1318267434835434, | |
| "learning_rate": 0.0002133442501008705, | |
| "loss": 2.6254, | |
| "step": 61000 | |
| }, | |
| { | |
| "epoch": 0.5477416582295999, | |
| "grad_norm": 0.13691310584545135, | |
| "learning_rate": 0.00021264791539517486, | |
| "loss": 2.6211, | |
| "step": 61100 | |
| }, | |
| { | |
| "epoch": 0.5486381257553439, | |
| "grad_norm": 0.12379534542560577, | |
| "learning_rate": 0.00021195187696209072, | |
| "loss": 2.6248, | |
| "step": 61200 | |
| }, | |
| { | |
| "epoch": 0.549534593281088, | |
| "grad_norm": 0.13373887538909912, | |
| "learning_rate": 0.00021125614032251822, | |
| "loss": 2.6304, | |
| "step": 61300 | |
| }, | |
| { | |
| "epoch": 0.550431060806832, | |
| "grad_norm": 0.13529092073440552, | |
| "learning_rate": 0.00021056071099496333, | |
| "loss": 2.6268, | |
| "step": 61400 | |
| }, | |
| { | |
| "epoch": 0.551327528332576, | |
| "grad_norm": 0.13044504821300507, | |
| "learning_rate": 0.0002098655944954948, | |
| "loss": 2.6265, | |
| "step": 61500 | |
| }, | |
| { | |
| "epoch": 0.55222399585832, | |
| "grad_norm": 0.12602747976779938, | |
| "learning_rate": 0.00020917079633770005, | |
| "loss": 2.6248, | |
| "step": 61600 | |
| }, | |
| { | |
| "epoch": 0.5531204633840641, | |
| "grad_norm": 0.12802156805992126, | |
| "learning_rate": 0.00020847632203264128, | |
| "loss": 2.6194, | |
| "step": 61700 | |
| }, | |
| { | |
| "epoch": 0.5540169309098081, | |
| "grad_norm": 0.1250976324081421, | |
| "learning_rate": 0.000207782177088812, | |
| "loss": 2.6297, | |
| "step": 61800 | |
| }, | |
| { | |
| "epoch": 0.5549133984355521, | |
| "grad_norm": 0.13608118891716003, | |
| "learning_rate": 0.00020708836701209332, | |
| "loss": 2.6286, | |
| "step": 61900 | |
| }, | |
| { | |
| "epoch": 0.5558098659612961, | |
| "grad_norm": 0.1274801641702652, | |
| "learning_rate": 0.00020639489730571014, | |
| "loss": 2.6283, | |
| "step": 62000 | |
| }, | |
| { | |
| "epoch": 0.5567063334870401, | |
| "grad_norm": 0.12696389853954315, | |
| "learning_rate": 0.0002057017734701877, | |
| "loss": 2.6262, | |
| "step": 62100 | |
| }, | |
| { | |
| "epoch": 0.5576028010127841, | |
| "grad_norm": 0.12838256359100342, | |
| "learning_rate": 0.00020501592697065653, | |
| "loss": 2.6261, | |
| "step": 62200 | |
| }, | |
| { | |
| "epoch": 0.5584992685385282, | |
| "grad_norm": 0.12481874227523804, | |
| "learning_rate": 0.00020432350777158638, | |
| "loss": 2.6197, | |
| "step": 62300 | |
| }, | |
| { | |
| "epoch": 0.5593957360642723, | |
| "grad_norm": 0.13098376989364624, | |
| "learning_rate": 0.00020363145087340989, | |
| "loss": 2.6218, | |
| "step": 62400 | |
| }, | |
| { | |
| "epoch": 0.5602922035900163, | |
| "grad_norm": 0.1315186321735382, | |
| "learning_rate": 0.0002029397617654461, | |
| "loss": 2.6229, | |
| "step": 62500 | |
| }, | |
| { | |
| "epoch": 0.5611886711157603, | |
| "grad_norm": 0.13062401115894318, | |
| "learning_rate": 0.00020224844593409665, | |
| "loss": 2.6224, | |
| "step": 62600 | |
| }, | |
| { | |
| "epoch": 0.5620851386415043, | |
| "grad_norm": 0.12941262125968933, | |
| "learning_rate": 0.00020155750886280245, | |
| "loss": 2.6264, | |
| "step": 62700 | |
| }, | |
| { | |
| "epoch": 0.5629816061672483, | |
| "grad_norm": 0.12870369851589203, | |
| "learning_rate": 0.00020086695603199994, | |
| "loss": 2.6176, | |
| "step": 62800 | |
| }, | |
| { | |
| "epoch": 0.5638780736929924, | |
| "grad_norm": 0.12999162077903748, | |
| "learning_rate": 0.00020017679291907804, | |
| "loss": 2.6217, | |
| "step": 62900 | |
| }, | |
| { | |
| "epoch": 0.5647745412187364, | |
| "grad_norm": 0.12885361909866333, | |
| "learning_rate": 0.00019948702499833436, | |
| "loss": 2.6248, | |
| "step": 63000 | |
| }, | |
| { | |
| "epoch": 0.5656710087444804, | |
| "grad_norm": 0.12850242853164673, | |
| "learning_rate": 0.00019879765774093188, | |
| "loss": 2.6276, | |
| "step": 63100 | |
| }, | |
| { | |
| "epoch": 0.5665674762702244, | |
| "grad_norm": 0.12949827313423157, | |
| "learning_rate": 0.0001981086966148556, | |
| "loss": 2.6191, | |
| "step": 63200 | |
| }, | |
| { | |
| "epoch": 0.5674639437959684, | |
| "grad_norm": 0.1278415471315384, | |
| "learning_rate": 0.00019742014708486904, | |
| "loss": 2.6321, | |
| "step": 63300 | |
| }, | |
| { | |
| "epoch": 0.5683604113217124, | |
| "grad_norm": 0.12578001618385315, | |
| "learning_rate": 0.00019673201461247125, | |
| "loss": 2.616, | |
| "step": 63400 | |
| }, | |
| { | |
| "epoch": 0.5692568788474566, | |
| "grad_norm": 0.12869875133037567, | |
| "learning_rate": 0.00019604430465585277, | |
| "loss": 2.6195, | |
| "step": 63500 | |
| }, | |
| { | |
| "epoch": 0.5701533463732006, | |
| "grad_norm": 0.12808062136173248, | |
| "learning_rate": 0.0001953570226698532, | |
| "loss": 2.6231, | |
| "step": 63600 | |
| }, | |
| { | |
| "epoch": 0.5710498138989446, | |
| "grad_norm": 0.13366073369979858, | |
| "learning_rate": 0.0001946701741059174, | |
| "loss": 2.6234, | |
| "step": 63700 | |
| }, | |
| { | |
| "epoch": 0.5719462814246886, | |
| "grad_norm": 0.12796130776405334, | |
| "learning_rate": 0.00019398376441205227, | |
| "loss": 2.6268, | |
| "step": 63800 | |
| }, | |
| { | |
| "epoch": 0.5728427489504326, | |
| "grad_norm": 0.13759686052799225, | |
| "learning_rate": 0.00019329779903278375, | |
| "loss": 2.619, | |
| "step": 63900 | |
| }, | |
| { | |
| "epoch": 0.5737392164761766, | |
| "grad_norm": 0.12661823630332947, | |
| "learning_rate": 0.00019261228340911352, | |
| "loss": 2.6152, | |
| "step": 64000 | |
| }, | |
| { | |
| "epoch": 0.5746356840019207, | |
| "grad_norm": 0.13385765254497528, | |
| "learning_rate": 0.00019192722297847587, | |
| "loss": 2.6217, | |
| "step": 64100 | |
| }, | |
| { | |
| "epoch": 0.5755321515276647, | |
| "grad_norm": 0.1298641860485077, | |
| "learning_rate": 0.00019124262317469443, | |
| "loss": 2.622, | |
| "step": 64200 | |
| }, | |
| { | |
| "epoch": 0.5764286190534087, | |
| "grad_norm": 0.12806908786296844, | |
| "learning_rate": 0.00019056532844060364, | |
| "loss": 2.6181, | |
| "step": 64300 | |
| }, | |
| { | |
| "epoch": 0.5773250865791527, | |
| "grad_norm": 0.12761425971984863, | |
| "learning_rate": 0.00018988849575433026, | |
| "loss": 2.6178, | |
| "step": 64400 | |
| }, | |
| { | |
| "epoch": 0.5782215541048968, | |
| "grad_norm": 0.1286136656999588, | |
| "learning_rate": 0.00018920530080606497, | |
| "loss": 2.6202, | |
| "step": 64500 | |
| }, | |
| { | |
| "epoch": 0.5791180216306409, | |
| "grad_norm": 0.12874256074428558, | |
| "learning_rate": 0.0001885225880746421, | |
| "loss": 2.6064, | |
| "step": 64600 | |
| }, | |
| { | |
| "epoch": 0.5800144891563849, | |
| "grad_norm": 0.12678521871566772, | |
| "learning_rate": 0.00018784036297526366, | |
| "loss": 2.6256, | |
| "step": 64700 | |
| }, | |
| { | |
| "epoch": 0.5809109566821289, | |
| "grad_norm": 0.1318819671869278, | |
| "learning_rate": 0.0001871586309192639, | |
| "loss": 2.6207, | |
| "step": 64800 | |
| }, | |
| { | |
| "epoch": 0.5818074242078729, | |
| "grad_norm": 0.13057227432727814, | |
| "learning_rate": 0.00018647739731406606, | |
| "loss": 2.6188, | |
| "step": 64900 | |
| }, | |
| { | |
| "epoch": 0.5827038917336169, | |
| "grad_norm": 0.1309261918067932, | |
| "learning_rate": 0.00018579666756314, | |
| "loss": 2.6225, | |
| "step": 65000 | |
| }, | |
| { | |
| "epoch": 0.5836003592593609, | |
| "grad_norm": 0.12833261489868164, | |
| "learning_rate": 0.0001851164470659589, | |
| "loss": 2.6151, | |
| "step": 65100 | |
| }, | |
| { | |
| "epoch": 0.584496826785105, | |
| "grad_norm": 0.1268189400434494, | |
| "learning_rate": 0.00018443674121795678, | |
| "loss": 2.6177, | |
| "step": 65200 | |
| }, | |
| { | |
| "epoch": 0.585393294310849, | |
| "grad_norm": 0.1255330890417099, | |
| "learning_rate": 0.00018375755541048526, | |
| "loss": 2.6206, | |
| "step": 65300 | |
| }, | |
| { | |
| "epoch": 0.586289761836593, | |
| "grad_norm": 0.1327618807554245, | |
| "learning_rate": 0.0001830788950307713, | |
| "loss": 2.6236, | |
| "step": 65400 | |
| }, | |
| { | |
| "epoch": 0.587186229362337, | |
| "grad_norm": 0.12842676043510437, | |
| "learning_rate": 0.0001824007654618739, | |
| "loss": 2.6176, | |
| "step": 65500 | |
| }, | |
| { | |
| "epoch": 0.5880826968880811, | |
| "grad_norm": 0.1308692991733551, | |
| "learning_rate": 0.00018172317208264212, | |
| "loss": 2.6171, | |
| "step": 65600 | |
| }, | |
| { | |
| "epoch": 0.5889791644138251, | |
| "grad_norm": 0.12851682305335999, | |
| "learning_rate": 0.0001810461202676717, | |
| "loss": 2.6164, | |
| "step": 65700 | |
| }, | |
| { | |
| "epoch": 0.5898756319395692, | |
| "grad_norm": 0.14155201613903046, | |
| "learning_rate": 0.00018036961538726314, | |
| "loss": 2.6198, | |
| "step": 65800 | |
| }, | |
| { | |
| "epoch": 0.5907720994653132, | |
| "grad_norm": 0.1263512820005417, | |
| "learning_rate": 0.0001796936628073782, | |
| "loss": 2.6202, | |
| "step": 65900 | |
| }, | |
| { | |
| "epoch": 0.5916685669910572, | |
| "grad_norm": 0.12849842011928558, | |
| "learning_rate": 0.00017901826788959825, | |
| "loss": 2.6154, | |
| "step": 66000 | |
| }, | |
| { | |
| "epoch": 0.5925650345168012, | |
| "grad_norm": 0.12791509926319122, | |
| "learning_rate": 0.0001783434359910811, | |
| "loss": 2.6201, | |
| "step": 66100 | |
| }, | |
| { | |
| "epoch": 0.5934615020425452, | |
| "grad_norm": 0.12755149602890015, | |
| "learning_rate": 0.00017766917246451902, | |
| "loss": 2.6166, | |
| "step": 66200 | |
| }, | |
| { | |
| "epoch": 0.5943579695682892, | |
| "grad_norm": 0.13783405721187592, | |
| "learning_rate": 0.00017699548265809578, | |
| "loss": 2.6165, | |
| "step": 66300 | |
| }, | |
| { | |
| "epoch": 0.5952544370940333, | |
| "grad_norm": 0.1294228881597519, | |
| "learning_rate": 0.00017632237191544462, | |
| "loss": 2.6155, | |
| "step": 66400 | |
| }, | |
| { | |
| "epoch": 0.5961509046197773, | |
| "grad_norm": 0.13272565603256226, | |
| "learning_rate": 0.00017564984557560555, | |
| "loss": 2.6116, | |
| "step": 66500 | |
| }, | |
| { | |
| "epoch": 0.5970473721455213, | |
| "grad_norm": 0.13658390939235687, | |
| "learning_rate": 0.00017497790897298327, | |
| "loss": 2.6159, | |
| "step": 66600 | |
| }, | |
| { | |
| "epoch": 0.5979438396712654, | |
| "grad_norm": 0.13049334287643433, | |
| "learning_rate": 0.00017430656743730477, | |
| "loss": 2.6191, | |
| "step": 66700 | |
| }, | |
| { | |
| "epoch": 0.5988403071970094, | |
| "grad_norm": 0.13098062574863434, | |
| "learning_rate": 0.00017363582629357705, | |
| "loss": 2.618, | |
| "step": 66800 | |
| }, | |
| { | |
| "epoch": 0.5997367747227534, | |
| "grad_norm": 0.12643533945083618, | |
| "learning_rate": 0.00017296569086204479, | |
| "loss": 2.6172, | |
| "step": 66900 | |
| }, | |
| { | |
| "epoch": 0.6006332422484975, | |
| "grad_norm": 0.12678323686122894, | |
| "learning_rate": 0.00017229616645814813, | |
| "loss": 2.6185, | |
| "step": 67000 | |
| }, | |
| { | |
| "epoch": 0.6015297097742415, | |
| "grad_norm": 0.12956561148166656, | |
| "learning_rate": 0.00017162725839248077, | |
| "loss": 2.6089, | |
| "step": 67100 | |
| }, | |
| { | |
| "epoch": 0.6024261772999855, | |
| "grad_norm": 0.129245787858963, | |
| "learning_rate": 0.00017095897197074773, | |
| "loss": 2.6043, | |
| "step": 67200 | |
| }, | |
| { | |
| "epoch": 0.6033226448257295, | |
| "grad_norm": 0.12439849972724915, | |
| "learning_rate": 0.00017029131249372305, | |
| "loss": 2.6174, | |
| "step": 67300 | |
| }, | |
| { | |
| "epoch": 0.6042191123514735, | |
| "grad_norm": 0.1366392970085144, | |
| "learning_rate": 0.00016962428525720807, | |
| "loss": 2.623, | |
| "step": 67400 | |
| }, | |
| { | |
| "epoch": 0.6051155798772175, | |
| "grad_norm": 0.12857592105865479, | |
| "learning_rate": 0.00016895789555198908, | |
| "loss": 2.61, | |
| "step": 67500 | |
| }, | |
| { | |
| "epoch": 0.6060120474029616, | |
| "grad_norm": 0.1319228857755661, | |
| "learning_rate": 0.00016829214866379572, | |
| "loss": 2.611, | |
| "step": 67600 | |
| }, | |
| { | |
| "epoch": 0.6069085149287056, | |
| "grad_norm": 0.1241278126835823, | |
| "learning_rate": 0.00016762704987325884, | |
| "loss": 2.6049, | |
| "step": 67700 | |
| }, | |
| { | |
| "epoch": 0.6078049824544497, | |
| "grad_norm": 0.13301433622837067, | |
| "learning_rate": 0.00016696260445586865, | |
| "loss": 2.6089, | |
| "step": 67800 | |
| }, | |
| { | |
| "epoch": 0.6087014499801937, | |
| "grad_norm": 0.1289006769657135, | |
| "learning_rate": 0.00016629881768193271, | |
| "loss": 2.6123, | |
| "step": 67900 | |
| }, | |
| { | |
| "epoch": 0.6095979175059377, | |
| "grad_norm": 0.1281334012746811, | |
| "learning_rate": 0.00016563569481653457, | |
| "loss": 2.6086, | |
| "step": 68000 | |
| }, | |
| { | |
| "epoch": 0.6104943850316817, | |
| "grad_norm": 0.13358446955680847, | |
| "learning_rate": 0.00016497324111949146, | |
| "loss": 2.6162, | |
| "step": 68100 | |
| }, | |
| { | |
| "epoch": 0.6113908525574258, | |
| "grad_norm": 0.12993699312210083, | |
| "learning_rate": 0.00016431146184531304, | |
| "loss": 2.6077, | |
| "step": 68200 | |
| }, | |
| { | |
| "epoch": 0.6122873200831698, | |
| "grad_norm": 0.12562768161296844, | |
| "learning_rate": 0.00016365036224315948, | |
| "loss": 2.6099, | |
| "step": 68300 | |
| }, | |
| { | |
| "epoch": 0.6131837876089138, | |
| "grad_norm": 0.12500709295272827, | |
| "learning_rate": 0.00016298994755679986, | |
| "loss": 2.6186, | |
| "step": 68400 | |
| }, | |
| { | |
| "epoch": 0.6140802551346578, | |
| "grad_norm": 0.13538555800914764, | |
| "learning_rate": 0.00016233681683644314, | |
| "loss": 2.6057, | |
| "step": 68500 | |
| }, | |
| { | |
| "epoch": 0.6149767226604018, | |
| "grad_norm": 0.13071191310882568, | |
| "learning_rate": 0.00016167778071145164, | |
| "loss": 2.6108, | |
| "step": 68600 | |
| }, | |
| { | |
| "epoch": 0.6158731901861458, | |
| "grad_norm": 0.12502072751522064, | |
| "learning_rate": 0.00016101944514855299, | |
| "loss": 2.6098, | |
| "step": 68700 | |
| }, | |
| { | |
| "epoch": 0.6167696577118899, | |
| "grad_norm": 0.130848690867424, | |
| "learning_rate": 0.0001603618153695922, | |
| "loss": 2.6067, | |
| "step": 68800 | |
| }, | |
| { | |
| "epoch": 0.617666125237634, | |
| "grad_norm": 0.12595491111278534, | |
| "learning_rate": 0.0001597048965908161, | |
| "loss": 2.6009, | |
| "step": 68900 | |
| }, | |
| { | |
| "epoch": 0.618562592763378, | |
| "grad_norm": 0.13049688935279846, | |
| "learning_rate": 0.0001590486940228319, | |
| "loss": 2.6069, | |
| "step": 69000 | |
| }, | |
| { | |
| "epoch": 0.619459060289122, | |
| "grad_norm": 0.1312711089849472, | |
| "learning_rate": 0.00015839321287056585, | |
| "loss": 2.6088, | |
| "step": 69100 | |
| }, | |
| { | |
| "epoch": 0.620355527814866, | |
| "grad_norm": 0.13168184459209442, | |
| "learning_rate": 0.00015773845833322208, | |
| "loss": 2.6188, | |
| "step": 69200 | |
| }, | |
| { | |
| "epoch": 0.62125199534061, | |
| "grad_norm": 0.12749746441841125, | |
| "learning_rate": 0.00015708443560424141, | |
| "loss": 2.6151, | |
| "step": 69300 | |
| }, | |
| { | |
| "epoch": 0.6221484628663541, | |
| "grad_norm": 0.13357709348201752, | |
| "learning_rate": 0.0001564311498712599, | |
| "loss": 2.6064, | |
| "step": 69400 | |
| }, | |
| { | |
| "epoch": 0.6230449303920981, | |
| "grad_norm": 0.12775705754756927, | |
| "learning_rate": 0.00015577860631606783, | |
| "loss": 2.6118, | |
| "step": 69500 | |
| }, | |
| { | |
| "epoch": 0.6239413979178421, | |
| "grad_norm": 0.13166511058807373, | |
| "learning_rate": 0.00015512681011456859, | |
| "loss": 2.602, | |
| "step": 69600 | |
| }, | |
| { | |
| "epoch": 0.6248378654435861, | |
| "grad_norm": 0.1261209398508072, | |
| "learning_rate": 0.00015447576643673778, | |
| "loss": 2.61, | |
| "step": 69700 | |
| }, | |
| { | |
| "epoch": 0.6257343329693301, | |
| "grad_norm": 0.12618456780910492, | |
| "learning_rate": 0.00015382548044658194, | |
| "loss": 2.6102, | |
| "step": 69800 | |
| }, | |
| { | |
| "epoch": 0.6266308004950741, | |
| "grad_norm": 0.12650653719902039, | |
| "learning_rate": 0.00015317595730209783, | |
| "loss": 2.6137, | |
| "step": 69900 | |
| }, | |
| { | |
| "epoch": 0.6275272680208183, | |
| "grad_norm": 0.12881776690483093, | |
| "learning_rate": 0.0001525272021552313, | |
| "loss": 2.6134, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 0.6284237355465623, | |
| "grad_norm": 0.1257714331150055, | |
| "learning_rate": 0.00015187922015183638, | |
| "loss": 2.6172, | |
| "step": 70100 | |
| }, | |
| { | |
| "epoch": 0.6293202030723063, | |
| "grad_norm": 0.12860926985740662, | |
| "learning_rate": 0.00015123201643163495, | |
| "loss": 2.6093, | |
| "step": 70200 | |
| }, | |
| { | |
| "epoch": 0.6302166705980503, | |
| "grad_norm": 0.1268460750579834, | |
| "learning_rate": 0.0001505855961281754, | |
| "loss": 2.6065, | |
| "step": 70300 | |
| }, | |
| { | |
| "epoch": 0.6311131381237943, | |
| "grad_norm": 0.1278506964445114, | |
| "learning_rate": 0.00014993996436879232, | |
| "loss": 2.6117, | |
| "step": 70400 | |
| }, | |
| { | |
| "epoch": 0.6320096056495383, | |
| "grad_norm": 0.1345282644033432, | |
| "learning_rate": 0.00014930157071006556, | |
| "loss": 2.6175, | |
| "step": 70500 | |
| }, | |
| { | |
| "epoch": 0.6329060731752824, | |
| "grad_norm": 0.13263201713562012, | |
| "learning_rate": 0.0001486575233826827, | |
| "loss": 2.6045, | |
| "step": 70600 | |
| }, | |
| { | |
| "epoch": 0.6338025407010264, | |
| "grad_norm": 0.12888827919960022, | |
| "learning_rate": 0.00014801427989263638, | |
| "loss": 2.606, | |
| "step": 70700 | |
| }, | |
| { | |
| "epoch": 0.6346990082267704, | |
| "grad_norm": 0.12627027928829193, | |
| "learning_rate": 0.00014737184534206293, | |
| "loss": 2.6056, | |
| "step": 70800 | |
| }, | |
| { | |
| "epoch": 0.6355954757525144, | |
| "grad_norm": 0.12864083051681519, | |
| "learning_rate": 0.00014673022482668242, | |
| "loss": 2.6085, | |
| "step": 70900 | |
| }, | |
| { | |
| "epoch": 0.6364919432782585, | |
| "grad_norm": 0.1318010538816452, | |
| "learning_rate": 0.00014608942343575788, | |
| "loss": 2.6087, | |
| "step": 71000 | |
| }, | |
| { | |
| "epoch": 0.6373884108040025, | |
| "grad_norm": 0.12685802578926086, | |
| "learning_rate": 0.0001454494462520551, | |
| "loss": 2.6069, | |
| "step": 71100 | |
| }, | |
| { | |
| "epoch": 0.6382848783297466, | |
| "grad_norm": 0.12846295535564423, | |
| "learning_rate": 0.0001448102983518025, | |
| "loss": 2.6005, | |
| "step": 71200 | |
| }, | |
| { | |
| "epoch": 0.6391813458554906, | |
| "grad_norm": 0.1338125467300415, | |
| "learning_rate": 0.00014417198480465068, | |
| "loss": 2.6043, | |
| "step": 71300 | |
| }, | |
| { | |
| "epoch": 0.6400778133812346, | |
| "grad_norm": 0.1277831345796585, | |
| "learning_rate": 0.00014353451067363227, | |
| "loss": 2.6048, | |
| "step": 71400 | |
| }, | |
| { | |
| "epoch": 0.6409742809069786, | |
| "grad_norm": 0.1271323561668396, | |
| "learning_rate": 0.0001428978810151216, | |
| "loss": 2.6066, | |
| "step": 71500 | |
| }, | |
| { | |
| "epoch": 0.6418707484327226, | |
| "grad_norm": 0.12640222907066345, | |
| "learning_rate": 0.00014226210087879494, | |
| "loss": 2.6082, | |
| "step": 71600 | |
| }, | |
| { | |
| "epoch": 0.6427672159584666, | |
| "grad_norm": 0.12706422805786133, | |
| "learning_rate": 0.0001416271753075902, | |
| "loss": 2.6041, | |
| "step": 71700 | |
| }, | |
| { | |
| "epoch": 0.6436636834842107, | |
| "grad_norm": 0.12946538627147675, | |
| "learning_rate": 0.0001409931093376667, | |
| "loss": 2.6042, | |
| "step": 71800 | |
| }, | |
| { | |
| "epoch": 0.6445601510099547, | |
| "grad_norm": 0.13379494845867157, | |
| "learning_rate": 0.00014035990799836584, | |
| "loss": 2.6064, | |
| "step": 71900 | |
| }, | |
| { | |
| "epoch": 0.6454566185356987, | |
| "grad_norm": 0.13055512309074402, | |
| "learning_rate": 0.00013972757631217075, | |
| "loss": 2.6067, | |
| "step": 72000 | |
| } | |
| ], | |
| "logging_steps": 100, | |
| "max_steps": 111548, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 2000, | |
| "total_flos": 3.857626077069312e+18, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |