| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.0, | |
| "eval_steps": 1660, | |
| "global_step": 33185, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 3.0134096730450506e-05, | |
| "grad_norm": 8.0, | |
| "learning_rate": 2e-06, | |
| "loss": 1.334, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0030134096730450506, | |
| "grad_norm": 0.34765625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9372, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.006026819346090101, | |
| "grad_norm": 1.2734375, | |
| "learning_rate": 0.0004, | |
| "loss": 0.6477, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.009040229019135152, | |
| "grad_norm": 0.59375, | |
| "learning_rate": 0.0006, | |
| "loss": 0.5967, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.012053638692180202, | |
| "grad_norm": 1.40625, | |
| "learning_rate": 0.0008, | |
| "loss": 0.5828, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.015067048365225252, | |
| "grad_norm": 0.453125, | |
| "learning_rate": 0.001, | |
| "loss": 0.5774, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.018080458038270304, | |
| "grad_norm": 0.890625, | |
| "learning_rate": 0.0012, | |
| "loss": 0.557, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.021093867711315353, | |
| "grad_norm": 0.396484375, | |
| "learning_rate": 0.0014, | |
| "loss": 0.5573, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.024107277384360404, | |
| "grad_norm": 0.416015625, | |
| "learning_rate": 0.0016, | |
| "loss": 0.5577, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.027120687057405453, | |
| "grad_norm": 0.375, | |
| "learning_rate": 0.0018000000000000002, | |
| "loss": 0.563, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.030134096730450505, | |
| "grad_norm": 0.58984375, | |
| "learning_rate": 0.002, | |
| "loss": 0.568, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.03314750640349556, | |
| "grad_norm": 0.33203125, | |
| "learning_rate": 0.0019999571252319053, | |
| "loss": 0.5969, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.03616091607654061, | |
| "grad_norm": 0.5234375, | |
| "learning_rate": 0.0019998285050126107, | |
| "loss": 0.6365, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.03917432574958565, | |
| "grad_norm": 0.302734375, | |
| "learning_rate": 0.0019996141515967, | |
| "loss": 0.6351, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.042187735422630705, | |
| "grad_norm": 0.375, | |
| "learning_rate": 0.001999314085407178, | |
| "loss": 0.5537, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.04520114509567576, | |
| "grad_norm": 0.271484375, | |
| "learning_rate": 0.0019989283350335314, | |
| "loss": 0.5484, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.04821455476872081, | |
| "grad_norm": 0.375, | |
| "learning_rate": 0.0019984569372289993, | |
| "loss": 0.5583, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.050022600572547836, | |
| "eval_peoplespeech-clean-transcription_loss": 4.166935443878174, | |
| "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, | |
| "eval_peoplespeech-clean-transcription_runtime": 15.2594, | |
| "eval_peoplespeech-clean-transcription_samples_per_second": 4.194, | |
| "eval_peoplespeech-clean-transcription_steps_per_second": 0.066, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.05122796444176586, | |
| "grad_norm": 0.2578125, | |
| "learning_rate": 0.0019978999369070737, | |
| "loss": 0.5722, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.054241374114810906, | |
| "grad_norm": 0.3203125, | |
| "learning_rate": 0.001997257387137221, | |
| "loss": 0.5699, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.05725478378785596, | |
| "grad_norm": 0.27734375, | |
| "learning_rate": 0.0019965293491398237, | |
| "loss": 0.595, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.06026819346090101, | |
| "grad_norm": 0.455078125, | |
| "learning_rate": 0.001995715892280349, | |
| "loss": 0.561, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.06328160313394605, | |
| "grad_norm": 0.26171875, | |
| "learning_rate": 0.00199481709406274, | |
| "loss": 0.5553, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.06629501280699111, | |
| "grad_norm": 0.27734375, | |
| "learning_rate": 0.0019938330401220307, | |
| "loss": 0.5668, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.06930842248003616, | |
| "grad_norm": 0.318359375, | |
| "learning_rate": 0.0019927638242161864, | |
| "loss": 0.574, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.07232183215308122, | |
| "grad_norm": 0.1767578125, | |
| "learning_rate": 0.001991609548217171, | |
| "loss": 0.6076, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.07533524182612626, | |
| "grad_norm": 0.18359375, | |
| "learning_rate": 0.001990370322101242, | |
| "loss": 0.6369, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.0783486514991713, | |
| "grad_norm": 0.314453125, | |
| "learning_rate": 0.001989046263938472, | |
| "loss": 0.6106, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.08136206117221637, | |
| "grad_norm": 0.240234375, | |
| "learning_rate": 0.0019876374998814973, | |
| "loss": 0.5993, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.08437547084526141, | |
| "grad_norm": 0.51953125, | |
| "learning_rate": 0.0019861441641535007, | |
| "loss": 0.5933, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.08738888051830647, | |
| "grad_norm": 0.609375, | |
| "learning_rate": 0.001984566399035423, | |
| "loss": 0.5937, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.09040229019135151, | |
| "grad_norm": 0.21484375, | |
| "learning_rate": 0.001982904354852404, | |
| "loss": 0.5881, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.09341569986439656, | |
| "grad_norm": 0.26171875, | |
| "learning_rate": 0.0019811581899594646, | |
| "loss": 0.5817, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.09642910953744162, | |
| "grad_norm": 0.2138671875, | |
| "learning_rate": 0.0019793280707264154, | |
| "loss": 0.588, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.09944251921048666, | |
| "grad_norm": 0.23828125, | |
| "learning_rate": 0.0019774141715220065, | |
| "loss": 0.5813, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.10004520114509567, | |
| "eval_peoplespeech-clean-transcription_loss": 4.31866455078125, | |
| "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, | |
| "eval_peoplespeech-clean-transcription_runtime": 14.293, | |
| "eval_peoplespeech-clean-transcription_samples_per_second": 4.478, | |
| "eval_peoplespeech-clean-transcription_steps_per_second": 0.07, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 0.10245592888353172, | |
| "grad_norm": 0.2080078125, | |
| "learning_rate": 0.0019754166746973156, | |
| "loss": 0.5784, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.10546933855657677, | |
| "grad_norm": 0.2119140625, | |
| "learning_rate": 0.0019733357705683705, | |
| "loss": 0.582, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.10848274822962181, | |
| "grad_norm": 0.248046875, | |
| "learning_rate": 0.001971171657398021, | |
| "loss": 0.5877, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.11149615790266687, | |
| "grad_norm": 0.255859375, | |
| "learning_rate": 0.001968924541377045, | |
| "loss": 0.5788, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.11450956757571192, | |
| "grad_norm": 0.271484375, | |
| "learning_rate": 0.001966594636604506, | |
| "loss": 0.5739, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.11752297724875697, | |
| "grad_norm": 0.26171875, | |
| "learning_rate": 0.001964182165067352, | |
| "loss": 0.5748, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.12053638692180202, | |
| "grad_norm": 7.28125, | |
| "learning_rate": 0.001961687356619266, | |
| "loss": 0.5746, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.12354979659484706, | |
| "grad_norm": 0.25, | |
| "learning_rate": 0.001959110448958769, | |
| "loss": 0.5877, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.1265632062678921, | |
| "grad_norm": 0.2421875, | |
| "learning_rate": 0.001956451687606567, | |
| "loss": 0.5652, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.12957661594093717, | |
| "grad_norm": 0.26171875, | |
| "learning_rate": 0.0019537113258821636, | |
| "loss": 0.5842, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.13259002561398223, | |
| "grad_norm": 0.2158203125, | |
| "learning_rate": 0.001950889624879722, | |
| "loss": 0.5687, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.13560343528702729, | |
| "grad_norm": 0.32421875, | |
| "learning_rate": 0.0019479868534431892, | |
| "loss": 0.5715, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.13861684496007232, | |
| "grad_norm": 0.251953125, | |
| "learning_rate": 0.001945003288140681, | |
| "loss": 0.5811, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.14163025463311738, | |
| "grad_norm": 0.44921875, | |
| "learning_rate": 0.0019419392132381317, | |
| "loss": 0.5936, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.14464366430616243, | |
| "grad_norm": 0.24609375, | |
| "learning_rate": 0.0019387949206722099, | |
| "loss": 0.5861, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.14765707397920746, | |
| "grad_norm": 0.255859375, | |
| "learning_rate": 0.0019355707100225034, | |
| "loss": 0.5867, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.1500678017176435, | |
| "eval_peoplespeech-clean-transcription_loss": 4.08488130569458, | |
| "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, | |
| "eval_peoplespeech-clean-transcription_runtime": 14.7738, | |
| "eval_peoplespeech-clean-transcription_samples_per_second": 4.332, | |
| "eval_peoplespeech-clean-transcription_steps_per_second": 0.068, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 0.15067048365225252, | |
| "grad_norm": 0.22265625, | |
| "learning_rate": 0.0019322668884829768, | |
| "loss": 0.5827, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.15368389332529758, | |
| "grad_norm": 0.267578125, | |
| "learning_rate": 0.0019288837708327019, | |
| "loss": 0.5829, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.1566973029983426, | |
| "grad_norm": 0.2236328125, | |
| "learning_rate": 0.0019254216794058665, | |
| "loss": 0.574, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.15971071267138767, | |
| "grad_norm": 0.232421875, | |
| "learning_rate": 0.0019218809440610645, | |
| "loss": 0.5907, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 0.16272412234443273, | |
| "grad_norm": 0.234375, | |
| "learning_rate": 0.0019182619021498664, | |
| "loss": 0.5736, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 0.1657375320174778, | |
| "grad_norm": 0.2451171875, | |
| "learning_rate": 0.001914564898484678, | |
| "loss": 0.586, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.16875094169052282, | |
| "grad_norm": 0.2080078125, | |
| "learning_rate": 0.0019107902853058875, | |
| "loss": 0.583, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 0.17176435136356788, | |
| "grad_norm": 0.2314453125, | |
| "learning_rate": 0.0019069384222483061, | |
| "loss": 0.589, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 0.17477776103661294, | |
| "grad_norm": 0.390625, | |
| "learning_rate": 0.0019030096763069007, | |
| "loss": 0.569, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 0.17779117070965797, | |
| "grad_norm": 0.24609375, | |
| "learning_rate": 0.0018990044218018295, | |
| "loss": 0.5914, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 0.18080458038270303, | |
| "grad_norm": 0.205078125, | |
| "learning_rate": 0.0018949230403427768, | |
| "loss": 0.5936, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.1838179900557481, | |
| "grad_norm": 0.2353515625, | |
| "learning_rate": 0.0018907659207925951, | |
| "loss": 0.5959, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 0.18683139972879312, | |
| "grad_norm": 0.220703125, | |
| "learning_rate": 0.0018865334592302553, | |
| "loss": 0.5734, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 0.18984480940183818, | |
| "grad_norm": 0.2431640625, | |
| "learning_rate": 0.0018822260589131075, | |
| "loss": 0.5815, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 0.19285821907488324, | |
| "grad_norm": 0.2158203125, | |
| "learning_rate": 0.0018778441302384629, | |
| "loss": 0.58, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 0.1958716287479283, | |
| "grad_norm": 0.2734375, | |
| "learning_rate": 0.0018733880907044892, | |
| "loss": 0.5807, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.19888503842097333, | |
| "grad_norm": 0.2216796875, | |
| "learning_rate": 0.0018688583648704348, | |
| "loss": 0.5741, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 0.20009040229019134, | |
| "eval_peoplespeech-clean-transcription_loss": 4.054948806762695, | |
| "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, | |
| "eval_peoplespeech-clean-transcription_runtime": 14.5864, | |
| "eval_peoplespeech-clean-transcription_samples_per_second": 4.388, | |
| "eval_peoplespeech-clean-transcription_steps_per_second": 0.069, | |
| "step": 6640 | |
| }, | |
| { | |
| "epoch": 0.20189844809401838, | |
| "grad_norm": 0.263671875, | |
| "learning_rate": 0.0018642553843161765, | |
| "loss": 0.5808, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 0.20491185776706344, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 0.0018595795876011011, | |
| "loss": 0.572, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 0.20792526744010847, | |
| "grad_norm": 0.224609375, | |
| "learning_rate": 0.001854831420222319, | |
| "loss": 0.5738, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 0.21093867711315353, | |
| "grad_norm": 0.18359375, | |
| "learning_rate": 0.001850011334572219, | |
| "loss": 0.5631, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.2139520867861986, | |
| "grad_norm": 0.2001953125, | |
| "learning_rate": 0.0018451197898953675, | |
| "loss": 0.5656, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 0.21696549645924362, | |
| "grad_norm": 0.1708984375, | |
| "learning_rate": 0.0018401572522447499, | |
| "loss": 0.5501, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 0.21997890613228868, | |
| "grad_norm": 0.1845703125, | |
| "learning_rate": 0.0018351241944373684, | |
| "loss": 0.5487, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 0.22299231580533374, | |
| "grad_norm": 0.14453125, | |
| "learning_rate": 0.0018300210960091926, | |
| "loss": 0.535, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 0.2260057254783788, | |
| "grad_norm": 0.111328125, | |
| "learning_rate": 0.0018248484431694705, | |
| "loss": 0.5265, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.22901913515142383, | |
| "grad_norm": 0.05126953125, | |
| "learning_rate": 0.0018196067287544043, | |
| "loss": 0.4819, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 0.2320325448244689, | |
| "grad_norm": 0.04296875, | |
| "learning_rate": 0.0018142964521801936, | |
| "loss": 0.4168, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 0.23504595449751395, | |
| "grad_norm": 0.033447265625, | |
| "learning_rate": 0.001808918119395454, | |
| "loss": 0.3548, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 0.23805936417055898, | |
| "grad_norm": 0.03271484375, | |
| "learning_rate": 0.0018034722428330089, | |
| "loss": 0.3206, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 0.24107277384360404, | |
| "grad_norm": 0.0274658203125, | |
| "learning_rate": 0.0017979593413610688, | |
| "loss": 0.3043, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.2440861835166491, | |
| "grad_norm": 0.0286865234375, | |
| "learning_rate": 0.0017923799402337944, | |
| "loss": 0.2899, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 0.24709959318969413, | |
| "grad_norm": 0.028076171875, | |
| "learning_rate": 0.0017867345710412504, | |
| "loss": 0.2772, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 0.2501130028627392, | |
| "grad_norm": 0.030029296875, | |
| "learning_rate": 0.00178102377165876, | |
| "loss": 0.2692, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 0.2501130028627392, | |
| "eval_peoplespeech-clean-transcription_loss": 1.788245677947998, | |
| "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, | |
| "eval_peoplespeech-clean-transcription_runtime": 13.7068, | |
| "eval_peoplespeech-clean-transcription_samples_per_second": 4.669, | |
| "eval_peoplespeech-clean-transcription_steps_per_second": 0.073, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 0.2531264125357842, | |
| "grad_norm": 0.0262451171875, | |
| "learning_rate": 0.0017752480861956536, | |
| "loss": 0.2649, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 0.2561398222088293, | |
| "grad_norm": 0.0272216796875, | |
| "learning_rate": 0.0017694080649434314, | |
| "loss": 0.2574, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.25915323188187434, | |
| "grad_norm": 0.0272216796875, | |
| "learning_rate": 0.0017635042643233307, | |
| "loss": 0.2522, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 0.26216664155491937, | |
| "grad_norm": 0.02880859375, | |
| "learning_rate": 0.0017575372468333127, | |
| "loss": 0.2487, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 0.26518005122796445, | |
| "grad_norm": 0.0230712890625, | |
| "learning_rate": 0.001751507580994468, | |
| "loss": 0.2407, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 0.2681934609010095, | |
| "grad_norm": 0.031005859375, | |
| "learning_rate": 0.0017454158412968522, | |
| "loss": 0.238, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 0.27120687057405457, | |
| "grad_norm": 0.0235595703125, | |
| "learning_rate": 0.0017392626081447465, | |
| "loss": 0.2347, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.2742202802470996, | |
| "grad_norm": 0.0306396484375, | |
| "learning_rate": 0.0017330484678013609, | |
| "loss": 0.2343, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 0.27723368992014463, | |
| "grad_norm": 0.0262451171875, | |
| "learning_rate": 0.0017267740123329753, | |
| "loss": 0.2324, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 0.2802470995931897, | |
| "grad_norm": 0.02880859375, | |
| "learning_rate": 0.0017204398395525308, | |
| "loss": 0.2294, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 0.28326050926623475, | |
| "grad_norm": 0.025390625, | |
| "learning_rate": 0.0017140465529626692, | |
| "loss": 0.2278, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 0.2862739189392798, | |
| "grad_norm": 0.027587890625, | |
| "learning_rate": 0.0017075947616982349, | |
| "loss": 0.2247, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.28928732861232487, | |
| "grad_norm": 0.024658203125, | |
| "learning_rate": 0.001701085080468237, | |
| "loss": 0.2204, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 0.2923007382853699, | |
| "grad_norm": 0.030029296875, | |
| "learning_rate": 0.0016945181294972828, | |
| "loss": 0.2201, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 0.29531414795841493, | |
| "grad_norm": 0.025634765625, | |
| "learning_rate": 0.0016878945344664831, | |
| "loss": 0.2196, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 0.29832755763146, | |
| "grad_norm": 0.0279541015625, | |
| "learning_rate": 0.0016812149264538402, | |
| "loss": 0.2163, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 0.300135603435287, | |
| "eval_peoplespeech-clean-transcription_loss": 1.6014918088912964, | |
| "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, | |
| "eval_peoplespeech-clean-transcription_runtime": 13.7355, | |
| "eval_peoplespeech-clean-transcription_samples_per_second": 4.659, | |
| "eval_peoplespeech-clean-transcription_steps_per_second": 0.073, | |
| "step": 9960 | |
| }, | |
| { | |
| "epoch": 0.30134096730450505, | |
| "grad_norm": 0.0257568359375, | |
| "learning_rate": 0.0016744799418741193, | |
| "loss": 0.2143, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.3043543769775501, | |
| "grad_norm": 0.027099609375, | |
| "learning_rate": 0.001667690222418214, | |
| "loss": 0.214, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 0.30736778665059517, | |
| "grad_norm": 0.0260009765625, | |
| "learning_rate": 0.0016608464149920064, | |
| "loss": 0.2111, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 0.3103811963236402, | |
| "grad_norm": 0.02734375, | |
| "learning_rate": 0.0016539491716547332, | |
| "loss": 0.2124, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 0.3133946059966852, | |
| "grad_norm": 0.025146484375, | |
| "learning_rate": 0.0016469991495568573, | |
| "loss": 0.2071, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 0.3164080156697303, | |
| "grad_norm": 0.0286865234375, | |
| "learning_rate": 0.0016399970108774587, | |
| "loss": 0.2106, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.31942142534277534, | |
| "grad_norm": 0.025634765625, | |
| "learning_rate": 0.001632943422761141, | |
| "loss": 0.2075, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 0.3224348350158204, | |
| "grad_norm": 0.0269775390625, | |
| "learning_rate": 0.0016258390572544716, | |
| "loss": 0.2065, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 0.32544824468886546, | |
| "grad_norm": 0.024169921875, | |
| "learning_rate": 0.001618684591241946, | |
| "loss": 0.2065, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 0.3284616543619105, | |
| "grad_norm": 0.026123046875, | |
| "learning_rate": 0.0016114807063815008, | |
| "loss": 0.2055, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 0.3314750640349556, | |
| "grad_norm": 0.0272216796875, | |
| "learning_rate": 0.0016042280890395642, | |
| "loss": 0.2043, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.3344884737080006, | |
| "grad_norm": 0.02685546875, | |
| "learning_rate": 0.0015969274302256621, | |
| "loss": 0.2006, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 0.33750188338104564, | |
| "grad_norm": 0.0245361328125, | |
| "learning_rate": 0.00158957942552658, | |
| "loss": 0.2021, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 0.34051529305409073, | |
| "grad_norm": 0.02783203125, | |
| "learning_rate": 0.00158218477504009, | |
| "loss": 0.2042, | |
| "step": 11300 | |
| }, | |
| { | |
| "epoch": 0.34352870272713576, | |
| "grad_norm": 0.0257568359375, | |
| "learning_rate": 0.0015747441833082476, | |
| "loss": 0.2043, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 0.3465421124001808, | |
| "grad_norm": 0.0263671875, | |
| "learning_rate": 0.0015672583592502632, | |
| "loss": 0.1991, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.3495555220732259, | |
| "grad_norm": 0.0281982421875, | |
| "learning_rate": 0.0015597280160949602, | |
| "loss": 0.1994, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 0.3501582040078349, | |
| "eval_peoplespeech-clean-transcription_loss": 1.5406583547592163, | |
| "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, | |
| "eval_peoplespeech-clean-transcription_runtime": 14.5635, | |
| "eval_peoplespeech-clean-transcription_samples_per_second": 4.395, | |
| "eval_peoplespeech-clean-transcription_steps_per_second": 0.069, | |
| "step": 11620 | |
| }, | |
| { | |
| "epoch": 0.3525689317462709, | |
| "grad_norm": 0.0257568359375, | |
| "learning_rate": 0.0015521538713128204, | |
| "loss": 0.2, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 0.35558234141931594, | |
| "grad_norm": 0.0244140625, | |
| "learning_rate": 0.001544536646547623, | |
| "loss": 0.1978, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 0.358595751092361, | |
| "grad_norm": 0.0255126953125, | |
| "learning_rate": 0.0015368770675476915, | |
| "loss": 0.1974, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 0.36160916076540606, | |
| "grad_norm": 0.025390625, | |
| "learning_rate": 0.001529175864096744, | |
| "loss": 0.1963, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.3646225704384511, | |
| "grad_norm": 0.027587890625, | |
| "learning_rate": 0.0015214337699443632, | |
| "loss": 0.1958, | |
| "step": 12100 | |
| }, | |
| { | |
| "epoch": 0.3676359801114962, | |
| "grad_norm": 0.0235595703125, | |
| "learning_rate": 0.0015136515227360855, | |
| "loss": 0.1974, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 0.3706493897845412, | |
| "grad_norm": 0.029052734375, | |
| "learning_rate": 0.0015058298639431193, | |
| "loss": 0.1974, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 0.37366279945758624, | |
| "grad_norm": 0.0245361328125, | |
| "learning_rate": 0.0014979695387917036, | |
| "loss": 0.1924, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 0.3766762091306313, | |
| "grad_norm": 0.024169921875, | |
| "learning_rate": 0.0014900712961920999, | |
| "loss": 0.1925, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 0.37968961880367635, | |
| "grad_norm": 0.021484375, | |
| "learning_rate": 0.0014821358886672414, | |
| "loss": 0.1935, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 0.3827030284767214, | |
| "grad_norm": 0.0252685546875, | |
| "learning_rate": 0.0014741640722810332, | |
| "loss": 0.1925, | |
| "step": 12700 | |
| }, | |
| { | |
| "epoch": 0.38571643814976647, | |
| "grad_norm": 0.024658203125, | |
| "learning_rate": 0.0014661566065663168, | |
| "loss": 0.1936, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 0.3887298478228115, | |
| "grad_norm": 0.0263671875, | |
| "learning_rate": 0.0014581142544525052, | |
| "loss": 0.1928, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 0.3917432574958566, | |
| "grad_norm": 0.025390625, | |
| "learning_rate": 0.0014500377821928911, | |
| "loss": 0.1927, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.3947566671689016, | |
| "grad_norm": 0.0245361328125, | |
| "learning_rate": 0.0014419279592916417, | |
| "loss": 0.1931, | |
| "step": 13100 | |
| }, | |
| { | |
| "epoch": 0.39777007684194665, | |
| "grad_norm": 0.025390625, | |
| "learning_rate": 0.001433785558430481, | |
| "loss": 0.1903, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 0.4001808045803827, | |
| "eval_peoplespeech-clean-transcription_loss": 1.5337910652160645, | |
| "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, | |
| "eval_peoplespeech-clean-transcription_runtime": 14.9866, | |
| "eval_peoplespeech-clean-transcription_samples_per_second": 4.27, | |
| "eval_peoplespeech-clean-transcription_steps_per_second": 0.067, | |
| "step": 13280 | |
| }, | |
| { | |
| "epoch": 0.40078348651499174, | |
| "grad_norm": 0.0257568359375, | |
| "learning_rate": 0.0014256113553950739, | |
| "loss": 0.1917, | |
| "step": 13300 | |
| }, | |
| { | |
| "epoch": 0.40379689618803677, | |
| "grad_norm": 0.023681640625, | |
| "learning_rate": 0.0014174061290011075, | |
| "loss": 0.1893, | |
| "step": 13400 | |
| }, | |
| { | |
| "epoch": 0.4068103058610818, | |
| "grad_norm": 0.02685546875, | |
| "learning_rate": 0.0014091706610200902, | |
| "loss": 0.1909, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 0.4098237155341269, | |
| "grad_norm": 0.02294921875, | |
| "learning_rate": 0.0014009057361048665, | |
| "loss": 0.19, | |
| "step": 13600 | |
| }, | |
| { | |
| "epoch": 0.4128371252071719, | |
| "grad_norm": 0.0274658203125, | |
| "learning_rate": 0.001392612141714856, | |
| "loss": 0.1913, | |
| "step": 13700 | |
| }, | |
| { | |
| "epoch": 0.41585053488021695, | |
| "grad_norm": 0.02294921875, | |
| "learning_rate": 0.0013842906680410286, | |
| "loss": 0.1898, | |
| "step": 13800 | |
| }, | |
| { | |
| "epoch": 0.41886394455326204, | |
| "grad_norm": 0.026611328125, | |
| "learning_rate": 0.0013759421079306145, | |
| "loss": 0.1892, | |
| "step": 13900 | |
| }, | |
| { | |
| "epoch": 0.42187735422630707, | |
| "grad_norm": 0.0252685546875, | |
| "learning_rate": 0.001367567256811567, | |
| "loss": 0.1893, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.4248907638993521, | |
| "grad_norm": 0.02783203125, | |
| "learning_rate": 0.0013591669126167736, | |
| "loss": 0.1898, | |
| "step": 14100 | |
| }, | |
| { | |
| "epoch": 0.4279041735723972, | |
| "grad_norm": 0.0235595703125, | |
| "learning_rate": 0.001350741875708033, | |
| "loss": 0.1874, | |
| "step": 14200 | |
| }, | |
| { | |
| "epoch": 0.4309175832454422, | |
| "grad_norm": 0.029541015625, | |
| "learning_rate": 0.0013422929487997973, | |
| "loss": 0.188, | |
| "step": 14300 | |
| }, | |
| { | |
| "epoch": 0.43393099291848725, | |
| "grad_norm": 0.0252685546875, | |
| "learning_rate": 0.0013338209368826933, | |
| "loss": 0.1879, | |
| "step": 14400 | |
| }, | |
| { | |
| "epoch": 0.43694440259153233, | |
| "grad_norm": 0.0260009765625, | |
| "learning_rate": 0.0013253266471468235, | |
| "loss": 0.1865, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 0.43995781226457736, | |
| "grad_norm": 0.0213623046875, | |
| "learning_rate": 0.0013168108889048602, | |
| "loss": 0.1859, | |
| "step": 14600 | |
| }, | |
| { | |
| "epoch": 0.4429712219376224, | |
| "grad_norm": 0.02685546875, | |
| "learning_rate": 0.0013082744735149366, | |
| "loss": 0.1872, | |
| "step": 14700 | |
| }, | |
| { | |
| "epoch": 0.4459846316106675, | |
| "grad_norm": 0.0223388671875, | |
| "learning_rate": 0.0012997182143033416, | |
| "loss": 0.1867, | |
| "step": 14800 | |
| }, | |
| { | |
| "epoch": 0.4489980412837125, | |
| "grad_norm": 0.0279541015625, | |
| "learning_rate": 0.00129114292648703, | |
| "loss": 0.1867, | |
| "step": 14900 | |
| }, | |
| { | |
| "epoch": 0.45020340515293056, | |
| "eval_peoplespeech-clean-transcription_loss": 1.5441259145736694, | |
| "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, | |
| "eval_peoplespeech-clean-transcription_runtime": 14.9015, | |
| "eval_peoplespeech-clean-transcription_samples_per_second": 4.295, | |
| "eval_peoplespeech-clean-transcription_steps_per_second": 0.067, | |
| "step": 14940 | |
| }, | |
| { | |
| "epoch": 0.4520114509567576, | |
| "grad_norm": 0.0211181640625, | |
| "learning_rate": 0.001282549427095949, | |
| "loss": 0.1866, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.45502486062980263, | |
| "grad_norm": 0.029541015625, | |
| "learning_rate": 0.0012739385348951955, | |
| "loss": 0.1852, | |
| "step": 15100 | |
| }, | |
| { | |
| "epoch": 0.45803827030284766, | |
| "grad_norm": 0.026123046875, | |
| "learning_rate": 0.0012653110703070055, | |
| "loss": 0.1849, | |
| "step": 15200 | |
| }, | |
| { | |
| "epoch": 0.46105167997589275, | |
| "grad_norm": 0.030029296875, | |
| "learning_rate": 0.001256667855332587, | |
| "loss": 0.1846, | |
| "step": 15300 | |
| }, | |
| { | |
| "epoch": 0.4640650896489378, | |
| "grad_norm": 0.0225830078125, | |
| "learning_rate": 0.0012480097134738009, | |
| "loss": 0.185, | |
| "step": 15400 | |
| }, | |
| { | |
| "epoch": 0.4670784993219828, | |
| "grad_norm": 0.02490234375, | |
| "learning_rate": 0.0012393374696547015, | |
| "loss": 0.1861, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 0.4700919089950279, | |
| "grad_norm": 0.0223388671875, | |
| "learning_rate": 0.0012306519501429395, | |
| "loss": 0.1877, | |
| "step": 15600 | |
| }, | |
| { | |
| "epoch": 0.4731053186680729, | |
| "grad_norm": 0.0286865234375, | |
| "learning_rate": 0.0012219539824710357, | |
| "loss": 0.1859, | |
| "step": 15700 | |
| }, | |
| { | |
| "epoch": 0.47611872834111796, | |
| "grad_norm": 0.02392578125, | |
| "learning_rate": 0.0012132443953575397, | |
| "loss": 0.1847, | |
| "step": 15800 | |
| }, | |
| { | |
| "epoch": 0.47913213801416304, | |
| "grad_norm": 0.0267333984375, | |
| "learning_rate": 0.0012045240186280676, | |
| "loss": 0.1853, | |
| "step": 15900 | |
| }, | |
| { | |
| "epoch": 0.4821455476872081, | |
| "grad_norm": 0.0211181640625, | |
| "learning_rate": 0.0011957936831362426, | |
| "loss": 0.185, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 0.4851589573602531, | |
| "grad_norm": 0.029541015625, | |
| "learning_rate": 0.0011870542206845298, | |
| "loss": 0.1849, | |
| "step": 16100 | |
| }, | |
| { | |
| "epoch": 0.4881723670332982, | |
| "grad_norm": 0.0213623046875, | |
| "learning_rate": 0.001178306463944987, | |
| "loss": 0.1835, | |
| "step": 16200 | |
| }, | |
| { | |
| "epoch": 0.4911857767063432, | |
| "grad_norm": 0.026123046875, | |
| "learning_rate": 0.0011695512463799286, | |
| "loss": 0.1837, | |
| "step": 16300 | |
| }, | |
| { | |
| "epoch": 0.49419918637938826, | |
| "grad_norm": 0.0234375, | |
| "learning_rate": 0.0011607894021625166, | |
| "loss": 0.1847, | |
| "step": 16400 | |
| }, | |
| { | |
| "epoch": 0.49721259605243334, | |
| "grad_norm": 0.0286865234375, | |
| "learning_rate": 0.0011520217660972811, | |
| "loss": 0.1853, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 0.5002260057254784, | |
| "grad_norm": 0.02392578125, | |
| "learning_rate": 0.0011432491735405852, | |
| "loss": 0.1827, | |
| "step": 16600 | |
| }, | |
| { | |
| "epoch": 0.5002260057254784, | |
| "eval_peoplespeech-clean-transcription_loss": 1.5211623907089233, | |
| "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, | |
| "eval_peoplespeech-clean-transcription_runtime": 14.7246, | |
| "eval_peoplespeech-clean-transcription_samples_per_second": 4.346, | |
| "eval_peoplespeech-clean-transcription_steps_per_second": 0.068, | |
| "step": 16600 | |
| }, | |
| { | |
| "epoch": 0.5032394153985235, | |
| "grad_norm": 0.0291748046875, | |
| "learning_rate": 0.0011344724603210318, | |
| "loss": 0.1818, | |
| "step": 16700 | |
| }, | |
| { | |
| "epoch": 0.5062528250715684, | |
| "grad_norm": 0.025146484375, | |
| "learning_rate": 0.0011256924626598297, | |
| "loss": 0.1831, | |
| "step": 16800 | |
| }, | |
| { | |
| "epoch": 0.5092662347446135, | |
| "grad_norm": 0.0286865234375, | |
| "learning_rate": 0.0011169100170911204, | |
| "loss": 0.184, | |
| "step": 16900 | |
| }, | |
| { | |
| "epoch": 0.5122796444176586, | |
| "grad_norm": 0.0235595703125, | |
| "learning_rate": 0.0011081259603822747, | |
| "loss": 0.1833, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 0.5152930540907036, | |
| "grad_norm": 0.0294189453125, | |
| "learning_rate": 0.0010993411294541694, | |
| "loss": 0.1841, | |
| "step": 17100 | |
| }, | |
| { | |
| "epoch": 0.5183064637637487, | |
| "grad_norm": 0.0223388671875, | |
| "learning_rate": 0.001090556361301446, | |
| "loss": 0.1849, | |
| "step": 17200 | |
| }, | |
| { | |
| "epoch": 0.5213198734367938, | |
| "grad_norm": 0.0274658203125, | |
| "learning_rate": 0.0010817724929127646, | |
| "loss": 0.1831, | |
| "step": 17300 | |
| }, | |
| { | |
| "epoch": 0.5243332831098387, | |
| "grad_norm": 0.021484375, | |
| "learning_rate": 0.00107299036119106, | |
| "loss": 0.1822, | |
| "step": 17400 | |
| }, | |
| { | |
| "epoch": 0.5273466927828838, | |
| "grad_norm": 0.02880859375, | |
| "learning_rate": 0.0010642108028738003, | |
| "loss": 0.1819, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 0.5303601024559289, | |
| "grad_norm": 0.021728515625, | |
| "learning_rate": 0.0010554346544532672, | |
| "loss": 0.1839, | |
| "step": 17600 | |
| }, | |
| { | |
| "epoch": 0.5333735121289739, | |
| "grad_norm": 0.031494140625, | |
| "learning_rate": 0.0010466627520968577, | |
| "loss": 0.1858, | |
| "step": 17700 | |
| }, | |
| { | |
| "epoch": 0.536386921802019, | |
| "grad_norm": 0.02197265625, | |
| "learning_rate": 0.001037895931567414, | |
| "loss": 0.1837, | |
| "step": 17800 | |
| }, | |
| { | |
| "epoch": 0.539400331475064, | |
| "grad_norm": 0.0272216796875, | |
| "learning_rate": 0.0010291350281435962, | |
| "loss": 0.1819, | |
| "step": 17900 | |
| }, | |
| { | |
| "epoch": 0.5424137411481091, | |
| "grad_norm": 0.024169921875, | |
| "learning_rate": 0.0010203808765402993, | |
| "loss": 0.1835, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 0.5454271508211541, | |
| "grad_norm": 0.0272216796875, | |
| "learning_rate": 0.0010116343108291233, | |
| "loss": 0.1828, | |
| "step": 18100 | |
| }, | |
| { | |
| "epoch": 0.5484405604941992, | |
| "grad_norm": 0.0234375, | |
| "learning_rate": 0.0010028961643589044, | |
| "loss": 0.1802, | |
| "step": 18200 | |
| }, | |
| { | |
| "epoch": 0.5502486062980262, | |
| "eval_peoplespeech-clean-transcription_loss": 1.5098843574523926, | |
| "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, | |
| "eval_peoplespeech-clean-transcription_runtime": 14.7248, | |
| "eval_peoplespeech-clean-transcription_samples_per_second": 4.346, | |
| "eval_peoplespeech-clean-transcription_steps_per_second": 0.068, | |
| "step": 18260 | |
| }, | |
| { | |
| "epoch": 0.5514539701672443, | |
| "grad_norm": 0.0277099609375, | |
| "learning_rate": 0.0009941672696763173, | |
| "loss": 0.1835, | |
| "step": 18300 | |
| }, | |
| { | |
| "epoch": 0.5544673798402893, | |
| "grad_norm": 0.0255126953125, | |
| "learning_rate": 0.0009854484584465506, | |
| "loss": 0.1815, | |
| "step": 18400 | |
| }, | |
| { | |
| "epoch": 0.5574807895133344, | |
| "grad_norm": 0.03076171875, | |
| "learning_rate": 0.0009767405613740716, | |
| "loss": 0.1817, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 0.5604941991863794, | |
| "grad_norm": 0.0230712890625, | |
| "learning_rate": 0.0009680444081234734, | |
| "loss": 0.1822, | |
| "step": 18600 | |
| }, | |
| { | |
| "epoch": 0.5635076088594244, | |
| "grad_norm": 0.0299072265625, | |
| "learning_rate": 0.0009593608272404317, | |
| "loss": 0.183, | |
| "step": 18700 | |
| }, | |
| { | |
| "epoch": 0.5665210185324695, | |
| "grad_norm": 0.0213623046875, | |
| "learning_rate": 0.0009506906460727618, | |
| "loss": 0.1813, | |
| "step": 18800 | |
| }, | |
| { | |
| "epoch": 0.5695344282055146, | |
| "grad_norm": 0.0277099609375, | |
| "learning_rate": 0.0009420346906915895, | |
| "loss": 0.1826, | |
| "step": 18900 | |
| }, | |
| { | |
| "epoch": 0.5725478378785596, | |
| "grad_norm": 0.021728515625, | |
| "learning_rate": 0.0009333937858126477, | |
| "loss": 0.1799, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 0.5755612475516046, | |
| "grad_norm": 0.0303955078125, | |
| "learning_rate": 0.0009247687547176979, | |
| "loss": 0.1819, | |
| "step": 19100 | |
| }, | |
| { | |
| "epoch": 0.5785746572246497, | |
| "grad_norm": 0.021728515625, | |
| "learning_rate": 0.0009161604191760915, | |
| "loss": 0.1804, | |
| "step": 19200 | |
| }, | |
| { | |
| "epoch": 0.5815880668976947, | |
| "grad_norm": 0.028564453125, | |
| "learning_rate": 0.000907569599366473, | |
| "loss": 0.181, | |
| "step": 19300 | |
| }, | |
| { | |
| "epoch": 0.5846014765707398, | |
| "grad_norm": 0.0223388671875, | |
| "learning_rate": 0.000898997113798635, | |
| "loss": 0.1798, | |
| "step": 19400 | |
| }, | |
| { | |
| "epoch": 0.5876148862437849, | |
| "grad_norm": 0.0279541015625, | |
| "learning_rate": 0.0008904437792355364, | |
| "loss": 0.179, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 0.5906282959168299, | |
| "grad_norm": 0.02392578125, | |
| "learning_rate": 0.0008819104106154776, | |
| "loss": 0.1808, | |
| "step": 19600 | |
| }, | |
| { | |
| "epoch": 0.593641705589875, | |
| "grad_norm": 0.028076171875, | |
| "learning_rate": 0.0008733978209744609, | |
| "loss": 0.1802, | |
| "step": 19700 | |
| }, | |
| { | |
| "epoch": 0.59665511526292, | |
| "grad_norm": 0.0235595703125, | |
| "learning_rate": 0.0008649068213687225, | |
| "loss": 0.1813, | |
| "step": 19800 | |
| }, | |
| { | |
| "epoch": 0.599668524935965, | |
| "grad_norm": 0.0279541015625, | |
| "learning_rate": 0.0008564382207974612, | |
| "loss": 0.1807, | |
| "step": 19900 | |
| }, | |
| { | |
| "epoch": 0.600271206870574, | |
| "eval_peoplespeech-clean-transcription_loss": 1.5082225799560547, | |
| "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, | |
| "eval_peoplespeech-clean-transcription_runtime": 14.0089, | |
| "eval_peoplespeech-clean-transcription_samples_per_second": 4.569, | |
| "eval_peoplespeech-clean-transcription_steps_per_second": 0.071, | |
| "step": 19920 | |
| }, | |
| { | |
| "epoch": 0.6026819346090101, | |
| "grad_norm": 0.0234375, | |
| "learning_rate": 0.0008479928261257557, | |
| "loss": 0.1807, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 0.6056953442820552, | |
| "grad_norm": 0.0284423828125, | |
| "learning_rate": 0.0008395714420076905, | |
| "loss": 0.1813, | |
| "step": 20100 | |
| }, | |
| { | |
| "epoch": 0.6087087539551002, | |
| "grad_norm": 0.0223388671875, | |
| "learning_rate": 0.0008311748708096898, | |
| "loss": 0.1794, | |
| "step": 20200 | |
| }, | |
| { | |
| "epoch": 0.6117221636281452, | |
| "grad_norm": 0.033935546875, | |
| "learning_rate": 0.0008228039125340721, | |
| "loss": 0.1809, | |
| "step": 20300 | |
| }, | |
| { | |
| "epoch": 0.6147355733011903, | |
| "grad_norm": 0.0234375, | |
| "learning_rate": 0.0008144593647428254, | |
| "loss": 0.1796, | |
| "step": 20400 | |
| }, | |
| { | |
| "epoch": 0.6177489829742353, | |
| "grad_norm": 0.0281982421875, | |
| "learning_rate": 0.0008061420224816187, | |
| "loss": 0.1807, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 0.6207623926472804, | |
| "grad_norm": 0.0242919921875, | |
| "learning_rate": 0.0007978526782040547, | |
| "loss": 0.1806, | |
| "step": 20600 | |
| }, | |
| { | |
| "epoch": 0.6237758023203255, | |
| "grad_norm": 0.0308837890625, | |
| "learning_rate": 0.0007895921216961628, | |
| "loss": 0.1802, | |
| "step": 20700 | |
| }, | |
| { | |
| "epoch": 0.6267892119933705, | |
| "grad_norm": 0.0244140625, | |
| "learning_rate": 0.0007813611400011535, | |
| "loss": 0.1806, | |
| "step": 20800 | |
| }, | |
| { | |
| "epoch": 0.6298026216664155, | |
| "grad_norm": 0.0322265625, | |
| "learning_rate": 0.0007731605173444294, | |
| "loss": 0.1799, | |
| "step": 20900 | |
| }, | |
| { | |
| "epoch": 0.6328160313394606, | |
| "grad_norm": 0.0228271484375, | |
| "learning_rate": 0.0007649910350588683, | |
| "loss": 0.1797, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 0.6358294410125056, | |
| "grad_norm": 0.025146484375, | |
| "learning_rate": 0.000756853471510377, | |
| "loss": 0.1779, | |
| "step": 21100 | |
| }, | |
| { | |
| "epoch": 0.6388428506855507, | |
| "grad_norm": 0.0240478515625, | |
| "learning_rate": 0.0007487486020237337, | |
| "loss": 0.1786, | |
| "step": 21200 | |
| }, | |
| { | |
| "epoch": 0.6418562603585958, | |
| "grad_norm": 0.0302734375, | |
| "learning_rate": 0.0007406771988087153, | |
| "loss": 0.1795, | |
| "step": 21300 | |
| }, | |
| { | |
| "epoch": 0.6448696700316408, | |
| "grad_norm": 0.02392578125, | |
| "learning_rate": 0.0007326400308865245, | |
| "loss": 0.1827, | |
| "step": 21400 | |
| }, | |
| { | |
| "epoch": 0.6478830797046858, | |
| "grad_norm": 0.02978515625, | |
| "learning_rate": 0.0007246378640165184, | |
| "loss": 0.1797, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 0.6502938074431219, | |
| "eval_peoplespeech-clean-transcription_loss": 1.4861868619918823, | |
| "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, | |
| "eval_peoplespeech-clean-transcription_runtime": 14.4228, | |
| "eval_peoplespeech-clean-transcription_samples_per_second": 4.437, | |
| "eval_peoplespeech-clean-transcription_steps_per_second": 0.069, | |
| "step": 21580 | |
| }, | |
| { | |
| "epoch": 0.6508964893777309, | |
| "grad_norm": 0.0218505859375, | |
| "learning_rate": 0.0007166714606232492, | |
| "loss": 0.1812, | |
| "step": 21600 | |
| }, | |
| { | |
| "epoch": 0.6539098990507759, | |
| "grad_norm": 0.0291748046875, | |
| "learning_rate": 0.0007087415797238248, | |
| "loss": 0.1826, | |
| "step": 21700 | |
| }, | |
| { | |
| "epoch": 0.656923308723821, | |
| "grad_norm": 0.0260009765625, | |
| "learning_rate": 0.0007008489768555886, | |
| "loss": 0.18, | |
| "step": 21800 | |
| }, | |
| { | |
| "epoch": 0.6599367183968661, | |
| "grad_norm": 0.0286865234375, | |
| "learning_rate": 0.0006929944040041347, | |
| "loss": 0.1786, | |
| "step": 21900 | |
| }, | |
| { | |
| "epoch": 0.6629501280699112, | |
| "grad_norm": 0.02294921875, | |
| "learning_rate": 0.0006851786095316618, | |
| "loss": 0.1805, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 0.6659635377429561, | |
| "grad_norm": 0.0257568359375, | |
| "learning_rate": 0.000677402338105672, | |
| "loss": 0.1786, | |
| "step": 22100 | |
| }, | |
| { | |
| "epoch": 0.6689769474160012, | |
| "grad_norm": 0.024658203125, | |
| "learning_rate": 0.0006696663306280182, | |
| "loss": 0.181, | |
| "step": 22200 | |
| }, | |
| { | |
| "epoch": 0.6719903570890463, | |
| "grad_norm": 0.026123046875, | |
| "learning_rate": 0.0006619713241643147, | |
| "loss": 0.1797, | |
| "step": 22300 | |
| }, | |
| { | |
| "epoch": 0.6750037667620913, | |
| "grad_norm": 0.0263671875, | |
| "learning_rate": 0.0006543180518737122, | |
| "loss": 0.1793, | |
| "step": 22400 | |
| }, | |
| { | |
| "epoch": 0.6780171764351364, | |
| "grad_norm": 0.02685546875, | |
| "learning_rate": 0.0006467072429390431, | |
| "loss": 0.178, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 0.6810305861081815, | |
| "grad_norm": 0.02392578125, | |
| "learning_rate": 0.0006391396224973473, | |
| "loss": 0.1793, | |
| "step": 22600 | |
| }, | |
| { | |
| "epoch": 0.6840439957812264, | |
| "grad_norm": 0.026123046875, | |
| "learning_rate": 0.0006316159115707838, | |
| "loss": 0.1793, | |
| "step": 22700 | |
| }, | |
| { | |
| "epoch": 0.6870574054542715, | |
| "grad_norm": 0.0235595703125, | |
| "learning_rate": 0.0006241368269979337, | |
| "loss": 0.177, | |
| "step": 22800 | |
| }, | |
| { | |
| "epoch": 0.6900708151273166, | |
| "grad_norm": 0.028076171875, | |
| "learning_rate": 0.0006167030813654996, | |
| "loss": 0.1781, | |
| "step": 22900 | |
| }, | |
| { | |
| "epoch": 0.6930842248003616, | |
| "grad_norm": 0.02685546875, | |
| "learning_rate": 0.0006093153829404155, | |
| "loss": 0.1782, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 0.6960976344734067, | |
| "grad_norm": 0.0296630859375, | |
| "learning_rate": 0.0006019744356023627, | |
| "loss": 0.179, | |
| "step": 23100 | |
| }, | |
| { | |
| "epoch": 0.6991110441464518, | |
| "grad_norm": 0.0250244140625, | |
| "learning_rate": 0.0005946809387767075, | |
| "loss": 0.1788, | |
| "step": 23200 | |
| }, | |
| { | |
| "epoch": 0.7003164080156697, | |
| "eval_peoplespeech-clean-transcription_loss": 1.503227710723877, | |
| "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, | |
| "eval_peoplespeech-clean-transcription_runtime": 14.6551, | |
| "eval_peoplespeech-clean-transcription_samples_per_second": 4.367, | |
| "eval_peoplespeech-clean-transcription_steps_per_second": 0.068, | |
| "step": 23240 | |
| }, | |
| { | |
| "epoch": 0.7021244538194967, | |
| "grad_norm": 0.03125, | |
| "learning_rate": 0.000587435587367861, | |
| "loss": 0.1789, | |
| "step": 23300 | |
| }, | |
| { | |
| "epoch": 0.7051378634925418, | |
| "grad_norm": 0.0235595703125, | |
| "learning_rate": 0.0005802390716930713, | |
| "loss": 0.1771, | |
| "step": 23400 | |
| }, | |
| { | |
| "epoch": 0.7081512731655869, | |
| "grad_norm": 0.0311279296875, | |
| "learning_rate": 0.0005730920774166495, | |
| "loss": 0.1793, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 0.7111646828386319, | |
| "grad_norm": 0.02587890625, | |
| "learning_rate": 0.0005659952854846461, | |
| "loss": 0.1773, | |
| "step": 23600 | |
| }, | |
| { | |
| "epoch": 0.714178092511677, | |
| "grad_norm": 0.0299072265625, | |
| "learning_rate": 0.0005589493720599683, | |
| "loss": 0.1785, | |
| "step": 23700 | |
| }, | |
| { | |
| "epoch": 0.717191502184722, | |
| "grad_norm": 0.02587890625, | |
| "learning_rate": 0.0005519550084579583, | |
| "loss": 0.1776, | |
| "step": 23800 | |
| }, | |
| { | |
| "epoch": 0.720204911857767, | |
| "grad_norm": 0.0279541015625, | |
| "learning_rate": 0.0005450128610824328, | |
| "loss": 0.179, | |
| "step": 23900 | |
| }, | |
| { | |
| "epoch": 0.7232183215308121, | |
| "grad_norm": 0.0224609375, | |
| "learning_rate": 0.0005381235913621889, | |
| "loss": 0.1779, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 0.7262317312038572, | |
| "grad_norm": 0.02783203125, | |
| "learning_rate": 0.0005312878556879856, | |
| "loss": 0.1776, | |
| "step": 24100 | |
| }, | |
| { | |
| "epoch": 0.7292451408769022, | |
| "grad_norm": 0.0267333984375, | |
| "learning_rate": 0.0005245063053500047, | |
| "loss": 0.1796, | |
| "step": 24200 | |
| }, | |
| { | |
| "epoch": 0.7322585505499473, | |
| "grad_norm": 0.0272216796875, | |
| "learning_rate": 0.0005177795864757979, | |
| "loss": 0.179, | |
| "step": 24300 | |
| }, | |
| { | |
| "epoch": 0.7352719602229923, | |
| "grad_norm": 0.024169921875, | |
| "learning_rate": 0.0005111083399687246, | |
| "loss": 0.179, | |
| "step": 24400 | |
| }, | |
| { | |
| "epoch": 0.7382853698960373, | |
| "grad_norm": 0.0264892578125, | |
| "learning_rate": 0.0005044932014468884, | |
| "loss": 0.178, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 0.7412987795690824, | |
| "grad_norm": 0.0238037109375, | |
| "learning_rate": 0.0004979348011825788, | |
| "loss": 0.1797, | |
| "step": 24600 | |
| }, | |
| { | |
| "epoch": 0.7443121892421275, | |
| "grad_norm": 0.0264892578125, | |
| "learning_rate": 0.000491433764042219, | |
| "loss": 0.1793, | |
| "step": 24700 | |
| }, | |
| { | |
| "epoch": 0.7473255989151725, | |
| "grad_norm": 0.025390625, | |
| "learning_rate": 0.0004849907094268304, | |
| "loss": 0.1788, | |
| "step": 24800 | |
| }, | |
| { | |
| "epoch": 0.7503390085882176, | |
| "grad_norm": 0.02490234375, | |
| "learning_rate": 0.0004786062512130186, | |
| "loss": 0.1784, | |
| "step": 24900 | |
| }, | |
| { | |
| "epoch": 0.7503390085882176, | |
| "eval_peoplespeech-clean-transcription_loss": 1.5016210079193115, | |
| "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, | |
| "eval_peoplespeech-clean-transcription_runtime": 13.8104, | |
| "eval_peoplespeech-clean-transcription_samples_per_second": 4.634, | |
| "eval_peoplespeech-clean-transcription_steps_per_second": 0.072, | |
| "step": 24900 | |
| }, | |
| { | |
| "epoch": 0.7533524182612626, | |
| "grad_norm": 0.0234375, | |
| "learning_rate": 0.00047228099769448437, | |
| "loss": 0.1769, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 0.7563658279343076, | |
| "grad_norm": 0.0277099609375, | |
| "learning_rate": 0.00046601555152406694, | |
| "loss": 0.1781, | |
| "step": 25100 | |
| }, | |
| { | |
| "epoch": 0.7593792376073527, | |
| "grad_norm": 0.0238037109375, | |
| "learning_rate": 0.0004598105096563256, | |
| "loss": 0.1773, | |
| "step": 25200 | |
| }, | |
| { | |
| "epoch": 0.7623926472803978, | |
| "grad_norm": 0.0262451171875, | |
| "learning_rate": 0.00045366646329066243, | |
| "loss": 0.1782, | |
| "step": 25300 | |
| }, | |
| { | |
| "epoch": 0.7654060569534428, | |
| "grad_norm": 0.0242919921875, | |
| "learning_rate": 0.0004475839978149959, | |
| "loss": 0.1768, | |
| "step": 25400 | |
| }, | |
| { | |
| "epoch": 0.7684194666264879, | |
| "grad_norm": 0.0269775390625, | |
| "learning_rate": 0.00044156369274998554, | |
| "loss": 0.1776, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 0.7714328762995329, | |
| "grad_norm": 0.0262451171875, | |
| "learning_rate": 0.00043560612169381583, | |
| "loss": 0.1763, | |
| "step": 25600 | |
| }, | |
| { | |
| "epoch": 0.7744462859725779, | |
| "grad_norm": 0.0260009765625, | |
| "learning_rate": 0.00042971185226754895, | |
| "loss": 0.1775, | |
| "step": 25700 | |
| }, | |
| { | |
| "epoch": 0.777459695645623, | |
| "grad_norm": 0.0252685546875, | |
| "learning_rate": 0.00042388144606103926, | |
| "loss": 0.1791, | |
| "step": 25800 | |
| }, | |
| { | |
| "epoch": 0.7804731053186681, | |
| "grad_norm": 0.0277099609375, | |
| "learning_rate": 0.00041811545857942936, | |
| "loss": 0.1802, | |
| "step": 25900 | |
| }, | |
| { | |
| "epoch": 0.7834865149917132, | |
| "grad_norm": 0.02392578125, | |
| "learning_rate": 0.00041241443919022124, | |
| "loss": 0.1768, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 0.7864999246647582, | |
| "grad_norm": 0.02978515625, | |
| "learning_rate": 0.0004067789310709359, | |
| "loss": 0.181, | |
| "step": 26100 | |
| }, | |
| { | |
| "epoch": 0.7895133343378032, | |
| "grad_norm": 0.02490234375, | |
| "learning_rate": 0.0004012094711573591, | |
| "loss": 0.1794, | |
| "step": 26200 | |
| }, | |
| { | |
| "epoch": 0.7925267440108483, | |
| "grad_norm": 0.027587890625, | |
| "learning_rate": 0.0003957065900923845, | |
| "loss": 0.1784, | |
| "step": 26300 | |
| }, | |
| { | |
| "epoch": 0.7955401536838933, | |
| "grad_norm": 0.0245361328125, | |
| "learning_rate": 0.00039027081217545554, | |
| "loss": 0.1777, | |
| "step": 26400 | |
| }, | |
| { | |
| "epoch": 0.7985535633569384, | |
| "grad_norm": 0.02880859375, | |
| "learning_rate": 0.0003849026553126118, | |
| "loss": 0.1762, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 0.8003616091607654, | |
| "eval_peoplespeech-clean-transcription_loss": 1.4861080646514893, | |
| "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, | |
| "eval_peoplespeech-clean-transcription_runtime": 13.7251, | |
| "eval_peoplespeech-clean-transcription_samples_per_second": 4.663, | |
| "eval_peoplespeech-clean-transcription_steps_per_second": 0.073, | |
| "step": 26560 | |
| }, | |
| { | |
| "epoch": 0.8015669730299835, | |
| "grad_norm": 0.0257568359375, | |
| "learning_rate": 0.0003796026309671429, | |
| "loss": 0.1792, | |
| "step": 26600 | |
| }, | |
| { | |
| "epoch": 0.8045803827030285, | |
| "grad_norm": 0.0267333984375, | |
| "learning_rate": 0.0003743712441108592, | |
| "loss": 0.1795, | |
| "step": 26700 | |
| }, | |
| { | |
| "epoch": 0.8075937923760735, | |
| "grad_norm": 0.0250244140625, | |
| "learning_rate": 0.00036920899317597976, | |
| "loss": 0.1764, | |
| "step": 26800 | |
| }, | |
| { | |
| "epoch": 0.8106072020491186, | |
| "grad_norm": 0.0269775390625, | |
| "learning_rate": 0.00036411637000764133, | |
| "loss": 0.18, | |
| "step": 26900 | |
| }, | |
| { | |
| "epoch": 0.8136206117221636, | |
| "grad_norm": 0.0240478515625, | |
| "learning_rate": 0.00035909385981703777, | |
| "loss": 0.1783, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 0.8166340213952087, | |
| "grad_norm": 0.02587890625, | |
| "learning_rate": 0.0003541419411351909, | |
| "loss": 0.1801, | |
| "step": 27100 | |
| }, | |
| { | |
| "epoch": 0.8196474310682538, | |
| "grad_norm": 0.02734375, | |
| "learning_rate": 0.0003492610857673564, | |
| "loss": 0.1781, | |
| "step": 27200 | |
| }, | |
| { | |
| "epoch": 0.8226608407412987, | |
| "grad_norm": 0.0263671875, | |
| "learning_rate": 0.000344451758748072, | |
| "loss": 0.1772, | |
| "step": 27300 | |
| }, | |
| { | |
| "epoch": 0.8256742504143438, | |
| "grad_norm": 0.0279541015625, | |
| "learning_rate": 0.00033971441829685036, | |
| "loss": 0.1762, | |
| "step": 27400 | |
| }, | |
| { | |
| "epoch": 0.8286876600873889, | |
| "grad_norm": 0.0274658203125, | |
| "learning_rate": 0.0003350495157745207, | |
| "loss": 0.1785, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 0.8317010697604339, | |
| "grad_norm": 0.0245361328125, | |
| "learning_rate": 0.00033045749564022497, | |
| "loss": 0.1778, | |
| "step": 27600 | |
| }, | |
| { | |
| "epoch": 0.834714479433479, | |
| "grad_norm": 0.0289306640625, | |
| "learning_rate": 0.00032593879540907076, | |
| "loss": 0.1773, | |
| "step": 27700 | |
| }, | |
| { | |
| "epoch": 0.8377278891065241, | |
| "grad_norm": 0.022705078125, | |
| "learning_rate": 0.0003214938456104454, | |
| "loss": 0.1785, | |
| "step": 27800 | |
| }, | |
| { | |
| "epoch": 0.840741298779569, | |
| "grad_norm": 0.0274658203125, | |
| "learning_rate": 0.0003171230697469978, | |
| "loss": 0.179, | |
| "step": 27900 | |
| }, | |
| { | |
| "epoch": 0.8437547084526141, | |
| "grad_norm": 0.026611328125, | |
| "learning_rate": 0.00031282688425428686, | |
| "loss": 0.1778, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 0.8467681181256592, | |
| "grad_norm": 0.0260009765625, | |
| "learning_rate": 0.0003086056984611053, | |
| "loss": 0.1782, | |
| "step": 28100 | |
| }, | |
| { | |
| "epoch": 0.8497815277987042, | |
| "grad_norm": 0.0257568359375, | |
| "learning_rate": 0.00030445991455047927, | |
| "loss": 0.1764, | |
| "step": 28200 | |
| }, | |
| { | |
| "epoch": 0.8503842097333132, | |
| "eval_peoplespeech-clean-transcription_loss": 1.485144853591919, | |
| "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, | |
| "eval_peoplespeech-clean-transcription_runtime": 13.3295, | |
| "eval_peoplespeech-clean-transcription_samples_per_second": 4.801, | |
| "eval_peoplespeech-clean-transcription_steps_per_second": 0.075, | |
| "step": 28220 | |
| }, | |
| { | |
| "epoch": 0.8527949374717493, | |
| "grad_norm": 0.0279541015625, | |
| "learning_rate": 0.000300389927521351, | |
| "loss": 0.178, | |
| "step": 28300 | |
| }, | |
| { | |
| "epoch": 0.8558083471447944, | |
| "grad_norm": 0.0255126953125, | |
| "learning_rate": 0.0002963961251509423, | |
| "loss": 0.1794, | |
| "step": 28400 | |
| }, | |
| { | |
| "epoch": 0.8588217568178393, | |
| "grad_norm": 0.0289306640625, | |
| "learning_rate": 0.0002924788879578099, | |
| "loss": 0.1774, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 0.8618351664908844, | |
| "grad_norm": 0.02734375, | |
| "learning_rate": 0.00028863858916559, | |
| "loss": 0.178, | |
| "step": 28600 | |
| }, | |
| { | |
| "epoch": 0.8648485761639295, | |
| "grad_norm": 0.0247802734375, | |
| "learning_rate": 0.0002848755946674383, | |
| "loss": 0.1756, | |
| "step": 28700 | |
| }, | |
| { | |
| "epoch": 0.8678619858369745, | |
| "grad_norm": 0.0264892578125, | |
| "learning_rate": 0.00028119026299116905, | |
| "loss": 0.1783, | |
| "step": 28800 | |
| }, | |
| { | |
| "epoch": 0.8708753955100196, | |
| "grad_norm": 0.029052734375, | |
| "learning_rate": 0.0002775829452650956, | |
| "loss": 0.1789, | |
| "step": 28900 | |
| }, | |
| { | |
| "epoch": 0.8738888051830647, | |
| "grad_norm": 0.0262451171875, | |
| "learning_rate": 0.00027405398518457575, | |
| "loss": 0.1763, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 0.8769022148561096, | |
| "grad_norm": 0.0281982421875, | |
| "learning_rate": 0.0002706037189792652, | |
| "loss": 0.1771, | |
| "step": 29100 | |
| }, | |
| { | |
| "epoch": 0.8799156245291547, | |
| "grad_norm": 0.0263671875, | |
| "learning_rate": 0.00026723247538108254, | |
| "loss": 0.1795, | |
| "step": 29200 | |
| }, | |
| { | |
| "epoch": 0.8829290342021998, | |
| "grad_norm": 0.0277099609375, | |
| "learning_rate": 0.00026394057559288856, | |
| "loss": 0.1779, | |
| "step": 29300 | |
| }, | |
| { | |
| "epoch": 0.8859424438752448, | |
| "grad_norm": 0.02685546875, | |
| "learning_rate": 0.00026072833325788375, | |
| "loss": 0.1779, | |
| "step": 29400 | |
| }, | |
| { | |
| "epoch": 0.8889558535482899, | |
| "grad_norm": 0.02587890625, | |
| "learning_rate": 0.0002575960544297239, | |
| "loss": 0.1783, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 0.891969263221335, | |
| "grad_norm": 0.0235595703125, | |
| "learning_rate": 0.0002545440375433609, | |
| "loss": 0.1766, | |
| "step": 29600 | |
| }, | |
| { | |
| "epoch": 0.8949826728943799, | |
| "grad_norm": 0.030517578125, | |
| "learning_rate": 0.0002515725733866084, | |
| "loss": 0.1782, | |
| "step": 29700 | |
| }, | |
| { | |
| "epoch": 0.897996082567425, | |
| "grad_norm": 0.024658203125, | |
| "learning_rate": 0.000248681945072437, | |
| "loss": 0.1769, | |
| "step": 29800 | |
| }, | |
| { | |
| "epoch": 0.9004068103058611, | |
| "eval_peoplespeech-clean-transcription_loss": 1.481724500656128, | |
| "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, | |
| "eval_peoplespeech-clean-transcription_runtime": 13.9492, | |
| "eval_peoplespeech-clean-transcription_samples_per_second": 4.588, | |
| "eval_peoplespeech-clean-transcription_steps_per_second": 0.072, | |
| "step": 29880 | |
| }, | |
| { | |
| "epoch": 0.9010094922404701, | |
| "grad_norm": 0.0272216796875, | |
| "learning_rate": 0.000245872428011999, | |
| "loss": 0.1763, | |
| "step": 29900 | |
| }, | |
| { | |
| "epoch": 0.9040229019135152, | |
| "grad_norm": 0.0255126953125, | |
| "learning_rate": 0.00024314428988838856, | |
| "loss": 0.1764, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 0.9070363115865602, | |
| "grad_norm": 0.0274658203125, | |
| "learning_rate": 0.000240497790631138, | |
| "loss": 0.1754, | |
| "step": 30100 | |
| }, | |
| { | |
| "epoch": 0.9100497212596053, | |
| "grad_norm": 0.02587890625, | |
| "learning_rate": 0.00023793318239145138, | |
| "loss": 0.1781, | |
| "step": 30200 | |
| }, | |
| { | |
| "epoch": 0.9130631309326503, | |
| "grad_norm": 0.0299072265625, | |
| "learning_rate": 0.00023545070951818084, | |
| "loss": 0.1776, | |
| "step": 30300 | |
| }, | |
| { | |
| "epoch": 0.9160765406056953, | |
| "grad_norm": 0.02392578125, | |
| "learning_rate": 0.00023305060853454597, | |
| "loss": 0.1789, | |
| "step": 30400 | |
| }, | |
| { | |
| "epoch": 0.9190899502787404, | |
| "grad_norm": 0.0277099609375, | |
| "learning_rate": 0.00023073310811559807, | |
| "loss": 0.1793, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 0.9221033599517855, | |
| "grad_norm": 0.02685546875, | |
| "learning_rate": 0.00022849842906643277, | |
| "loss": 0.1793, | |
| "step": 30600 | |
| }, | |
| { | |
| "epoch": 0.9251167696248305, | |
| "grad_norm": 0.024169921875, | |
| "learning_rate": 0.00022634678430115206, | |
| "loss": 0.177, | |
| "step": 30700 | |
| }, | |
| { | |
| "epoch": 0.9281301792978756, | |
| "grad_norm": 0.0247802734375, | |
| "learning_rate": 0.0002242783788225793, | |
| "loss": 0.1788, | |
| "step": 30800 | |
| }, | |
| { | |
| "epoch": 0.9311435889709206, | |
| "grad_norm": 0.0257568359375, | |
| "learning_rate": 0.00022229340970272572, | |
| "loss": 0.1777, | |
| "step": 30900 | |
| }, | |
| { | |
| "epoch": 0.9341569986439656, | |
| "grad_norm": 0.02392578125, | |
| "learning_rate": 0.00022039206606401526, | |
| "loss": 0.1776, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 0.9371704083170107, | |
| "grad_norm": 0.02880859375, | |
| "learning_rate": 0.0002185745290612646, | |
| "loss": 0.1774, | |
| "step": 31100 | |
| }, | |
| { | |
| "epoch": 0.9401838179900558, | |
| "grad_norm": 0.0286865234375, | |
| "learning_rate": 0.00021684097186442405, | |
| "loss": 0.1786, | |
| "step": 31200 | |
| }, | |
| { | |
| "epoch": 0.9431972276631008, | |
| "grad_norm": 0.026123046875, | |
| "learning_rate": 0.0002151915596420774, | |
| "loss": 0.1777, | |
| "step": 31300 | |
| }, | |
| { | |
| "epoch": 0.9462106373361459, | |
| "grad_norm": 0.0252685546875, | |
| "learning_rate": 0.0002136264495457057, | |
| "loss": 0.1781, | |
| "step": 31400 | |
| }, | |
| { | |
| "epoch": 0.9492240470091909, | |
| "grad_norm": 0.028076171875, | |
| "learning_rate": 0.00021214579069471447, | |
| "loss": 0.1772, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 0.9504294108784089, | |
| "eval_peoplespeech-clean-transcription_loss": 1.4812726974487305, | |
| "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, | |
| "eval_peoplespeech-clean-transcription_runtime": 13.7554, | |
| "eval_peoplespeech-clean-transcription_samples_per_second": 4.653, | |
| "eval_peoplespeech-clean-transcription_steps_per_second": 0.073, | |
| "step": 31540 | |
| }, | |
| { | |
| "epoch": 0.9522374566822359, | |
| "grad_norm": 0.0230712890625, | |
| "learning_rate": 0.0002107497241622257, | |
| "loss": 0.1767, | |
| "step": 31600 | |
| }, | |
| { | |
| "epoch": 0.955250866355281, | |
| "grad_norm": 0.02685546875, | |
| "learning_rate": 0.00020943838296163657, | |
| "loss": 0.1788, | |
| "step": 31700 | |
| }, | |
| { | |
| "epoch": 0.9582642760283261, | |
| "grad_norm": 0.026123046875, | |
| "learning_rate": 0.00020821189203394706, | |
| "loss": 0.1783, | |
| "step": 31800 | |
| }, | |
| { | |
| "epoch": 0.9612776857013711, | |
| "grad_norm": 0.0238037109375, | |
| "learning_rate": 0.00020707036823585488, | |
| "loss": 0.1753, | |
| "step": 31900 | |
| }, | |
| { | |
| "epoch": 0.9642910953744162, | |
| "grad_norm": 0.027587890625, | |
| "learning_rate": 0.00020601392032862275, | |
| "loss": 0.1752, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 0.9673045050474612, | |
| "grad_norm": 0.0255126953125, | |
| "learning_rate": 0.00020504264896771505, | |
| "loss": 0.177, | |
| "step": 32100 | |
| }, | |
| { | |
| "epoch": 0.9703179147205062, | |
| "grad_norm": 0.025634765625, | |
| "learning_rate": 0.00020415664669320817, | |
| "loss": 0.1786, | |
| "step": 32200 | |
| }, | |
| { | |
| "epoch": 0.9733313243935513, | |
| "grad_norm": 0.0244140625, | |
| "learning_rate": 0.00020335599792097327, | |
| "loss": 0.1764, | |
| "step": 32300 | |
| }, | |
| { | |
| "epoch": 0.9763447340665964, | |
| "grad_norm": 0.0240478515625, | |
| "learning_rate": 0.00020264077893463362, | |
| "loss": 0.1781, | |
| "step": 32400 | |
| }, | |
| { | |
| "epoch": 0.9793581437396414, | |
| "grad_norm": 0.0277099609375, | |
| "learning_rate": 0.00020201105787829627, | |
| "loss": 0.1761, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 0.9823715534126864, | |
| "grad_norm": 0.027099609375, | |
| "learning_rate": 0.00020146689475005947, | |
| "loss": 0.1786, | |
| "step": 32600 | |
| }, | |
| { | |
| "epoch": 0.9853849630857315, | |
| "grad_norm": 0.0242919921875, | |
| "learning_rate": 0.00020100834139629646, | |
| "loss": 0.1776, | |
| "step": 32700 | |
| }, | |
| { | |
| "epoch": 0.9883983727587765, | |
| "grad_norm": 0.02685546875, | |
| "learning_rate": 0.00020063544150671555, | |
| "loss": 0.1765, | |
| "step": 32800 | |
| }, | |
| { | |
| "epoch": 0.9914117824318216, | |
| "grad_norm": 0.026123046875, | |
| "learning_rate": 0.00020034823061019724, | |
| "loss": 0.1768, | |
| "step": 32900 | |
| }, | |
| { | |
| "epoch": 0.9944251921048667, | |
| "grad_norm": 0.026611328125, | |
| "learning_rate": 0.00020014673607140958, | |
| "loss": 0.1771, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 0.9974386017779117, | |
| "grad_norm": 0.02490234375, | |
| "learning_rate": 0.00020003097708820057, | |
| "loss": 0.1781, | |
| "step": 33100 | |
| } | |
| ], | |
| "logging_steps": 100, | |
| "max_steps": 33185, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 8297, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 5.732431512509769e+19, | |
| "train_batch_size": 672, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |