{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 1925, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0, "step": 0, "train/loss_ctc": 0.4116014242172241, "train/loss_error": 0.332448273897171, "train/loss_total": 0.3482789397239685 }, { "epoch": 0.0025974025974025974, "step": 1, "train/loss_ctc": 0.4357004165649414, "train/loss_error": 0.3162831664085388, "train/loss_total": 0.3401666283607483 }, { "epoch": 0.005194805194805195, "step": 2, "train/loss_ctc": 0.28240150213241577, "train/loss_error": 0.28672486543655396, "train/loss_total": 0.28586018085479736 }, { "epoch": 0.007792207792207792, "step": 3, "train/loss_ctc": 0.3549893796443939, "train/loss_error": 0.36616504192352295, "train/loss_total": 0.3639299273490906 }, { "epoch": 0.01038961038961039, "step": 4, "train/loss_ctc": 0.5775766968727112, "train/loss_error": 0.3967369496822357, "train/loss_total": 0.4329049289226532 }, { "epoch": 0.012987012987012988, "step": 5, "train/loss_ctc": 0.6276828646659851, "train/loss_error": 0.41341862082481384, "train/loss_total": 0.4562714695930481 }, { "epoch": 0.015584415584415584, "step": 6, "train/loss_ctc": 0.5604949593544006, "train/loss_error": 0.4100603759288788, "train/loss_total": 0.4401473104953766 }, { "epoch": 0.01818181818181818, "step": 7, "train/loss_ctc": 0.40248942375183105, "train/loss_error": 0.4141484498977661, "train/loss_total": 0.41181665658950806 }, { "epoch": 0.02077922077922078, "step": 8, "train/loss_ctc": 0.4439212381839752, "train/loss_error": 0.3577412962913513, "train/loss_total": 0.3749772906303406 }, { "epoch": 0.023376623376623377, "step": 9, "train/loss_ctc": 0.6861023306846619, "train/loss_error": 0.434171199798584, "train/loss_total": 0.48455744981765747 }, { "epoch": 0.025974025974025976, "grad_norm": 57599.13671875, "learning_rate": 2.985974025974026e-05, "loss": 0.3939, "step": 10 }, { "epoch": 0.025974025974025976, "step": 10, "train/loss_ctc": 0.4697543680667877, "train/loss_error": 0.3732757270336151, "train/loss_total": 0.39257144927978516 }, { "epoch": 0.02857142857142857, "step": 11, "train/loss_ctc": 0.6932277679443359, "train/loss_error": 0.3513205349445343, "train/loss_total": 0.4197019934654236 }, { "epoch": 0.03116883116883117, "step": 12, "train/loss_ctc": 0.6727705001831055, "train/loss_error": 0.34733328223228455, "train/loss_total": 0.41242074966430664 }, { "epoch": 0.033766233766233764, "step": 13, "train/loss_ctc": 0.6456692218780518, "train/loss_error": 0.38286638259887695, "train/loss_total": 0.4354269504547119 }, { "epoch": 0.03636363636363636, "step": 14, "train/loss_ctc": 0.3389711380004883, "train/loss_error": 0.36149710416793823, "train/loss_total": 0.3569919168949127 }, { "epoch": 0.03896103896103896, "step": 15, "train/loss_ctc": 0.2877556085586548, "train/loss_error": 0.2379094660282135, "train/loss_total": 0.24787870049476624 }, { "epoch": 0.04155844155844156, "step": 16, "train/loss_ctc": 0.450775682926178, "train/loss_error": 0.37737318873405457, "train/loss_total": 0.3920537233352661 }, { "epoch": 0.04415584415584416, "step": 17, "train/loss_ctc": 0.6092091798782349, "train/loss_error": 0.39170917868614197, "train/loss_total": 0.435209184885025 }, { "epoch": 0.046753246753246755, "step": 18, "train/loss_ctc": 0.31641823053359985, "train/loss_error": 0.28537076711654663, "train/loss_total": 0.2915802597999573 }, { "epoch": 0.04935064935064935, "step": 19, "train/loss_ctc": 0.38181692361831665, "train/loss_error": 0.31530845165252686, "train/loss_total": 0.3286101520061493 }, { "epoch": 0.05194805194805195, "grad_norm": 61361.9765625, "learning_rate": 2.9703896103896104e-05, "loss": 0.3712, "step": 20 }, { "epoch": 0.05194805194805195, "step": 20, "train/loss_ctc": 0.5233941674232483, "train/loss_error": 0.4615887701511383, "train/loss_total": 0.4739498496055603 }, { "epoch": 0.05454545454545454, "step": 21, "train/loss_ctc": 0.5830339789390564, "train/loss_error": 0.36400675773620605, "train/loss_total": 0.4078122079372406 }, { "epoch": 0.05714285714285714, "step": 22, "train/loss_ctc": 0.5160292387008667, "train/loss_error": 0.3902731239795685, "train/loss_total": 0.4154243767261505 }, { "epoch": 0.05974025974025974, "step": 23, "train/loss_ctc": 0.9006620645523071, "train/loss_error": 0.4078509211540222, "train/loss_total": 0.5064131617546082 }, { "epoch": 0.06233766233766234, "step": 24, "train/loss_ctc": 0.42039981484413147, "train/loss_error": 0.3121728301048279, "train/loss_total": 0.3338182270526886 }, { "epoch": 0.06493506493506493, "step": 25, "train/loss_ctc": 0.5100263357162476, "train/loss_error": 0.399075448513031, "train/loss_total": 0.4212656319141388 }, { "epoch": 0.06753246753246753, "step": 26, "train/loss_ctc": 0.7521955966949463, "train/loss_error": 0.41407957673072815, "train/loss_total": 0.4817028045654297 }, { "epoch": 0.07012987012987013, "step": 27, "train/loss_ctc": 0.4105278253555298, "train/loss_error": 0.3359241485595703, "train/loss_total": 0.3508448898792267 }, { "epoch": 0.07272727272727272, "step": 28, "train/loss_ctc": 0.45979321002960205, "train/loss_error": 0.36397266387939453, "train/loss_total": 0.3831367790699005 }, { "epoch": 0.07532467532467532, "step": 29, "train/loss_ctc": 0.5811054706573486, "train/loss_error": 0.3652283549308777, "train/loss_total": 0.40840378403663635 }, { "epoch": 0.07792207792207792, "grad_norm": 48431.34375, "learning_rate": 2.954805194805195e-05, "loss": 0.4183, "step": 30 }, { "epoch": 0.07792207792207792, "step": 30, "train/loss_ctc": 0.797727644443512, "train/loss_error": 0.41290122270584106, "train/loss_total": 0.4898664951324463 }, { "epoch": 0.08051948051948052, "step": 31, "train/loss_ctc": 0.37050336599349976, "train/loss_error": 0.3306114375591278, "train/loss_total": 0.3385898172855377 }, { "epoch": 0.08311688311688312, "step": 32, "train/loss_ctc": 0.4473273754119873, "train/loss_error": 0.442990779876709, "train/loss_total": 0.4438580870628357 }, { "epoch": 0.08571428571428572, "step": 33, "train/loss_ctc": 0.6308501958847046, "train/loss_error": 0.40553924441337585, "train/loss_total": 0.4506014287471771 }, { "epoch": 0.08831168831168831, "step": 34, "train/loss_ctc": 0.2426990121603012, "train/loss_error": 0.3679273724555969, "train/loss_total": 0.3428816795349121 }, { "epoch": 0.09090909090909091, "step": 35, "train/loss_ctc": 0.3917582035064697, "train/loss_error": 0.3398832678794861, "train/loss_total": 0.3502582609653473 }, { "epoch": 0.09350649350649351, "step": 36, "train/loss_ctc": 0.459766685962677, "train/loss_error": 0.330552875995636, "train/loss_total": 0.3563956320285797 }, { "epoch": 0.09610389610389611, "step": 37, "train/loss_ctc": 0.3685152232646942, "train/loss_error": 0.35461390018463135, "train/loss_total": 0.35739418864250183 }, { "epoch": 0.0987012987012987, "step": 38, "train/loss_ctc": 0.5375202298164368, "train/loss_error": 0.3150472640991211, "train/loss_total": 0.3595418632030487 }, { "epoch": 0.1012987012987013, "step": 39, "train/loss_ctc": 0.45801931619644165, "train/loss_error": 0.3970620930194855, "train/loss_total": 0.4092535674571991 }, { "epoch": 0.1038961038961039, "grad_norm": 41047.3125, "learning_rate": 2.939220779220779e-05, "loss": 0.3899, "step": 40 }, { "epoch": 0.1038961038961039, "step": 40, "train/loss_ctc": 0.3428259789943695, "train/loss_error": 0.28243133425712585, "train/loss_total": 0.29451027512550354 }, { "epoch": 0.10649350649350649, "step": 41, "train/loss_ctc": 0.399419903755188, "train/loss_error": 0.3434898257255554, "train/loss_total": 0.35467585921287537 }, { "epoch": 0.10909090909090909, "step": 42, "train/loss_ctc": 0.5130240321159363, "train/loss_error": 0.362753689289093, "train/loss_total": 0.3928077518939972 }, { "epoch": 0.11168831168831168, "step": 43, "train/loss_ctc": 0.40623360872268677, "train/loss_error": 0.3237414062023163, "train/loss_total": 0.34023985266685486 }, { "epoch": 0.11428571428571428, "step": 44, "train/loss_ctc": 0.5207386612892151, "train/loss_error": 0.3806527256965637, "train/loss_total": 0.40866991877555847 }, { "epoch": 0.11688311688311688, "step": 45, "train/loss_ctc": 0.6864684820175171, "train/loss_error": 0.3309851884841919, "train/loss_total": 0.40208184719085693 }, { "epoch": 0.11948051948051948, "step": 46, "train/loss_ctc": 0.32172420620918274, "train/loss_error": 0.23394574224948883, "train/loss_total": 0.2515014410018921 }, { "epoch": 0.12207792207792208, "step": 47, "train/loss_ctc": 0.6153249740600586, "train/loss_error": 0.38875019550323486, "train/loss_total": 0.43406516313552856 }, { "epoch": 0.12467532467532468, "step": 48, "train/loss_ctc": 0.48542410135269165, "train/loss_error": 0.3428468704223633, "train/loss_total": 0.3713623285293579 }, { "epoch": 0.12727272727272726, "step": 49, "train/loss_ctc": 0.4493388831615448, "train/loss_error": 0.3907068073749542, "train/loss_total": 0.40243321657180786 }, { "epoch": 0.12987012987012986, "grad_norm": 49088.109375, "learning_rate": 2.9236363636363635e-05, "loss": 0.3652, "step": 50 }, { "epoch": 0.12987012987012986, "step": 50, "train/loss_ctc": 0.3979894816875458, "train/loss_error": 0.4049939215183258, "train/loss_total": 0.4035930335521698 }, { "epoch": 0.13246753246753246, "step": 51, "train/loss_ctc": 0.3446880578994751, "train/loss_error": 0.328390896320343, "train/loss_total": 0.3316503167152405 }, { "epoch": 0.13506493506493505, "step": 52, "train/loss_ctc": 0.527186930179596, "train/loss_error": 0.3816911280155182, "train/loss_total": 0.4107902944087982 }, { "epoch": 0.13766233766233765, "step": 53, "train/loss_ctc": 0.46613389253616333, "train/loss_error": 0.2890409827232361, "train/loss_total": 0.32445958256721497 }, { "epoch": 0.14025974025974025, "step": 54, "train/loss_ctc": 0.4377257823944092, "train/loss_error": 0.39849117398262024, "train/loss_total": 0.406338095664978 }, { "epoch": 0.14285714285714285, "step": 55, "train/loss_ctc": 0.6229503154754639, "train/loss_error": 0.3944684863090515, "train/loss_total": 0.44016486406326294 }, { "epoch": 0.14545454545454545, "step": 56, "train/loss_ctc": 0.4153672158718109, "train/loss_error": 0.38049399852752686, "train/loss_total": 0.3874686360359192 }, { "epoch": 0.14805194805194805, "step": 57, "train/loss_ctc": 0.6522865295410156, "train/loss_error": 0.38088324666023254, "train/loss_total": 0.4351639151573181 }, { "epoch": 0.15064935064935064, "step": 58, "train/loss_ctc": 0.44472700357437134, "train/loss_error": 0.41176894307136536, "train/loss_total": 0.4183605909347534 }, { "epoch": 0.15324675324675324, "step": 59, "train/loss_ctc": 0.3114878535270691, "train/loss_error": 0.32288768887519836, "train/loss_total": 0.3206077516078949 }, { "epoch": 0.15584415584415584, "grad_norm": 36141.2265625, "learning_rate": 2.908051948051948e-05, "loss": 0.3879, "step": 60 }, { "epoch": 0.15584415584415584, "step": 60, "train/loss_ctc": 0.3701961934566498, "train/loss_error": 0.3479975759983063, "train/loss_total": 0.3524373173713684 }, { "epoch": 0.15844155844155844, "step": 61, "train/loss_ctc": 0.278195321559906, "train/loss_error": 0.307664692401886, "train/loss_total": 0.3017708361148834 }, { "epoch": 0.16103896103896104, "step": 62, "train/loss_ctc": 0.39014941453933716, "train/loss_error": 0.33666208386421204, "train/loss_total": 0.3473595380783081 }, { "epoch": 0.16363636363636364, "step": 63, "train/loss_ctc": 0.3652295768260956, "train/loss_error": 0.4002580940723419, "train/loss_total": 0.3932524025440216 }, { "epoch": 0.16623376623376623, "step": 64, "train/loss_ctc": 0.24218463897705078, "train/loss_error": 0.3218824565410614, "train/loss_total": 0.30594292283058167 }, { "epoch": 0.16883116883116883, "step": 65, "train/loss_ctc": 0.29601722955703735, "train/loss_error": 0.2907305657863617, "train/loss_total": 0.29178792238235474 }, { "epoch": 0.17142857142857143, "step": 66, "train/loss_ctc": 0.38351738452911377, "train/loss_error": 0.27062761783599854, "train/loss_total": 0.293205589056015 }, { "epoch": 0.17402597402597403, "step": 67, "train/loss_ctc": 0.3898458480834961, "train/loss_error": 0.39703795313835144, "train/loss_total": 0.3955995440483093 }, { "epoch": 0.17662337662337663, "step": 68, "train/loss_ctc": 0.33587509393692017, "train/loss_error": 0.31256863474845886, "train/loss_total": 0.3172299563884735 }, { "epoch": 0.17922077922077922, "step": 69, "train/loss_ctc": 0.38021987676620483, "train/loss_error": 0.35395875573158264, "train/loss_total": 0.3592109680175781 }, { "epoch": 0.18181818181818182, "grad_norm": 43420.20703125, "learning_rate": 2.8924675324675328e-05, "loss": 0.3358, "step": 70 }, { "epoch": 0.18181818181818182, "step": 70, "train/loss_ctc": 0.33189529180526733, "train/loss_error": 0.32725992798805237, "train/loss_total": 0.3281870186328888 }, { "epoch": 0.18441558441558442, "step": 71, "train/loss_ctc": 0.3390176296234131, "train/loss_error": 0.3369383215904236, "train/loss_total": 0.3373541831970215 }, { "epoch": 0.18701298701298702, "step": 72, "train/loss_ctc": 0.46008431911468506, "train/loss_error": 0.33292990922927856, "train/loss_total": 0.35836079716682434 }, { "epoch": 0.18961038961038962, "step": 73, "train/loss_ctc": 0.41684362292289734, "train/loss_error": 0.4331229329109192, "train/loss_total": 0.42986705899238586 }, { "epoch": 0.19220779220779222, "step": 74, "train/loss_ctc": 0.6347250938415527, "train/loss_error": 0.28208574652671814, "train/loss_total": 0.352613627910614 }, { "epoch": 0.19480519480519481, "step": 75, "train/loss_ctc": 0.29652082920074463, "train/loss_error": 0.3357245922088623, "train/loss_total": 0.32788386940956116 }, { "epoch": 0.1974025974025974, "step": 76, "train/loss_ctc": 0.21564115583896637, "train/loss_error": 0.2916478216648102, "train/loss_total": 0.27644649147987366 }, { "epoch": 0.2, "step": 77, "train/loss_ctc": 0.524330198764801, "train/loss_error": 0.3828580975532532, "train/loss_total": 0.41115254163742065 }, { "epoch": 0.2025974025974026, "step": 78, "train/loss_ctc": 0.3440263271331787, "train/loss_error": 0.2989775240421295, "train/loss_total": 0.3079873025417328 }, { "epoch": 0.2051948051948052, "step": 79, "train/loss_ctc": 0.8088093996047974, "train/loss_error": 0.4102171063423157, "train/loss_total": 0.48993557691574097 }, { "epoch": 0.2077922077922078, "grad_norm": 62876.8125, "learning_rate": 2.8768831168831172e-05, "loss": 0.362, "step": 80 }, { "epoch": 0.2077922077922078, "step": 80, "train/loss_ctc": 0.822921633720398, "train/loss_error": 0.34466230869293213, "train/loss_total": 0.4403141736984253 }, { "epoch": 0.21038961038961038, "step": 81, "train/loss_ctc": 0.6505895853042603, "train/loss_error": 0.38228675723075867, "train/loss_total": 0.43594732880592346 }, { "epoch": 0.21298701298701297, "step": 82, "train/loss_ctc": 0.3891351819038391, "train/loss_error": 0.40767544507980347, "train/loss_total": 0.40396741032600403 }, { "epoch": 0.21558441558441557, "step": 83, "train/loss_ctc": 0.46761423349380493, "train/loss_error": 0.32665008306503296, "train/loss_total": 0.3548429310321808 }, { "epoch": 0.21818181818181817, "step": 84, "train/loss_ctc": 0.4454967677593231, "train/loss_error": 0.34421437978744507, "train/loss_total": 0.36447086930274963 }, { "epoch": 0.22077922077922077, "step": 85, "train/loss_ctc": 0.30694109201431274, "train/loss_error": 0.33732521533966064, "train/loss_total": 0.33124840259552 }, { "epoch": 0.22337662337662337, "step": 86, "train/loss_ctc": 0.2999690771102905, "train/loss_error": 0.3078445792198181, "train/loss_total": 0.30626946687698364 }, { "epoch": 0.22597402597402597, "step": 87, "train/loss_ctc": 0.29124748706817627, "train/loss_error": 0.27525416016578674, "train/loss_total": 0.2784528136253357 }, { "epoch": 0.22857142857142856, "step": 88, "train/loss_ctc": 0.35056883096694946, "train/loss_error": 0.31894928216934204, "train/loss_total": 0.32527321577072144 }, { "epoch": 0.23116883116883116, "step": 89, "train/loss_ctc": 0.5163109302520752, "train/loss_error": 0.34144696593284607, "train/loss_total": 0.3764197528362274 }, { "epoch": 0.23376623376623376, "grad_norm": 43970.1640625, "learning_rate": 2.8612987012987014e-05, "loss": 0.3617, "step": 90 }, { "epoch": 0.23376623376623376, "step": 90, "train/loss_ctc": 0.5958362817764282, "train/loss_error": 0.41340890526771545, "train/loss_total": 0.44989439845085144 }, { "epoch": 0.23636363636363636, "step": 91, "train/loss_ctc": 0.5627836585044861, "train/loss_error": 0.39134472608566284, "train/loss_total": 0.425632506608963 }, { "epoch": 0.23896103896103896, "step": 92, "train/loss_ctc": 0.34862348437309265, "train/loss_error": 0.3793436884880066, "train/loss_total": 0.3731996715068817 }, { "epoch": 0.24155844155844156, "step": 93, "train/loss_ctc": 0.48762643337249756, "train/loss_error": 0.39299100637435913, "train/loss_total": 0.41191810369491577 }, { "epoch": 0.24415584415584415, "step": 94, "train/loss_ctc": 0.41612160205841064, "train/loss_error": 0.32320836186408997, "train/loss_total": 0.341791033744812 }, { "epoch": 0.24675324675324675, "step": 95, "train/loss_ctc": 0.3737485110759735, "train/loss_error": 0.29520383477211, "train/loss_total": 0.31091275811195374 }, { "epoch": 0.24935064935064935, "step": 96, "train/loss_ctc": 0.5075569152832031, "train/loss_error": 0.3025394082069397, "train/loss_total": 0.3435429036617279 }, { "epoch": 0.2519480519480519, "step": 97, "train/loss_ctc": 0.3919292688369751, "train/loss_error": 0.32132989168167114, "train/loss_total": 0.33544978499412537 }, { "epoch": 0.2545454545454545, "step": 98, "train/loss_ctc": 0.38454335927963257, "train/loss_error": 0.2964280843734741, "train/loss_total": 0.31405115127563477 }, { "epoch": 0.2571428571428571, "step": 99, "train/loss_ctc": 0.5836817026138306, "train/loss_error": 0.3591878414154053, "train/loss_total": 0.4040866196155548 }, { "epoch": 0.2597402597402597, "grad_norm": 43693.6640625, "learning_rate": 2.845714285714286e-05, "loss": 0.371, "step": 100 }, { "epoch": 0.2597402597402597, "step": 100, "train/loss_ctc": 0.38685762882232666, "train/loss_error": 0.3638920783996582, "train/loss_total": 0.3684852123260498 }, { "epoch": 0.2623376623376623, "step": 101, "train/loss_ctc": 0.43993502855300903, "train/loss_error": 0.38747450709342957, "train/loss_total": 0.3979666233062744 }, { "epoch": 0.2649350649350649, "step": 102, "train/loss_ctc": 0.34165990352630615, "train/loss_error": 0.33025309443473816, "train/loss_total": 0.33253446221351624 }, { "epoch": 0.2675324675324675, "step": 103, "train/loss_ctc": 0.27870988845825195, "train/loss_error": 0.3013240098953247, "train/loss_total": 0.2968011796474457 }, { "epoch": 0.2701298701298701, "step": 104, "train/loss_ctc": 0.40297096967697144, "train/loss_error": 0.3455545902252197, "train/loss_total": 0.35703790187835693 }, { "epoch": 0.2727272727272727, "step": 105, "train/loss_ctc": 0.21287783980369568, "train/loss_error": 0.2694125473499298, "train/loss_total": 0.258105605840683 }, { "epoch": 0.2753246753246753, "step": 106, "train/loss_ctc": 0.393646776676178, "train/loss_error": 0.3885370194911957, "train/loss_total": 0.38955897092819214 }, { "epoch": 0.2779220779220779, "step": 107, "train/loss_ctc": 0.33315491676330566, "train/loss_error": 0.29310253262519836, "train/loss_total": 0.3011130094528198 }, { "epoch": 0.2805194805194805, "step": 108, "train/loss_ctc": 0.48825713992118835, "train/loss_error": 0.39007624983787537, "train/loss_total": 0.40971243381500244 }, { "epoch": 0.2831168831168831, "step": 109, "train/loss_ctc": 0.45271867513656616, "train/loss_error": 0.3535824716091156, "train/loss_total": 0.3734097182750702 }, { "epoch": 0.2857142857142857, "grad_norm": 54204.41015625, "learning_rate": 2.8301298701298703e-05, "loss": 0.3485, "step": 110 }, { "epoch": 0.2857142857142857, "step": 110, "train/loss_ctc": 0.3022448718547821, "train/loss_error": 0.3544032573699951, "train/loss_total": 0.3439715802669525 }, { "epoch": 0.2883116883116883, "step": 111, "train/loss_ctc": 0.44910770654678345, "train/loss_error": 0.45525676012039185, "train/loss_total": 0.4540269672870636 }, { "epoch": 0.2909090909090909, "step": 112, "train/loss_ctc": 0.4126552641391754, "train/loss_error": 0.32306230068206787, "train/loss_total": 0.3409809172153473 }, { "epoch": 0.2935064935064935, "step": 113, "train/loss_ctc": 0.4695359766483307, "train/loss_error": 0.3421376943588257, "train/loss_total": 0.3676173686981201 }, { "epoch": 0.2961038961038961, "step": 114, "train/loss_ctc": 0.30297547578811646, "train/loss_error": 0.29868125915527344, "train/loss_total": 0.29954010248184204 }, { "epoch": 0.2987012987012987, "step": 115, "train/loss_ctc": 0.31851401925086975, "train/loss_error": 0.2443922907114029, "train/loss_total": 0.25921663641929626 }, { "epoch": 0.3012987012987013, "step": 116, "train/loss_ctc": 0.46976742148399353, "train/loss_error": 0.31301918625831604, "train/loss_total": 0.3443688452243805 }, { "epoch": 0.3038961038961039, "step": 117, "train/loss_ctc": 0.5292235612869263, "train/loss_error": 0.3999556303024292, "train/loss_total": 0.42580920457839966 }, { "epoch": 0.3064935064935065, "step": 118, "train/loss_ctc": 0.5127226710319519, "train/loss_error": 0.405835896730423, "train/loss_total": 0.42721328139305115 }, { "epoch": 0.3090909090909091, "step": 119, "train/loss_ctc": 0.5361043214797974, "train/loss_error": 0.43759867548942566, "train/loss_total": 0.4572997987270355 }, { "epoch": 0.3116883116883117, "grad_norm": 33459.99609375, "learning_rate": 2.816103896103896e-05, "loss": 0.372, "step": 120 }, { "epoch": 0.3116883116883117, "step": 120, "train/loss_ctc": 0.6951860785484314, "train/loss_error": 0.3637552559375763, "train/loss_total": 0.43004143238067627 }, { "epoch": 0.3142857142857143, "step": 121, "train/loss_ctc": 0.3513140082359314, "train/loss_error": 0.3766666352748871, "train/loss_total": 0.371596097946167 }, { "epoch": 0.3168831168831169, "step": 122, "train/loss_ctc": 0.3478519916534424, "train/loss_error": 0.4304841160774231, "train/loss_total": 0.4139576852321625 }, { "epoch": 0.3194805194805195, "step": 123, "train/loss_ctc": 0.5345361232757568, "train/loss_error": 0.41396859288215637, "train/loss_total": 0.43808209896087646 }, { "epoch": 0.3220779220779221, "step": 124, "train/loss_ctc": 0.3605356514453888, "train/loss_error": 0.2556537091732025, "train/loss_total": 0.27663010358810425 }, { "epoch": 0.3246753246753247, "step": 125, "train/loss_ctc": 0.3995080888271332, "train/loss_error": 0.2946002781391144, "train/loss_total": 0.3155818581581116 }, { "epoch": 0.32727272727272727, "step": 126, "train/loss_ctc": 0.5746212005615234, "train/loss_error": 0.31450504064559937, "train/loss_total": 0.36652830243110657 }, { "epoch": 0.32987012987012987, "step": 127, "train/loss_ctc": 0.3267635107040405, "train/loss_error": 0.32922235131263733, "train/loss_total": 0.32873058319091797 }, { "epoch": 0.33246753246753247, "step": 128, "train/loss_ctc": 0.43755388259887695, "train/loss_error": 0.41872474551200867, "train/loss_total": 0.42249059677124023 }, { "epoch": 0.33506493506493507, "step": 129, "train/loss_ctc": 0.3290424346923828, "train/loss_error": 0.24273301661014557, "train/loss_total": 0.2599949240684509 }, { "epoch": 0.33766233766233766, "grad_norm": 25083.173828125, "learning_rate": 2.8005194805194806e-05, "loss": 0.3624, "step": 130 }, { "epoch": 0.33766233766233766, "step": 130, "train/loss_ctc": 0.3149247169494629, "train/loss_error": 0.3664567470550537, "train/loss_total": 0.356150358915329 }, { "epoch": 0.34025974025974026, "step": 131, "train/loss_ctc": 0.3875763416290283, "train/loss_error": 0.24478475749492645, "train/loss_total": 0.2733430862426758 }, { "epoch": 0.34285714285714286, "step": 132, "train/loss_ctc": 0.48273617029190063, "train/loss_error": 0.3795401453971863, "train/loss_total": 0.40017935633659363 }, { "epoch": 0.34545454545454546, "step": 133, "train/loss_ctc": 0.36872756481170654, "train/loss_error": 0.3389447033405304, "train/loss_total": 0.34490129351615906 }, { "epoch": 0.34805194805194806, "step": 134, "train/loss_ctc": 0.5748626589775085, "train/loss_error": 0.3169487714767456, "train/loss_total": 0.36853155493736267 }, { "epoch": 0.35064935064935066, "step": 135, "train/loss_ctc": 0.3694392442703247, "train/loss_error": 0.3358103334903717, "train/loss_total": 0.3425361216068268 }, { "epoch": 0.35324675324675325, "step": 136, "train/loss_ctc": 0.5238616466522217, "train/loss_error": 0.3595260679721832, "train/loss_total": 0.39239320158958435 }, { "epoch": 0.35584415584415585, "step": 137, "train/loss_ctc": 0.35033726692199707, "train/loss_error": 0.405957967042923, "train/loss_total": 0.39483383297920227 }, { "epoch": 0.35844155844155845, "step": 138, "train/loss_ctc": 0.46711310744285583, "train/loss_error": 0.343797892332077, "train/loss_total": 0.3684609532356262 }, { "epoch": 0.36103896103896105, "step": 139, "train/loss_ctc": 0.422011137008667, "train/loss_error": 0.42424502968788147, "train/loss_total": 0.42379826307296753 }, { "epoch": 0.36363636363636365, "grad_norm": 34499.0625, "learning_rate": 2.784935064935065e-05, "loss": 0.3665, "step": 140 }, { "epoch": 0.36363636363636365, "step": 140, "train/loss_ctc": 0.4040990471839905, "train/loss_error": 0.34424543380737305, "train/loss_total": 0.356216162443161 }, { "epoch": 0.36623376623376624, "step": 141, "train/loss_ctc": 0.3783193826675415, "train/loss_error": 0.2844621539115906, "train/loss_total": 0.30323362350463867 }, { "epoch": 0.36883116883116884, "step": 142, "train/loss_ctc": 0.5976817607879639, "train/loss_error": 0.29766377806663513, "train/loss_total": 0.35766738653182983 }, { "epoch": 0.37142857142857144, "step": 143, "train/loss_ctc": 0.43515974283218384, "train/loss_error": 0.2875743806362152, "train/loss_total": 0.3170914649963379 }, { "epoch": 0.37402597402597404, "step": 144, "train/loss_ctc": 0.4646715223789215, "train/loss_error": 0.3784927427768707, "train/loss_total": 0.3957284986972809 }, { "epoch": 0.37662337662337664, "step": 145, "train/loss_ctc": 0.46109071373939514, "train/loss_error": 0.3137670159339905, "train/loss_total": 0.343231737613678 }, { "epoch": 0.37922077922077924, "step": 146, "train/loss_ctc": 0.5450319051742554, "train/loss_error": 0.37279266119003296, "train/loss_total": 0.40724050998687744 }, { "epoch": 0.38181818181818183, "step": 147, "train/loss_ctc": 0.36578041315078735, "train/loss_error": 0.31546351313591003, "train/loss_total": 0.3255268931388855 }, { "epoch": 0.38441558441558443, "step": 148, "train/loss_ctc": 0.299034059047699, "train/loss_error": 0.3027099370956421, "train/loss_total": 0.3019747734069824 }, { "epoch": 0.38701298701298703, "step": 149, "train/loss_ctc": 0.47114503383636475, "train/loss_error": 0.3409874737262726, "train/loss_total": 0.36701899766921997 }, { "epoch": 0.38961038961038963, "grad_norm": 21780.0546875, "learning_rate": 2.7693506493506492e-05, "loss": 0.3475, "step": 150 }, { "epoch": 0.38961038961038963, "step": 150, "train/loss_ctc": 0.30311858654022217, "train/loss_error": 0.32538825273513794, "train/loss_total": 0.32093432545661926 }, { "epoch": 0.3922077922077922, "step": 151, "train/loss_ctc": 0.2883327901363373, "train/loss_error": 0.3776102364063263, "train/loss_total": 0.3597547709941864 }, { "epoch": 0.3948051948051948, "step": 152, "train/loss_ctc": 0.5493779182434082, "train/loss_error": 0.39639025926589966, "train/loss_total": 0.42698779702186584 }, { "epoch": 0.3974025974025974, "step": 153, "train/loss_ctc": 0.45601528882980347, "train/loss_error": 0.3879132568836212, "train/loss_total": 0.40153366327285767 }, { "epoch": 0.4, "step": 154, "train/loss_ctc": 0.33605626225471497, "train/loss_error": 0.3217656910419464, "train/loss_total": 0.32462382316589355 }, { "epoch": 0.4025974025974026, "step": 155, "train/loss_ctc": 0.5990520715713501, "train/loss_error": 0.3232201933860779, "train/loss_total": 0.37838655710220337 }, { "epoch": 0.4051948051948052, "step": 156, "train/loss_ctc": 0.38615238666534424, "train/loss_error": 0.3373470604419708, "train/loss_total": 0.3471081256866455 }, { "epoch": 0.4077922077922078, "step": 157, "train/loss_ctc": 0.2281651496887207, "train/loss_error": 0.23885144293308258, "train/loss_total": 0.2367141842842102 }, { "epoch": 0.4103896103896104, "step": 158, "train/loss_ctc": 0.6120427250862122, "train/loss_error": 0.362976998090744, "train/loss_total": 0.4127901494503021 }, { "epoch": 0.412987012987013, "step": 159, "train/loss_ctc": 0.5026404857635498, "train/loss_error": 0.3861441910266876, "train/loss_total": 0.4094434380531311 }, { "epoch": 0.4155844155844156, "grad_norm": 28741.830078125, "learning_rate": 2.7537662337662337e-05, "loss": 0.3618, "step": 160 }, { "epoch": 0.4155844155844156, "step": 160, "train/loss_ctc": 0.3356563448905945, "train/loss_error": 0.3423846364021301, "train/loss_total": 0.3410390019416809 }, { "epoch": 0.41818181818181815, "step": 161, "train/loss_ctc": 0.5309107303619385, "train/loss_error": 0.4260532259941101, "train/loss_total": 0.44702473282814026 }, { "epoch": 0.42077922077922075, "step": 162, "train/loss_ctc": 0.5684621930122375, "train/loss_error": 0.43928012251853943, "train/loss_total": 0.4651165306568146 }, { "epoch": 0.42337662337662335, "step": 163, "train/loss_ctc": 0.34636783599853516, "train/loss_error": 0.3198454976081848, "train/loss_total": 0.3251499533653259 }, { "epoch": 0.42597402597402595, "step": 164, "train/loss_ctc": 0.5158129930496216, "train/loss_error": 0.3869795799255371, "train/loss_total": 0.41274625062942505 }, { "epoch": 0.42857142857142855, "step": 165, "train/loss_ctc": 0.40120089054107666, "train/loss_error": 0.35244354605674744, "train/loss_total": 0.36219504475593567 }, { "epoch": 0.43116883116883115, "step": 166, "train/loss_ctc": 0.3377457559108734, "train/loss_error": 0.3971347510814667, "train/loss_total": 0.38525694608688354 }, { "epoch": 0.43376623376623374, "step": 167, "train/loss_ctc": 0.39200446009635925, "train/loss_error": 0.3081461191177368, "train/loss_total": 0.3249177932739258 }, { "epoch": 0.43636363636363634, "step": 168, "train/loss_ctc": 0.389478474855423, "train/loss_error": 0.40895530581474304, "train/loss_total": 0.40505996346473694 }, { "epoch": 0.43896103896103894, "step": 169, "train/loss_ctc": 0.5164867639541626, "train/loss_error": 0.42359402775764465, "train/loss_total": 0.4421725869178772 }, { "epoch": 0.44155844155844154, "grad_norm": 23031.404296875, "learning_rate": 2.7381818181818182e-05, "loss": 0.3911, "step": 170 }, { "epoch": 0.44155844155844154, "step": 170, "train/loss_ctc": 0.35602739453315735, "train/loss_error": 0.2175227552652359, "train/loss_total": 0.24522368609905243 }, { "epoch": 0.44415584415584414, "step": 171, "train/loss_ctc": 0.614433228969574, "train/loss_error": 0.3743084669113159, "train/loss_total": 0.4223334491252899 }, { "epoch": 0.44675324675324674, "step": 172, "train/loss_ctc": 0.33496394753456116, "train/loss_error": 0.3128748834133148, "train/loss_total": 0.3172926902770996 }, { "epoch": 0.44935064935064933, "step": 173, "train/loss_ctc": 0.40994027256965637, "train/loss_error": 0.3382743000984192, "train/loss_total": 0.35260751843452454 }, { "epoch": 0.45194805194805193, "step": 174, "train/loss_ctc": 0.40523260831832886, "train/loss_error": 0.278840571641922, "train/loss_total": 0.3041189908981323 }, { "epoch": 0.45454545454545453, "step": 175, "train/loss_ctc": 0.6315127015113831, "train/loss_error": 0.3712315559387207, "train/loss_total": 0.4232877790927887 }, { "epoch": 0.45714285714285713, "step": 176, "train/loss_ctc": 0.3229829967021942, "train/loss_error": 0.40724894404411316, "train/loss_total": 0.39039576053619385 }, { "epoch": 0.4597402597402597, "step": 177, "train/loss_ctc": 0.3392813205718994, "train/loss_error": 0.3146410882472992, "train/loss_total": 0.3195691704750061 }, { "epoch": 0.4623376623376623, "step": 178, "train/loss_ctc": 0.4536489248275757, "train/loss_error": 0.38239654898643494, "train/loss_total": 0.39664703607559204 }, { "epoch": 0.4649350649350649, "step": 179, "train/loss_ctc": 0.4217790961265564, "train/loss_error": 0.2806907892227173, "train/loss_total": 0.30890846252441406 }, { "epoch": 0.4675324675324675, "grad_norm": 27156.9375, "learning_rate": 2.7225974025974027e-05, "loss": 0.348, "step": 180 }, { "epoch": 0.4675324675324675, "step": 180, "train/loss_ctc": 0.3659857213497162, "train/loss_error": 0.30686989426612854, "train/loss_total": 0.318693071603775 }, { "epoch": 0.4701298701298701, "step": 181, "train/loss_ctc": 0.4668356776237488, "train/loss_error": 0.4048503637313843, "train/loss_total": 0.4172474145889282 }, { "epoch": 0.4727272727272727, "step": 182, "train/loss_ctc": 0.437044620513916, "train/loss_error": 0.3788928985595703, "train/loss_total": 0.3905232548713684 }, { "epoch": 0.4753246753246753, "step": 183, "train/loss_ctc": 0.37603759765625, "train/loss_error": 0.35779333114624023, "train/loss_total": 0.3614422082901001 }, { "epoch": 0.4779220779220779, "step": 184, "train/loss_ctc": 1.2886958122253418, "train/loss_error": 0.4061935544013977, "train/loss_total": 0.5826939940452576 }, { "epoch": 0.4805194805194805, "step": 185, "train/loss_ctc": 0.585574746131897, "train/loss_error": 0.3772028088569641, "train/loss_total": 0.4188772141933441 }, { "epoch": 0.4831168831168831, "step": 186, "train/loss_ctc": 0.44567131996154785, "train/loss_error": 0.37361079454421997, "train/loss_total": 0.38802292943000793 }, { "epoch": 0.4857142857142857, "step": 187, "train/loss_ctc": 0.3594258725643158, "train/loss_error": 0.3020654618740082, "train/loss_total": 0.3135375380516052 }, { "epoch": 0.4883116883116883, "step": 188, "train/loss_ctc": 0.34104424715042114, "train/loss_error": 0.3769902288913727, "train/loss_total": 0.36980104446411133 }, { "epoch": 0.4909090909090909, "step": 189, "train/loss_ctc": 0.47630825638771057, "train/loss_error": 0.36443567276000977, "train/loss_total": 0.38681021332740784 }, { "epoch": 0.4935064935064935, "grad_norm": 27346.6640625, "learning_rate": 2.7070129870129872e-05, "loss": 0.3948, "step": 190 }, { "epoch": 0.4935064935064935, "step": 190, "train/loss_ctc": 0.2919171452522278, "train/loss_error": 0.3062206506729126, "train/loss_total": 0.3033599555492401 }, { "epoch": 0.4961038961038961, "step": 191, "train/loss_ctc": 0.46932506561279297, "train/loss_error": 0.3895440399646759, "train/loss_total": 0.40550023317337036 }, { "epoch": 0.4987012987012987, "step": 192, "train/loss_ctc": 0.39248478412628174, "train/loss_error": 0.37490010261535645, "train/loss_total": 0.378417044878006 }, { "epoch": 0.5012987012987012, "step": 193, "train/loss_ctc": 0.3928970694541931, "train/loss_error": 0.2922249138355255, "train/loss_total": 0.31235936284065247 }, { "epoch": 0.5038961038961038, "step": 194, "train/loss_ctc": 0.4539378881454468, "train/loss_error": 0.3689957857131958, "train/loss_total": 0.3859842121601105 }, { "epoch": 0.5064935064935064, "step": 195, "train/loss_ctc": 0.3483414351940155, "train/loss_error": 0.2720561921596527, "train/loss_total": 0.2873132526874542 }, { "epoch": 0.509090909090909, "step": 196, "train/loss_ctc": 0.36276158690452576, "train/loss_error": 0.3573598563671112, "train/loss_total": 0.35844022035598755 }, { "epoch": 0.5116883116883116, "step": 197, "train/loss_ctc": 0.6235524415969849, "train/loss_error": 0.3245532810688019, "train/loss_total": 0.3843531310558319 }, { "epoch": 0.5142857142857142, "step": 198, "train/loss_ctc": 0.4543779790401459, "train/loss_error": 0.385804146528244, "train/loss_total": 0.3995189070701599 }, { "epoch": 0.5168831168831168, "step": 199, "train/loss_ctc": 0.3976207375526428, "train/loss_error": 0.29327359795570374, "train/loss_total": 0.31414303183555603 }, { "epoch": 0.5194805194805194, "grad_norm": 22767.00390625, "learning_rate": 2.6914285714285713e-05, "loss": 0.3529, "step": 200 }, { "epoch": 0.5194805194805194, "step": 200, "train/loss_ctc": 0.3632248342037201, "train/loss_error": 0.31166815757751465, "train/loss_total": 0.32197949290275574 }, { "epoch": 0.522077922077922, "step": 201, "train/loss_ctc": 0.5665127038955688, "train/loss_error": 0.42717644572257996, "train/loss_total": 0.4550437331199646 }, { "epoch": 0.5246753246753246, "step": 202, "train/loss_ctc": 0.316131055355072, "train/loss_error": 0.3018094301223755, "train/loss_total": 0.3046737611293793 }, { "epoch": 0.5272727272727272, "step": 203, "train/loss_ctc": 0.46885448694229126, "train/loss_error": 0.33362331986427307, "train/loss_total": 0.3606695532798767 }, { "epoch": 0.5298701298701298, "step": 204, "train/loss_ctc": 0.34516167640686035, "train/loss_error": 0.26915067434310913, "train/loss_total": 0.2843528687953949 }, { "epoch": 0.5324675324675324, "step": 205, "train/loss_ctc": 0.3473971486091614, "train/loss_error": 0.2237280309200287, "train/loss_total": 0.24846184253692627 }, { "epoch": 0.535064935064935, "step": 206, "train/loss_ctc": 0.28850260376930237, "train/loss_error": 0.33666419982910156, "train/loss_total": 0.3270318806171417 }, { "epoch": 0.5376623376623376, "step": 207, "train/loss_ctc": 0.8976432085037231, "train/loss_error": 0.4337098002433777, "train/loss_total": 0.5264964699745178 }, { "epoch": 0.5402597402597402, "step": 208, "train/loss_ctc": 0.37336426973342896, "train/loss_error": 0.3560982644557953, "train/loss_total": 0.35955145955085754 }, { "epoch": 0.5428571428571428, "step": 209, "train/loss_ctc": 0.5330466032028198, "train/loss_error": 0.43520575761795044, "train/loss_total": 0.4547739326953888 }, { "epoch": 0.5454545454545454, "grad_norm": 28691.42578125, "learning_rate": 2.6758441558441558e-05, "loss": 0.3643, "step": 210 }, { "epoch": 0.5454545454545454, "step": 210, "train/loss_ctc": 0.4001771807670593, "train/loss_error": 0.3226031959056854, "train/loss_total": 0.3381180167198181 }, { "epoch": 0.548051948051948, "step": 211, "train/loss_ctc": 0.44906607270240784, "train/loss_error": 0.3897551894187927, "train/loss_total": 0.4016173481941223 }, { "epoch": 0.5506493506493506, "step": 212, "train/loss_ctc": 0.5262753963470459, "train/loss_error": 0.48672255873680115, "train/loss_total": 0.49463313817977905 }, { "epoch": 0.5532467532467532, "step": 213, "train/loss_ctc": 0.4676845669746399, "train/loss_error": 0.4096420705318451, "train/loss_total": 0.421250581741333 }, { "epoch": 0.5558441558441558, "step": 214, "train/loss_ctc": 0.5162551403045654, "train/loss_error": 0.4350341856479645, "train/loss_total": 0.4512783885002136 }, { "epoch": 0.5584415584415584, "step": 215, "train/loss_ctc": 0.40487146377563477, "train/loss_error": 0.4125872850418091, "train/loss_total": 0.4110441207885742 }, { "epoch": 0.561038961038961, "step": 216, "train/loss_ctc": 0.3994683623313904, "train/loss_error": 0.3751136064529419, "train/loss_total": 0.3799845576286316 }, { "epoch": 0.5636363636363636, "step": 217, "train/loss_ctc": 0.4129399061203003, "train/loss_error": 0.35818907618522644, "train/loss_total": 0.36913925409317017 }, { "epoch": 0.5662337662337662, "step": 218, "train/loss_ctc": 0.30207696557044983, "train/loss_error": 0.32694298028945923, "train/loss_total": 0.32196977734565735 }, { "epoch": 0.5688311688311688, "step": 219, "train/loss_ctc": 0.420742392539978, "train/loss_error": 0.3494946360588074, "train/loss_total": 0.36374419927597046 }, { "epoch": 0.5714285714285714, "grad_norm": 25303.607421875, "learning_rate": 2.6602597402597403e-05, "loss": 0.3953, "step": 220 }, { "epoch": 0.5714285714285714, "step": 220, "train/loss_ctc": 0.27296724915504456, "train/loss_error": 0.3153311312198639, "train/loss_total": 0.3068583607673645 }, { "epoch": 0.574025974025974, "step": 221, "train/loss_ctc": 0.35099485516548157, "train/loss_error": 0.3048248887062073, "train/loss_total": 0.31405889987945557 }, { "epoch": 0.5766233766233766, "step": 222, "train/loss_ctc": 0.45413821935653687, "train/loss_error": 0.37537604570388794, "train/loss_total": 0.3911284804344177 }, { "epoch": 0.5792207792207792, "step": 223, "train/loss_ctc": 0.4859244227409363, "train/loss_error": 0.3853004276752472, "train/loss_total": 0.4054252505302429 }, { "epoch": 0.5818181818181818, "step": 224, "train/loss_ctc": 0.34131506085395813, "train/loss_error": 0.28819337487220764, "train/loss_total": 0.2988177239894867 }, { "epoch": 0.5844155844155844, "step": 225, "train/loss_ctc": 0.5639625787734985, "train/loss_error": 0.34172767400741577, "train/loss_total": 0.38617467880249023 }, { "epoch": 0.587012987012987, "step": 226, "train/loss_ctc": 0.8429668545722961, "train/loss_error": 0.2764878571033478, "train/loss_total": 0.38978368043899536 }, { "epoch": 0.5896103896103896, "step": 227, "train/loss_ctc": 0.4332185387611389, "train/loss_error": 0.351394921541214, "train/loss_total": 0.367759644985199 }, { "epoch": 0.5922077922077922, "step": 228, "train/loss_ctc": 0.47701945900917053, "train/loss_error": 0.4423590302467346, "train/loss_total": 0.4492911100387573 }, { "epoch": 0.5948051948051948, "step": 229, "train/loss_ctc": 0.45697465538978577, "train/loss_error": 0.30529898405075073, "train/loss_total": 0.33563411235809326 }, { "epoch": 0.5974025974025974, "grad_norm": 23568.68359375, "learning_rate": 2.6446753246753248e-05, "loss": 0.3645, "step": 230 }, { "epoch": 0.5974025974025974, "step": 230, "train/loss_ctc": 0.4049168825149536, "train/loss_error": 0.3368241786956787, "train/loss_total": 0.35044270753860474 }, { "epoch": 0.6, "step": 231, "train/loss_ctc": 0.40864211320877075, "train/loss_error": 0.33976665139198303, "train/loss_total": 0.353541761636734 }, { "epoch": 0.6025974025974026, "step": 232, "train/loss_ctc": 0.28140905499458313, "train/loss_error": 0.24122655391693115, "train/loss_total": 0.24926304817199707 }, { "epoch": 0.6051948051948052, "step": 233, "train/loss_ctc": 0.4438045620918274, "train/loss_error": 0.4265562891960144, "train/loss_total": 0.4300059378147125 }, { "epoch": 0.6077922077922078, "step": 234, "train/loss_ctc": 0.43554985523223877, "train/loss_error": 0.30568408966064453, "train/loss_total": 0.3316572606563568 }, { "epoch": 0.6103896103896104, "step": 235, "train/loss_ctc": 0.48383578658103943, "train/loss_error": 0.4783856272697449, "train/loss_total": 0.4794756770133972 }, { "epoch": 0.612987012987013, "step": 236, "train/loss_ctc": 0.3309652507305145, "train/loss_error": 0.2869800925254822, "train/loss_total": 0.2957771420478821 }, { "epoch": 0.6155844155844156, "step": 237, "train/loss_ctc": 0.7144870162010193, "train/loss_error": 0.4156522750854492, "train/loss_total": 0.47541922330856323 }, { "epoch": 0.6181818181818182, "step": 238, "train/loss_ctc": 0.41246694326400757, "train/loss_error": 0.40978875756263733, "train/loss_total": 0.4103243947029114 }, { "epoch": 0.6207792207792208, "step": 239, "train/loss_ctc": 0.3006684482097626, "train/loss_error": 0.2530428469181061, "train/loss_total": 0.2625679671764374 }, { "epoch": 0.6233766233766234, "grad_norm": 21444.177734375, "learning_rate": 2.6290909090909092e-05, "loss": 0.3638, "step": 240 }, { "epoch": 0.6233766233766234, "step": 240, "train/loss_ctc": 0.38136082887649536, "train/loss_error": 0.3084414303302765, "train/loss_total": 0.32302531599998474 }, { "epoch": 0.625974025974026, "step": 241, "train/loss_ctc": 0.39003682136535645, "train/loss_error": 0.28413617610931396, "train/loss_total": 0.305316299200058 }, { "epoch": 0.6285714285714286, "step": 242, "train/loss_ctc": 0.3273082971572876, "train/loss_error": 0.31314921379089355, "train/loss_total": 0.31598103046417236 }, { "epoch": 0.6311688311688312, "step": 243, "train/loss_ctc": 0.5596866607666016, "train/loss_error": 0.3712642192840576, "train/loss_total": 0.40894871950149536 }, { "epoch": 0.6337662337662338, "step": 244, "train/loss_ctc": 0.33036717772483826, "train/loss_error": 0.37094706296920776, "train/loss_total": 0.36283111572265625 }, { "epoch": 0.6363636363636364, "step": 245, "train/loss_ctc": 0.4024241864681244, "train/loss_error": 0.30766361951828003, "train/loss_total": 0.32661575078964233 }, { "epoch": 0.638961038961039, "step": 246, "train/loss_ctc": 0.28036728501319885, "train/loss_error": 0.267008513212204, "train/loss_total": 0.2696802616119385 }, { "epoch": 0.6415584415584416, "step": 247, "train/loss_ctc": 0.4785882234573364, "train/loss_error": 0.4065519869327545, "train/loss_total": 0.4209592342376709 }, { "epoch": 0.6441558441558441, "step": 248, "train/loss_ctc": 0.483078271150589, "train/loss_error": 0.43147146701812744, "train/loss_total": 0.4417928457260132 }, { "epoch": 0.6467532467532467, "step": 249, "train/loss_ctc": 0.7345058917999268, "train/loss_error": 0.3221513628959656, "train/loss_total": 0.40462225675582886 }, { "epoch": 0.6493506493506493, "grad_norm": 28623.013671875, "learning_rate": 2.6135064935064934e-05, "loss": 0.358, "step": 250 }, { "epoch": 0.6493506493506493, "step": 250, "train/loss_ctc": 0.43413758277893066, "train/loss_error": 0.35181954503059387, "train/loss_total": 0.36828315258026123 }, { "epoch": 0.6519480519480519, "step": 251, "train/loss_ctc": 0.46927833557128906, "train/loss_error": 0.31805333495140076, "train/loss_total": 0.3482983410358429 }, { "epoch": 0.6545454545454545, "step": 252, "train/loss_ctc": 0.5144405364990234, "train/loss_error": 0.3659208118915558, "train/loss_total": 0.3956247568130493 }, { "epoch": 0.6571428571428571, "step": 253, "train/loss_ctc": 0.2802880108356476, "train/loss_error": 0.3198605477809906, "train/loss_total": 0.3119460344314575 }, { "epoch": 0.6597402597402597, "step": 254, "train/loss_ctc": 0.36363720893859863, "train/loss_error": 0.31777867674827576, "train/loss_total": 0.32695040106773376 }, { "epoch": 0.6623376623376623, "step": 255, "train/loss_ctc": 0.48388856649398804, "train/loss_error": 0.325737327337265, "train/loss_total": 0.3573675751686096 }, { "epoch": 0.6649350649350649, "step": 256, "train/loss_ctc": 0.39503034949302673, "train/loss_error": 0.365969717502594, "train/loss_total": 0.3717818558216095 }, { "epoch": 0.6675324675324675, "step": 257, "train/loss_ctc": 0.4514634609222412, "train/loss_error": 0.4045567810535431, "train/loss_total": 0.41393813490867615 }, { "epoch": 0.6701298701298701, "step": 258, "train/loss_ctc": 0.30695366859436035, "train/loss_error": 0.2940598428249359, "train/loss_total": 0.2966386079788208 }, { "epoch": 0.6727272727272727, "step": 259, "train/loss_ctc": 0.4479135274887085, "train/loss_error": 0.45762887597084045, "train/loss_total": 0.4556858241558075 }, { "epoch": 0.6753246753246753, "grad_norm": 23297.76171875, "learning_rate": 2.597922077922078e-05, "loss": 0.3647, "step": 260 }, { "epoch": 0.6753246753246753, "step": 260, "train/loss_ctc": 0.4023851156234741, "train/loss_error": 0.2815113663673401, "train/loss_total": 0.3056861162185669 }, { "epoch": 0.6779220779220779, "step": 261, "train/loss_ctc": 0.3361915946006775, "train/loss_error": 0.31679844856262207, "train/loss_total": 0.32067710161209106 }, { "epoch": 0.6805194805194805, "step": 262, "train/loss_ctc": 0.4177173674106598, "train/loss_error": 0.3607991635799408, "train/loss_total": 0.37218281626701355 }, { "epoch": 0.6831168831168831, "step": 263, "train/loss_ctc": 0.5532109141349792, "train/loss_error": 0.35775360465049744, "train/loss_total": 0.3968450725078583 }, { "epoch": 0.6857142857142857, "step": 264, "train/loss_ctc": 0.6537083983421326, "train/loss_error": 0.3501454293727875, "train/loss_total": 0.41085803508758545 }, { "epoch": 0.6883116883116883, "step": 265, "train/loss_ctc": 0.5003011226654053, "train/loss_error": 0.36121705174446106, "train/loss_total": 0.38903388381004333 }, { "epoch": 0.6909090909090909, "step": 266, "train/loss_ctc": 0.2834610342979431, "train/loss_error": 0.31835266947746277, "train/loss_total": 0.31137436628341675 }, { "epoch": 0.6935064935064935, "step": 267, "train/loss_ctc": 0.5387441515922546, "train/loss_error": 0.3887421786785126, "train/loss_total": 0.4187425971031189 }, { "epoch": 0.6961038961038961, "step": 268, "train/loss_ctc": 0.29010018706321716, "train/loss_error": 0.2870390713214874, "train/loss_total": 0.28765130043029785 }, { "epoch": 0.6987012987012987, "step": 269, "train/loss_ctc": 0.44061967730522156, "train/loss_error": 0.37183260917663574, "train/loss_total": 0.3855900466442108 }, { "epoch": 0.7012987012987013, "grad_norm": 27820.59375, "learning_rate": 2.5823376623376623e-05, "loss": 0.3599, "step": 270 }, { "epoch": 0.7012987012987013, "step": 270, "train/loss_ctc": 0.3791232705116272, "train/loss_error": 0.3768686056137085, "train/loss_total": 0.3773195445537567 }, { "epoch": 0.7038961038961039, "step": 271, "train/loss_ctc": 0.22008810937404633, "train/loss_error": 0.22818946838378906, "train/loss_total": 0.22656920552253723 }, { "epoch": 0.7064935064935065, "step": 272, "train/loss_ctc": 0.3351387083530426, "train/loss_error": 0.3553205728530884, "train/loss_total": 0.3512842059135437 }, { "epoch": 0.7090909090909091, "step": 273, "train/loss_ctc": 0.5356822609901428, "train/loss_error": 0.3578815162181854, "train/loss_total": 0.39344167709350586 }, { "epoch": 0.7116883116883117, "step": 274, "train/loss_ctc": 0.5884527564048767, "train/loss_error": 0.4627361297607422, "train/loss_total": 0.4878794848918915 }, { "epoch": 0.7142857142857143, "step": 275, "train/loss_ctc": 0.22393588721752167, "train/loss_error": 0.3829960227012634, "train/loss_total": 0.3511839807033539 }, { "epoch": 0.7168831168831169, "step": 276, "train/loss_ctc": 0.3860454559326172, "train/loss_error": 0.38049259781837463, "train/loss_total": 0.3816031813621521 }, { "epoch": 0.7194805194805195, "step": 277, "train/loss_ctc": 0.6169874668121338, "train/loss_error": 0.3535917103290558, "train/loss_total": 0.4062708616256714 }, { "epoch": 0.7220779220779221, "step": 278, "train/loss_ctc": 0.342138409614563, "train/loss_error": 0.35316717624664307, "train/loss_total": 0.3509614169597626 }, { "epoch": 0.7246753246753247, "step": 279, "train/loss_ctc": 0.3814559578895569, "train/loss_error": 0.2909921705722809, "train/loss_total": 0.3090849220752716 }, { "epoch": 0.7272727272727273, "grad_norm": 22526.96484375, "learning_rate": 2.5667532467532468e-05, "loss": 0.3636, "step": 280 }, { "epoch": 0.7272727272727273, "step": 280, "train/loss_ctc": 0.4916198253631592, "train/loss_error": 0.4286263585090637, "train/loss_total": 0.4412250518798828 }, { "epoch": 0.7298701298701299, "step": 281, "train/loss_ctc": 0.2942402958869934, "train/loss_error": 0.34471017122268677, "train/loss_total": 0.33461618423461914 }, { "epoch": 0.7324675324675325, "step": 282, "train/loss_ctc": 0.25999826192855835, "train/loss_error": 0.2869489789009094, "train/loss_total": 0.2815588414669037 }, { "epoch": 0.7350649350649351, "step": 283, "train/loss_ctc": 0.38915467262268066, "train/loss_error": 0.3066401779651642, "train/loss_total": 0.3231430649757385 }, { "epoch": 0.7376623376623377, "step": 284, "train/loss_ctc": 0.6703136563301086, "train/loss_error": 0.41176050901412964, "train/loss_total": 0.4634711444377899 }, { "epoch": 0.7402597402597403, "step": 285, "train/loss_ctc": 0.2975463271141052, "train/loss_error": 0.3499086797237396, "train/loss_total": 0.33943623304367065 }, { "epoch": 0.7428571428571429, "step": 286, "train/loss_ctc": 0.4595615863800049, "train/loss_error": 0.372159481048584, "train/loss_total": 0.3896399140357971 }, { "epoch": 0.7454545454545455, "step": 287, "train/loss_ctc": 0.45898115634918213, "train/loss_error": 0.32023754715919495, "train/loss_total": 0.34798628091812134 }, { "epoch": 0.7480519480519481, "step": 288, "train/loss_ctc": 0.36516812443733215, "train/loss_error": 0.31672269105911255, "train/loss_total": 0.32641178369522095 }, { "epoch": 0.7506493506493507, "step": 289, "train/loss_ctc": 0.3871649503707886, "train/loss_error": 0.3956572413444519, "train/loss_total": 0.39395877718925476 }, { "epoch": 0.7532467532467533, "grad_norm": 23180.814453125, "learning_rate": 2.5511688311688313e-05, "loss": 0.3641, "step": 290 }, { "epoch": 0.7532467532467533, "step": 290, "train/loss_ctc": 0.5151258111000061, "train/loss_error": 0.36818456649780273, "train/loss_total": 0.3975728154182434 }, { "epoch": 0.7558441558441559, "step": 291, "train/loss_ctc": 0.42342329025268555, "train/loss_error": 0.33282357454299927, "train/loss_total": 0.35094353556632996 }, { "epoch": 0.7584415584415585, "step": 292, "train/loss_ctc": 0.3022388815879822, "train/loss_error": 0.3043747544288635, "train/loss_total": 0.3039475679397583 }, { "epoch": 0.7610389610389611, "step": 293, "train/loss_ctc": 0.46895045042037964, "train/loss_error": 0.3710747957229614, "train/loss_total": 0.3906499147415161 }, { "epoch": 0.7636363636363637, "step": 294, "train/loss_ctc": 0.35200050473213196, "train/loss_error": 0.3893146514892578, "train/loss_total": 0.38185185194015503 }, { "epoch": 0.7662337662337663, "step": 295, "train/loss_ctc": 0.3494299054145813, "train/loss_error": 0.3256676495075226, "train/loss_total": 0.3304201364517212 }, { "epoch": 0.7688311688311689, "step": 296, "train/loss_ctc": 0.7615123987197876, "train/loss_error": 0.41284388303756714, "train/loss_total": 0.4825776219367981 }, { "epoch": 0.7714285714285715, "step": 297, "train/loss_ctc": 0.6958169937133789, "train/loss_error": 0.40230637788772583, "train/loss_total": 0.4610085189342499 }, { "epoch": 0.7740259740259741, "step": 298, "train/loss_ctc": 0.47372153401374817, "train/loss_error": 0.3732481300830841, "train/loss_total": 0.3933427929878235 }, { "epoch": 0.7766233766233767, "step": 299, "train/loss_ctc": 0.4496305584907532, "train/loss_error": 0.36496153473854065, "train/loss_total": 0.38189536333084106 }, { "epoch": 0.7792207792207793, "grad_norm": 23907.88671875, "learning_rate": 2.5355844155844154e-05, "loss": 0.3874, "step": 300 }, { "epoch": 0.7792207792207793, "step": 300, "train/loss_ctc": 0.40647345781326294, "train/loss_error": 0.32093051075935364, "train/loss_total": 0.3380391001701355 }, { "epoch": 0.7818181818181819, "step": 301, "train/loss_ctc": 0.6000281572341919, "train/loss_error": 0.34365159273147583, "train/loss_total": 0.39492690563201904 }, { "epoch": 0.7844155844155845, "step": 302, "train/loss_ctc": 0.510002613067627, "train/loss_error": 0.25482118129730225, "train/loss_total": 0.30585747957229614 }, { "epoch": 0.787012987012987, "step": 303, "train/loss_ctc": 0.29983875155448914, "train/loss_error": 0.2960861921310425, "train/loss_total": 0.2968367040157318 }, { "epoch": 0.7896103896103897, "step": 304, "train/loss_ctc": 0.4203146994113922, "train/loss_error": 0.35243508219718933, "train/loss_total": 0.36601099371910095 }, { "epoch": 0.7922077922077922, "step": 305, "train/loss_ctc": 0.4009416401386261, "train/loss_error": 0.396989643573761, "train/loss_total": 0.39778006076812744 }, { "epoch": 0.7948051948051948, "step": 306, "train/loss_ctc": 0.2508049011230469, "train/loss_error": 0.26163843274116516, "train/loss_total": 0.25947174429893494 }, { "epoch": 0.7974025974025974, "step": 307, "train/loss_ctc": 0.4857748746871948, "train/loss_error": 0.3838246464729309, "train/loss_total": 0.4042147099971771 }, { "epoch": 0.8, "step": 308, "train/loss_ctc": 0.3055418133735657, "train/loss_error": 0.2459239810705185, "train/loss_total": 0.25784754753112793 }, { "epoch": 0.8025974025974026, "step": 309, "train/loss_ctc": 0.481447696685791, "train/loss_error": 0.45272716879844666, "train/loss_total": 0.45847129821777344 }, { "epoch": 0.8051948051948052, "grad_norm": 35387.921875, "learning_rate": 2.52e-05, "loss": 0.3479, "step": 310 }, { "epoch": 0.8051948051948052, "step": 310, "train/loss_ctc": 0.49516427516937256, "train/loss_error": 0.3490211069583893, "train/loss_total": 0.37824973464012146 }, { "epoch": 0.8077922077922078, "step": 311, "train/loss_ctc": 0.2695940136909485, "train/loss_error": 0.3571712076663971, "train/loss_total": 0.3396557867527008 }, { "epoch": 0.8103896103896104, "step": 312, "train/loss_ctc": 0.42255452275276184, "train/loss_error": 0.35100433230400085, "train/loss_total": 0.3653143644332886 }, { "epoch": 0.812987012987013, "step": 313, "train/loss_ctc": 0.5683062076568604, "train/loss_error": 0.44193777441978455, "train/loss_total": 0.4672114849090576 }, { "epoch": 0.8155844155844156, "step": 314, "train/loss_ctc": 0.7459508180618286, "train/loss_error": 0.4092414379119873, "train/loss_total": 0.4765833020210266 }, { "epoch": 0.8181818181818182, "step": 315, "train/loss_ctc": 0.36052772402763367, "train/loss_error": 0.29495641589164734, "train/loss_total": 0.30807068943977356 }, { "epoch": 0.8207792207792208, "step": 316, "train/loss_ctc": 0.39381128549575806, "train/loss_error": 0.3674761652946472, "train/loss_total": 0.3727431893348694 }, { "epoch": 0.8233766233766234, "step": 317, "train/loss_ctc": 0.26728999614715576, "train/loss_error": 0.21219880878925323, "train/loss_total": 0.22321705520153046 }, { "epoch": 0.825974025974026, "step": 318, "train/loss_ctc": 0.4433158040046692, "train/loss_error": 0.27095192670822144, "train/loss_total": 0.30542469024658203 }, { "epoch": 0.8285714285714286, "step": 319, "train/loss_ctc": 0.56886887550354, "train/loss_error": 0.3197595775127411, "train/loss_total": 0.3695814609527588 }, { "epoch": 0.8311688311688312, "grad_norm": 31442.958984375, "learning_rate": 2.5044155844155844e-05, "loss": 0.3606, "step": 320 }, { "epoch": 0.8311688311688312, "step": 320, "train/loss_ctc": 0.5201083421707153, "train/loss_error": 0.4188494384288788, "train/loss_total": 0.4391012191772461 }, { "epoch": 0.8337662337662337, "step": 321, "train/loss_ctc": 0.29672133922576904, "train/loss_error": 0.31551843881607056, "train/loss_total": 0.31175902485847473 }, { "epoch": 0.8363636363636363, "step": 322, "train/loss_ctc": 0.4204404354095459, "train/loss_error": 0.38120904564857483, "train/loss_total": 0.3890553414821625 }, { "epoch": 0.8389610389610389, "step": 323, "train/loss_ctc": 0.3928925395011902, "train/loss_error": 0.29011043906211853, "train/loss_total": 0.31066685914993286 }, { "epoch": 0.8415584415584415, "step": 324, "train/loss_ctc": 0.5168473720550537, "train/loss_error": 0.38749954104423523, "train/loss_total": 0.4133691191673279 }, { "epoch": 0.8441558441558441, "step": 325, "train/loss_ctc": 0.4026387333869934, "train/loss_error": 0.37570133805274963, "train/loss_total": 0.38108882308006287 }, { "epoch": 0.8467532467532467, "step": 326, "train/loss_ctc": 0.29657360911369324, "train/loss_error": 0.40010902285575867, "train/loss_total": 0.37940195202827454 }, { "epoch": 0.8493506493506493, "step": 327, "train/loss_ctc": 0.5478739738464355, "train/loss_error": 0.3265385925769806, "train/loss_total": 0.37080568075180054 }, { "epoch": 0.8519480519480519, "step": 328, "train/loss_ctc": 0.4745749235153198, "train/loss_error": 0.3826153874397278, "train/loss_total": 0.4010072946548462 }, { "epoch": 0.8545454545454545, "step": 329, "train/loss_ctc": 0.48814016580581665, "train/loss_error": 0.32991209626197815, "train/loss_total": 0.3615577220916748 }, { "epoch": 0.8571428571428571, "grad_norm": 21101.873046875, "learning_rate": 2.488831168831169e-05, "loss": 0.3758, "step": 330 }, { "epoch": 0.8571428571428571, "step": 330, "train/loss_ctc": 0.3427393436431885, "train/loss_error": 0.2800573408603668, "train/loss_total": 0.29259374737739563 }, { "epoch": 0.8597402597402597, "step": 331, "train/loss_ctc": 0.5480760335922241, "train/loss_error": 0.356105774641037, "train/loss_total": 0.39449983835220337 }, { "epoch": 0.8623376623376623, "step": 332, "train/loss_ctc": 0.4449375867843628, "train/loss_error": 0.34756365418434143, "train/loss_total": 0.36703845858573914 }, { "epoch": 0.8649350649350649, "step": 333, "train/loss_ctc": 0.3572324514389038, "train/loss_error": 0.31246161460876465, "train/loss_total": 0.3214157819747925 }, { "epoch": 0.8675324675324675, "step": 334, "train/loss_ctc": 0.19999679923057556, "train/loss_error": 0.29441767930984497, "train/loss_total": 0.2755334973335266 }, { "epoch": 0.8701298701298701, "step": 335, "train/loss_ctc": 0.6673377752304077, "train/loss_error": 0.32522642612457275, "train/loss_total": 0.3936487138271332 }, { "epoch": 0.8727272727272727, "step": 336, "train/loss_ctc": 0.46393126249313354, "train/loss_error": 0.3437224328517914, "train/loss_total": 0.3677642047405243 }, { "epoch": 0.8753246753246753, "step": 337, "train/loss_ctc": 0.6437364816665649, "train/loss_error": 0.3593495190143585, "train/loss_total": 0.41622692346572876 }, { "epoch": 0.8779220779220779, "step": 338, "train/loss_ctc": 0.27966487407684326, "train/loss_error": 0.3127220571041107, "train/loss_total": 0.3061106204986572 }, { "epoch": 0.8805194805194805, "step": 339, "train/loss_ctc": 0.526718258857727, "train/loss_error": 0.2913677990436554, "train/loss_total": 0.33843791484832764 }, { "epoch": 0.8831168831168831, "grad_norm": 29283.08203125, "learning_rate": 2.473246753246753e-05, "loss": 0.3473, "step": 340 }, { "epoch": 0.8831168831168831, "step": 340, "train/loss_ctc": 0.29090777039527893, "train/loss_error": 0.31618446111679077, "train/loss_total": 0.3111291229724884 }, { "epoch": 0.8857142857142857, "step": 341, "train/loss_ctc": 0.42237183451652527, "train/loss_error": 0.43595752120018005, "train/loss_total": 0.4332404136657715 }, { "epoch": 0.8883116883116883, "step": 342, "train/loss_ctc": 0.3439171314239502, "train/loss_error": 0.24306032061576843, "train/loss_total": 0.26323169469833374 }, { "epoch": 0.8909090909090909, "step": 343, "train/loss_ctc": 0.540580153465271, "train/loss_error": 0.3215090334415436, "train/loss_total": 0.3653232753276825 }, { "epoch": 0.8935064935064935, "step": 344, "train/loss_ctc": 0.37305963039398193, "train/loss_error": 0.32820114493370056, "train/loss_total": 0.33717286586761475 }, { "epoch": 0.8961038961038961, "step": 345, "train/loss_ctc": 0.5762365460395813, "train/loss_error": 0.38577136397361755, "train/loss_total": 0.4238643944263458 }, { "epoch": 0.8987012987012987, "step": 346, "train/loss_ctc": 0.4368545413017273, "train/loss_error": 0.37993431091308594, "train/loss_total": 0.39131835103034973 }, { "epoch": 0.9012987012987013, "step": 347, "train/loss_ctc": 0.5169637799263, "train/loss_error": 0.3412202298641205, "train/loss_total": 0.3763689398765564 }, { "epoch": 0.9038961038961039, "step": 348, "train/loss_ctc": 0.6133180856704712, "train/loss_error": 0.2932462990283966, "train/loss_total": 0.35726064443588257 }, { "epoch": 0.9064935064935065, "step": 349, "train/loss_ctc": 0.3360356092453003, "train/loss_error": 0.3669890761375427, "train/loss_total": 0.3607983887195587 }, { "epoch": 0.9090909090909091, "grad_norm": 24231.083984375, "learning_rate": 2.4576623376623375e-05, "loss": 0.362, "step": 350 }, { "epoch": 0.9090909090909091, "step": 350, "train/loss_ctc": 0.4170071482658386, "train/loss_error": 0.19790297746658325, "train/loss_total": 0.24172380566596985 }, { "epoch": 0.9116883116883117, "step": 351, "train/loss_ctc": 0.4292232096195221, "train/loss_error": 0.39602819085121155, "train/loss_total": 0.40266719460487366 }, { "epoch": 0.9142857142857143, "step": 352, "train/loss_ctc": 0.47929608821868896, "train/loss_error": 0.3309504985809326, "train/loss_total": 0.3606196343898773 }, { "epoch": 0.9168831168831169, "step": 353, "train/loss_ctc": 0.30566102266311646, "train/loss_error": 0.28978413343429565, "train/loss_total": 0.2929595112800598 }, { "epoch": 0.9194805194805195, "step": 354, "train/loss_ctc": 0.278481662273407, "train/loss_error": 0.3265637457370758, "train/loss_total": 0.316947340965271 }, { "epoch": 0.922077922077922, "step": 355, "train/loss_ctc": 0.33733952045440674, "train/loss_error": 0.2496909648180008, "train/loss_total": 0.267220675945282 }, { "epoch": 0.9246753246753247, "step": 356, "train/loss_ctc": 0.2528141140937805, "train/loss_error": 0.3711775541305542, "train/loss_total": 0.3475048840045929 }, { "epoch": 0.9272727272727272, "step": 357, "train/loss_ctc": 0.6311486959457397, "train/loss_error": 0.41369837522506714, "train/loss_total": 0.4571884274482727 }, { "epoch": 0.9298701298701298, "step": 358, "train/loss_ctc": 0.35564902424812317, "train/loss_error": 0.24787288904190063, "train/loss_total": 0.2694281339645386 }, { "epoch": 0.9324675324675324, "step": 359, "train/loss_ctc": 0.331866979598999, "train/loss_error": 0.28180283308029175, "train/loss_total": 0.2918156683444977 }, { "epoch": 0.935064935064935, "grad_norm": 23446.185546875, "learning_rate": 2.442077922077922e-05, "loss": 0.3248, "step": 360 }, { "epoch": 0.935064935064935, "step": 360, "train/loss_ctc": 0.5157185792922974, "train/loss_error": 0.31431061029434204, "train/loss_total": 0.3545922040939331 }, { "epoch": 0.9376623376623376, "step": 361, "train/loss_ctc": 0.4876220226287842, "train/loss_error": 0.3875833749771118, "train/loss_total": 0.4075911045074463 }, { "epoch": 0.9402597402597402, "step": 362, "train/loss_ctc": 0.28839194774627686, "train/loss_error": 0.2801903486251831, "train/loss_total": 0.28183066844940186 }, { "epoch": 0.9428571428571428, "step": 363, "train/loss_ctc": 0.4500102400779724, "train/loss_error": 0.372587651014328, "train/loss_total": 0.3880721926689148 }, { "epoch": 0.9454545454545454, "step": 364, "train/loss_ctc": 0.5426642894744873, "train/loss_error": 0.39797243475914, "train/loss_total": 0.4269108176231384 }, { "epoch": 0.948051948051948, "step": 365, "train/loss_ctc": 0.40926483273506165, "train/loss_error": 0.396403044462204, "train/loss_total": 0.3989754021167755 }, { "epoch": 0.9506493506493506, "step": 366, "train/loss_ctc": 0.30025699734687805, "train/loss_error": 0.3442377746105194, "train/loss_total": 0.3354416489601135 }, { "epoch": 0.9532467532467532, "step": 367, "train/loss_ctc": 0.507075309753418, "train/loss_error": 0.3636031746864319, "train/loss_total": 0.392297625541687 }, { "epoch": 0.9558441558441558, "step": 368, "train/loss_ctc": 0.5480566620826721, "train/loss_error": 0.43843328952789307, "train/loss_total": 0.4603579640388489 }, { "epoch": 0.9584415584415584, "step": 369, "train/loss_ctc": 0.32788485288619995, "train/loss_error": 0.3922174572944641, "train/loss_total": 0.3793509304523468 }, { "epoch": 0.961038961038961, "grad_norm": 21662.490234375, "learning_rate": 2.4264935064935065e-05, "loss": 0.3825, "step": 370 }, { "epoch": 0.961038961038961, "step": 370, "train/loss_ctc": 0.4804488718509674, "train/loss_error": 0.36086201667785645, "train/loss_total": 0.3847793936729431 }, { "epoch": 0.9636363636363636, "step": 371, "train/loss_ctc": 0.37824827432632446, "train/loss_error": 0.36399510502815247, "train/loss_total": 0.3668457269668579 }, { "epoch": 0.9662337662337662, "step": 372, "train/loss_ctc": 0.34025055170059204, "train/loss_error": 0.33508768677711487, "train/loss_total": 0.33612027764320374 }, { "epoch": 0.9688311688311688, "step": 373, "train/loss_ctc": 0.45758840441703796, "train/loss_error": 0.2792995572090149, "train/loss_total": 0.31495732069015503 }, { "epoch": 0.9714285714285714, "step": 374, "train/loss_ctc": 0.49011343717575073, "train/loss_error": 0.40441903471946716, "train/loss_total": 0.4215579330921173 }, { "epoch": 0.974025974025974, "step": 375, "train/loss_ctc": 0.3895321488380432, "train/loss_error": 0.33829614520072937, "train/loss_total": 0.34854334592819214 }, { "epoch": 0.9766233766233766, "step": 376, "train/loss_ctc": 0.5740460157394409, "train/loss_error": 0.37587371468544006, "train/loss_total": 0.4155081808567047 }, { "epoch": 0.9792207792207792, "step": 377, "train/loss_ctc": 0.3161547780036926, "train/loss_error": 0.31022292375564575, "train/loss_total": 0.3114092946052551 }, { "epoch": 0.9818181818181818, "step": 378, "train/loss_ctc": 0.46037817001342773, "train/loss_error": 0.35867980122566223, "train/loss_total": 0.37901949882507324 }, { "epoch": 0.9844155844155844, "step": 379, "train/loss_ctc": 0.43379488587379456, "train/loss_error": 0.4301861524581909, "train/loss_total": 0.4309079051017761 }, { "epoch": 0.987012987012987, "grad_norm": 27087.23828125, "learning_rate": 2.410909090909091e-05, "loss": 0.371, "step": 380 }, { "epoch": 0.987012987012987, "step": 380, "train/loss_ctc": 0.34591156244277954, "train/loss_error": 0.30305615067481995, "train/loss_total": 0.31162723898887634 }, { "epoch": 0.9896103896103896, "step": 381, "train/loss_ctc": 0.49189668893814087, "train/loss_error": 0.3612860441207886, "train/loss_total": 0.38740819692611694 }, { "epoch": 0.9922077922077922, "step": 382, "train/loss_ctc": 0.39670222997665405, "train/loss_error": 0.4762187600135803, "train/loss_total": 0.460315465927124 }, { "epoch": 0.9948051948051948, "step": 383, "train/loss_ctc": 0.3331208825111389, "train/loss_error": 0.37567001581192017, "train/loss_total": 0.36716020107269287 }, { "epoch": 0.9974025974025974, "step": 384, "train/loss_ctc": 0.4160964787006378, "train/loss_error": 0.3856041729450226, "train/loss_total": 0.3917026221752167 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.6875490546226501, "val/loss_error": 0.15470756590366364, "val/loss_total": 0.26127585768699646 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.5419287085533142, "val/loss_error": 0.3653072118759155, "val/loss_total": 0.40063151717185974 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.10468504577875137, "val/loss_error": 0.25799596309661865, "val/loss_total": 0.22733378410339355 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.21671800315380096, "val/loss_error": 0.1281515210866928, "val/loss_total": 0.1458648145198822 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.14890135824680328, "val/loss_error": 0.20448344945907593, "val/loss_total": 0.19336703419685364 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.11583980172872543, "val/loss_error": 0.26794567704200745, "val/loss_total": 0.23752450942993164 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.19505809247493744, "val/loss_error": 0.36295086145401, "val/loss_total": 0.3293723165988922 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.3307742178440094, "val/loss_error": 0.28392457962036133, "val/loss_total": 0.2932945191860199 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.3851718306541443, "val/loss_error": 0.310342937707901, "val/loss_total": 0.32530874013900757 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.3180360496044159, "val/loss_error": 0.3122689425945282, "val/loss_total": 0.3134223520755768 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.13046607375144958, "val/loss_error": 0.13045719265937805, "val/loss_total": 0.1304589807987213 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.2529490292072296, "val/loss_error": 0.30723270773887634, "val/loss_total": 0.29637598991394043 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.8228862881660461, "val/loss_error": 0.40968477725982666, "val/loss_total": 0.4923250675201416 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.8182926177978516, "val/loss_error": 0.5661148428916931, "val/loss_total": 0.6165504455566406 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.5859315395355225, "val/loss_error": 0.42359596490859985, "val/loss_total": 0.45606309175491333 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.6527995467185974, "val/loss_error": 0.6376646161079407, "val/loss_total": 0.6406916379928589 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.07840032875537872, "val/loss_error": 0.17165622115135193, "val/loss_total": 0.15300503373146057 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.1327848583459854, "val/loss_error": 0.2390715628862381, "val/loss_total": 0.21781422197818756 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.10864680260419846, "val/loss_error": 0.2764890193939209, "val/loss_total": 0.24292057752609253 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.6803518533706665, "val/loss_error": 0.3778061270713806, "val/loss_total": 0.4383152723312378 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.9094017744064331, "val/loss_error": 0.5752231478691101, "val/loss_total": 0.6420588493347168 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.3002661466598511, "val/loss_error": 0.5806401968002319, "val/loss_total": 0.5245653986930847 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.35945963859558105, "val/loss_error": 0.3856211304664612, "val/loss_total": 0.38038885593414307 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.17952077090740204, "val/loss_error": 0.16836290061473846, "val/loss_total": 0.1705944836139679 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.28724151849746704, "val/loss_error": 0.5540834069252014, "val/loss_total": 0.5007150173187256 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.4934302270412445, "val/loss_error": 0.4774593114852905, "val/loss_total": 0.4806534945964813 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.2901739776134491, "val/loss_error": 0.329725980758667, "val/loss_total": 0.3218156099319458 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.20987029373645782, "val/loss_error": 0.30417105555534363, "val/loss_total": 0.28531089425086975 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.026751656085252762, "val/loss_error": 0.07376936078071594, "val/loss_total": 0.06436581909656525 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.27957189083099365, "val/loss_error": 0.49581557512283325, "val/loss_total": 0.45256683230400085 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.10496821254491806, "val/loss_error": 0.11273551732301712, "val/loss_total": 0.11118205636739731 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.1278497278690338, "val/loss_error": 0.2850320041179657, "val/loss_total": 0.2535955309867859 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.11460237205028534, "val/loss_error": 0.17554089426994324, "val/loss_total": 0.16335318982601166 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.632587194442749, "val/loss_error": 0.6077760457992554, "val/loss_total": 0.6127382516860962 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.09339753538370132, "val/loss_error": 0.28031614422798157, "val/loss_total": 0.24293243885040283 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.18964652717113495, "val/loss_error": 0.3016185760498047, "val/loss_total": 0.279224157333374 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.5150977373123169, "val/loss_error": 0.5472825765609741, "val/loss_total": 0.5408456325531006 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.431504487991333, "val/loss_error": 0.39616668224334717, "val/loss_total": 0.4032342731952667 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.2717972695827484, "val/loss_error": 0.48320531845092773, "val/loss_total": 0.4409237205982208 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.6259618997573853, "val/loss_error": 0.644675076007843, "val/loss_total": 0.6409325003623962 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.3001953363418579, "val/loss_error": 0.2561388909816742, "val/loss_total": 0.2649501860141754 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.2698942720890045, "val/loss_error": 0.24625356495380402, "val/loss_total": 0.25098171830177307 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.48462846875190735, "val/loss_error": 0.898536205291748, "val/loss_total": 0.8157546520233154 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.6518195867538452, "val/loss_error": 0.5889329314231873, "val/loss_total": 0.6015102863311768 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.3526431620121002, "val/loss_error": 0.21988694369792938, "val/loss_total": 0.24643820524215698 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.39250171184539795, "val/loss_error": 0.6564154028892517, "val/loss_total": 0.6036326885223389 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.704780101776123, "val/loss_error": 0.40366241335868835, "val/loss_total": 0.46388596296310425 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.33294856548309326, "val/loss_error": 0.45671191811561584, "val/loss_total": 0.43195924162864685 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.4124176502227783, "val/loss_error": 0.2529188096523285, "val/loss_total": 0.2848185896873474 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.627615749835968, "val/loss_error": 0.7070494890213013, "val/loss_total": 0.6911627650260925 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.3716880977153778, "val/loss_error": 0.899031400680542, "val/loss_total": 0.7935627102851868 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.3868110179901123, "val/loss_error": 0.580732524394989, "val/loss_total": 0.5419481992721558 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.2674768269062042, "val/loss_error": 0.39884820580482483, "val/loss_total": 0.37257394194602966 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.6733666062355042, "val/loss_error": 0.8181118965148926, "val/loss_total": 0.7891628742218018 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.46843478083610535, "val/loss_error": 0.3836829662322998, "val/loss_total": 0.4006333351135254 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.17817817628383636, "val/loss_error": 0.41714513301849365, "val/loss_total": 0.36935174465179443 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.1327437460422516, "val/loss_error": 0.26699137687683105, "val/loss_total": 0.2401418536901474 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.4530733525753021, "val/loss_error": 0.4367244839668274, "val/loss_total": 0.43999427556991577 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.5494645237922668, "val/loss_error": 0.6103328466415405, "val/loss_total": 0.5981591939926147 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.03713636472821236, "val/loss_error": 0.10577840358018875, "val/loss_total": 0.09205000102519989 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.8317826390266418, "val/loss_error": 0.6528055667877197, "val/loss_total": 0.688601016998291 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.5797836184501648, "val/loss_error": 0.38235223293304443, "val/loss_total": 0.42183852195739746 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.04070484638214111, "val/loss_error": 0.07236619293689728, "val/loss_total": 0.06603392213582993 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.18936409056186676, "val/loss_error": 0.28935083746910095, "val/loss_total": 0.2693534791469574 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.17685364186763763, "val/loss_error": 0.20053841173648834, "val/loss_total": 0.19580146670341492 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.18646998703479767, "val/loss_error": 0.36401763558387756, "val/loss_total": 0.3285081088542938 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.28441548347473145, "val/loss_error": 0.26875388622283936, "val/loss_total": 0.2718862295150757 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.5324382185935974, "val/loss_error": 0.5479824542999268, "val/loss_total": 0.5448735952377319 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.4895276427268982, "val/loss_error": 0.42573779821395874, "val/loss_total": 0.4384957551956177 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.4607556164264679, "val/loss_error": 0.4936349093914032, "val/loss_total": 0.4870590567588806 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.243263840675354, "val/loss_error": 0.25895872712135315, "val/loss_total": 0.25581973791122437 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.6913214325904846, "val/loss_error": 0.5270300507545471, "val/loss_total": 0.5598883032798767 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.4889543354511261, "val/loss_error": 0.6071286201477051, "val/loss_total": 0.5834937691688538 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.22779910266399384, "val/loss_error": 0.33982712030410767, "val/loss_total": 0.3174215257167816 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.20680837333202362, "val/loss_error": 0.23858439922332764, "val/loss_total": 0.23222920298576355 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.05276520550251007, "val/loss_error": 0.2704625427722931, "val/loss_total": 0.22692307829856873 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.8108874559402466, "val/loss_error": 0.455204576253891, "val/loss_total": 0.5263411402702332 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.654788613319397, "val/loss_error": 0.2736782729625702, "val/loss_total": 0.34990036487579346 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.6034975051879883, "val/loss_error": 0.3079814016819, "val/loss_total": 0.3670846223831177 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.005365017335861921, "val/loss_error": 0.37414878606796265, "val/loss_total": 0.3003920316696167 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.42052727937698364, "val/loss_error": 0.28148624300956726, "val/loss_total": 0.3092944622039795 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.049528926610946655, "val/loss_error": 0.0933159813284874, "val/loss_total": 0.08455856889486313 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.24353942275047302, "val/loss_error": 0.2647419273853302, "val/loss_total": 0.2605014145374298 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.3650975525379181, "val/loss_error": 0.3202224373817444, "val/loss_total": 0.3291974663734436 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.3358005881309509, "val/loss_error": 0.5321823358535767, "val/loss_total": 0.49290600419044495 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.5018070340156555, "val/loss_error": 0.5208809971809387, "val/loss_total": 0.517066240310669 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.06023162230849266, "val/loss_error": 0.13020570576190948, "val/loss_total": 0.11621088534593582 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.4284183084964752, "val/loss_error": 0.4465292990207672, "val/loss_total": 0.4429071247577667 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.29431572556495667, "val/loss_error": 0.2671334743499756, "val/loss_total": 0.2725699245929718 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.38660627603530884, "val/loss_error": 0.4068857431411743, "val/loss_total": 0.4028298556804657 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.2624673545360565, "val/loss_error": 0.4182335436344147, "val/loss_total": 0.3870803117752075 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.9632376432418823, "val/loss_error": 0.5537149310112, "val/loss_total": 0.6356194615364075 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.21505296230316162, "val/loss_error": 0.08917678892612457, "val/loss_total": 0.1143520250916481 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.40243232250213623, "val/loss_error": 0.28110232949256897, "val/loss_total": 0.3053683340549469 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.6340964436531067, "val/loss_error": 0.5629323720932007, "val/loss_total": 0.5771651864051819 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.1914648860692978, "val/loss_error": 0.20905034244060516, "val/loss_total": 0.2055332511663437 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.1390804946422577, "val/loss_error": 0.3153555393218994, "val/loss_total": 0.2801005244255066 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.662627637386322, "val/loss_error": 0.5256022810935974, "val/loss_total": 0.5530073642730713 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.07821541279554367, "val/loss_error": 0.32296571135520935, "val/loss_total": 0.2740156650543213 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.19146975874900818, "val/loss_error": 0.24724963307380676, "val/loss_total": 0.236093670129776 }, { "epoch": 1.0, "eval_eval/f1_0": 0.5206185579299927, "eval_eval/f1_1": 0.9346221685409546, "eval_eval/precision_0": 0.5976331233978271, "eval_eval/precision_1": 0.9184801578521729, "eval_eval/recall_0": 0.4611872136592865, "eval_eval/recall_1": 0.9513416886329651, "eval_eval/wer": 0.1524587762434874, "eval_loss": 0.3744491636753082, "eval_runtime": 7.3445, "eval_samples_per_second": 13.616, "eval_steps_per_second": 13.616, "step": 385 }, { "epoch": 1.0, "step": 385, "train/loss_ctc": 0.5853636264801025, "train/loss_error": 0.40152978897094727, "train/loss_total": 0.4382965564727783 }, { "epoch": 1.0025974025974025, "step": 386, "train/loss_ctc": 0.3838910460472107, "train/loss_error": 0.3613385260105133, "train/loss_total": 0.36584901809692383 }, { "epoch": 1.0051948051948052, "step": 387, "train/loss_ctc": 0.43483930826187134, "train/loss_error": 0.3084992468357086, "train/loss_total": 0.33376726508140564 }, { "epoch": 1.0077922077922077, "step": 388, "train/loss_ctc": 0.49927592277526855, "train/loss_error": 0.40097248554229736, "train/loss_total": 0.4206331670284271 }, { "epoch": 1.0103896103896104, "step": 389, "train/loss_ctc": 0.3926433324813843, "train/loss_error": 0.2970089018344879, "train/loss_total": 0.31613579392433167 }, { "epoch": 1.0129870129870129, "grad_norm": 33701.703125, "learning_rate": 2.395324675324675e-05, "loss": 0.3793, "step": 390 }, { "epoch": 1.0129870129870129, "step": 390, "train/loss_ctc": 0.6314498782157898, "train/loss_error": 0.3913571238517761, "train/loss_total": 0.43937569856643677 }, { "epoch": 1.0155844155844156, "step": 391, "train/loss_ctc": 0.4459434747695923, "train/loss_error": 0.3293592631816864, "train/loss_total": 0.352676123380661 }, { "epoch": 1.018181818181818, "step": 392, "train/loss_ctc": 0.5169634222984314, "train/loss_error": 0.3083614706993103, "train/loss_total": 0.3500818610191345 }, { "epoch": 1.0207792207792208, "step": 393, "train/loss_ctc": 0.8963823318481445, "train/loss_error": 0.39474624395370483, "train/loss_total": 0.49507346749305725 }, { "epoch": 1.0233766233766233, "step": 394, "train/loss_ctc": 0.35786834359169006, "train/loss_error": 0.3163049519062042, "train/loss_total": 0.3246176540851593 }, { "epoch": 1.025974025974026, "step": 395, "train/loss_ctc": 0.44512176513671875, "train/loss_error": 0.4150662422180176, "train/loss_total": 0.4210773706436157 }, { "epoch": 1.0285714285714285, "step": 396, "train/loss_ctc": 0.5392573475837708, "train/loss_error": 0.31057047843933105, "train/loss_total": 0.35630786418914795 }, { "epoch": 1.0311688311688312, "step": 397, "train/loss_ctc": 0.4054194688796997, "train/loss_error": 0.37014514207839966, "train/loss_total": 0.37720000743865967 }, { "epoch": 1.0337662337662337, "step": 398, "train/loss_ctc": 0.5018097162246704, "train/loss_error": 0.36128222942352295, "train/loss_total": 0.38938772678375244 }, { "epoch": 1.0363636363636364, "step": 399, "train/loss_ctc": 0.4913959503173828, "train/loss_error": 0.37161457538604736, "train/loss_total": 0.39557087421417236 }, { "epoch": 1.0389610389610389, "grad_norm": 21996.34765625, "learning_rate": 2.3797402597402596e-05, "loss": 0.3901, "step": 400 }, { "epoch": 1.0389610389610389, "step": 400, "train/loss_ctc": 0.27533596754074097, "train/loss_error": 0.2966810464859009, "train/loss_total": 0.29241204261779785 }, { "epoch": 1.0415584415584416, "step": 401, "train/loss_ctc": 0.4445362985134125, "train/loss_error": 0.33218109607696533, "train/loss_total": 0.35465216636657715 }, { "epoch": 1.044155844155844, "step": 402, "train/loss_ctc": 0.3768073618412018, "train/loss_error": 0.2952827215194702, "train/loss_total": 0.3115876615047455 }, { "epoch": 1.0467532467532468, "step": 403, "train/loss_ctc": 0.42005956172943115, "train/loss_error": 0.3912069499492645, "train/loss_total": 0.3969774842262268 }, { "epoch": 1.0493506493506493, "step": 404, "train/loss_ctc": 0.5692473649978638, "train/loss_error": 0.36427152156829834, "train/loss_total": 0.4052667021751404 }, { "epoch": 1.051948051948052, "step": 405, "train/loss_ctc": 0.4374919533729553, "train/loss_error": 0.37530317902565, "train/loss_total": 0.38774093985557556 }, { "epoch": 1.0545454545454545, "step": 406, "train/loss_ctc": 0.25539690256118774, "train/loss_error": 0.31664639711380005, "train/loss_total": 0.30439651012420654 }, { "epoch": 1.0571428571428572, "step": 407, "train/loss_ctc": 0.35246095061302185, "train/loss_error": 0.3290826678276062, "train/loss_total": 0.3337583541870117 }, { "epoch": 1.0597402597402596, "step": 408, "train/loss_ctc": 0.36511826515197754, "train/loss_error": 0.3711419999599457, "train/loss_total": 0.3699372410774231 }, { "epoch": 1.0623376623376624, "step": 409, "train/loss_ctc": 0.5860915184020996, "train/loss_error": 0.3433401882648468, "train/loss_total": 0.3918904662132263 }, { "epoch": 1.0649350649350648, "grad_norm": 33554.17578125, "learning_rate": 2.364155844155844e-05, "loss": 0.3549, "step": 410 }, { "epoch": 1.0649350649350648, "step": 410, "train/loss_ctc": 0.37539732456207275, "train/loss_error": 0.36330845952033997, "train/loss_total": 0.3657262325286865 }, { "epoch": 1.0675324675324676, "step": 411, "train/loss_ctc": 0.4544655680656433, "train/loss_error": 0.323740690946579, "train/loss_total": 0.3498856723308563 }, { "epoch": 1.07012987012987, "step": 412, "train/loss_ctc": 0.7055639028549194, "train/loss_error": 0.4001661241054535, "train/loss_total": 0.46124571561813354 }, { "epoch": 1.0727272727272728, "step": 413, "train/loss_ctc": 0.4977979063987732, "train/loss_error": 0.35371020436286926, "train/loss_total": 0.38252773880958557 }, { "epoch": 1.0753246753246752, "step": 414, "train/loss_ctc": 0.2634584307670593, "train/loss_error": 0.3117663860321045, "train/loss_total": 0.30210480093955994 }, { "epoch": 1.077922077922078, "step": 415, "train/loss_ctc": 0.521848201751709, "train/loss_error": 0.3934501111507416, "train/loss_total": 0.41912972927093506 }, { "epoch": 1.0805194805194804, "step": 416, "train/loss_ctc": 0.5539393424987793, "train/loss_error": 0.41707977652549744, "train/loss_total": 0.4444516897201538 }, { "epoch": 1.0831168831168831, "step": 417, "train/loss_ctc": 0.3220188617706299, "train/loss_error": 0.3501807153224945, "train/loss_total": 0.3445483446121216 }, { "epoch": 1.0857142857142856, "step": 418, "train/loss_ctc": 1.2741672992706299, "train/loss_error": 0.5199360847473145, "train/loss_total": 0.6707823276519775 }, { "epoch": 1.0883116883116883, "step": 419, "train/loss_ctc": 0.4512460231781006, "train/loss_error": 0.3597969114780426, "train/loss_total": 0.37808674573898315 }, { "epoch": 1.0909090909090908, "grad_norm": 22130.96875, "learning_rate": 2.3485714285714285e-05, "loss": 0.4118, "step": 420 }, { "epoch": 1.0909090909090908, "step": 420, "train/loss_ctc": 0.4989060163497925, "train/loss_error": 0.3529156744480133, "train/loss_total": 0.3821137547492981 }, { "epoch": 1.0935064935064935, "step": 421, "train/loss_ctc": 0.6273645758628845, "train/loss_error": 0.3806304335594177, "train/loss_total": 0.42997729778289795 }, { "epoch": 1.096103896103896, "step": 422, "train/loss_ctc": 0.37419813871383667, "train/loss_error": 0.36738550662994385, "train/loss_total": 0.3687480390071869 }, { "epoch": 1.0987012987012987, "step": 423, "train/loss_ctc": 0.46581411361694336, "train/loss_error": 0.3733687996864319, "train/loss_total": 0.39185789227485657 }, { "epoch": 1.1012987012987012, "step": 424, "train/loss_ctc": 0.5627496838569641, "train/loss_error": 0.4291229546070099, "train/loss_total": 0.4558483064174652 }, { "epoch": 1.103896103896104, "step": 425, "train/loss_ctc": 0.745029091835022, "train/loss_error": 0.3746761381626129, "train/loss_total": 0.4487467408180237 }, { "epoch": 1.1064935064935064, "step": 426, "train/loss_ctc": 0.4350806772708893, "train/loss_error": 0.3983886241912842, "train/loss_total": 0.4057270288467407 }, { "epoch": 1.1090909090909091, "step": 427, "train/loss_ctc": 0.42540454864501953, "train/loss_error": 0.3679608702659607, "train/loss_total": 0.37944963574409485 }, { "epoch": 1.1116883116883116, "step": 428, "train/loss_ctc": 0.5110723376274109, "train/loss_error": 0.43649592995643616, "train/loss_total": 0.45141124725341797 }, { "epoch": 1.1142857142857143, "step": 429, "train/loss_ctc": 0.3751834034919739, "train/loss_error": 0.34227848052978516, "train/loss_total": 0.3488594591617584 }, { "epoch": 1.1168831168831168, "grad_norm": 19577.427734375, "learning_rate": 2.332987012987013e-05, "loss": 0.4063, "step": 430 }, { "epoch": 1.1168831168831168, "step": 430, "train/loss_ctc": 0.624438464641571, "train/loss_error": 0.3240596652030945, "train/loss_total": 0.3841354548931122 }, { "epoch": 1.1194805194805195, "step": 431, "train/loss_ctc": 0.33847981691360474, "train/loss_error": 0.3463086485862732, "train/loss_total": 0.34474289417266846 }, { "epoch": 1.122077922077922, "step": 432, "train/loss_ctc": 0.6059143543243408, "train/loss_error": 0.3977591097354889, "train/loss_total": 0.4393901824951172 }, { "epoch": 1.1246753246753247, "step": 433, "train/loss_ctc": 0.5785468816757202, "train/loss_error": 0.35982558131217957, "train/loss_total": 0.40356987714767456 }, { "epoch": 1.1272727272727272, "step": 434, "train/loss_ctc": 0.228939488530159, "train/loss_error": 0.24698448181152344, "train/loss_total": 0.2433754950761795 }, { "epoch": 1.12987012987013, "step": 435, "train/loss_ctc": 0.48356467485427856, "train/loss_error": 0.3827206492424011, "train/loss_total": 0.4028894603252411 }, { "epoch": 1.1324675324675324, "step": 436, "train/loss_ctc": 0.4323671758174896, "train/loss_error": 0.350084125995636, "train/loss_total": 0.36654072999954224 }, { "epoch": 1.135064935064935, "step": 437, "train/loss_ctc": 0.3744457960128784, "train/loss_error": 0.35455262660980225, "train/loss_total": 0.35853126645088196 }, { "epoch": 1.1376623376623376, "step": 438, "train/loss_ctc": 0.4958599805831909, "train/loss_error": 0.3684372007846832, "train/loss_total": 0.39392176270484924 }, { "epoch": 1.1402597402597403, "step": 439, "train/loss_ctc": 0.5126945972442627, "train/loss_error": 0.3985946476459503, "train/loss_total": 0.42141464352607727 }, { "epoch": 1.1428571428571428, "grad_norm": 29565.8828125, "learning_rate": 2.317402597402597e-05, "loss": 0.3759, "step": 440 }, { "epoch": 1.1428571428571428, "step": 440, "train/loss_ctc": 0.3659365773200989, "train/loss_error": 0.39238762855529785, "train/loss_total": 0.38709741830825806 }, { "epoch": 1.1454545454545455, "step": 441, "train/loss_ctc": 0.5543156266212463, "train/loss_error": 0.390970915555954, "train/loss_total": 0.42363986372947693 }, { "epoch": 1.148051948051948, "step": 442, "train/loss_ctc": 0.45742160081863403, "train/loss_error": 0.3650299906730652, "train/loss_total": 0.3835083246231079 }, { "epoch": 1.1506493506493507, "step": 443, "train/loss_ctc": 0.3457763195037842, "train/loss_error": 0.34668561816215515, "train/loss_total": 0.34650376439094543 }, { "epoch": 1.1532467532467532, "step": 444, "train/loss_ctc": 0.5227066874504089, "train/loss_error": 0.3528929352760315, "train/loss_total": 0.38685569167137146 }, { "epoch": 1.155844155844156, "step": 445, "train/loss_ctc": 0.3539370894432068, "train/loss_error": 0.2445438951253891, "train/loss_total": 0.2664225399494171 }, { "epoch": 1.1584415584415584, "step": 446, "train/loss_ctc": 0.4153331220149994, "train/loss_error": 0.30717116594314575, "train/loss_total": 0.32880356907844543 }, { "epoch": 1.161038961038961, "step": 447, "train/loss_ctc": 0.4397730827331543, "train/loss_error": 0.3533308506011963, "train/loss_total": 0.3706192970275879 }, { "epoch": 1.1636363636363636, "step": 448, "train/loss_ctc": 0.5707885026931763, "train/loss_error": 0.28771892189979553, "train/loss_total": 0.34433284401893616 }, { "epoch": 1.1662337662337663, "step": 449, "train/loss_ctc": 0.6262338161468506, "train/loss_error": 0.37352487444877625, "train/loss_total": 0.4240666627883911 }, { "epoch": 1.1688311688311688, "grad_norm": 32407.126953125, "learning_rate": 2.3018181818181816e-05, "loss": 0.3662, "step": 450 }, { "epoch": 1.1688311688311688, "step": 450, "train/loss_ctc": 0.5456961393356323, "train/loss_error": 0.4354025423526764, "train/loss_total": 0.45746126770973206 }, { "epoch": 1.1714285714285715, "step": 451, "train/loss_ctc": 0.5305813550949097, "train/loss_error": 0.37506288290023804, "train/loss_total": 0.40616658329963684 }, { "epoch": 1.174025974025974, "step": 452, "train/loss_ctc": 0.47689691185951233, "train/loss_error": 0.45003989338874817, "train/loss_total": 0.45541131496429443 }, { "epoch": 1.1766233766233767, "step": 453, "train/loss_ctc": 0.45872074365615845, "train/loss_error": 0.4730827212333679, "train/loss_total": 0.47021034359931946 }, { "epoch": 1.1792207792207792, "step": 454, "train/loss_ctc": 0.5242329835891724, "train/loss_error": 0.4351136088371277, "train/loss_total": 0.4529374837875366 }, { "epoch": 1.1818181818181819, "step": 455, "train/loss_ctc": 0.47314080595970154, "train/loss_error": 0.37163299322128296, "train/loss_total": 0.3919345438480377 }, { "epoch": 1.1844155844155844, "step": 456, "train/loss_ctc": 0.3516017496585846, "train/loss_error": 0.34268245100975037, "train/loss_total": 0.34446632862091064 }, { "epoch": 1.187012987012987, "step": 457, "train/loss_ctc": 0.5192954540252686, "train/loss_error": 0.35079848766326904, "train/loss_total": 0.38449788093566895 }, { "epoch": 1.1896103896103896, "step": 458, "train/loss_ctc": 0.682726263999939, "train/loss_error": 0.30412355065345764, "train/loss_total": 0.3798440992832184 }, { "epoch": 1.1922077922077923, "step": 459, "train/loss_ctc": 0.5977647304534912, "train/loss_error": 0.4832066595554352, "train/loss_total": 0.5061182975769043 }, { "epoch": 1.1948051948051948, "grad_norm": 34430.203125, "learning_rate": 2.2862337662337665e-05, "loss": 0.4249, "step": 460 }, { "epoch": 1.1948051948051948, "step": 460, "train/loss_ctc": 0.40467244386672974, "train/loss_error": 0.36665675044059753, "train/loss_total": 0.374259889125824 }, { "epoch": 1.1974025974025975, "step": 461, "train/loss_ctc": 0.5387927293777466, "train/loss_error": 0.3987988531589508, "train/loss_total": 0.42679762840270996 }, { "epoch": 1.2, "step": 462, "train/loss_ctc": 0.3705289661884308, "train/loss_error": 0.30253341794013977, "train/loss_total": 0.316132515668869 }, { "epoch": 1.2025974025974027, "step": 463, "train/loss_ctc": 0.529025673866272, "train/loss_error": 0.2914603650569916, "train/loss_total": 0.33897343277931213 }, { "epoch": 1.2051948051948052, "step": 464, "train/loss_ctc": 0.9099610447883606, "train/loss_error": 0.3833993971347809, "train/loss_total": 0.48871171474456787 }, { "epoch": 1.2077922077922079, "step": 465, "train/loss_ctc": 0.34250521659851074, "train/loss_error": 0.27111878991127014, "train/loss_total": 0.2853960692882538 }, { "epoch": 1.2103896103896103, "step": 466, "train/loss_ctc": 0.6191755533218384, "train/loss_error": 0.378248006105423, "train/loss_total": 0.4264335334300995 }, { "epoch": 1.212987012987013, "step": 467, "train/loss_ctc": 0.6608357429504395, "train/loss_error": 0.40693846344947815, "train/loss_total": 0.4577178955078125 }, { "epoch": 1.2155844155844155, "step": 468, "train/loss_ctc": 0.5403280258178711, "train/loss_error": 0.40327557921409607, "train/loss_total": 0.4306860864162445 }, { "epoch": 1.2181818181818183, "step": 469, "train/loss_ctc": 0.48625773191452026, "train/loss_error": 0.33293473720550537, "train/loss_total": 0.36359935998916626 }, { "epoch": 1.2207792207792207, "grad_norm": 33204.6953125, "learning_rate": 2.270649350649351e-05, "loss": 0.3909, "step": 470 }, { "epoch": 1.2207792207792207, "step": 470, "train/loss_ctc": 0.400088906288147, "train/loss_error": 0.39315539598464966, "train/loss_total": 0.3945420980453491 }, { "epoch": 1.2233766233766235, "step": 471, "train/loss_ctc": 0.42655935883522034, "train/loss_error": 0.26309487223625183, "train/loss_total": 0.2957877814769745 }, { "epoch": 1.225974025974026, "step": 472, "train/loss_ctc": 0.6024197340011597, "train/loss_error": 0.3574691116809845, "train/loss_total": 0.4064592719078064 }, { "epoch": 1.2285714285714286, "step": 473, "train/loss_ctc": 0.312089204788208, "train/loss_error": 0.2071411907672882, "train/loss_total": 0.2281307876110077 }, { "epoch": 1.2311688311688311, "step": 474, "train/loss_ctc": 0.5025306940078735, "train/loss_error": 0.3801623284816742, "train/loss_total": 0.404636025428772 }, { "epoch": 1.2337662337662338, "step": 475, "train/loss_ctc": 0.306465744972229, "train/loss_error": 0.3091370761394501, "train/loss_total": 0.30860280990600586 }, { "epoch": 1.2363636363636363, "step": 476, "train/loss_ctc": 0.3469736576080322, "train/loss_error": 0.412492960691452, "train/loss_total": 0.3993891179561615 }, { "epoch": 1.238961038961039, "step": 477, "train/loss_ctc": 0.4119279980659485, "train/loss_error": 0.3711227774620056, "train/loss_total": 0.3792838156223297 }, { "epoch": 1.2415584415584415, "step": 478, "train/loss_ctc": 0.28067755699157715, "train/loss_error": 0.2992175221443176, "train/loss_total": 0.29550954699516296 }, { "epoch": 1.2441558441558442, "step": 479, "train/loss_ctc": 0.3895191550254822, "train/loss_error": 0.345146119594574, "train/loss_total": 0.35402074456214905 }, { "epoch": 1.2467532467532467, "grad_norm": 24304.974609375, "learning_rate": 2.2550649350649354e-05, "loss": 0.3466, "step": 480 }, { "epoch": 1.2467532467532467, "step": 480, "train/loss_ctc": 0.5745303630828857, "train/loss_error": 0.31007614731788635, "train/loss_total": 0.36296701431274414 }, { "epoch": 1.2493506493506494, "step": 481, "train/loss_ctc": 0.39912474155426025, "train/loss_error": 0.3434598445892334, "train/loss_total": 0.35459282994270325 }, { "epoch": 1.251948051948052, "step": 482, "train/loss_ctc": 0.32126590609550476, "train/loss_error": 0.32359758019447327, "train/loss_total": 0.323131263256073 }, { "epoch": 1.2545454545454544, "step": 483, "train/loss_ctc": 0.483406662940979, "train/loss_error": 0.3188616633415222, "train/loss_total": 0.35177066922187805 }, { "epoch": 1.2571428571428571, "step": 484, "train/loss_ctc": 0.2783830463886261, "train/loss_error": 0.276475191116333, "train/loss_total": 0.27685678005218506 }, { "epoch": 1.2597402597402598, "step": 485, "train/loss_ctc": 0.3069860339164734, "train/loss_error": 0.3205130994319916, "train/loss_total": 0.317807674407959 }, { "epoch": 1.2623376623376623, "step": 486, "train/loss_ctc": 0.5526745319366455, "train/loss_error": 0.3258383870124817, "train/loss_total": 0.3712056279182434 }, { "epoch": 1.2649350649350648, "step": 487, "train/loss_ctc": 0.47273147106170654, "train/loss_error": 0.3555142879486084, "train/loss_total": 0.37895771861076355 }, { "epoch": 1.2675324675324675, "step": 488, "train/loss_ctc": 0.36916959285736084, "train/loss_error": 0.2824450731277466, "train/loss_total": 0.29978999495506287 }, { "epoch": 1.2701298701298702, "step": 489, "train/loss_ctc": 0.4806148111820221, "train/loss_error": 0.37338319420814514, "train/loss_total": 0.39482951164245605 }, { "epoch": 1.2727272727272727, "grad_norm": 25088.1484375, "learning_rate": 2.2394805194805196e-05, "loss": 0.3432, "step": 490 }, { "epoch": 1.2727272727272727, "step": 490, "train/loss_ctc": 0.41657617688179016, "train/loss_error": 0.28348010778427124, "train/loss_total": 0.310099333524704 }, { "epoch": 1.2753246753246752, "step": 491, "train/loss_ctc": 0.6466580033302307, "train/loss_error": 0.4460374712944031, "train/loss_total": 0.48616158962249756 }, { "epoch": 1.277922077922078, "step": 492, "train/loss_ctc": 0.4244529604911804, "train/loss_error": 0.3845522403717041, "train/loss_total": 0.3925324082374573 }, { "epoch": 1.2805194805194806, "step": 493, "train/loss_ctc": 0.41018080711364746, "train/loss_error": 0.3832576274871826, "train/loss_total": 0.388642281293869 }, { "epoch": 1.283116883116883, "step": 494, "train/loss_ctc": 0.38550931215286255, "train/loss_error": 0.3465186655521393, "train/loss_total": 0.3543168008327484 }, { "epoch": 1.2857142857142856, "step": 495, "train/loss_ctc": 0.41059282422065735, "train/loss_error": 0.2871870994567871, "train/loss_total": 0.31186825037002563 }, { "epoch": 1.2883116883116883, "step": 496, "train/loss_ctc": 0.513845682144165, "train/loss_error": 0.3137076795101166, "train/loss_total": 0.3537352979183197 }, { "epoch": 1.290909090909091, "step": 497, "train/loss_ctc": 0.32651665806770325, "train/loss_error": 0.36932459473609924, "train/loss_total": 0.3607630133628845 }, { "epoch": 1.2935064935064935, "step": 498, "train/loss_ctc": 0.558557391166687, "train/loss_error": 0.363200306892395, "train/loss_total": 0.40227171778678894 }, { "epoch": 1.296103896103896, "step": 499, "train/loss_ctc": 0.352846622467041, "train/loss_error": 0.41289058327674866, "train/loss_total": 0.4008817970752716 }, { "epoch": 1.2987012987012987, "grad_norm": 32329.26953125, "learning_rate": 2.223896103896104e-05, "loss": 0.3761, "step": 500 }, { "epoch": 1.2987012987012987, "step": 500, "train/loss_ctc": 0.5815698504447937, "train/loss_error": 0.4775105118751526, "train/loss_total": 0.49832236766815186 }, { "epoch": 1.3012987012987014, "step": 501, "train/loss_ctc": 0.244430810213089, "train/loss_error": 0.27878981828689575, "train/loss_total": 0.27191802859306335 }, { "epoch": 1.3038961038961039, "step": 502, "train/loss_ctc": 0.4118507504463196, "train/loss_error": 0.37044277787208557, "train/loss_total": 0.3787243962287903 }, { "epoch": 1.3064935064935064, "step": 503, "train/loss_ctc": 0.4818904399871826, "train/loss_error": 0.3755156397819519, "train/loss_total": 0.39679059386253357 }, { "epoch": 1.309090909090909, "step": 504, "train/loss_ctc": 0.5107941031455994, "train/loss_error": 0.3025610148906708, "train/loss_total": 0.34420764446258545 }, { "epoch": 1.3116883116883118, "step": 505, "train/loss_ctc": 0.3046233654022217, "train/loss_error": 0.31914031505584717, "train/loss_total": 0.3162369430065155 }, { "epoch": 1.3142857142857143, "step": 506, "train/loss_ctc": 0.49479371309280396, "train/loss_error": 0.32772451639175415, "train/loss_total": 0.36113834381103516 }, { "epoch": 1.3168831168831168, "step": 507, "train/loss_ctc": 0.3333094120025635, "train/loss_error": 0.2894943654537201, "train/loss_total": 0.29825738072395325 }, { "epoch": 1.3194805194805195, "step": 508, "train/loss_ctc": 0.6401631832122803, "train/loss_error": 0.48176273703575134, "train/loss_total": 0.5134428143501282 }, { "epoch": 1.3220779220779222, "step": 509, "train/loss_ctc": 0.49886763095855713, "train/loss_error": 0.3521420359611511, "train/loss_total": 0.3814871609210968 }, { "epoch": 1.3246753246753247, "grad_norm": 24445.13671875, "learning_rate": 2.2083116883116885e-05, "loss": 0.3761, "step": 510 }, { "epoch": 1.3246753246753247, "step": 510, "train/loss_ctc": 0.526260256767273, "train/loss_error": 0.44229310750961304, "train/loss_total": 0.459086537361145 }, { "epoch": 1.3272727272727272, "step": 511, "train/loss_ctc": 0.31853312253952026, "train/loss_error": 0.28105106949806213, "train/loss_total": 0.28854748606681824 }, { "epoch": 1.3298701298701299, "step": 512, "train/loss_ctc": 0.35999447107315063, "train/loss_error": 0.31769222021102905, "train/loss_total": 0.3261526823043823 }, { "epoch": 1.3324675324675326, "step": 513, "train/loss_ctc": 0.40157201886177063, "train/loss_error": 0.388263463973999, "train/loss_total": 0.39092516899108887 }, { "epoch": 1.335064935064935, "step": 514, "train/loss_ctc": 0.4162537455558777, "train/loss_error": 0.3144802451133728, "train/loss_total": 0.3348349630832672 }, { "epoch": 1.3376623376623376, "step": 515, "train/loss_ctc": 0.3729410171508789, "train/loss_error": 0.3048003613948822, "train/loss_total": 0.31842848658561707 }, { "epoch": 1.3402597402597403, "step": 516, "train/loss_ctc": 0.4520706534385681, "train/loss_error": 0.3405296206474304, "train/loss_total": 0.36283785104751587 }, { "epoch": 1.342857142857143, "step": 517, "train/loss_ctc": 0.7336049675941467, "train/loss_error": 0.48467859625816345, "train/loss_total": 0.5344638824462891 }, { "epoch": 1.3454545454545455, "step": 518, "train/loss_ctc": 0.3689846396446228, "train/loss_error": 0.3327777087688446, "train/loss_total": 0.3400191068649292 }, { "epoch": 1.348051948051948, "step": 519, "train/loss_ctc": 0.3231569528579712, "train/loss_error": 0.27955153584480286, "train/loss_total": 0.2882726192474365 }, { "epoch": 1.3506493506493507, "grad_norm": 20373.326171875, "learning_rate": 2.192727272727273e-05, "loss": 0.3644, "step": 520 }, { "epoch": 1.3506493506493507, "step": 520, "train/loss_ctc": 0.47635966539382935, "train/loss_error": 0.23997639119625092, "train/loss_total": 0.2872530519962311 }, { "epoch": 1.3532467532467534, "step": 521, "train/loss_ctc": 0.6403058767318726, "train/loss_error": 0.40285882353782654, "train/loss_total": 0.45034822821617126 }, { "epoch": 1.3558441558441559, "step": 522, "train/loss_ctc": 0.5145407319068909, "train/loss_error": 0.3287656605243683, "train/loss_total": 0.36592066287994385 }, { "epoch": 1.3584415584415583, "step": 523, "train/loss_ctc": 0.3095550537109375, "train/loss_error": 0.2959955632686615, "train/loss_total": 0.2987074553966522 }, { "epoch": 1.361038961038961, "step": 524, "train/loss_ctc": 0.43043315410614014, "train/loss_error": 0.4221132695674896, "train/loss_total": 0.4237772524356842 }, { "epoch": 1.3636363636363638, "step": 525, "train/loss_ctc": 0.550750732421875, "train/loss_error": 0.3897378146648407, "train/loss_total": 0.421940416097641 }, { "epoch": 1.3662337662337662, "step": 526, "train/loss_ctc": 0.5065853595733643, "train/loss_error": 0.3736878037452698, "train/loss_total": 0.4002673327922821 }, { "epoch": 1.3688311688311687, "step": 527, "train/loss_ctc": 0.5267056226730347, "train/loss_error": 0.3703691363334656, "train/loss_total": 0.4016364514827728 }, { "epoch": 1.3714285714285714, "step": 528, "train/loss_ctc": 0.4137175977230072, "train/loss_error": 0.3647737205028534, "train/loss_total": 0.37456250190734863 }, { "epoch": 1.3740259740259742, "step": 529, "train/loss_ctc": 0.6564739942550659, "train/loss_error": 0.4340324103832245, "train/loss_total": 0.4785207509994507 }, { "epoch": 1.3766233766233766, "grad_norm": 25237.591796875, "learning_rate": 2.177142857142857e-05, "loss": 0.3903, "step": 530 }, { "epoch": 1.3766233766233766, "step": 530, "train/loss_ctc": 0.7766618132591248, "train/loss_error": 0.3917471766471863, "train/loss_total": 0.468730092048645 }, { "epoch": 1.3792207792207791, "step": 531, "train/loss_ctc": 0.4649583697319031, "train/loss_error": 0.3229016065597534, "train/loss_total": 0.3513129651546478 }, { "epoch": 1.3818181818181818, "step": 532, "train/loss_ctc": 0.6808677911758423, "train/loss_error": 0.3627151548862457, "train/loss_total": 0.42634570598602295 }, { "epoch": 1.3844155844155845, "step": 533, "train/loss_ctc": 0.6176801919937134, "train/loss_error": 0.34598028659820557, "train/loss_total": 0.40032029151916504 }, { "epoch": 1.387012987012987, "step": 534, "train/loss_ctc": 0.5112459659576416, "train/loss_error": 0.37046921253204346, "train/loss_total": 0.39862456917762756 }, { "epoch": 1.3896103896103895, "step": 535, "train/loss_ctc": 0.4279436767101288, "train/loss_error": 0.3845837712287903, "train/loss_total": 0.3932557702064514 }, { "epoch": 1.3922077922077922, "step": 536, "train/loss_ctc": 0.5026466250419617, "train/loss_error": 0.4199739098548889, "train/loss_total": 0.4365084767341614 }, { "epoch": 1.394805194805195, "step": 537, "train/loss_ctc": 0.4348308742046356, "train/loss_error": 0.3621087074279785, "train/loss_total": 0.37665316462516785 }, { "epoch": 1.3974025974025974, "step": 538, "train/loss_ctc": 0.7982702255249023, "train/loss_error": 0.36673709750175476, "train/loss_total": 0.45304372906684875 }, { "epoch": 1.4, "step": 539, "train/loss_ctc": 0.43917855620384216, "train/loss_error": 0.4644852578639984, "train/loss_total": 0.45942389965057373 }, { "epoch": 1.4025974025974026, "grad_norm": 25821.44140625, "learning_rate": 2.1615584415584416e-05, "loss": 0.4164, "step": 540 }, { "epoch": 1.4025974025974026, "step": 540, "train/loss_ctc": 0.47974610328674316, "train/loss_error": 0.3568398058414459, "train/loss_total": 0.3814210891723633 }, { "epoch": 1.4051948051948053, "step": 541, "train/loss_ctc": 0.6158280372619629, "train/loss_error": 0.44826099276542664, "train/loss_total": 0.4817744195461273 }, { "epoch": 1.4077922077922078, "step": 542, "train/loss_ctc": 0.518024742603302, "train/loss_error": 0.3736872971057892, "train/loss_total": 0.40255478024482727 }, { "epoch": 1.4103896103896103, "step": 543, "train/loss_ctc": 0.5785751342773438, "train/loss_error": 0.42919108271598816, "train/loss_total": 0.4590679109096527 }, { "epoch": 1.412987012987013, "step": 544, "train/loss_ctc": 0.34806323051452637, "train/loss_error": 0.2766466736793518, "train/loss_total": 0.29092997312545776 }, { "epoch": 1.4155844155844157, "step": 545, "train/loss_ctc": 0.2571503221988678, "train/loss_error": 0.2709127962589264, "train/loss_total": 0.26816031336784363 }, { "epoch": 1.4181818181818182, "step": 546, "train/loss_ctc": 0.42751842737197876, "train/loss_error": 0.3501787781715393, "train/loss_total": 0.36564671993255615 }, { "epoch": 1.4207792207792207, "step": 547, "train/loss_ctc": 0.297592431306839, "train/loss_error": 0.2511996030807495, "train/loss_total": 0.2604781687259674 }, { "epoch": 1.4233766233766234, "step": 548, "train/loss_ctc": 0.44350796937942505, "train/loss_error": 0.3439272940158844, "train/loss_total": 0.3638434410095215 }, { "epoch": 1.425974025974026, "step": 549, "train/loss_ctc": 0.6448813080787659, "train/loss_error": 0.3039188086986542, "train/loss_total": 0.37211132049560547 }, { "epoch": 1.4285714285714286, "grad_norm": 39503.76171875, "learning_rate": 2.145974025974026e-05, "loss": 0.3646, "step": 550 }, { "epoch": 1.4285714285714286, "step": 550, "train/loss_ctc": 0.2745121717453003, "train/loss_error": 0.3776659667491913, "train/loss_total": 0.35703521966934204 }, { "epoch": 1.431168831168831, "step": 551, "train/loss_ctc": 0.3867128789424896, "train/loss_error": 0.3563111126422882, "train/loss_total": 0.36239147186279297 }, { "epoch": 1.4337662337662338, "step": 552, "train/loss_ctc": 0.48125898838043213, "train/loss_error": 0.26771676540374756, "train/loss_total": 0.3104252219200134 }, { "epoch": 1.4363636363636363, "step": 553, "train/loss_ctc": 0.37709349393844604, "train/loss_error": 0.3537208139896393, "train/loss_total": 0.35839536786079407 }, { "epoch": 1.438961038961039, "step": 554, "train/loss_ctc": 0.8853598237037659, "train/loss_error": 0.4737718999385834, "train/loss_total": 0.5560895204544067 }, { "epoch": 1.4415584415584415, "step": 555, "train/loss_ctc": 0.44846397638320923, "train/loss_error": 0.3169558644294739, "train/loss_total": 0.34325748682022095 }, { "epoch": 1.4441558441558442, "step": 556, "train/loss_ctc": 0.6139488816261292, "train/loss_error": 0.3545096814632416, "train/loss_total": 0.4063975214958191 }, { "epoch": 1.4467532467532467, "step": 557, "train/loss_ctc": 0.46743327379226685, "train/loss_error": 0.37020570039749146, "train/loss_total": 0.38965123891830444 }, { "epoch": 1.4493506493506494, "step": 558, "train/loss_ctc": 0.3499261140823364, "train/loss_error": 0.3336247205734253, "train/loss_total": 0.3368850350379944 }, { "epoch": 1.4519480519480519, "step": 559, "train/loss_ctc": 0.6725403070449829, "train/loss_error": 0.3633289039134979, "train/loss_total": 0.4251711964607239 }, { "epoch": 1.4545454545454546, "grad_norm": 31417.599609375, "learning_rate": 2.1303896103896106e-05, "loss": 0.3846, "step": 560 }, { "epoch": 1.4545454545454546, "step": 560, "train/loss_ctc": 0.5647168755531311, "train/loss_error": 0.3771699368953705, "train/loss_total": 0.4146793484687805 }, { "epoch": 1.457142857142857, "step": 561, "train/loss_ctc": 0.6764385104179382, "train/loss_error": 0.4546443521976471, "train/loss_total": 0.49900320172309875 }, { "epoch": 1.4597402597402598, "step": 562, "train/loss_ctc": 0.45636487007141113, "train/loss_error": 0.32776275277137756, "train/loss_total": 0.3534832000732422 }, { "epoch": 1.4623376623376623, "step": 563, "train/loss_ctc": 0.40439745783805847, "train/loss_error": 0.352735698223114, "train/loss_total": 0.3630680441856384 }, { "epoch": 1.464935064935065, "step": 564, "train/loss_ctc": 0.43290746212005615, "train/loss_error": 0.32331642508506775, "train/loss_total": 0.34523463249206543 }, { "epoch": 1.4675324675324675, "step": 565, "train/loss_ctc": 0.6041193008422852, "train/loss_error": 0.33908140659332275, "train/loss_total": 0.39208897948265076 }, { "epoch": 1.4701298701298702, "step": 566, "train/loss_ctc": 0.3653728663921356, "train/loss_error": 0.19342471659183502, "train/loss_total": 0.22781434655189514 }, { "epoch": 1.4727272727272727, "step": 567, "train/loss_ctc": 0.3679927587509155, "train/loss_error": 0.3356268107891083, "train/loss_total": 0.34210002422332764 }, { "epoch": 1.4753246753246754, "step": 568, "train/loss_ctc": 0.3203829526901245, "train/loss_error": 0.3294046223163605, "train/loss_total": 0.32760030031204224 }, { "epoch": 1.4779220779220779, "step": 569, "train/loss_ctc": 0.2808234393596649, "train/loss_error": 0.29980579018592834, "train/loss_total": 0.2960093319416046 }, { "epoch": 1.4805194805194806, "grad_norm": 21160.8828125, "learning_rate": 2.114805194805195e-05, "loss": 0.3561, "step": 570 }, { "epoch": 1.4805194805194806, "step": 570, "train/loss_ctc": 0.4623369574546814, "train/loss_error": 0.3506431579589844, "train/loss_total": 0.3729819357395172 }, { "epoch": 1.483116883116883, "step": 571, "train/loss_ctc": 0.45957016944885254, "train/loss_error": 0.3984665870666504, "train/loss_total": 0.41068729758262634 }, { "epoch": 1.4857142857142858, "step": 572, "train/loss_ctc": 0.38131338357925415, "train/loss_error": 0.3012106716632843, "train/loss_total": 0.3172312080860138 }, { "epoch": 1.4883116883116883, "step": 573, "train/loss_ctc": 0.27788984775543213, "train/loss_error": 0.3141897916793823, "train/loss_total": 0.3069297969341278 }, { "epoch": 1.490909090909091, "step": 574, "train/loss_ctc": 0.373485803604126, "train/loss_error": 0.391730934381485, "train/loss_total": 0.3880819082260132 }, { "epoch": 1.4935064935064934, "step": 575, "train/loss_ctc": 0.4064125418663025, "train/loss_error": 0.3797913193702698, "train/loss_total": 0.3851155638694763 }, { "epoch": 1.4961038961038962, "step": 576, "train/loss_ctc": 0.4438096284866333, "train/loss_error": 0.33372628688812256, "train/loss_total": 0.3557429611682892 }, { "epoch": 1.4987012987012986, "step": 577, "train/loss_ctc": 0.3920140862464905, "train/loss_error": 0.3838873505592346, "train/loss_total": 0.38551270961761475 }, { "epoch": 1.5012987012987011, "step": 578, "train/loss_ctc": 0.35915693640708923, "train/loss_error": 0.3036881983280182, "train/loss_total": 0.31478196382522583 }, { "epoch": 1.5038961038961038, "step": 579, "train/loss_ctc": 0.44084012508392334, "train/loss_error": 0.26023170351982117, "train/loss_total": 0.29635339975357056 }, { "epoch": 1.5064935064935066, "grad_norm": 20535.037109375, "learning_rate": 2.0992207792207792e-05, "loss": 0.3533, "step": 580 }, { "epoch": 1.5064935064935066, "step": 580, "train/loss_ctc": 0.36642593145370483, "train/loss_error": 0.30712059140205383, "train/loss_total": 0.3189816474914551 }, { "epoch": 1.509090909090909, "step": 581, "train/loss_ctc": 0.49016109108924866, "train/loss_error": 0.354915052652359, "train/loss_total": 0.3819642663002014 }, { "epoch": 1.5116883116883115, "step": 582, "train/loss_ctc": 0.7985116243362427, "train/loss_error": 0.3054192364215851, "train/loss_total": 0.4040377140045166 }, { "epoch": 1.5142857142857142, "step": 583, "train/loss_ctc": 0.3342759311199188, "train/loss_error": 0.2779410779476166, "train/loss_total": 0.2892080545425415 }, { "epoch": 1.516883116883117, "step": 584, "train/loss_ctc": 0.47025299072265625, "train/loss_error": 0.3738058805465698, "train/loss_total": 0.39309531450271606 }, { "epoch": 1.5194805194805194, "step": 585, "train/loss_ctc": 0.5997492074966431, "train/loss_error": 0.2837756872177124, "train/loss_total": 0.3469703793525696 }, { "epoch": 1.522077922077922, "step": 586, "train/loss_ctc": 0.7511489391326904, "train/loss_error": 0.310947448015213, "train/loss_total": 0.3989877700805664 }, { "epoch": 1.5246753246753246, "step": 587, "train/loss_ctc": 0.28710687160491943, "train/loss_error": 0.36534708738327026, "train/loss_total": 0.3496990501880646 }, { "epoch": 1.5272727272727273, "step": 588, "train/loss_ctc": 0.39148375391960144, "train/loss_error": 0.34730055928230286, "train/loss_total": 0.356137216091156 }, { "epoch": 1.5298701298701298, "step": 589, "train/loss_ctc": 0.4220133125782013, "train/loss_error": 0.34858620166778564, "train/loss_total": 0.36327165365219116 }, { "epoch": 1.5324675324675323, "grad_norm": 22689.46875, "learning_rate": 2.0836363636363637e-05, "loss": 0.3602, "step": 590 }, { "epoch": 1.5324675324675323, "step": 590, "train/loss_ctc": 0.3089958429336548, "train/loss_error": 0.3162086308002472, "train/loss_total": 0.3147660791873932 }, { "epoch": 1.535064935064935, "step": 591, "train/loss_ctc": 0.44208255410194397, "train/loss_error": 0.36180537939071655, "train/loss_total": 0.37786081433296204 }, { "epoch": 1.5376623376623377, "step": 592, "train/loss_ctc": 0.4225612282752991, "train/loss_error": 0.2630390226840973, "train/loss_total": 0.2949434518814087 }, { "epoch": 1.5402597402597402, "step": 593, "train/loss_ctc": 0.3356591761112213, "train/loss_error": 0.3646678328514099, "train/loss_total": 0.3588661253452301 }, { "epoch": 1.5428571428571427, "step": 594, "train/loss_ctc": 0.4986133575439453, "train/loss_error": 0.3737378418445587, "train/loss_total": 0.39871296286582947 }, { "epoch": 1.5454545454545454, "step": 595, "train/loss_ctc": 0.4947509765625, "train/loss_error": 0.3107672333717346, "train/loss_total": 0.3475639820098877 }, { "epoch": 1.5480519480519481, "step": 596, "train/loss_ctc": 0.46357569098472595, "train/loss_error": 0.3321352005004883, "train/loss_total": 0.3584233224391937 }, { "epoch": 1.5506493506493506, "step": 597, "train/loss_ctc": 0.5350378155708313, "train/loss_error": 0.3528476655483246, "train/loss_total": 0.38928571343421936 }, { "epoch": 1.553246753246753, "step": 598, "train/loss_ctc": 0.6011439561843872, "train/loss_error": 0.3995976150035858, "train/loss_total": 0.43990689516067505 }, { "epoch": 1.5558441558441558, "step": 599, "train/loss_ctc": 0.3365955352783203, "train/loss_error": 0.31113842129707336, "train/loss_total": 0.31622985005378723 }, { "epoch": 1.5584415584415585, "grad_norm": 22577.5859375, "learning_rate": 2.0680519480519482e-05, "loss": 0.3597, "step": 600 }, { "epoch": 1.5584415584415585, "step": 600, "train/loss_ctc": 0.4804026484489441, "train/loss_error": 0.44256842136383057, "train/loss_total": 0.4501352906227112 }, { "epoch": 1.561038961038961, "step": 601, "train/loss_ctc": 0.3501112461090088, "train/loss_error": 0.3687506318092346, "train/loss_total": 0.36502277851104736 }, { "epoch": 1.5636363636363635, "step": 602, "train/loss_ctc": 0.4862782657146454, "train/loss_error": 0.40138745307922363, "train/loss_total": 0.41836562752723694 }, { "epoch": 1.5662337662337662, "step": 603, "train/loss_ctc": 0.47905147075653076, "train/loss_error": 0.3943447768688202, "train/loss_total": 0.4112861156463623 }, { "epoch": 1.568831168831169, "step": 604, "train/loss_ctc": 0.38971611857414246, "train/loss_error": 0.32447031140327454, "train/loss_total": 0.33751949667930603 }, { "epoch": 1.5714285714285714, "step": 605, "train/loss_ctc": 0.5850565433502197, "train/loss_error": 0.4602178633213043, "train/loss_total": 0.4851855933666229 }, { "epoch": 1.5740259740259739, "step": 606, "train/loss_ctc": 0.6147217750549316, "train/loss_error": 0.28961434960365295, "train/loss_total": 0.3546358346939087 }, { "epoch": 1.5766233766233766, "step": 607, "train/loss_ctc": 0.4593358635902405, "train/loss_error": 0.3522673547267914, "train/loss_total": 0.37368106842041016 }, { "epoch": 1.5792207792207793, "step": 608, "train/loss_ctc": 0.37586304545402527, "train/loss_error": 0.32533371448516846, "train/loss_total": 0.3354395925998688 }, { "epoch": 1.5818181818181818, "step": 609, "train/loss_ctc": 0.4638831615447998, "train/loss_error": 0.3761054575443268, "train/loss_total": 0.3936609923839569 }, { "epoch": 1.5844155844155843, "grad_norm": 33480.9921875, "learning_rate": 2.0524675324675327e-05, "loss": 0.3925, "step": 610 }, { "epoch": 1.5844155844155843, "step": 610, "train/loss_ctc": 0.5266464948654175, "train/loss_error": 0.37042611837387085, "train/loss_total": 0.4016702175140381 }, { "epoch": 1.587012987012987, "step": 611, "train/loss_ctc": 0.39771825075149536, "train/loss_error": 0.3900771141052246, "train/loss_total": 0.39160534739494324 }, { "epoch": 1.5896103896103897, "step": 612, "train/loss_ctc": 0.29159677028656006, "train/loss_error": 0.2906040549278259, "train/loss_total": 0.29080259799957275 }, { "epoch": 1.5922077922077922, "step": 613, "train/loss_ctc": 0.6145275831222534, "train/loss_error": 0.4048175811767578, "train/loss_total": 0.44675958156585693 }, { "epoch": 1.5948051948051947, "step": 614, "train/loss_ctc": 0.3222199082374573, "train/loss_error": 0.31062090396881104, "train/loss_total": 0.31294071674346924 }, { "epoch": 1.5974025974025974, "step": 615, "train/loss_ctc": 0.2895471453666687, "train/loss_error": 0.30168861150741577, "train/loss_total": 0.29926031827926636 }, { "epoch": 1.6, "step": 616, "train/loss_ctc": 0.41312915086746216, "train/loss_error": 0.383373886346817, "train/loss_total": 0.38932496309280396 }, { "epoch": 1.6025974025974026, "step": 617, "train/loss_ctc": 0.4393923878669739, "train/loss_error": 0.33019375801086426, "train/loss_total": 0.35203349590301514 }, { "epoch": 1.605194805194805, "step": 618, "train/loss_ctc": 0.3988490700721741, "train/loss_error": 0.3168860375881195, "train/loss_total": 0.3332786560058594 }, { "epoch": 1.6077922077922078, "step": 619, "train/loss_ctc": 0.6094553470611572, "train/loss_error": 0.4029877483844757, "train/loss_total": 0.44428128004074097 }, { "epoch": 1.6103896103896105, "grad_norm": 23218.345703125, "learning_rate": 2.036883116883117e-05, "loss": 0.3662, "step": 620 }, { "epoch": 1.6103896103896105, "step": 620, "train/loss_ctc": 0.3938644230365753, "train/loss_error": 0.2820076644420624, "train/loss_total": 0.30437901616096497 }, { "epoch": 1.612987012987013, "step": 621, "train/loss_ctc": 0.722612202167511, "train/loss_error": 0.3093867897987366, "train/loss_total": 0.39203187823295593 }, { "epoch": 1.6155844155844155, "step": 622, "train/loss_ctc": 0.4684518277645111, "train/loss_error": 0.40602150559425354, "train/loss_total": 0.41850757598876953 }, { "epoch": 1.6181818181818182, "step": 623, "train/loss_ctc": 0.7310417294502258, "train/loss_error": 0.4488747715950012, "train/loss_total": 0.5053081512451172 }, { "epoch": 1.6207792207792209, "step": 624, "train/loss_ctc": 0.8381940126419067, "train/loss_error": 0.3086043894290924, "train/loss_total": 0.41452232003211975 }, { "epoch": 1.6233766233766234, "step": 625, "train/loss_ctc": 0.4404849112033844, "train/loss_error": 0.33364439010620117, "train/loss_total": 0.3550125062465668 }, { "epoch": 1.6259740259740258, "step": 626, "train/loss_ctc": 0.5171626806259155, "train/loss_error": 0.30206355452537537, "train/loss_total": 0.3450833857059479 }, { "epoch": 1.6285714285714286, "step": 627, "train/loss_ctc": 0.37134456634521484, "train/loss_error": 0.4135684669017792, "train/loss_total": 0.40512368083000183 }, { "epoch": 1.6311688311688313, "step": 628, "train/loss_ctc": 0.3618020713329315, "train/loss_error": 0.33688920736312866, "train/loss_total": 0.34187179803848267 }, { "epoch": 1.6337662337662338, "step": 629, "train/loss_ctc": 0.41278043389320374, "train/loss_error": 0.3011680245399475, "train/loss_total": 0.3234905004501343 }, { "epoch": 1.6363636363636362, "grad_norm": 20954.095703125, "learning_rate": 2.0212987012987013e-05, "loss": 0.3805, "step": 630 }, { "epoch": 1.6363636363636362, "step": 630, "train/loss_ctc": 0.4965079426765442, "train/loss_error": 0.28736573457717896, "train/loss_total": 0.32919418811798096 }, { "epoch": 1.638961038961039, "step": 631, "train/loss_ctc": 0.4669243097305298, "train/loss_error": 0.4678165316581726, "train/loss_total": 0.4676381051540375 }, { "epoch": 1.6415584415584417, "step": 632, "train/loss_ctc": 0.5782583951950073, "train/loss_error": 0.3619556427001953, "train/loss_total": 0.4052162170410156 }, { "epoch": 1.6441558441558441, "step": 633, "train/loss_ctc": 0.43930283188819885, "train/loss_error": 0.2863599956035614, "train/loss_total": 0.3169485628604889 }, { "epoch": 1.6467532467532466, "step": 634, "train/loss_ctc": 0.29916590452194214, "train/loss_error": 0.29829874634742737, "train/loss_total": 0.29847216606140137 }, { "epoch": 1.6493506493506493, "step": 635, "train/loss_ctc": 0.6101655960083008, "train/loss_error": 0.38879725337028503, "train/loss_total": 0.43307092785835266 }, { "epoch": 1.651948051948052, "step": 636, "train/loss_ctc": 0.3763948082923889, "train/loss_error": 0.33953818678855896, "train/loss_total": 0.3469095230102539 }, { "epoch": 1.6545454545454545, "step": 637, "train/loss_ctc": 0.3898642361164093, "train/loss_error": 0.3847740590572357, "train/loss_total": 0.3857921063899994 }, { "epoch": 1.657142857142857, "step": 638, "train/loss_ctc": 0.5015693306922913, "train/loss_error": 0.3814793825149536, "train/loss_total": 0.40549737215042114 }, { "epoch": 1.6597402597402597, "step": 639, "train/loss_ctc": 0.47200873494148254, "train/loss_error": 0.40192103385925293, "train/loss_total": 0.4159385859966278 }, { "epoch": 1.6623376623376624, "grad_norm": 26866.44140625, "learning_rate": 2.0057142857142858e-05, "loss": 0.3805, "step": 640 }, { "epoch": 1.6623376623376624, "step": 640, "train/loss_ctc": 0.42847102880477905, "train/loss_error": 0.3641585111618042, "train/loss_total": 0.37702101469039917 }, { "epoch": 1.664935064935065, "step": 641, "train/loss_ctc": 0.460157573223114, "train/loss_error": 0.34179794788360596, "train/loss_total": 0.36546987295150757 }, { "epoch": 1.6675324675324674, "step": 642, "train/loss_ctc": 0.5584791898727417, "train/loss_error": 0.3697797358036041, "train/loss_total": 0.4075196385383606 }, { "epoch": 1.6701298701298701, "step": 643, "train/loss_ctc": 0.37974458932876587, "train/loss_error": 0.26931384205818176, "train/loss_total": 0.2913999855518341 }, { "epoch": 1.6727272727272728, "step": 644, "train/loss_ctc": 0.4050508737564087, "train/loss_error": 0.3884073495864868, "train/loss_total": 0.39173609018325806 }, { "epoch": 1.6753246753246753, "step": 645, "train/loss_ctc": 0.42125704884529114, "train/loss_error": 0.35263335704803467, "train/loss_total": 0.36635810136795044 }, { "epoch": 1.6779220779220778, "step": 646, "train/loss_ctc": 0.32887858152389526, "train/loss_error": 0.3570130467414856, "train/loss_total": 0.351386159658432 }, { "epoch": 1.6805194805194805, "step": 647, "train/loss_ctc": 0.5278000235557556, "train/loss_error": 0.3587814271450043, "train/loss_total": 0.3925851583480835 }, { "epoch": 1.6831168831168832, "step": 648, "train/loss_ctc": 0.420854389667511, "train/loss_error": 0.41141900420188904, "train/loss_total": 0.4133060872554779 }, { "epoch": 1.6857142857142857, "step": 649, "train/loss_ctc": 0.4627649188041687, "train/loss_error": 0.2546018660068512, "train/loss_total": 0.29623448848724365 }, { "epoch": 1.6883116883116882, "grad_norm": 24620.748046875, "learning_rate": 1.9901298701298703e-05, "loss": 0.3653, "step": 650 }, { "epoch": 1.6883116883116882, "step": 650, "train/loss_ctc": 0.49434030055999756, "train/loss_error": 0.35049062967300415, "train/loss_total": 0.3792605698108673 }, { "epoch": 1.690909090909091, "step": 651, "train/loss_ctc": 0.4027806520462036, "train/loss_error": 0.2916449010372162, "train/loss_total": 0.3138720691204071 }, { "epoch": 1.6935064935064936, "step": 652, "train/loss_ctc": 0.42459362745285034, "train/loss_error": 0.44788658618927, "train/loss_total": 0.44322800636291504 }, { "epoch": 1.6961038961038961, "step": 653, "train/loss_ctc": 0.4803949296474457, "train/loss_error": 0.30155739188194275, "train/loss_total": 0.3373248875141144 }, { "epoch": 1.6987012987012986, "step": 654, "train/loss_ctc": 0.4578898549079895, "train/loss_error": 0.382743239402771, "train/loss_total": 0.39777258038520813 }, { "epoch": 1.7012987012987013, "step": 655, "train/loss_ctc": 0.3040686249732971, "train/loss_error": 0.32486945390701294, "train/loss_total": 0.3207092881202698 }, { "epoch": 1.703896103896104, "step": 656, "train/loss_ctc": 0.40505820512771606, "train/loss_error": 0.4619594216346741, "train/loss_total": 0.4505791962146759 }, { "epoch": 1.7064935064935065, "step": 657, "train/loss_ctc": 0.5022231340408325, "train/loss_error": 0.38110998272895813, "train/loss_total": 0.40533262491226196 }, { "epoch": 1.709090909090909, "step": 658, "train/loss_ctc": 0.3588781952857971, "train/loss_error": 0.26409685611724854, "train/loss_total": 0.28305312991142273 }, { "epoch": 1.7116883116883117, "step": 659, "train/loss_ctc": 0.5157111287117004, "train/loss_error": 0.4104006886482239, "train/loss_total": 0.4314627945423126 }, { "epoch": 1.7142857142857144, "grad_norm": 23249.701171875, "learning_rate": 1.9745454545454547e-05, "loss": 0.3763, "step": 660 }, { "epoch": 1.7142857142857144, "step": 660, "train/loss_ctc": 0.36119332909584045, "train/loss_error": 0.35669004917144775, "train/loss_total": 0.3575907349586487 }, { "epoch": 1.716883116883117, "step": 661, "train/loss_ctc": 0.44175755977630615, "train/loss_error": 0.39273425936698914, "train/loss_total": 0.402538925409317 }, { "epoch": 1.7194805194805194, "step": 662, "train/loss_ctc": 0.5454682111740112, "train/loss_error": 0.388078898191452, "train/loss_total": 0.41955676674842834 }, { "epoch": 1.722077922077922, "step": 663, "train/loss_ctc": 0.7057255506515503, "train/loss_error": 0.3942088186740875, "train/loss_total": 0.4565121829509735 }, { "epoch": 1.7246753246753248, "step": 664, "train/loss_ctc": 0.4834209978580475, "train/loss_error": 0.297025591135025, "train/loss_total": 0.33430469036102295 }, { "epoch": 1.7272727272727273, "step": 665, "train/loss_ctc": 0.6364103555679321, "train/loss_error": 0.42136821150779724, "train/loss_total": 0.46437662839889526 }, { "epoch": 1.7298701298701298, "step": 666, "train/loss_ctc": 0.5694994926452637, "train/loss_error": 0.38182586431503296, "train/loss_total": 0.41936057806015015 }, { "epoch": 1.7324675324675325, "step": 667, "train/loss_ctc": 0.5823494791984558, "train/loss_error": 0.36259353160858154, "train/loss_total": 0.4065447151660919 }, { "epoch": 1.7350649350649352, "step": 668, "train/loss_ctc": 0.34297871589660645, "train/loss_error": 0.3202867805957794, "train/loss_total": 0.3248251676559448 }, { "epoch": 1.7376623376623377, "step": 669, "train/loss_ctc": 0.5323396921157837, "train/loss_error": 0.41221481561660767, "train/loss_total": 0.4362397789955139 }, { "epoch": 1.7402597402597402, "grad_norm": 29550.693359375, "learning_rate": 1.9589610389610392e-05, "loss": 0.4022, "step": 670 }, { "epoch": 1.7402597402597402, "step": 670, "train/loss_ctc": 0.35499510169029236, "train/loss_error": 0.3575451076030731, "train/loss_total": 0.3570351302623749 }, { "epoch": 1.7428571428571429, "step": 671, "train/loss_ctc": 0.33852213621139526, "train/loss_error": 0.30431944131851196, "train/loss_total": 0.31115999817848206 }, { "epoch": 1.7454545454545456, "step": 672, "train/loss_ctc": 0.4789576232433319, "train/loss_error": 0.33637377619743347, "train/loss_total": 0.36489054560661316 }, { "epoch": 1.748051948051948, "step": 673, "train/loss_ctc": 0.3982687294483185, "train/loss_error": 0.3403289020061493, "train/loss_total": 0.3519168794155121 }, { "epoch": 1.7506493506493506, "step": 674, "train/loss_ctc": 0.38729795813560486, "train/loss_error": 0.3103243410587311, "train/loss_total": 0.32571905851364136 }, { "epoch": 1.7532467532467533, "step": 675, "train/loss_ctc": 0.3726652264595032, "train/loss_error": 0.3404005765914917, "train/loss_total": 0.3468535244464874 }, { "epoch": 1.755844155844156, "step": 676, "train/loss_ctc": 0.5627468824386597, "train/loss_error": 0.3936110734939575, "train/loss_total": 0.42743825912475586 }, { "epoch": 1.7584415584415585, "step": 677, "train/loss_ctc": 0.5752575397491455, "train/loss_error": 0.40516167879104614, "train/loss_total": 0.439180850982666 }, { "epoch": 1.761038961038961, "step": 678, "train/loss_ctc": 0.8364459276199341, "train/loss_error": 0.4056621491909027, "train/loss_total": 0.491818904876709 }, { "epoch": 1.7636363636363637, "step": 679, "train/loss_ctc": 0.45233070850372314, "train/loss_error": 0.3338974714279175, "train/loss_total": 0.3575841188430786 }, { "epoch": 1.7662337662337664, "grad_norm": 21989.16015625, "learning_rate": 1.9433766233766234e-05, "loss": 0.3774, "step": 680 }, { "epoch": 1.7662337662337664, "step": 680, "train/loss_ctc": 0.48761579394340515, "train/loss_error": 0.27213963866233826, "train/loss_total": 0.31523486971855164 }, { "epoch": 1.7688311688311689, "step": 681, "train/loss_ctc": 0.5354026556015015, "train/loss_error": 0.3795827329158783, "train/loss_total": 0.4107467532157898 }, { "epoch": 1.7714285714285714, "step": 682, "train/loss_ctc": 0.47225284576416016, "train/loss_error": 0.26603952050209045, "train/loss_total": 0.3072821795940399 }, { "epoch": 1.774025974025974, "step": 683, "train/loss_ctc": 0.3011409640312195, "train/loss_error": 0.4017075300216675, "train/loss_total": 0.3815942406654358 }, { "epoch": 1.7766233766233768, "step": 684, "train/loss_ctc": 0.5700555443763733, "train/loss_error": 0.36943599581718445, "train/loss_total": 0.4095599055290222 }, { "epoch": 1.7792207792207793, "step": 685, "train/loss_ctc": 0.48380357027053833, "train/loss_error": 0.3116387724876404, "train/loss_total": 0.346071720123291 }, { "epoch": 1.7818181818181817, "step": 686, "train/loss_ctc": 0.33573585748672485, "train/loss_error": 0.3586776852607727, "train/loss_total": 0.35408931970596313 }, { "epoch": 1.7844155844155845, "step": 687, "train/loss_ctc": 0.46109387278556824, "train/loss_error": 0.3359508216381073, "train/loss_total": 0.36097943782806396 }, { "epoch": 1.7870129870129872, "step": 688, "train/loss_ctc": 0.5689878463745117, "train/loss_error": 0.4262869954109192, "train/loss_total": 0.4548271894454956 }, { "epoch": 1.7896103896103897, "step": 689, "train/loss_ctc": 0.33582794666290283, "train/loss_error": 0.29196029901504517, "train/loss_total": 0.3007338345050812 }, { "epoch": 1.7922077922077921, "grad_norm": 23842.3125, "learning_rate": 1.927792207792208e-05, "loss": 0.3641, "step": 690 }, { "epoch": 1.7922077922077921, "step": 690, "train/loss_ctc": 0.4258405566215515, "train/loss_error": 0.3746350407600403, "train/loss_total": 0.38487616181373596 }, { "epoch": 1.7948051948051948, "step": 691, "train/loss_ctc": 0.36284226179122925, "train/loss_error": 0.28924721479415894, "train/loss_total": 0.303966224193573 }, { "epoch": 1.7974025974025976, "step": 692, "train/loss_ctc": 0.44862958788871765, "train/loss_error": 0.3230965733528137, "train/loss_total": 0.348203182220459 }, { "epoch": 1.8, "step": 693, "train/loss_ctc": 0.5153921842575073, "train/loss_error": 0.36403924226760864, "train/loss_total": 0.3943098187446594 }, { "epoch": 1.8025974025974025, "step": 694, "train/loss_ctc": 0.2925747036933899, "train/loss_error": 0.3089377284049988, "train/loss_total": 0.30566513538360596 }, { "epoch": 1.8051948051948052, "step": 695, "train/loss_ctc": 0.49462372064590454, "train/loss_error": 0.3366323411464691, "train/loss_total": 0.3682306408882141 }, { "epoch": 1.807792207792208, "step": 696, "train/loss_ctc": 0.4037085771560669, "train/loss_error": 0.38076120615005493, "train/loss_total": 0.38535070419311523 }, { "epoch": 1.8103896103896104, "step": 697, "train/loss_ctc": 0.3968261778354645, "train/loss_error": 0.37679946422576904, "train/loss_total": 0.3808048367500305 }, { "epoch": 1.812987012987013, "step": 698, "train/loss_ctc": 0.49402010440826416, "train/loss_error": 0.3373732268810272, "train/loss_total": 0.36870262026786804 }, { "epoch": 1.8155844155844156, "step": 699, "train/loss_ctc": 0.4268750250339508, "train/loss_error": 0.36546072363853455, "train/loss_total": 0.37774360179901123 }, { "epoch": 1.8181818181818183, "grad_norm": 24812.71875, "learning_rate": 1.9122077922077923e-05, "loss": 0.3618, "step": 700 }, { "epoch": 1.8181818181818183, "step": 700, "train/loss_ctc": 0.4505705237388611, "train/loss_error": 0.2882344126701355, "train/loss_total": 0.32070162892341614 }, { "epoch": 1.8207792207792208, "step": 701, "train/loss_ctc": 0.31947633624076843, "train/loss_error": 0.2908881902694702, "train/loss_total": 0.29660582542419434 }, { "epoch": 1.8233766233766233, "step": 702, "train/loss_ctc": 0.6876518130302429, "train/loss_error": 0.36323603987693787, "train/loss_total": 0.4281191825866699 }, { "epoch": 1.825974025974026, "step": 703, "train/loss_ctc": 0.42431318759918213, "train/loss_error": 0.34547367691993713, "train/loss_total": 0.3612416088581085 }, { "epoch": 1.8285714285714287, "step": 704, "train/loss_ctc": 0.8714624643325806, "train/loss_error": 0.4296688735485077, "train/loss_total": 0.5180276036262512 }, { "epoch": 1.8311688311688312, "step": 705, "train/loss_ctc": 0.6081867218017578, "train/loss_error": 0.3758505880832672, "train/loss_total": 0.42231783270835876 }, { "epoch": 1.8337662337662337, "step": 706, "train/loss_ctc": 0.395913302898407, "train/loss_error": 0.3316766619682312, "train/loss_total": 0.34452399611473083 }, { "epoch": 1.8363636363636364, "step": 707, "train/loss_ctc": 0.2917942702770233, "train/loss_error": 0.21540133655071259, "train/loss_total": 0.2306799292564392 }, { "epoch": 1.838961038961039, "step": 708, "train/loss_ctc": 0.4703623950481415, "train/loss_error": 0.3490507900714874, "train/loss_total": 0.37331312894821167 }, { "epoch": 1.8415584415584414, "step": 709, "train/loss_ctc": 0.5649371147155762, "train/loss_error": 0.32161781191825867, "train/loss_total": 0.3702816963195801 }, { "epoch": 1.844155844155844, "grad_norm": 30704.751953125, "learning_rate": 1.8966233766233768e-05, "loss": 0.3666, "step": 710 }, { "epoch": 1.844155844155844, "step": 710, "train/loss_ctc": 0.3874501883983612, "train/loss_error": 0.3208937346935272, "train/loss_total": 0.3342050313949585 }, { "epoch": 1.8467532467532468, "step": 711, "train/loss_ctc": 0.37763816118240356, "train/loss_error": 0.3835083842277527, "train/loss_total": 0.3823343515396118 }, { "epoch": 1.8493506493506493, "step": 712, "train/loss_ctc": 0.33831602334976196, "train/loss_error": 0.2733326554298401, "train/loss_total": 0.28632932901382446 }, { "epoch": 1.8519480519480518, "step": 713, "train/loss_ctc": 0.37331998348236084, "train/loss_error": 0.352749228477478, "train/loss_total": 0.3568633794784546 }, { "epoch": 1.8545454545454545, "step": 714, "train/loss_ctc": 0.3413107097148895, "train/loss_error": 0.3731815218925476, "train/loss_total": 0.36680734157562256 }, { "epoch": 1.8571428571428572, "step": 715, "train/loss_ctc": 0.6431249380111694, "train/loss_error": 0.337596595287323, "train/loss_total": 0.3987022638320923 }, { "epoch": 1.8597402597402597, "step": 716, "train/loss_ctc": 0.38250046968460083, "train/loss_error": 0.26312166452407837, "train/loss_total": 0.2869974374771118 }, { "epoch": 1.8623376623376622, "step": 717, "train/loss_ctc": 0.9388673305511475, "train/loss_error": 0.33705851435661316, "train/loss_total": 0.457420289516449 }, { "epoch": 1.864935064935065, "step": 718, "train/loss_ctc": 0.4134286642074585, "train/loss_error": 0.3080805540084839, "train/loss_total": 0.32915017008781433 }, { "epoch": 1.8675324675324676, "step": 719, "train/loss_ctc": 0.4649423062801361, "train/loss_error": 0.3587118089199066, "train/loss_total": 0.379957914352417 }, { "epoch": 1.87012987012987, "grad_norm": 23560.796875, "learning_rate": 1.881038961038961e-05, "loss": 0.3579, "step": 720 }, { "epoch": 1.87012987012987, "step": 720, "train/loss_ctc": 0.5207735300064087, "train/loss_error": 0.4238133132457733, "train/loss_total": 0.4432053565979004 }, { "epoch": 1.8727272727272726, "step": 721, "train/loss_ctc": 0.36984261870384216, "train/loss_error": 0.3255079984664917, "train/loss_total": 0.33437493443489075 }, { "epoch": 1.8753246753246753, "step": 722, "train/loss_ctc": 0.40598630905151367, "train/loss_error": 0.3499596416950226, "train/loss_total": 0.36116498708724976 }, { "epoch": 1.877922077922078, "step": 723, "train/loss_ctc": 0.5382587313652039, "train/loss_error": 0.36448022723197937, "train/loss_total": 0.39923593401908875 }, { "epoch": 1.8805194805194805, "step": 724, "train/loss_ctc": 0.5815345048904419, "train/loss_error": 0.3843955993652344, "train/loss_total": 0.42382338643074036 }, { "epoch": 1.883116883116883, "step": 725, "train/loss_ctc": 0.7180856466293335, "train/loss_error": 0.3551926612854004, "train/loss_total": 0.42777127027511597 }, { "epoch": 1.8857142857142857, "step": 726, "train/loss_ctc": 0.5073139667510986, "train/loss_error": 0.34594568610191345, "train/loss_total": 0.3782193660736084 }, { "epoch": 1.8883116883116884, "step": 727, "train/loss_ctc": 0.3454534113407135, "train/loss_error": 0.2988293468952179, "train/loss_total": 0.3081541657447815 }, { "epoch": 1.8909090909090909, "step": 728, "train/loss_ctc": 0.5222856998443604, "train/loss_error": 0.4625934362411499, "train/loss_total": 0.474531888961792 }, { "epoch": 1.8935064935064934, "step": 729, "train/loss_ctc": 0.24968042969703674, "train/loss_error": 0.2510108947753906, "train/loss_total": 0.2507448196411133 }, { "epoch": 1.896103896103896, "grad_norm": 19627.6171875, "learning_rate": 1.8654545454545454e-05, "loss": 0.3801, "step": 730 }, { "epoch": 1.896103896103896, "step": 730, "train/loss_ctc": 0.4575114846229553, "train/loss_error": 0.3037249445915222, "train/loss_total": 0.33448225259780884 }, { "epoch": 1.8987012987012988, "step": 731, "train/loss_ctc": 0.4426967203617096, "train/loss_error": 0.35825613141059875, "train/loss_total": 0.37514424324035645 }, { "epoch": 1.9012987012987013, "step": 732, "train/loss_ctc": 0.32556506991386414, "train/loss_error": 0.36465343832969666, "train/loss_total": 0.3568357527256012 }, { "epoch": 1.9038961038961038, "step": 733, "train/loss_ctc": 0.8685684204101562, "train/loss_error": 0.39102181792259216, "train/loss_total": 0.486531138420105 }, { "epoch": 1.9064935064935065, "step": 734, "train/loss_ctc": 0.35776132345199585, "train/loss_error": 0.2862175703048706, "train/loss_total": 0.30052632093429565 }, { "epoch": 1.9090909090909092, "step": 735, "train/loss_ctc": 0.49757006764411926, "train/loss_error": 0.42359843850135803, "train/loss_total": 0.4383927583694458 }, { "epoch": 1.9116883116883117, "step": 736, "train/loss_ctc": 0.8593733906745911, "train/loss_error": 0.3768109679222107, "train/loss_total": 0.4733234643936157 }, { "epoch": 1.9142857142857141, "step": 737, "train/loss_ctc": 0.44772931933403015, "train/loss_error": 0.3028361201286316, "train/loss_total": 0.3318147659301758 }, { "epoch": 1.9168831168831169, "step": 738, "train/loss_ctc": 0.44815173745155334, "train/loss_error": 0.39155465364456177, "train/loss_total": 0.40287405252456665 }, { "epoch": 1.9194805194805196, "step": 739, "train/loss_ctc": 0.5501865744590759, "train/loss_error": 0.30605989694595337, "train/loss_total": 0.3548852503299713 }, { "epoch": 1.922077922077922, "grad_norm": 24062.744140625, "learning_rate": 1.84987012987013e-05, "loss": 0.3855, "step": 740 }, { "epoch": 1.922077922077922, "step": 740, "train/loss_ctc": 0.48371538519859314, "train/loss_error": 0.3126247227191925, "train/loss_total": 0.34684285521507263 }, { "epoch": 1.9246753246753245, "step": 741, "train/loss_ctc": 0.7266988754272461, "train/loss_error": 0.4048958718776703, "train/loss_total": 0.4692564606666565 }, { "epoch": 1.9272727272727272, "step": 742, "train/loss_ctc": 0.35786324739456177, "train/loss_error": 0.2433750331401825, "train/loss_total": 0.2662726938724518 }, { "epoch": 1.92987012987013, "step": 743, "train/loss_ctc": 0.27519887685775757, "train/loss_error": 0.3701172471046448, "train/loss_total": 0.3511335551738739 }, { "epoch": 1.9324675324675324, "step": 744, "train/loss_ctc": 0.5145028829574585, "train/loss_error": 0.39843520522117615, "train/loss_total": 0.4216487407684326 }, { "epoch": 1.935064935064935, "step": 745, "train/loss_ctc": 0.3432126045227051, "train/loss_error": 0.32749074697494507, "train/loss_total": 0.330635130405426 }, { "epoch": 1.9376623376623376, "step": 746, "train/loss_ctc": 0.3141961395740509, "train/loss_error": 0.2869825065135956, "train/loss_total": 0.2924252450466156 }, { "epoch": 1.9402597402597404, "step": 747, "train/loss_ctc": 0.5582700967788696, "train/loss_error": 0.37955841422080994, "train/loss_total": 0.41530075669288635 }, { "epoch": 1.9428571428571428, "step": 748, "train/loss_ctc": 0.4026537835597992, "train/loss_error": 0.3965802788734436, "train/loss_total": 0.3977949917316437 }, { "epoch": 1.9454545454545453, "step": 749, "train/loss_ctc": 0.39477694034576416, "train/loss_error": 0.2755521535873413, "train/loss_total": 0.2993971109390259 }, { "epoch": 1.948051948051948, "grad_norm": 20162.84765625, "learning_rate": 1.8342857142857144e-05, "loss": 0.3591, "step": 750 }, { "epoch": 1.948051948051948, "step": 750, "train/loss_ctc": 0.5513525605201721, "train/loss_error": 0.37366554141044617, "train/loss_total": 0.4092029333114624 }, { "epoch": 1.9506493506493507, "step": 751, "train/loss_ctc": 0.7355890274047852, "train/loss_error": 0.4002092182636261, "train/loss_total": 0.4672852158546448 }, { "epoch": 1.9532467532467532, "step": 752, "train/loss_ctc": 0.31352096796035767, "train/loss_error": 0.3077150881290436, "train/loss_total": 0.30887627601623535 }, { "epoch": 1.9558441558441557, "step": 753, "train/loss_ctc": 0.4724349081516266, "train/loss_error": 0.38253048062324524, "train/loss_total": 0.40051138401031494 }, { "epoch": 1.9584415584415584, "step": 754, "train/loss_ctc": 0.5896301865577698, "train/loss_error": 0.3230689465999603, "train/loss_total": 0.37638118863105774 }, { "epoch": 1.9610389610389611, "step": 755, "train/loss_ctc": 0.43446671962738037, "train/loss_error": 0.3472963869571686, "train/loss_total": 0.36473047733306885 }, { "epoch": 1.9636363636363636, "step": 756, "train/loss_ctc": 0.3615896701812744, "train/loss_error": 0.3015562891960144, "train/loss_total": 0.31356295943260193 }, { "epoch": 1.9662337662337661, "step": 757, "train/loss_ctc": 0.6032392978668213, "train/loss_error": 0.442081481218338, "train/loss_total": 0.47431308031082153 }, { "epoch": 1.9688311688311688, "step": 758, "train/loss_ctc": 0.5529225468635559, "train/loss_error": 0.3616954982280731, "train/loss_total": 0.3999409079551697 }, { "epoch": 1.9714285714285715, "step": 759, "train/loss_ctc": 0.651772141456604, "train/loss_error": 0.3561409115791321, "train/loss_total": 0.4152671694755554 }, { "epoch": 1.974025974025974, "grad_norm": 30352.984375, "learning_rate": 1.818701298701299e-05, "loss": 0.393, "step": 760 }, { "epoch": 1.974025974025974, "step": 760, "train/loss_ctc": 0.43325597047805786, "train/loss_error": 0.2897983491420746, "train/loss_total": 0.3184898793697357 }, { "epoch": 1.9766233766233765, "step": 761, "train/loss_ctc": 0.43930351734161377, "train/loss_error": 0.3330192565917969, "train/loss_total": 0.3542761206626892 }, { "epoch": 1.9792207792207792, "step": 762, "train/loss_ctc": 0.30837714672088623, "train/loss_error": 0.29732412099838257, "train/loss_total": 0.29953473806381226 }, { "epoch": 1.981818181818182, "step": 763, "train/loss_ctc": 0.5434048175811768, "train/loss_error": 0.361221581697464, "train/loss_total": 0.39765822887420654 }, { "epoch": 1.9844155844155844, "step": 764, "train/loss_ctc": 0.32000696659088135, "train/loss_error": 0.30416491627693176, "train/loss_total": 0.3073333501815796 }, { "epoch": 1.987012987012987, "step": 765, "train/loss_ctc": 0.3093123435974121, "train/loss_error": 0.3028091788291931, "train/loss_total": 0.3041098117828369 }, { "epoch": 1.9896103896103896, "step": 766, "train/loss_ctc": 0.49106255173683167, "train/loss_error": 0.378399521112442, "train/loss_total": 0.4009321331977844 }, { "epoch": 1.9922077922077923, "step": 767, "train/loss_ctc": 0.47295647859573364, "train/loss_error": 0.36410290002822876, "train/loss_total": 0.38587361574172974 }, { "epoch": 1.9948051948051948, "step": 768, "train/loss_ctc": 0.37513935565948486, "train/loss_error": 0.42311030626296997, "train/loss_total": 0.4135161340236664 }, { "epoch": 1.9974025974025973, "step": 769, "train/loss_ctc": 0.7483372092247009, "train/loss_error": 0.399945467710495, "train/loss_total": 0.4696238338947296 }, { "epoch": 2.0, "grad_norm": 48162.359375, "learning_rate": 1.803116883116883e-05, "loss": 0.3651, "step": 770 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.6212524771690369, "val/loss_error": 0.07040655612945557, "val/loss_total": 0.18057574331760406 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5641827583312988, "val/loss_error": 0.46774816513061523, "val/loss_total": 0.4870350956916809 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.08122025430202484, "val/loss_error": 0.25361424684524536, "val/loss_total": 0.21913544833660126 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.1838136464357376, "val/loss_error": 0.0659593716263771, "val/loss_total": 0.08953022956848145 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.16408589482307434, "val/loss_error": 0.20068122446537018, "val/loss_total": 0.19336216151714325 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.14686042070388794, "val/loss_error": 0.24997790157794952, "val/loss_total": 0.22935441136360168 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.213694229722023, "val/loss_error": 0.4911298453807831, "val/loss_total": 0.4356427490711212 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.38619381189346313, "val/loss_error": 0.2665678560733795, "val/loss_total": 0.29049304127693176 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.3238905370235443, "val/loss_error": 0.2809361517429352, "val/loss_total": 0.289527028799057 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.29229432344436646, "val/loss_error": 0.269528865814209, "val/loss_total": 0.2740819454193115 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.1295689046382904, "val/loss_error": 0.16926100850105286, "val/loss_total": 0.16132257878780365 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.20032210648059845, "val/loss_error": 0.2680743634700775, "val/loss_total": 0.254523903131485 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.8009371757507324, "val/loss_error": 0.38533589243888855, "val/loss_total": 0.4684561491012573 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.7501871585845947, "val/loss_error": 0.5076416730880737, "val/loss_total": 0.5561507940292358 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.6157832145690918, "val/loss_error": 0.4080537259578705, "val/loss_total": 0.44959962368011475 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.49836596846580505, "val/loss_error": 0.6043725609779358, "val/loss_total": 0.5831712484359741 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.09362031519412994, "val/loss_error": 0.26188668608665466, "val/loss_total": 0.2282334268093109 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.21052472293376923, "val/loss_error": 0.26420462131500244, "val/loss_total": 0.2534686326980591 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.14820709824562073, "val/loss_error": 0.28483936190605164, "val/loss_total": 0.2575129270553589 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5595799684524536, "val/loss_error": 0.35331279039382935, "val/loss_total": 0.39456623792648315 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.8914788961410522, "val/loss_error": 0.4747396409511566, "val/loss_total": 0.5580874681472778 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.2253076583147049, "val/loss_error": 0.5052464008331299, "val/loss_total": 0.4492586553096771 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.27350929379463196, "val/loss_error": 0.39923179149627686, "val/loss_total": 0.37408730387687683 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.26869645714759827, "val/loss_error": 0.2149008810520172, "val/loss_total": 0.22565999627113342 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.3059903383255005, "val/loss_error": 0.5805414915084839, "val/loss_total": 0.5256312489509583 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.4374830424785614, "val/loss_error": 0.5360971093177795, "val/loss_total": 0.5163742899894714 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.29972341656684875, "val/loss_error": 0.3435831367969513, "val/loss_total": 0.3348112106323242 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.11130885034799576, "val/loss_error": 0.2055967003107071, "val/loss_total": 0.18673913180828094 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.014011613093316555, "val/loss_error": 0.03985102102160454, "val/loss_total": 0.034683141857385635 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.25541141629219055, "val/loss_error": 0.5095804333686829, "val/loss_total": 0.45874664187431335 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.08171942085027695, "val/loss_error": 0.10864593088626862, "val/loss_total": 0.10326062887907028 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.146670401096344, "val/loss_error": 0.2713693678379059, "val/loss_total": 0.24642957746982574 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.0761578232049942, "val/loss_error": 0.14257733523845673, "val/loss_total": 0.12929344177246094 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.6315022110939026, "val/loss_error": 0.6775221824645996, "val/loss_total": 0.6683182120323181 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.12327930331230164, "val/loss_error": 0.30720677971839905, "val/loss_total": 0.2704212963581085 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.1935950070619583, "val/loss_error": 0.3364352881908417, "val/loss_total": 0.30786722898483276 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.49117201566696167, "val/loss_error": 0.5615693926811218, "val/loss_total": 0.5474899411201477 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.46130651235580444, "val/loss_error": 0.37872791290283203, "val/loss_total": 0.39524364471435547 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.26587674021720886, "val/loss_error": 0.530285656452179, "val/loss_total": 0.4774038791656494 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5271091461181641, "val/loss_error": 0.6918538212776184, "val/loss_total": 0.6589049100875854 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.21048709750175476, "val/loss_error": 0.197299525141716, "val/loss_total": 0.19993704557418823 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.2731887400150299, "val/loss_error": 0.3073316216468811, "val/loss_total": 0.30050304532051086 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.6042062640190125, "val/loss_error": 0.947188675403595, "val/loss_total": 0.8785921931266785 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.521141529083252, "val/loss_error": 0.5990126729011536, "val/loss_total": 0.5834384560585022 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.48457249999046326, "val/loss_error": 0.3547654449939728, "val/loss_total": 0.3807268738746643 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.3813009560108185, "val/loss_error": 0.7163338661193848, "val/loss_total": 0.6493273377418518 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.6791588068008423, "val/loss_error": 0.4617833197116852, "val/loss_total": 0.5052584409713745 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.34650057554244995, "val/loss_error": 0.5571224093437195, "val/loss_total": 0.5149980783462524 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5475804805755615, "val/loss_error": 0.24710439145565033, "val/loss_total": 0.3071995973587036 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.6814380884170532, "val/loss_error": 0.6457170248031616, "val/loss_total": 0.6528612375259399 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.29710814356803894, "val/loss_error": 0.782929003238678, "val/loss_total": 0.6857647895812988 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.2583695948123932, "val/loss_error": 0.3881555497646332, "val/loss_total": 0.3621983528137207 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.3956788182258606, "val/loss_error": 0.42563873529434204, "val/loss_total": 0.4196467697620392 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5030521154403687, "val/loss_error": 0.9431535601615906, "val/loss_total": 0.8551332950592041 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.4116271734237671, "val/loss_error": 0.4361271858215332, "val/loss_total": 0.4312271773815155 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.20258107781410217, "val/loss_error": 0.4658236801624298, "val/loss_total": 0.41317516565322876 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.13977645337581635, "val/loss_error": 0.2563163936138153, "val/loss_total": 0.23300841450691223 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.4463246464729309, "val/loss_error": 0.26200070977211, "val/loss_total": 0.29886549711227417 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.6945659518241882, "val/loss_error": 0.610458254814148, "val/loss_total": 0.6272798180580139 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.10253960639238358, "val/loss_error": 0.11241838335990906, "val/loss_total": 0.1104426309466362 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.8543664813041687, "val/loss_error": 0.5813599228858948, "val/loss_total": 0.6359612345695496 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.7063799500465393, "val/loss_error": 0.34287405014038086, "val/loss_total": 0.41557520627975464 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.07072682678699493, "val/loss_error": 0.09006337821483612, "val/loss_total": 0.08619607239961624 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.2508507966995239, "val/loss_error": 0.2915966510772705, "val/loss_total": 0.2834474742412567 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.1294764131307602, "val/loss_error": 0.20240899920463562, "val/loss_total": 0.18782249093055725 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.23565176129341125, "val/loss_error": 0.3747059404850006, "val/loss_total": 0.34689509868621826 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.3149636685848236, "val/loss_error": 0.28151389956474304, "val/loss_total": 0.2882038652896881 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5782596468925476, "val/loss_error": 0.58573979139328, "val/loss_total": 0.5842437744140625 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.45180001854896545, "val/loss_error": 0.5928553342819214, "val/loss_total": 0.5646442770957947 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.4721611440181732, "val/loss_error": 0.49266794323921204, "val/loss_total": 0.4885666072368622 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.18733324110507965, "val/loss_error": 0.23270484805107117, "val/loss_total": 0.22363053262233734 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.45128771662712097, "val/loss_error": 0.6148039698600769, "val/loss_total": 0.5821007490158081 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.6158726215362549, "val/loss_error": 0.5962280035018921, "val/loss_total": 0.6001569628715515 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.16351334750652313, "val/loss_error": 0.3354180157184601, "val/loss_total": 0.301037073135376 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.25489506125450134, "val/loss_error": 0.26064297556877136, "val/loss_total": 0.2594933807849884 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.04301539063453674, "val/loss_error": 0.2017109990119934, "val/loss_total": 0.16997188329696655 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.6893584132194519, "val/loss_error": 0.49725520610809326, "val/loss_total": 0.5356758832931519 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.7915071845054626, "val/loss_error": 0.39527732133865356, "val/loss_total": 0.47452330589294434 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5799186825752258, "val/loss_error": 0.2938801646232605, "val/loss_total": 0.35108786821365356 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.004956813994795084, "val/loss_error": 0.31261488795280457, "val/loss_total": 0.25108328461647034 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5224615335464478, "val/loss_error": 0.29184213280677795, "val/loss_total": 0.33796602487564087 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.03209036961197853, "val/loss_error": 0.10064525157213211, "val/loss_total": 0.08693427592515945 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.1880158931016922, "val/loss_error": 0.24509508907794952, "val/loss_total": 0.23367926478385925 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.43621084094047546, "val/loss_error": 0.432693749666214, "val/loss_total": 0.43339717388153076 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.3738035261631012, "val/loss_error": 0.6517160534858704, "val/loss_total": 0.5961335897445679 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.4566536843776703, "val/loss_error": 0.5999904274940491, "val/loss_total": 0.5713230967521667 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.0437878742814064, "val/loss_error": 0.12508875131607056, "val/loss_total": 0.1088285818696022 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.46363988518714905, "val/loss_error": 0.4002024531364441, "val/loss_total": 0.4128899574279785 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.28916677832603455, "val/loss_error": 0.2801087498664856, "val/loss_total": 0.28192034363746643 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.28374505043029785, "val/loss_error": 0.3596665859222412, "val/loss_total": 0.34448230266571045 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.2839352786540985, "val/loss_error": 0.4611583650112152, "val/loss_total": 0.42571374773979187 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.9060429334640503, "val/loss_error": 0.41005873680114746, "val/loss_total": 0.509255588054657 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.194244384765625, "val/loss_error": 0.05360177904367447, "val/loss_total": 0.08173030614852905 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.4223482310771942, "val/loss_error": 0.27553582191467285, "val/loss_total": 0.30489832162857056 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5960039496421814, "val/loss_error": 0.5597643256187439, "val/loss_total": 0.5670122504234314 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.1727776676416397, "val/loss_error": 0.2041119933128357, "val/loss_total": 0.19784513115882874 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.11337430775165558, "val/loss_error": 0.4563121199607849, "val/loss_total": 0.38772454857826233 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.6440345644950867, "val/loss_error": 0.4926941394805908, "val/loss_total": 0.522962212562561 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.09782572835683823, "val/loss_error": 0.24919287860393524, "val/loss_total": 0.21891945600509644 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.20558835566043854, "val/loss_error": 0.20761054754257202, "val/loss_total": 0.2072061151266098 }, { "epoch": 2.0, "eval_eval/f1_0": 0.5331664681434631, "eval_eval/f1_1": 0.9341803193092346, "eval_eval/precision_0": 0.5900276899337769, "eval_eval/precision_1": 0.921657383441925, "eval_eval/recall_0": 0.48630136251449585, "eval_eval/recall_1": 0.9470483064651489, "eval_eval/wer": 0.1535138657408022, "eval_loss": 0.37635180354118347, "eval_runtime": 7.2584, "eval_samples_per_second": 13.777, "eval_steps_per_second": 13.777, "step": 770 }, { "epoch": 2.0, "step": 770, "train/loss_ctc": 0.5506945848464966, "train/loss_error": 0.3837851583957672, "train/loss_total": 0.417167067527771 }, { "epoch": 2.0025974025974027, "step": 771, "train/loss_ctc": 0.3490327000617981, "train/loss_error": 0.3211553990840912, "train/loss_total": 0.326730877161026 }, { "epoch": 2.005194805194805, "step": 772, "train/loss_ctc": 0.3958051800727844, "train/loss_error": 0.3126053214073181, "train/loss_total": 0.32924529910087585 }, { "epoch": 2.0077922077922077, "step": 773, "train/loss_ctc": 0.33227384090423584, "train/loss_error": 0.331447571516037, "train/loss_total": 0.33161282539367676 }, { "epoch": 2.0103896103896104, "step": 774, "train/loss_ctc": 0.6834820508956909, "train/loss_error": 0.32353195548057556, "train/loss_total": 0.39552199840545654 }, { "epoch": 2.012987012987013, "step": 775, "train/loss_ctc": 0.6063044667243958, "train/loss_error": 0.36985161900520325, "train/loss_total": 0.41714221239089966 }, { "epoch": 2.0155844155844154, "step": 776, "train/loss_ctc": 0.3437439203262329, "train/loss_error": 0.33400702476501465, "train/loss_total": 0.3359544277191162 }, { "epoch": 2.018181818181818, "step": 777, "train/loss_ctc": 0.4893108010292053, "train/loss_error": 0.41343098878860474, "train/loss_total": 0.42860695719718933 }, { "epoch": 2.020779220779221, "step": 778, "train/loss_ctc": 0.3075157403945923, "train/loss_error": 0.3256153166294098, "train/loss_total": 0.32199540734291077 }, { "epoch": 2.0233766233766235, "step": 779, "train/loss_ctc": 0.5592908263206482, "train/loss_error": 0.39105328917503357, "train/loss_total": 0.4247007966041565 }, { "epoch": 2.0259740259740258, "grad_norm": 26477.287109375, "learning_rate": 1.7875324675324675e-05, "loss": 0.3729, "step": 780 }, { "epoch": 2.0259740259740258, "step": 780, "train/loss_ctc": 0.2886391282081604, "train/loss_error": 0.33437755703926086, "train/loss_total": 0.32522985339164734 }, { "epoch": 2.0285714285714285, "step": 781, "train/loss_ctc": 0.332396537065506, "train/loss_error": 0.3184034526348114, "train/loss_total": 0.3212020993232727 }, { "epoch": 2.031168831168831, "step": 782, "train/loss_ctc": 0.4479905962944031, "train/loss_error": 0.29566869139671326, "train/loss_total": 0.3261330723762512 }, { "epoch": 2.033766233766234, "step": 783, "train/loss_ctc": 0.6973990797996521, "train/loss_error": 0.4506222605705261, "train/loss_total": 0.49997761845588684 }, { "epoch": 2.036363636363636, "step": 784, "train/loss_ctc": 0.41316601634025574, "train/loss_error": 0.30453306436538696, "train/loss_total": 0.32625967264175415 }, { "epoch": 2.038961038961039, "step": 785, "train/loss_ctc": 0.4219191372394562, "train/loss_error": 0.3201497197151184, "train/loss_total": 0.34050363302230835 }, { "epoch": 2.0415584415584416, "step": 786, "train/loss_ctc": 0.44849881529808044, "train/loss_error": 0.3982158303260803, "train/loss_total": 0.4082724452018738 }, { "epoch": 2.0441558441558443, "step": 787, "train/loss_ctc": 0.3314429521560669, "train/loss_error": 0.3167494833469391, "train/loss_total": 0.31968817114830017 }, { "epoch": 2.0467532467532465, "step": 788, "train/loss_ctc": 0.4260593056678772, "train/loss_error": 0.2959684133529663, "train/loss_total": 0.321986585855484 }, { "epoch": 2.0493506493506493, "step": 789, "train/loss_ctc": 0.38089942932128906, "train/loss_error": 0.3501519560813904, "train/loss_total": 0.3563014566898346 }, { "epoch": 2.051948051948052, "grad_norm": 11112.05078125, "learning_rate": 1.7735064935064933e-05, "loss": 0.3546, "step": 790 }, { "epoch": 2.051948051948052, "step": 790, "train/loss_ctc": 0.4361821413040161, "train/loss_error": 0.34517431259155273, "train/loss_total": 0.3633759021759033 }, { "epoch": 2.0545454545454547, "step": 791, "train/loss_ctc": 0.4366905689239502, "train/loss_error": 0.36047476530075073, "train/loss_total": 0.3757179379463196 }, { "epoch": 2.057142857142857, "step": 792, "train/loss_ctc": 0.3096434772014618, "train/loss_error": 0.30500927567481995, "train/loss_total": 0.30593612790107727 }, { "epoch": 2.0597402597402596, "step": 793, "train/loss_ctc": 0.3329349160194397, "train/loss_error": 0.2748412489891052, "train/loss_total": 0.2864599823951721 }, { "epoch": 2.0623376623376624, "step": 794, "train/loss_ctc": 0.6503974795341492, "train/loss_error": 0.35280144214630127, "train/loss_total": 0.41232067346572876 }, { "epoch": 2.064935064935065, "step": 795, "train/loss_ctc": 0.9980349540710449, "train/loss_error": 0.4725582003593445, "train/loss_total": 0.5776535868644714 }, { "epoch": 2.0675324675324673, "step": 796, "train/loss_ctc": 0.3507482707500458, "train/loss_error": 0.3725424110889435, "train/loss_total": 0.36818358302116394 }, { "epoch": 2.07012987012987, "step": 797, "train/loss_ctc": 0.44926074147224426, "train/loss_error": 0.32243987917900085, "train/loss_total": 0.34780406951904297 }, { "epoch": 2.0727272727272728, "step": 798, "train/loss_ctc": 0.4207616448402405, "train/loss_error": 0.439129501581192, "train/loss_total": 0.43545594811439514 }, { "epoch": 2.0753246753246755, "step": 799, "train/loss_ctc": 0.45460522174835205, "train/loss_error": 0.32638755440711975, "train/loss_total": 0.35203108191490173 }, { "epoch": 2.0779220779220777, "grad_norm": 12686.251953125, "learning_rate": 1.7579220779220778e-05, "loss": 0.3825, "step": 800 }, { "epoch": 2.0779220779220777, "step": 800, "train/loss_ctc": 0.38571709394454956, "train/loss_error": 0.26246994733810425, "train/loss_total": 0.28711938858032227 }, { "epoch": 2.0805194805194804, "step": 801, "train/loss_ctc": 0.406166136264801, "train/loss_error": 0.28729814291000366, "train/loss_total": 0.3110717535018921 }, { "epoch": 2.083116883116883, "step": 802, "train/loss_ctc": 0.32693251967430115, "train/loss_error": 0.28294867277145386, "train/loss_total": 0.2917454242706299 }, { "epoch": 2.085714285714286, "step": 803, "train/loss_ctc": 0.376781702041626, "train/loss_error": 0.3977927565574646, "train/loss_total": 0.3935905396938324 }, { "epoch": 2.088311688311688, "step": 804, "train/loss_ctc": 0.3488351106643677, "train/loss_error": 0.3706354796886444, "train/loss_total": 0.366275429725647 }, { "epoch": 2.090909090909091, "step": 805, "train/loss_ctc": 0.4165085554122925, "train/loss_error": 0.3130813241004944, "train/loss_total": 0.33376678824424744 }, { "epoch": 2.0935064935064935, "step": 806, "train/loss_ctc": 0.5484981536865234, "train/loss_error": 0.3721875250339508, "train/loss_total": 0.4074496626853943 }, { "epoch": 2.0961038961038962, "step": 807, "train/loss_ctc": 0.38697874546051025, "train/loss_error": 0.36148032546043396, "train/loss_total": 0.3665800094604492 }, { "epoch": 2.0987012987012985, "step": 808, "train/loss_ctc": 0.6808809638023376, "train/loss_error": 0.2864490747451782, "train/loss_total": 0.36533546447753906 }, { "epoch": 2.101298701298701, "step": 809, "train/loss_ctc": 0.2817627787590027, "train/loss_error": 0.28292030096054077, "train/loss_total": 0.28268879652023315 }, { "epoch": 2.103896103896104, "grad_norm": 9604.087890625, "learning_rate": 1.7423376623376623e-05, "loss": 0.3406, "step": 810 }, { "epoch": 2.103896103896104, "step": 810, "train/loss_ctc": 0.4463643431663513, "train/loss_error": 0.30396416783332825, "train/loss_total": 0.3324441909790039 }, { "epoch": 2.1064935064935066, "step": 811, "train/loss_ctc": 0.30069199204444885, "train/loss_error": 0.3134627938270569, "train/loss_total": 0.31090864539146423 }, { "epoch": 2.109090909090909, "step": 812, "train/loss_ctc": 0.6335312724113464, "train/loss_error": 0.3866156339645386, "train/loss_total": 0.435998797416687 }, { "epoch": 2.1116883116883116, "step": 813, "train/loss_ctc": 0.3324704170227051, "train/loss_error": 0.2981662452220917, "train/loss_total": 0.3050270974636078 }, { "epoch": 2.1142857142857143, "step": 814, "train/loss_ctc": 0.6655696630477905, "train/loss_error": 0.49002164602279663, "train/loss_total": 0.5251312851905823 }, { "epoch": 2.116883116883117, "step": 815, "train/loss_ctc": 0.2898806631565094, "train/loss_error": 0.2262437790632248, "train/loss_total": 0.23897117376327515 }, { "epoch": 2.1194805194805193, "step": 816, "train/loss_ctc": 0.47535350918769836, "train/loss_error": 0.32803407311439514, "train/loss_total": 0.3574979901313782 }, { "epoch": 2.122077922077922, "step": 817, "train/loss_ctc": 0.22685518860816956, "train/loss_error": 0.24766018986701965, "train/loss_total": 0.24349918961524963 }, { "epoch": 2.1246753246753247, "step": 818, "train/loss_ctc": 0.7005879878997803, "train/loss_error": 0.4091179668903351, "train/loss_total": 0.46741199493408203 }, { "epoch": 2.1272727272727274, "step": 819, "train/loss_ctc": 0.38029396533966064, "train/loss_error": 0.34351009130477905, "train/loss_total": 0.3508668839931488 }, { "epoch": 2.1298701298701297, "grad_norm": 10840.5693359375, "learning_rate": 1.7267532467532467e-05, "loss": 0.3568, "step": 820 }, { "epoch": 2.1298701298701297, "step": 820, "train/loss_ctc": 0.44937455654144287, "train/loss_error": 0.3302880525588989, "train/loss_total": 0.3541053831577301 }, { "epoch": 2.1324675324675324, "step": 821, "train/loss_ctc": 0.5681313276290894, "train/loss_error": 0.3654235303401947, "train/loss_total": 0.40596508979797363 }, { "epoch": 2.135064935064935, "step": 822, "train/loss_ctc": 0.5596339702606201, "train/loss_error": 0.40076372027397156, "train/loss_total": 0.4325377643108368 }, { "epoch": 2.137662337662338, "step": 823, "train/loss_ctc": 0.5228732824325562, "train/loss_error": 0.351002961397171, "train/loss_total": 0.38537701964378357 }, { "epoch": 2.14025974025974, "step": 824, "train/loss_ctc": 0.5031170845031738, "train/loss_error": 0.3568040132522583, "train/loss_total": 0.38606664538383484 }, { "epoch": 2.142857142857143, "step": 825, "train/loss_ctc": 0.3276178240776062, "train/loss_error": 0.3818340301513672, "train/loss_total": 0.3709907829761505 }, { "epoch": 2.1454545454545455, "step": 826, "train/loss_ctc": 0.42574992775917053, "train/loss_error": 0.4146285653114319, "train/loss_total": 0.41685283184051514 }, { "epoch": 2.148051948051948, "step": 827, "train/loss_ctc": 0.3520503044128418, "train/loss_error": 0.2619549334049225, "train/loss_total": 0.2799740135669708 }, { "epoch": 2.1506493506493505, "step": 828, "train/loss_ctc": 0.43559587001800537, "train/loss_error": 0.310391902923584, "train/loss_total": 0.3354327082633972 }, { "epoch": 2.153246753246753, "step": 829, "train/loss_ctc": 0.5317234992980957, "train/loss_error": 0.4235526919364929, "train/loss_total": 0.4451868534088135 }, { "epoch": 2.155844155844156, "grad_norm": 13089.2265625, "learning_rate": 1.7111688311688312e-05, "loss": 0.3812, "step": 830 }, { "epoch": 2.155844155844156, "step": 830, "train/loss_ctc": 0.42490026354789734, "train/loss_error": 0.38703611493110657, "train/loss_total": 0.3946089744567871 }, { "epoch": 2.1584415584415586, "step": 831, "train/loss_ctc": 0.5042603611946106, "train/loss_error": 0.34757575392723083, "train/loss_total": 0.37891268730163574 }, { "epoch": 2.161038961038961, "step": 832, "train/loss_ctc": 0.5441144108772278, "train/loss_error": 0.3915369510650635, "train/loss_total": 0.42205244302749634 }, { "epoch": 2.1636363636363636, "step": 833, "train/loss_ctc": 0.35199081897735596, "train/loss_error": 0.3583206832408905, "train/loss_total": 0.3570547103881836 }, { "epoch": 2.1662337662337663, "step": 834, "train/loss_ctc": 0.4970782399177551, "train/loss_error": 0.3381589353084564, "train/loss_total": 0.3699428141117096 }, { "epoch": 2.168831168831169, "step": 835, "train/loss_ctc": 0.49298417568206787, "train/loss_error": 0.39629384875297546, "train/loss_total": 0.4156319200992584 }, { "epoch": 2.1714285714285713, "step": 836, "train/loss_ctc": 0.4269568622112274, "train/loss_error": 0.4253832995891571, "train/loss_total": 0.42569801211357117 }, { "epoch": 2.174025974025974, "step": 837, "train/loss_ctc": 0.4792553782463074, "train/loss_error": 0.32093533873558044, "train/loss_total": 0.3525993824005127 }, { "epoch": 2.1766233766233767, "step": 838, "train/loss_ctc": 0.49615466594696045, "train/loss_error": 0.3405884802341461, "train/loss_total": 0.37170174717903137 }, { "epoch": 2.1792207792207794, "step": 839, "train/loss_ctc": 0.34943920373916626, "train/loss_error": 0.31919753551483154, "train/loss_total": 0.3252458870410919 }, { "epoch": 2.1818181818181817, "grad_norm": 13184.44921875, "learning_rate": 1.6955844155844157e-05, "loss": 0.3813, "step": 840 }, { "epoch": 2.1818181818181817, "step": 840, "train/loss_ctc": 0.3961713910102844, "train/loss_error": 0.2756741940975189, "train/loss_total": 0.299773633480072 }, { "epoch": 2.1844155844155844, "step": 841, "train/loss_ctc": 0.5361825823783875, "train/loss_error": 0.4103766083717346, "train/loss_total": 0.43553781509399414 }, { "epoch": 2.187012987012987, "step": 842, "train/loss_ctc": 0.6315559148788452, "train/loss_error": 0.4890437126159668, "train/loss_total": 0.5175461769104004 }, { "epoch": 2.18961038961039, "step": 843, "train/loss_ctc": 0.5515539646148682, "train/loss_error": 0.30606094002723694, "train/loss_total": 0.35515955090522766 }, { "epoch": 2.192207792207792, "step": 844, "train/loss_ctc": 0.3134952783584595, "train/loss_error": 0.29397475719451904, "train/loss_total": 0.29787886142730713 }, { "epoch": 2.1948051948051948, "step": 845, "train/loss_ctc": 0.5819541215896606, "train/loss_error": 0.4310627281665802, "train/loss_total": 0.4612410068511963 }, { "epoch": 2.1974025974025975, "step": 846, "train/loss_ctc": 0.21159878373146057, "train/loss_error": 0.2707543671131134, "train/loss_total": 0.2589232623577118 }, { "epoch": 2.2, "step": 847, "train/loss_ctc": 0.3624574542045593, "train/loss_error": 0.4025708734989166, "train/loss_total": 0.3945482075214386 }, { "epoch": 2.2025974025974024, "step": 848, "train/loss_ctc": 0.3507847487926483, "train/loss_error": 0.4327497184276581, "train/loss_total": 0.41635674238204956 }, { "epoch": 2.205194805194805, "step": 849, "train/loss_ctc": 0.4901050925254822, "train/loss_error": 0.43362361192703247, "train/loss_total": 0.4449199140071869 }, { "epoch": 2.207792207792208, "grad_norm": 13661.2431640625, "learning_rate": 1.6800000000000002e-05, "loss": 0.3882, "step": 850 }, { "epoch": 2.207792207792208, "step": 850, "train/loss_ctc": 0.4294441342353821, "train/loss_error": 0.39411601424217224, "train/loss_total": 0.4011816382408142 }, { "epoch": 2.2103896103896106, "step": 851, "train/loss_ctc": 0.5181917548179626, "train/loss_error": 0.33252981305122375, "train/loss_total": 0.36966219544410706 }, { "epoch": 2.212987012987013, "step": 852, "train/loss_ctc": 0.5366232991218567, "train/loss_error": 0.3340369164943695, "train/loss_total": 0.37455418705940247 }, { "epoch": 2.2155844155844155, "step": 853, "train/loss_ctc": 0.37579742074012756, "train/loss_error": 0.3664291501045227, "train/loss_total": 0.3683028221130371 }, { "epoch": 2.2181818181818183, "step": 854, "train/loss_ctc": 0.6295367479324341, "train/loss_error": 0.37412410974502563, "train/loss_total": 0.42520666122436523 }, { "epoch": 2.220779220779221, "step": 855, "train/loss_ctc": 0.34499505162239075, "train/loss_error": 0.36687856912612915, "train/loss_total": 0.3625018894672394 }, { "epoch": 2.2233766233766232, "step": 856, "train/loss_ctc": 0.3896849751472473, "train/loss_error": 0.3899049162864685, "train/loss_total": 0.38986095786094666 }, { "epoch": 2.225974025974026, "step": 857, "train/loss_ctc": 0.5774383544921875, "train/loss_error": 0.3297578990459442, "train/loss_total": 0.3792939782142639 }, { "epoch": 2.2285714285714286, "step": 858, "train/loss_ctc": 0.4423345923423767, "train/loss_error": 0.37989503145217896, "train/loss_total": 0.392382949590683 }, { "epoch": 2.2311688311688314, "step": 859, "train/loss_ctc": 0.36325615644454956, "train/loss_error": 0.30499351024627686, "train/loss_total": 0.3166460394859314 }, { "epoch": 2.2337662337662336, "grad_norm": 12258.779296875, "learning_rate": 1.6644155844155847e-05, "loss": 0.378, "step": 860 }, { "epoch": 2.2337662337662336, "step": 860, "train/loss_ctc": 0.2928610146045685, "train/loss_error": 0.3560744524002075, "train/loss_total": 0.3434317708015442 }, { "epoch": 2.2363636363636363, "step": 861, "train/loss_ctc": 0.47508883476257324, "train/loss_error": 0.33788126707077026, "train/loss_total": 0.3653227686882019 }, { "epoch": 2.238961038961039, "step": 862, "train/loss_ctc": 0.4401177763938904, "train/loss_error": 0.37680450081825256, "train/loss_total": 0.38946717977523804 }, { "epoch": 2.2415584415584417, "step": 863, "train/loss_ctc": 0.421631395816803, "train/loss_error": 0.35988906025886536, "train/loss_total": 0.37223756313323975 }, { "epoch": 2.244155844155844, "step": 864, "train/loss_ctc": 0.5284383296966553, "train/loss_error": 0.3511538803577423, "train/loss_total": 0.3866107761859894 }, { "epoch": 2.2467532467532467, "step": 865, "train/loss_ctc": 0.26777908205986023, "train/loss_error": 0.290813148021698, "train/loss_total": 0.28620633482933044 }, { "epoch": 2.2493506493506494, "step": 866, "train/loss_ctc": 0.5045652985572815, "train/loss_error": 0.36966440081596375, "train/loss_total": 0.39664459228515625 }, { "epoch": 2.2519480519480517, "step": 867, "train/loss_ctc": 0.44637423753738403, "train/loss_error": 0.33404791355133057, "train/loss_total": 0.35651320219039917 }, { "epoch": 2.2545454545454544, "step": 868, "train/loss_ctc": 0.8210136890411377, "train/loss_error": 0.3279368281364441, "train/loss_total": 0.4265521764755249 }, { "epoch": 2.257142857142857, "step": 869, "train/loss_ctc": 0.3852221667766571, "train/loss_error": 0.36304742097854614, "train/loss_total": 0.36748236417770386 }, { "epoch": 2.25974025974026, "grad_norm": 11049.7861328125, "learning_rate": 1.648831168831169e-05, "loss": 0.369, "step": 870 }, { "epoch": 2.25974025974026, "step": 870, "train/loss_ctc": 0.42352306842803955, "train/loss_error": 0.3305385112762451, "train/loss_total": 0.3491354286670685 }, { "epoch": 2.2623376623376625, "step": 871, "train/loss_ctc": 0.3417712450027466, "train/loss_error": 0.34853196144104004, "train/loss_total": 0.3471798300743103 }, { "epoch": 2.264935064935065, "step": 872, "train/loss_ctc": 0.36496543884277344, "train/loss_error": 0.4026876986026764, "train/loss_total": 0.3951432704925537 }, { "epoch": 2.2675324675324675, "step": 873, "train/loss_ctc": 0.3995744585990906, "train/loss_error": 0.3172396719455719, "train/loss_total": 0.33370664715766907 }, { "epoch": 2.27012987012987, "step": 874, "train/loss_ctc": 0.5595875382423401, "train/loss_error": 0.4408365488052368, "train/loss_total": 0.4645867347717285 }, { "epoch": 2.2727272727272725, "step": 875, "train/loss_ctc": 0.4141700863838196, "train/loss_error": 0.3018035888671875, "train/loss_total": 0.3242768943309784 }, { "epoch": 2.275324675324675, "step": 876, "train/loss_ctc": 0.5877038836479187, "train/loss_error": 0.38164153695106506, "train/loss_total": 0.4228540062904358 }, { "epoch": 2.277922077922078, "step": 877, "train/loss_ctc": 0.553611159324646, "train/loss_error": 0.4526807367801666, "train/loss_total": 0.47286683320999146 }, { "epoch": 2.2805194805194806, "step": 878, "train/loss_ctc": 0.4939487874507904, "train/loss_error": 0.38070613145828247, "train/loss_total": 0.403354674577713 }, { "epoch": 2.2831168831168833, "step": 879, "train/loss_ctc": 0.31544309854507446, "train/loss_error": 0.2910305857658386, "train/loss_total": 0.29591310024261475 }, { "epoch": 2.2857142857142856, "grad_norm": 10650.7451171875, "learning_rate": 1.6332467532467533e-05, "loss": 0.3809, "step": 880 }, { "epoch": 2.2857142857142856, "step": 880, "train/loss_ctc": 0.8552532196044922, "train/loss_error": 0.3227745592594147, "train/loss_total": 0.42927032709121704 }, { "epoch": 2.2883116883116883, "step": 881, "train/loss_ctc": 0.589035153388977, "train/loss_error": 0.4120142459869385, "train/loss_total": 0.4474184215068817 }, { "epoch": 2.290909090909091, "step": 882, "train/loss_ctc": 0.6258327960968018, "train/loss_error": 0.37049737572669983, "train/loss_total": 0.4215644598007202 }, { "epoch": 2.2935064935064933, "step": 883, "train/loss_ctc": 0.566501259803772, "train/loss_error": 0.36322852969169617, "train/loss_total": 0.40388309955596924 }, { "epoch": 2.296103896103896, "step": 884, "train/loss_ctc": 0.38451507687568665, "train/loss_error": 0.3566967844963074, "train/loss_total": 0.36226046085357666 }, { "epoch": 2.2987012987012987, "step": 885, "train/loss_ctc": 0.36263740062713623, "train/loss_error": 0.2960042953491211, "train/loss_total": 0.30933094024658203 }, { "epoch": 2.3012987012987014, "step": 886, "train/loss_ctc": 0.5001669526100159, "train/loss_error": 0.3541005253791809, "train/loss_total": 0.3833138346672058 }, { "epoch": 2.303896103896104, "step": 887, "train/loss_ctc": 0.40376096963882446, "train/loss_error": 0.355701208114624, "train/loss_total": 0.36531317234039307 }, { "epoch": 2.3064935064935064, "step": 888, "train/loss_ctc": 0.3274621069431305, "train/loss_error": 0.36602798104286194, "train/loss_total": 0.3583148121833801 }, { "epoch": 2.309090909090909, "step": 889, "train/loss_ctc": 0.34785449504852295, "train/loss_error": 0.31362733244895935, "train/loss_total": 0.320472776889801 }, { "epoch": 2.311688311688312, "grad_norm": 10386.9765625, "learning_rate": 1.6176623376623378e-05, "loss": 0.3801, "step": 890 }, { "epoch": 2.311688311688312, "step": 890, "train/loss_ctc": 0.5358231067657471, "train/loss_error": 0.4248698353767395, "train/loss_total": 0.4470604956150055 }, { "epoch": 2.314285714285714, "step": 891, "train/loss_ctc": 0.39486342668533325, "train/loss_error": 0.39850401878356934, "train/loss_total": 0.39777591824531555 }, { "epoch": 2.3168831168831168, "step": 892, "train/loss_ctc": 0.336963027715683, "train/loss_error": 0.33320772647857666, "train/loss_total": 0.33395880460739136 }, { "epoch": 2.3194805194805195, "step": 893, "train/loss_ctc": 0.4207421839237213, "train/loss_error": 0.3473644256591797, "train/loss_total": 0.3620399832725525 }, { "epoch": 2.322077922077922, "step": 894, "train/loss_ctc": 0.868154764175415, "train/loss_error": 0.4767680764198303, "train/loss_total": 0.5550454258918762 }, { "epoch": 2.324675324675325, "step": 895, "train/loss_ctc": 0.5644359588623047, "train/loss_error": 0.3733848035335541, "train/loss_total": 0.41159504652023315 }, { "epoch": 2.327272727272727, "step": 896, "train/loss_ctc": 0.24893462657928467, "train/loss_error": 0.26292872428894043, "train/loss_total": 0.2601299285888672 }, { "epoch": 2.32987012987013, "step": 897, "train/loss_ctc": 0.24928735196590424, "train/loss_error": 0.34960201382637024, "train/loss_total": 0.32953909039497375 }, { "epoch": 2.3324675324675326, "step": 898, "train/loss_ctc": 0.36676496267318726, "train/loss_error": 0.3222757875919342, "train/loss_total": 0.3311736285686493 }, { "epoch": 2.335064935064935, "step": 899, "train/loss_ctc": 0.6042141318321228, "train/loss_error": 0.4123739004135132, "train/loss_total": 0.4507419466972351 }, { "epoch": 2.3376623376623376, "grad_norm": 17261.544921875, "learning_rate": 1.6020779220779222e-05, "loss": 0.3879, "step": 900 }, { "epoch": 2.3376623376623376, "step": 900, "train/loss_ctc": 0.3042517900466919, "train/loss_error": 0.3104218542575836, "train/loss_total": 0.3091878592967987 }, { "epoch": 2.3402597402597403, "step": 901, "train/loss_ctc": 0.3075050115585327, "train/loss_error": 0.3639039993286133, "train/loss_total": 0.35262420773506165 }, { "epoch": 2.342857142857143, "step": 902, "train/loss_ctc": 0.5191371440887451, "train/loss_error": 0.41619253158569336, "train/loss_total": 0.43678146600723267 }, { "epoch": 2.3454545454545457, "step": 903, "train/loss_ctc": 0.6698498725891113, "train/loss_error": 0.30073294043540955, "train/loss_total": 0.3745563328266144 }, { "epoch": 2.348051948051948, "step": 904, "train/loss_ctc": 0.5230435132980347, "train/loss_error": 0.4539378881454468, "train/loss_total": 0.46775904297828674 }, { "epoch": 2.3506493506493507, "step": 905, "train/loss_ctc": 0.36879462003707886, "train/loss_error": 0.3833230137825012, "train/loss_total": 0.3804173469543457 }, { "epoch": 2.3532467532467534, "step": 906, "train/loss_ctc": 0.41836926341056824, "train/loss_error": 0.37661874294281006, "train/loss_total": 0.3849688768386841 }, { "epoch": 2.3558441558441556, "step": 907, "train/loss_ctc": 0.4777686893939972, "train/loss_error": 0.34318503737449646, "train/loss_total": 0.3701017498970032 }, { "epoch": 2.3584415584415583, "step": 908, "train/loss_ctc": 0.30809980630874634, "train/loss_error": 0.4174201190471649, "train/loss_total": 0.3955560624599457 }, { "epoch": 2.361038961038961, "step": 909, "train/loss_ctc": 0.3987725079059601, "train/loss_error": 0.3329008221626282, "train/loss_total": 0.346075177192688 }, { "epoch": 2.3636363636363638, "grad_norm": 11590.8994140625, "learning_rate": 1.5864935064935067e-05, "loss": 0.3818, "step": 910 }, { "epoch": 2.3636363636363638, "step": 910, "train/loss_ctc": 0.2585296630859375, "train/loss_error": 0.2889813780784607, "train/loss_total": 0.28289103507995605 }, { "epoch": 2.3662337662337665, "step": 911, "train/loss_ctc": 0.36967164278030396, "train/loss_error": 0.35053351521492004, "train/loss_total": 0.3543611764907837 }, { "epoch": 2.3688311688311687, "step": 912, "train/loss_ctc": 0.5156214833259583, "train/loss_error": 0.38265547156333923, "train/loss_total": 0.4092486798763275 }, { "epoch": 2.3714285714285714, "step": 913, "train/loss_ctc": 0.4536289870738983, "train/loss_error": 0.4199873208999634, "train/loss_total": 0.4267156720161438 }, { "epoch": 2.374025974025974, "step": 914, "train/loss_ctc": 0.405021607875824, "train/loss_error": 0.3256690800189972, "train/loss_total": 0.341539591550827 }, { "epoch": 2.3766233766233764, "step": 915, "train/loss_ctc": 0.4043554663658142, "train/loss_error": 0.4135401248931885, "train/loss_total": 0.4117031991481781 }, { "epoch": 2.379220779220779, "step": 916, "train/loss_ctc": 0.5190176963806152, "train/loss_error": 0.3533239960670471, "train/loss_total": 0.3864627480506897 }, { "epoch": 2.381818181818182, "step": 917, "train/loss_ctc": 0.49542713165283203, "train/loss_error": 0.3192252516746521, "train/loss_total": 0.35446563363075256 }, { "epoch": 2.3844155844155845, "step": 918, "train/loss_ctc": 0.4609020948410034, "train/loss_error": 0.32467424869537354, "train/loss_total": 0.35191982984542847 }, { "epoch": 2.3870129870129873, "step": 919, "train/loss_ctc": 0.39600056409835815, "train/loss_error": 0.34840673208236694, "train/loss_total": 0.35792550444602966 }, { "epoch": 2.3896103896103895, "grad_norm": 10488.96875, "learning_rate": 1.5709090909090912e-05, "loss": 0.3677, "step": 920 }, { "epoch": 2.3896103896103895, "step": 920, "train/loss_ctc": 0.49644187092781067, "train/loss_error": 0.3313000202178955, "train/loss_total": 0.36432838439941406 }, { "epoch": 2.3922077922077922, "step": 921, "train/loss_ctc": 0.3289913833141327, "train/loss_error": 0.2964610159397125, "train/loss_total": 0.3029671013355255 }, { "epoch": 2.394805194805195, "step": 922, "train/loss_ctc": 0.45928263664245605, "train/loss_error": 0.30344098806381226, "train/loss_total": 0.33460932970046997 }, { "epoch": 2.397402597402597, "step": 923, "train/loss_ctc": 0.37868231534957886, "train/loss_error": 0.36096495389938354, "train/loss_total": 0.36450842022895813 }, { "epoch": 2.4, "step": 924, "train/loss_ctc": 0.5226849317550659, "train/loss_error": 0.4583260416984558, "train/loss_total": 0.47119781374931335 }, { "epoch": 2.4025974025974026, "step": 925, "train/loss_ctc": 0.5038869380950928, "train/loss_error": 0.33191192150115967, "train/loss_total": 0.36630693078041077 }, { "epoch": 2.4051948051948053, "step": 926, "train/loss_ctc": 0.6973129510879517, "train/loss_error": 0.48890599608421326, "train/loss_total": 0.5305874347686768 }, { "epoch": 2.407792207792208, "step": 927, "train/loss_ctc": 0.5963142514228821, "train/loss_error": 0.4232082664966583, "train/loss_total": 0.45782947540283203 }, { "epoch": 2.4103896103896103, "step": 928, "train/loss_ctc": 0.5373433828353882, "train/loss_error": 0.438041090965271, "train/loss_total": 0.4579015374183655 }, { "epoch": 2.412987012987013, "step": 929, "train/loss_ctc": 0.39108288288116455, "train/loss_error": 0.29574787616729736, "train/loss_total": 0.31481486558914185 }, { "epoch": 2.4155844155844157, "grad_norm": 13252.0185546875, "learning_rate": 1.5553246753246753e-05, "loss": 0.3965, "step": 930 }, { "epoch": 2.4155844155844157, "step": 930, "train/loss_ctc": 0.3054746389389038, "train/loss_error": 0.3099878430366516, "train/loss_total": 0.3090852200984955 }, { "epoch": 2.418181818181818, "step": 931, "train/loss_ctc": 0.49812227487564087, "train/loss_error": 0.4103512763977051, "train/loss_total": 0.42790547013282776 }, { "epoch": 2.4207792207792207, "step": 932, "train/loss_ctc": 0.42130279541015625, "train/loss_error": 0.29572293162345886, "train/loss_total": 0.32083889842033386 }, { "epoch": 2.4233766233766234, "step": 933, "train/loss_ctc": 0.4151044189929962, "train/loss_error": 0.3334619402885437, "train/loss_total": 0.34979045391082764 }, { "epoch": 2.425974025974026, "step": 934, "train/loss_ctc": 0.5209363698959351, "train/loss_error": 0.37500062584877014, "train/loss_total": 0.40418779850006104 }, { "epoch": 2.4285714285714284, "step": 935, "train/loss_ctc": 0.44657832384109497, "train/loss_error": 0.31894436478614807, "train/loss_total": 0.34447115659713745 }, { "epoch": 2.431168831168831, "step": 936, "train/loss_ctc": 0.39212140440940857, "train/loss_error": 0.3281785547733307, "train/loss_total": 0.3409671187400818 }, { "epoch": 2.433766233766234, "step": 937, "train/loss_ctc": 0.3683585226535797, "train/loss_error": 0.32391655445098877, "train/loss_total": 0.33280494809150696 }, { "epoch": 2.4363636363636365, "step": 938, "train/loss_ctc": 0.3175581097602844, "train/loss_error": 0.2729666233062744, "train/loss_total": 0.28188490867614746 }, { "epoch": 2.4389610389610388, "step": 939, "train/loss_ctc": 0.43126559257507324, "train/loss_error": 0.29712948203086853, "train/loss_total": 0.3239567279815674 }, { "epoch": 2.4415584415584415, "grad_norm": 11682.5078125, "learning_rate": 1.53974025974026e-05, "loss": 0.3436, "step": 940 }, { "epoch": 2.4415584415584415, "step": 940, "train/loss_ctc": 0.23989704251289368, "train/loss_error": 0.30813315510749817, "train/loss_total": 0.2944859266281128 }, { "epoch": 2.444155844155844, "step": 941, "train/loss_ctc": 0.2894848585128784, "train/loss_error": 0.3239074647426605, "train/loss_total": 0.3170229494571686 }, { "epoch": 2.446753246753247, "step": 942, "train/loss_ctc": 0.3449432849884033, "train/loss_error": 0.2471223622560501, "train/loss_total": 0.2666865587234497 }, { "epoch": 2.449350649350649, "step": 943, "train/loss_ctc": 0.3473716974258423, "train/loss_error": 0.2952207326889038, "train/loss_total": 0.3056509494781494 }, { "epoch": 2.451948051948052, "step": 944, "train/loss_ctc": 0.499568909406662, "train/loss_error": 0.29812246561050415, "train/loss_total": 0.33841174840927124 }, { "epoch": 2.4545454545454546, "step": 945, "train/loss_ctc": 0.36810022592544556, "train/loss_error": 0.3498402237892151, "train/loss_total": 0.35349223017692566 }, { "epoch": 2.4571428571428573, "step": 946, "train/loss_ctc": 0.5339991450309753, "train/loss_error": 0.36081841588020325, "train/loss_total": 0.3954545855522156 }, { "epoch": 2.4597402597402596, "step": 947, "train/loss_ctc": 0.4120730757713318, "train/loss_error": 0.3625624477863312, "train/loss_total": 0.3724645972251892 }, { "epoch": 2.4623376623376623, "step": 948, "train/loss_ctc": 0.3367123007774353, "train/loss_error": 0.4186733067035675, "train/loss_total": 0.40228110551834106 }, { "epoch": 2.464935064935065, "step": 949, "train/loss_ctc": 0.6363128423690796, "train/loss_error": 0.42760953307151794, "train/loss_total": 0.4693502187728882 }, { "epoch": 2.4675324675324677, "grad_norm": 15376.2275390625, "learning_rate": 1.5241558441558443e-05, "loss": 0.3515, "step": 950 }, { "epoch": 2.4675324675324677, "step": 950, "train/loss_ctc": 0.4712243676185608, "train/loss_error": 0.36283108592033386, "train/loss_total": 0.38450974225997925 }, { "epoch": 2.47012987012987, "step": 951, "train/loss_ctc": 0.35674357414245605, "train/loss_error": 0.32991835474967957, "train/loss_total": 0.33528342843055725 }, { "epoch": 2.4727272727272727, "step": 952, "train/loss_ctc": 0.712581992149353, "train/loss_error": 0.3968988060951233, "train/loss_total": 0.46003544330596924 }, { "epoch": 2.4753246753246754, "step": 953, "train/loss_ctc": 0.37550288438796997, "train/loss_error": 0.2508161664009094, "train/loss_total": 0.27575352787971497 }, { "epoch": 2.477922077922078, "step": 954, "train/loss_ctc": 0.5048550367355347, "train/loss_error": 0.3890848159790039, "train/loss_total": 0.41223886609077454 }, { "epoch": 2.4805194805194803, "step": 955, "train/loss_ctc": 0.3353637754917145, "train/loss_error": 0.2975614368915558, "train/loss_total": 0.30512189865112305 }, { "epoch": 2.483116883116883, "step": 956, "train/loss_ctc": 0.40571239590644836, "train/loss_error": 0.3550129234790802, "train/loss_total": 0.36515283584594727 }, { "epoch": 2.4857142857142858, "step": 957, "train/loss_ctc": 0.4761955142021179, "train/loss_error": 0.29828447103500366, "train/loss_total": 0.333866685628891 }, { "epoch": 2.4883116883116885, "step": 958, "train/loss_ctc": 0.5744301676750183, "train/loss_error": 0.38737931847572327, "train/loss_total": 0.42478951811790466 }, { "epoch": 2.4909090909090907, "step": 959, "train/loss_ctc": 0.3141331672668457, "train/loss_error": 0.32524779438972473, "train/loss_total": 0.3230248689651489 }, { "epoch": 2.4935064935064934, "grad_norm": 12196.572265625, "learning_rate": 1.5085714285714288e-05, "loss": 0.362, "step": 960 }, { "epoch": 2.4935064935064934, "step": 960, "train/loss_ctc": 0.3864890933036804, "train/loss_error": 0.39570966362953186, "train/loss_total": 0.39386558532714844 }, { "epoch": 2.496103896103896, "step": 961, "train/loss_ctc": 0.21793481707572937, "train/loss_error": 0.37672919034957886, "train/loss_total": 0.34497031569480896 }, { "epoch": 2.498701298701299, "step": 962, "train/loss_ctc": 0.5268540382385254, "train/loss_error": 0.3073601722717285, "train/loss_total": 0.3512589633464813 }, { "epoch": 2.501298701298701, "step": 963, "train/loss_ctc": 0.5801652669906616, "train/loss_error": 0.3606026768684387, "train/loss_total": 0.40451520681381226 }, { "epoch": 2.503896103896104, "step": 964, "train/loss_ctc": 0.5102086663246155, "train/loss_error": 0.3311788737773895, "train/loss_total": 0.36698484420776367 }, { "epoch": 2.5064935064935066, "step": 965, "train/loss_ctc": 0.4433797597885132, "train/loss_error": 0.32293644547462463, "train/loss_total": 0.3470250964164734 }, { "epoch": 2.509090909090909, "step": 966, "train/loss_ctc": 0.45079463720321655, "train/loss_error": 0.4087880551815033, "train/loss_total": 0.4171893894672394 }, { "epoch": 2.5116883116883115, "step": 967, "train/loss_ctc": 0.32031160593032837, "train/loss_error": 0.3192206621170044, "train/loss_total": 0.3194388747215271 }, { "epoch": 2.5142857142857142, "step": 968, "train/loss_ctc": 0.45405441522598267, "train/loss_error": 0.3422890305519104, "train/loss_total": 0.36464211344718933 }, { "epoch": 2.516883116883117, "step": 969, "train/loss_ctc": 0.31223464012145996, "train/loss_error": 0.2887568473815918, "train/loss_total": 0.2934524118900299 }, { "epoch": 2.5194805194805197, "grad_norm": 10439.2880859375, "learning_rate": 1.492987012987013e-05, "loss": 0.3603, "step": 970 }, { "epoch": 2.5194805194805197, "step": 970, "train/loss_ctc": 0.31005120277404785, "train/loss_error": 0.30236130952835083, "train/loss_total": 0.30389928817749023 }, { "epoch": 2.522077922077922, "step": 971, "train/loss_ctc": 0.41103604435920715, "train/loss_error": 0.2916954457759857, "train/loss_total": 0.3155635595321655 }, { "epoch": 2.5246753246753246, "step": 972, "train/loss_ctc": 0.4235996603965759, "train/loss_error": 0.36243706941604614, "train/loss_total": 0.3746695816516876 }, { "epoch": 2.5272727272727273, "step": 973, "train/loss_ctc": 0.5435839891433716, "train/loss_error": 0.38526788353919983, "train/loss_total": 0.4169310927391052 }, { "epoch": 2.5298701298701296, "step": 974, "train/loss_ctc": 0.5114194750785828, "train/loss_error": 0.24596187472343445, "train/loss_total": 0.2990534007549286 }, { "epoch": 2.5324675324675323, "step": 975, "train/loss_ctc": 0.6060614585876465, "train/loss_error": 0.29173508286476135, "train/loss_total": 0.35460036993026733 }, { "epoch": 2.535064935064935, "step": 976, "train/loss_ctc": 0.36004841327667236, "train/loss_error": 0.2422797530889511, "train/loss_total": 0.2658334970474243 }, { "epoch": 2.5376623376623377, "step": 977, "train/loss_ctc": 0.46894896030426025, "train/loss_error": 0.29830247163772583, "train/loss_total": 0.33243176341056824 }, { "epoch": 2.5402597402597404, "step": 978, "train/loss_ctc": 0.5419163703918457, "train/loss_error": 0.3000902831554413, "train/loss_total": 0.3484555184841156 }, { "epoch": 2.5428571428571427, "step": 979, "train/loss_ctc": 0.3767039179801941, "train/loss_error": 0.3667455017566681, "train/loss_total": 0.36873719096183777 }, { "epoch": 2.5454545454545454, "grad_norm": 12114.365234375, "learning_rate": 1.4774025974025974e-05, "loss": 0.338, "step": 980 }, { "epoch": 2.5454545454545454, "step": 980, "train/loss_ctc": 0.31906867027282715, "train/loss_error": 0.33615654706954956, "train/loss_total": 0.332738995552063 }, { "epoch": 2.548051948051948, "step": 981, "train/loss_ctc": 0.48368558287620544, "train/loss_error": 0.3801955580711365, "train/loss_total": 0.40089356899261475 }, { "epoch": 2.5506493506493504, "step": 982, "train/loss_ctc": 0.5989481210708618, "train/loss_error": 0.3020153343677521, "train/loss_total": 0.36140188574790955 }, { "epoch": 2.553246753246753, "step": 983, "train/loss_ctc": 0.35907572507858276, "train/loss_error": 0.3173680603504181, "train/loss_total": 0.32570958137512207 }, { "epoch": 2.555844155844156, "step": 984, "train/loss_ctc": 0.5504027605056763, "train/loss_error": 0.35737699270248413, "train/loss_total": 0.39598214626312256 }, { "epoch": 2.5584415584415585, "step": 985, "train/loss_ctc": 0.5900946855545044, "train/loss_error": 0.3949063718318939, "train/loss_total": 0.43394404649734497 }, { "epoch": 2.5610389610389612, "step": 986, "train/loss_ctc": 0.5437309145927429, "train/loss_error": 0.3631269931793213, "train/loss_total": 0.39924776554107666 }, { "epoch": 2.5636363636363635, "step": 987, "train/loss_ctc": 0.684668779373169, "train/loss_error": 0.38786762952804565, "train/loss_total": 0.4472278952598572 }, { "epoch": 2.566233766233766, "step": 988, "train/loss_ctc": 0.5236945748329163, "train/loss_error": 0.3820498287677765, "train/loss_total": 0.4103787839412689 }, { "epoch": 2.568831168831169, "step": 989, "train/loss_ctc": 0.39974352717399597, "train/loss_error": 0.3041248917579651, "train/loss_total": 0.32324862480163574 }, { "epoch": 2.571428571428571, "grad_norm": 24975.705078125, "learning_rate": 1.4618181818181817e-05, "loss": 0.3831, "step": 990 }, { "epoch": 2.571428571428571, "step": 990, "train/loss_ctc": 0.5010501146316528, "train/loss_error": 0.387592077255249, "train/loss_total": 0.4102836847305298 }, { "epoch": 2.574025974025974, "step": 991, "train/loss_ctc": 0.5038841962814331, "train/loss_error": 0.3242257237434387, "train/loss_total": 0.36015743017196655 }, { "epoch": 2.5766233766233766, "step": 992, "train/loss_ctc": 0.4651426076889038, "train/loss_error": 0.41147780418395996, "train/loss_total": 0.4222107529640198 }, { "epoch": 2.5792207792207793, "step": 993, "train/loss_ctc": 0.4818176329135895, "train/loss_error": 0.31300824880599976, "train/loss_total": 0.34677010774612427 }, { "epoch": 2.581818181818182, "step": 994, "train/loss_ctc": 0.5965245366096497, "train/loss_error": 0.32545045018196106, "train/loss_total": 0.3796652555465698 }, { "epoch": 2.5844155844155843, "step": 995, "train/loss_ctc": 0.6103013753890991, "train/loss_error": 0.376872718334198, "train/loss_total": 0.42355844378471375 }, { "epoch": 2.587012987012987, "step": 996, "train/loss_ctc": 0.439606636762619, "train/loss_error": 0.30947065353393555, "train/loss_total": 0.3354978561401367 }, { "epoch": 2.5896103896103897, "step": 997, "train/loss_ctc": 0.33455801010131836, "train/loss_error": 0.37465059757232666, "train/loss_total": 0.3666321039199829 }, { "epoch": 2.592207792207792, "step": 998, "train/loss_ctc": 0.4021854102611542, "train/loss_error": 0.40826359391212463, "train/loss_total": 0.40704798698425293 }, { "epoch": 2.5948051948051947, "step": 999, "train/loss_ctc": 0.8027676939964294, "train/loss_error": 0.47852087020874023, "train/loss_total": 0.543370246887207 }, { "epoch": 2.5974025974025974, "grad_norm": 19377.7421875, "learning_rate": 1.4462337662337664e-05, "loss": 0.3995, "step": 1000 }, { "epoch": 2.5974025974025974, "step": 1000, "train/loss_ctc": 0.3359498977661133, "train/loss_error": 0.3643854856491089, "train/loss_total": 0.35869839787483215 }, { "epoch": 2.6, "step": 1001, "train/loss_ctc": 0.3431236743927002, "train/loss_error": 0.33923184871673584, "train/loss_total": 0.34001022577285767 }, { "epoch": 2.602597402597403, "step": 1002, "train/loss_ctc": 0.3237950801849365, "train/loss_error": 0.3454827666282654, "train/loss_total": 0.34114524722099304 }, { "epoch": 2.605194805194805, "step": 1003, "train/loss_ctc": 0.4676632881164551, "train/loss_error": 0.36020606756210327, "train/loss_total": 0.38169750571250916 }, { "epoch": 2.6077922077922078, "step": 1004, "train/loss_ctc": 0.33911603689193726, "train/loss_error": 0.3254324197769165, "train/loss_total": 0.3281691372394562 }, { "epoch": 2.6103896103896105, "step": 1005, "train/loss_ctc": 0.49431324005126953, "train/loss_error": 0.4437451660633087, "train/loss_total": 0.45385879278182983 }, { "epoch": 2.6129870129870127, "step": 1006, "train/loss_ctc": 0.20570050179958344, "train/loss_error": 0.2254614233970642, "train/loss_total": 0.22150924801826477 }, { "epoch": 2.6155844155844155, "step": 1007, "train/loss_ctc": 0.5519058704376221, "train/loss_error": 0.34444305300712585, "train/loss_total": 0.38593563437461853 }, { "epoch": 2.618181818181818, "step": 1008, "train/loss_ctc": 0.41213488578796387, "train/loss_error": 0.36870619654655457, "train/loss_total": 0.3773919343948364 }, { "epoch": 2.620779220779221, "step": 1009, "train/loss_ctc": 0.4886101186275482, "train/loss_error": 0.32390668988227844, "train/loss_total": 0.3568473756313324 }, { "epoch": 2.6233766233766236, "grad_norm": 13479.025390625, "learning_rate": 1.4306493506493507e-05, "loss": 0.3545, "step": 1010 }, { "epoch": 2.6233766233766236, "step": 1010, "train/loss_ctc": 0.4230859875679016, "train/loss_error": 0.37788882851600647, "train/loss_total": 0.3869282603263855 }, { "epoch": 2.625974025974026, "step": 1011, "train/loss_ctc": 0.4544793963432312, "train/loss_error": 0.3667750954627991, "train/loss_total": 0.38431596755981445 }, { "epoch": 2.6285714285714286, "step": 1012, "train/loss_ctc": 0.42728349566459656, "train/loss_error": 0.34447452425956726, "train/loss_total": 0.3610363304615021 }, { "epoch": 2.6311688311688313, "step": 1013, "train/loss_ctc": 0.7162882089614868, "train/loss_error": 0.3976109027862549, "train/loss_total": 0.4613463878631592 }, { "epoch": 2.6337662337662335, "step": 1014, "train/loss_ctc": 0.5741870403289795, "train/loss_error": 0.3786415755748749, "train/loss_total": 0.41775068640708923 }, { "epoch": 2.6363636363636362, "step": 1015, "train/loss_ctc": 0.32833266258239746, "train/loss_error": 0.31575995683670044, "train/loss_total": 0.31827449798583984 }, { "epoch": 2.638961038961039, "step": 1016, "train/loss_ctc": 0.3528439998626709, "train/loss_error": 0.38794949650764465, "train/loss_total": 0.3809283971786499 }, { "epoch": 2.6415584415584417, "step": 1017, "train/loss_ctc": 0.7795848250389099, "train/loss_error": 0.3405327796936035, "train/loss_total": 0.42834317684173584 }, { "epoch": 2.6441558441558444, "step": 1018, "train/loss_ctc": 0.41325774788856506, "train/loss_error": 0.377971351146698, "train/loss_total": 0.3850286304950714 }, { "epoch": 2.6467532467532466, "step": 1019, "train/loss_ctc": 0.4749829173088074, "train/loss_error": 0.4470319449901581, "train/loss_total": 0.4526221752166748 }, { "epoch": 2.6493506493506493, "grad_norm": 12586.201171875, "learning_rate": 1.4150649350649352e-05, "loss": 0.3977, "step": 1020 }, { "epoch": 2.6493506493506493, "step": 1020, "train/loss_ctc": 0.44906193017959595, "train/loss_error": 0.29496753215789795, "train/loss_total": 0.32578641176223755 }, { "epoch": 2.651948051948052, "step": 1021, "train/loss_ctc": 0.3824901580810547, "train/loss_error": 0.3230908215045929, "train/loss_total": 0.3349706828594208 }, { "epoch": 2.6545454545454543, "step": 1022, "train/loss_ctc": 0.33122843503952026, "train/loss_error": 0.2533060908317566, "train/loss_total": 0.2688905596733093 }, { "epoch": 2.657142857142857, "step": 1023, "train/loss_ctc": 0.7809158563613892, "train/loss_error": 0.4211518168449402, "train/loss_total": 0.49310463666915894 }, { "epoch": 2.6597402597402597, "step": 1024, "train/loss_ctc": 0.3384808301925659, "train/loss_error": 0.2656119465827942, "train/loss_total": 0.280185729265213 }, { "epoch": 2.6623376623376624, "step": 1025, "train/loss_ctc": 0.3991542458534241, "train/loss_error": 0.3319140672683716, "train/loss_total": 0.34536212682724 }, { "epoch": 2.664935064935065, "step": 1026, "train/loss_ctc": 0.4263971447944641, "train/loss_error": 0.27060630917549133, "train/loss_total": 0.30176448822021484 }, { "epoch": 2.6675324675324674, "step": 1027, "train/loss_ctc": 0.6230997443199158, "train/loss_error": 0.4409349262714386, "train/loss_total": 0.47736790776252747 }, { "epoch": 2.67012987012987, "step": 1028, "train/loss_ctc": 0.7268500328063965, "train/loss_error": 0.3007884621620178, "train/loss_total": 0.38600078225135803 }, { "epoch": 2.672727272727273, "step": 1029, "train/loss_ctc": 0.5477238893508911, "train/loss_error": 0.3456062078475952, "train/loss_total": 0.38602975010871887 }, { "epoch": 2.675324675324675, "grad_norm": 16531.73828125, "learning_rate": 1.3994805194805197e-05, "loss": 0.3599, "step": 1030 }, { "epoch": 2.675324675324675, "step": 1030, "train/loss_ctc": 0.6396021246910095, "train/loss_error": 0.39248448610305786, "train/loss_total": 0.44190800189971924 }, { "epoch": 2.677922077922078, "step": 1031, "train/loss_ctc": 0.397212415933609, "train/loss_error": 0.3524024486541748, "train/loss_total": 0.3613644242286682 }, { "epoch": 2.6805194805194805, "step": 1032, "train/loss_ctc": 0.9689115285873413, "train/loss_error": 0.4776173532009125, "train/loss_total": 0.5758761763572693 }, { "epoch": 2.6831168831168832, "step": 1033, "train/loss_ctc": 0.405029296875, "train/loss_error": 0.2887507677078247, "train/loss_total": 0.31200647354125977 }, { "epoch": 2.685714285714286, "step": 1034, "train/loss_ctc": 0.4063393771648407, "train/loss_error": 0.33835330605506897, "train/loss_total": 0.3519505262374878 }, { "epoch": 2.688311688311688, "step": 1035, "train/loss_ctc": 0.4409845173358917, "train/loss_error": 0.3529564440250397, "train/loss_total": 0.3705620765686035 }, { "epoch": 2.690909090909091, "step": 1036, "train/loss_ctc": 0.5681995749473572, "train/loss_error": 0.4013751149177551, "train/loss_total": 0.43474000692367554 }, { "epoch": 2.6935064935064936, "step": 1037, "train/loss_ctc": 0.3555120825767517, "train/loss_error": 0.3785693645477295, "train/loss_total": 0.37395790219306946 }, { "epoch": 2.696103896103896, "step": 1038, "train/loss_ctc": 0.389099657535553, "train/loss_error": 0.36740854382514954, "train/loss_total": 0.3717467784881592 }, { "epoch": 2.6987012987012986, "step": 1039, "train/loss_ctc": 0.31708022952079773, "train/loss_error": 0.3282821476459503, "train/loss_total": 0.3260417580604553 }, { "epoch": 2.7012987012987013, "grad_norm": 9943.48046875, "learning_rate": 1.383896103896104e-05, "loss": 0.392, "step": 1040 }, { "epoch": 2.7012987012987013, "step": 1040, "train/loss_ctc": 0.3958554267883301, "train/loss_error": 0.3058752715587616, "train/loss_total": 0.32387131452560425 }, { "epoch": 2.703896103896104, "step": 1041, "train/loss_ctc": 0.6608508825302124, "train/loss_error": 0.4292660653591156, "train/loss_total": 0.475583016872406 }, { "epoch": 2.7064935064935067, "step": 1042, "train/loss_ctc": 0.3551744818687439, "train/loss_error": 0.3446804881095886, "train/loss_total": 0.34677931666374207 }, { "epoch": 2.709090909090909, "step": 1043, "train/loss_ctc": 0.3650969862937927, "train/loss_error": 0.4160190522670746, "train/loss_total": 0.4058346748352051 }, { "epoch": 2.7116883116883117, "step": 1044, "train/loss_ctc": 0.4018913209438324, "train/loss_error": 0.3680332601070404, "train/loss_total": 0.37480488419532776 }, { "epoch": 2.7142857142857144, "step": 1045, "train/loss_ctc": 0.26926690340042114, "train/loss_error": 0.2548079788684845, "train/loss_total": 0.25769975781440735 }, { "epoch": 2.7168831168831167, "step": 1046, "train/loss_ctc": 0.285256564617157, "train/loss_error": 0.30223071575164795, "train/loss_total": 0.2988358736038208 }, { "epoch": 2.7194805194805194, "step": 1047, "train/loss_ctc": 0.6344508528709412, "train/loss_error": 0.36738482117652893, "train/loss_total": 0.42079800367355347 }, { "epoch": 2.722077922077922, "step": 1048, "train/loss_ctc": 0.46885985136032104, "train/loss_error": 0.3398069143295288, "train/loss_total": 0.3656175136566162 }, { "epoch": 2.724675324675325, "step": 1049, "train/loss_ctc": 0.5081998705863953, "train/loss_error": 0.34029528498649597, "train/loss_total": 0.3738762140274048 }, { "epoch": 2.7272727272727275, "grad_norm": 13697.8076171875, "learning_rate": 1.3683116883116884e-05, "loss": 0.3644, "step": 1050 }, { "epoch": 2.7272727272727275, "step": 1050, "train/loss_ctc": 1.3412374258041382, "train/loss_error": 0.4034794270992279, "train/loss_total": 0.5910310745239258 }, { "epoch": 2.72987012987013, "step": 1051, "train/loss_ctc": 0.5152893662452698, "train/loss_error": 0.36841729283332825, "train/loss_total": 0.3977917432785034 }, { "epoch": 2.7324675324675325, "step": 1052, "train/loss_ctc": 0.3485276699066162, "train/loss_error": 0.26765817403793335, "train/loss_total": 0.2838320732116699 }, { "epoch": 2.735064935064935, "step": 1053, "train/loss_ctc": 0.6056705117225647, "train/loss_error": 0.3612663745880127, "train/loss_total": 0.41014721989631653 }, { "epoch": 2.7376623376623375, "step": 1054, "train/loss_ctc": 0.3276284337043762, "train/loss_error": 0.27140092849731445, "train/loss_total": 0.28264644742012024 }, { "epoch": 2.74025974025974, "step": 1055, "train/loss_ctc": 0.4437589943408966, "train/loss_error": 0.41497474908828735, "train/loss_total": 0.4207316040992737 }, { "epoch": 2.742857142857143, "step": 1056, "train/loss_ctc": 0.34419000148773193, "train/loss_error": 0.3200010657310486, "train/loss_total": 0.3248388469219208 }, { "epoch": 2.7454545454545456, "step": 1057, "train/loss_ctc": 0.4461744427680969, "train/loss_error": 0.3395451307296753, "train/loss_total": 0.36087098717689514 }, { "epoch": 2.7480519480519483, "step": 1058, "train/loss_ctc": 0.28912895917892456, "train/loss_error": 0.28174155950546265, "train/loss_total": 0.28321903944015503 }, { "epoch": 2.7506493506493506, "step": 1059, "train/loss_ctc": 0.2615073025226593, "train/loss_error": 0.38956132531166077, "train/loss_total": 0.3639505207538605 }, { "epoch": 2.7532467532467533, "grad_norm": 7881.39892578125, "learning_rate": 1.3527272727272728e-05, "loss": 0.3719, "step": 1060 }, { "epoch": 2.7532467532467533, "step": 1060, "train/loss_ctc": 0.4909559190273285, "train/loss_error": 0.39947786927223206, "train/loss_total": 0.4177734851837158 }, { "epoch": 2.755844155844156, "step": 1061, "train/loss_ctc": 0.9446845650672913, "train/loss_error": 0.4011959731578827, "train/loss_total": 0.5098937153816223 }, { "epoch": 2.7584415584415583, "step": 1062, "train/loss_ctc": 0.6148326992988586, "train/loss_error": 0.4411488473415375, "train/loss_total": 0.47588562965393066 }, { "epoch": 2.761038961038961, "step": 1063, "train/loss_ctc": 0.4429991841316223, "train/loss_error": 0.4512679874897003, "train/loss_total": 0.4496142268180847 }, { "epoch": 2.7636363636363637, "step": 1064, "train/loss_ctc": 0.5832711458206177, "train/loss_error": 0.3111698627471924, "train/loss_total": 0.3655901253223419 }, { "epoch": 2.7662337662337664, "step": 1065, "train/loss_ctc": 0.5172616243362427, "train/loss_error": 0.40243634581565857, "train/loss_total": 0.4254014194011688 }, { "epoch": 2.768831168831169, "step": 1066, "train/loss_ctc": 0.3046184778213501, "train/loss_error": 0.32525426149368286, "train/loss_total": 0.32112711668014526 }, { "epoch": 2.7714285714285714, "step": 1067, "train/loss_ctc": 0.38172096014022827, "train/loss_error": 0.3168317973613739, "train/loss_total": 0.32980963587760925 }, { "epoch": 2.774025974025974, "step": 1068, "train/loss_ctc": 0.42061787843704224, "train/loss_error": 0.36899834871292114, "train/loss_total": 0.37932226061820984 }, { "epoch": 2.7766233766233768, "step": 1069, "train/loss_ctc": 0.5655828714370728, "train/loss_error": 0.34589412808418274, "train/loss_total": 0.38983190059661865 }, { "epoch": 2.779220779220779, "grad_norm": 14750.6416015625, "learning_rate": 1.3371428571428572e-05, "loss": 0.4064, "step": 1070 }, { "epoch": 2.779220779220779, "step": 1070, "train/loss_ctc": 0.4762229919433594, "train/loss_error": 0.39219897985458374, "train/loss_total": 0.4090037941932678 }, { "epoch": 2.7818181818181817, "step": 1071, "train/loss_ctc": 0.5772247314453125, "train/loss_error": 0.36075371503829956, "train/loss_total": 0.4040479362010956 }, { "epoch": 2.7844155844155845, "step": 1072, "train/loss_ctc": 0.27823176980018616, "train/loss_error": 0.30715063214302063, "train/loss_total": 0.3013668656349182 }, { "epoch": 2.787012987012987, "step": 1073, "train/loss_ctc": 0.5921212434768677, "train/loss_error": 0.3096419870853424, "train/loss_total": 0.36613786220550537 }, { "epoch": 2.78961038961039, "step": 1074, "train/loss_ctc": 0.4193434417247772, "train/loss_error": 0.35128942131996155, "train/loss_total": 0.36490023136138916 }, { "epoch": 2.792207792207792, "step": 1075, "train/loss_ctc": 0.3539276719093323, "train/loss_error": 0.355386346578598, "train/loss_total": 0.3550946116447449 }, { "epoch": 2.794805194805195, "step": 1076, "train/loss_ctc": 0.33785325288772583, "train/loss_error": 0.2863316833972931, "train/loss_total": 0.2966359853744507 }, { "epoch": 2.7974025974025976, "step": 1077, "train/loss_ctc": 0.2754999101161957, "train/loss_error": 0.3911743462085724, "train/loss_total": 0.36803948879241943 }, { "epoch": 2.8, "step": 1078, "train/loss_ctc": 0.6298069357872009, "train/loss_error": 0.4267227351665497, "train/loss_total": 0.46733957529067993 }, { "epoch": 2.8025974025974025, "step": 1079, "train/loss_ctc": 0.5144349336624146, "train/loss_error": 0.3637296259403229, "train/loss_total": 0.3938707113265991 }, { "epoch": 2.8051948051948052, "grad_norm": 14102.5791015625, "learning_rate": 1.3215584415584416e-05, "loss": 0.3726, "step": 1080 }, { "epoch": 2.8051948051948052, "step": 1080, "train/loss_ctc": 0.4010191559791565, "train/loss_error": 0.31220394372940063, "train/loss_total": 0.3299669921398163 }, { "epoch": 2.807792207792208, "step": 1081, "train/loss_ctc": 0.42148423194885254, "train/loss_error": 0.39566460251808167, "train/loss_total": 0.4008285403251648 }, { "epoch": 2.8103896103896107, "step": 1082, "train/loss_ctc": 0.3738340437412262, "train/loss_error": 0.39634841680526733, "train/loss_total": 0.39184555411338806 }, { "epoch": 2.812987012987013, "step": 1083, "train/loss_ctc": 0.2400420606136322, "train/loss_error": 0.2835612893104553, "train/loss_total": 0.27485746145248413 }, { "epoch": 2.8155844155844156, "step": 1084, "train/loss_ctc": 0.46195563673973083, "train/loss_error": 0.295743852853775, "train/loss_total": 0.32898619771003723 }, { "epoch": 2.8181818181818183, "step": 1085, "train/loss_ctc": 0.6207825541496277, "train/loss_error": 0.37224653363227844, "train/loss_total": 0.4219537377357483 }, { "epoch": 2.8207792207792206, "step": 1086, "train/loss_ctc": 0.2524707317352295, "train/loss_error": 0.2730770707130432, "train/loss_total": 0.268955796957016 }, { "epoch": 2.8233766233766233, "step": 1087, "train/loss_ctc": 0.3452378511428833, "train/loss_error": 0.3006857931613922, "train/loss_total": 0.3095962107181549 }, { "epoch": 2.825974025974026, "step": 1088, "train/loss_ctc": 0.6059728264808655, "train/loss_error": 0.38849517703056335, "train/loss_total": 0.43199071288108826 }, { "epoch": 2.8285714285714287, "step": 1089, "train/loss_ctc": 0.41945880651474, "train/loss_error": 0.43626460433006287, "train/loss_total": 0.4329034686088562 }, { "epoch": 2.8311688311688314, "grad_norm": 12539.021484375, "learning_rate": 1.305974025974026e-05, "loss": 0.3592, "step": 1090 }, { "epoch": 2.8311688311688314, "step": 1090, "train/loss_ctc": 0.255490243434906, "train/loss_error": 0.2613048255443573, "train/loss_total": 0.26014190912246704 }, { "epoch": 2.8337662337662337, "step": 1091, "train/loss_ctc": 0.5618157386779785, "train/loss_error": 0.47440415620803833, "train/loss_total": 0.4918864965438843 }, { "epoch": 2.8363636363636364, "step": 1092, "train/loss_ctc": 0.37609827518463135, "train/loss_error": 0.3363550305366516, "train/loss_total": 0.344303697347641 }, { "epoch": 2.838961038961039, "step": 1093, "train/loss_ctc": 0.29173487424850464, "train/loss_error": 0.24101954698562622, "train/loss_total": 0.2511626183986664 }, { "epoch": 2.8415584415584414, "step": 1094, "train/loss_ctc": 0.24591045081615448, "train/loss_error": 0.30535703897476196, "train/loss_total": 0.2934677302837372 }, { "epoch": 2.844155844155844, "step": 1095, "train/loss_ctc": 0.5275003910064697, "train/loss_error": 0.3744862377643585, "train/loss_total": 0.4050890803337097 }, { "epoch": 2.846753246753247, "step": 1096, "train/loss_ctc": 0.32905811071395874, "train/loss_error": 0.2873765826225281, "train/loss_total": 0.2957128882408142 }, { "epoch": 2.849350649350649, "step": 1097, "train/loss_ctc": 0.5667786002159119, "train/loss_error": 0.3789600431919098, "train/loss_total": 0.4165237545967102 }, { "epoch": 2.851948051948052, "step": 1098, "train/loss_ctc": 0.3077855110168457, "train/loss_error": 0.2724267840385437, "train/loss_total": 0.27949854731559753 }, { "epoch": 2.8545454545454545, "step": 1099, "train/loss_ctc": 0.4460987448692322, "train/loss_error": 0.37806469202041626, "train/loss_total": 0.3916715085506439 }, { "epoch": 2.857142857142857, "grad_norm": 10228.3818359375, "learning_rate": 1.2903896103896105e-05, "loss": 0.3429, "step": 1100 }, { "epoch": 2.857142857142857, "step": 1100, "train/loss_ctc": 0.43718844652175903, "train/loss_error": 0.3687076270580292, "train/loss_total": 0.38240379095077515 }, { "epoch": 2.85974025974026, "step": 1101, "train/loss_ctc": 0.4254368543624878, "train/loss_error": 0.27278557419776917, "train/loss_total": 0.30331581830978394 }, { "epoch": 2.862337662337662, "step": 1102, "train/loss_ctc": 0.3201931118965149, "train/loss_error": 0.29890385270118713, "train/loss_total": 0.30316171050071716 }, { "epoch": 2.864935064935065, "step": 1103, "train/loss_ctc": 0.3811696171760559, "train/loss_error": 0.27779510617256165, "train/loss_total": 0.29847002029418945 }, { "epoch": 2.8675324675324676, "step": 1104, "train/loss_ctc": 0.43547648191452026, "train/loss_error": 0.44553208351135254, "train/loss_total": 0.4435209631919861 }, { "epoch": 2.87012987012987, "step": 1105, "train/loss_ctc": 0.3854406476020813, "train/loss_error": 0.2978976368904114, "train/loss_total": 0.31540626287460327 }, { "epoch": 2.8727272727272726, "step": 1106, "train/loss_ctc": 0.5187088251113892, "train/loss_error": 0.3342038094997406, "train/loss_total": 0.37110480666160583 }, { "epoch": 2.8753246753246753, "step": 1107, "train/loss_ctc": 0.34003329277038574, "train/loss_error": 0.24998174607753754, "train/loss_total": 0.2679920494556427 }, { "epoch": 2.877922077922078, "step": 1108, "train/loss_ctc": 0.3158590793609619, "train/loss_error": 0.40207549929618835, "train/loss_total": 0.3848322033882141 }, { "epoch": 2.8805194805194807, "step": 1109, "train/loss_ctc": 0.41531047224998474, "train/loss_error": 0.376665323972702, "train/loss_total": 0.3843943476676941 }, { "epoch": 2.883116883116883, "grad_norm": 14236.7548828125, "learning_rate": 1.2748051948051948e-05, "loss": 0.3455, "step": 1110 }, { "epoch": 2.883116883116883, "step": 1110, "train/loss_ctc": 0.5211119651794434, "train/loss_error": 0.3929343521595001, "train/loss_total": 0.4185698628425598 }, { "epoch": 2.8857142857142857, "step": 1111, "train/loss_ctc": 0.4091382920742035, "train/loss_error": 0.3124958872795105, "train/loss_total": 0.3318243622779846 }, { "epoch": 2.8883116883116884, "step": 1112, "train/loss_ctc": 0.3656706213951111, "train/loss_error": 0.3014504015445709, "train/loss_total": 0.3142944574356079 }, { "epoch": 2.8909090909090907, "step": 1113, "train/loss_ctc": 0.33343249559402466, "train/loss_error": 0.306309312582016, "train/loss_total": 0.3117339611053467 }, { "epoch": 2.8935064935064934, "step": 1114, "train/loss_ctc": 0.6050374507904053, "train/loss_error": 0.35112464427948, "train/loss_total": 0.4019072353839874 }, { "epoch": 2.896103896103896, "step": 1115, "train/loss_ctc": 0.491107314825058, "train/loss_error": 0.34310606122016907, "train/loss_total": 0.3727062940597534 }, { "epoch": 2.898701298701299, "step": 1116, "train/loss_ctc": 0.3576526641845703, "train/loss_error": 0.34035131335258484, "train/loss_total": 0.343811571598053 }, { "epoch": 2.9012987012987015, "step": 1117, "train/loss_ctc": 0.6902976036071777, "train/loss_error": 0.32937633991241455, "train/loss_total": 0.40156060457229614 }, { "epoch": 2.9038961038961038, "step": 1118, "train/loss_ctc": 0.46538805961608887, "train/loss_error": 0.3724183440208435, "train/loss_total": 0.3910123109817505 }, { "epoch": 2.9064935064935065, "step": 1119, "train/loss_ctc": 0.5785304307937622, "train/loss_error": 0.3427664041519165, "train/loss_total": 0.3899192214012146 }, { "epoch": 2.909090909090909, "grad_norm": 16413.43359375, "learning_rate": 1.2592207792207793e-05, "loss": 0.3677, "step": 1120 }, { "epoch": 2.909090909090909, "step": 1120, "train/loss_ctc": 0.332150936126709, "train/loss_error": 0.3290652930736542, "train/loss_total": 0.3296824097633362 }, { "epoch": 2.9116883116883114, "step": 1121, "train/loss_ctc": 0.48885005712509155, "train/loss_error": 0.3803443908691406, "train/loss_total": 0.40204551815986633 }, { "epoch": 2.914285714285714, "step": 1122, "train/loss_ctc": 0.30326324701309204, "train/loss_error": 0.2894461452960968, "train/loss_total": 0.29220956563949585 }, { "epoch": 2.916883116883117, "step": 1123, "train/loss_ctc": 0.8381088972091675, "train/loss_error": 0.31392213702201843, "train/loss_total": 0.41875946521759033 }, { "epoch": 2.9194805194805196, "step": 1124, "train/loss_ctc": 0.3647855520248413, "train/loss_error": 0.29426315426826477, "train/loss_total": 0.30836763978004456 }, { "epoch": 2.9220779220779223, "step": 1125, "train/loss_ctc": 0.48739951848983765, "train/loss_error": 0.380413293838501, "train/loss_total": 0.40181055665016174 }, { "epoch": 2.9246753246753245, "step": 1126, "train/loss_ctc": 0.2286536991596222, "train/loss_error": 0.24641497433185577, "train/loss_total": 0.24286271631717682 }, { "epoch": 2.9272727272727272, "step": 1127, "train/loss_ctc": 0.3748511075973511, "train/loss_error": 0.37766045331954956, "train/loss_total": 0.37709859013557434 }, { "epoch": 2.92987012987013, "step": 1128, "train/loss_ctc": 0.3869403004646301, "train/loss_error": 0.3380477726459503, "train/loss_total": 0.3478263020515442 }, { "epoch": 2.9324675324675322, "step": 1129, "train/loss_ctc": 0.48469728231430054, "train/loss_error": 0.3803933560848236, "train/loss_total": 0.40125417709350586 }, { "epoch": 2.935064935064935, "grad_norm": 12587.390625, "learning_rate": 1.2436363636363636e-05, "loss": 0.3522, "step": 1130 }, { "epoch": 2.935064935064935, "step": 1130, "train/loss_ctc": 0.6830769777297974, "train/loss_error": 0.4817976951599121, "train/loss_total": 0.522053599357605 }, { "epoch": 2.9376623376623376, "step": 1131, "train/loss_ctc": 0.27480456233024597, "train/loss_error": 0.3310162425041199, "train/loss_total": 0.31977391242980957 }, { "epoch": 2.9402597402597404, "step": 1132, "train/loss_ctc": 0.31573235988616943, "train/loss_error": 0.2694210708141327, "train/loss_total": 0.2786833345890045 }, { "epoch": 2.942857142857143, "step": 1133, "train/loss_ctc": 0.5026955604553223, "train/loss_error": 0.3186335861682892, "train/loss_total": 0.3554459810256958 }, { "epoch": 2.9454545454545453, "step": 1134, "train/loss_ctc": 0.44162654876708984, "train/loss_error": 0.3074513375759125, "train/loss_total": 0.3342863917350769 }, { "epoch": 2.948051948051948, "step": 1135, "train/loss_ctc": 0.2865985035896301, "train/loss_error": 0.3367452323436737, "train/loss_total": 0.326715886592865 }, { "epoch": 2.9506493506493507, "step": 1136, "train/loss_ctc": 0.3986034393310547, "train/loss_error": 0.3223287761211395, "train/loss_total": 0.3375837206840515 }, { "epoch": 2.953246753246753, "step": 1137, "train/loss_ctc": 0.6021392345428467, "train/loss_error": 0.2872830033302307, "train/loss_total": 0.35025423765182495 }, { "epoch": 2.9558441558441557, "step": 1138, "train/loss_ctc": 0.5490780472755432, "train/loss_error": 0.3173528015613556, "train/loss_total": 0.36369788646698 }, { "epoch": 2.9584415584415584, "step": 1139, "train/loss_ctc": 0.37890517711639404, "train/loss_error": 0.3370150625705719, "train/loss_total": 0.3453930914402008 }, { "epoch": 2.961038961038961, "grad_norm": 10213.638671875, "learning_rate": 1.2280519480519481e-05, "loss": 0.3534, "step": 1140 }, { "epoch": 2.961038961038961, "step": 1140, "train/loss_ctc": 0.39650633931159973, "train/loss_error": 0.3598039746284485, "train/loss_total": 0.36714446544647217 }, { "epoch": 2.963636363636364, "step": 1141, "train/loss_ctc": 0.3849213421344757, "train/loss_error": 0.3123362362384796, "train/loss_total": 0.32685327529907227 }, { "epoch": 2.966233766233766, "step": 1142, "train/loss_ctc": 0.5356845855712891, "train/loss_error": 0.39868518710136414, "train/loss_total": 0.42608505487442017 }, { "epoch": 2.968831168831169, "step": 1143, "train/loss_ctc": 0.3530384302139282, "train/loss_error": 0.3107089102268219, "train/loss_total": 0.3191748261451721 }, { "epoch": 2.9714285714285715, "step": 1144, "train/loss_ctc": 0.6977394223213196, "train/loss_error": 0.4455972909927368, "train/loss_total": 0.4960257112979889 }, { "epoch": 2.974025974025974, "step": 1145, "train/loss_ctc": 0.49092382192611694, "train/loss_error": 0.4624221920967102, "train/loss_total": 0.4681225121021271 }, { "epoch": 2.9766233766233765, "step": 1146, "train/loss_ctc": 0.331593781709671, "train/loss_error": 0.3744965195655823, "train/loss_total": 0.365915983915329 }, { "epoch": 2.979220779220779, "step": 1147, "train/loss_ctc": 0.4550088047981262, "train/loss_error": 0.4232358932495117, "train/loss_total": 0.42959046363830566 }, { "epoch": 2.981818181818182, "step": 1148, "train/loss_ctc": 0.637073814868927, "train/loss_error": 0.37403789162635803, "train/loss_total": 0.42664507031440735 }, { "epoch": 2.9844155844155846, "step": 1149, "train/loss_ctc": 0.4746593236923218, "train/loss_error": 0.3213474750518799, "train/loss_total": 0.3520098626613617 }, { "epoch": 2.987012987012987, "grad_norm": 11156.0634765625, "learning_rate": 1.2124675324675326e-05, "loss": 0.3978, "step": 1150 }, { "epoch": 2.987012987012987, "step": 1150, "train/loss_ctc": 0.400226354598999, "train/loss_error": 0.34000593423843384, "train/loss_total": 0.3520500361919403 }, { "epoch": 2.9896103896103896, "step": 1151, "train/loss_ctc": 0.5040577054023743, "train/loss_error": 0.4025910794734955, "train/loss_total": 0.42288440465927124 }, { "epoch": 2.9922077922077923, "step": 1152, "train/loss_ctc": 0.44067493081092834, "train/loss_error": 0.3355536162853241, "train/loss_total": 0.35657787322998047 }, { "epoch": 2.9948051948051946, "step": 1153, "train/loss_ctc": 0.5008596181869507, "train/loss_error": 0.2806050777435303, "train/loss_total": 0.32465600967407227 }, { "epoch": 2.9974025974025973, "step": 1154, "train/loss_ctc": 0.3694598078727722, "train/loss_error": 0.38329190015792847, "train/loss_total": 0.38052549958229065 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.558677077293396, "val/loss_error": 0.13691769540309906, "val/loss_total": 0.22126957774162292 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5417988300323486, "val/loss_error": 0.4019646942615509, "val/loss_total": 0.42993152141571045 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.07170018553733826, "val/loss_error": 0.25661590695381165, "val/loss_total": 0.21963277459144592 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.2218426913022995, "val/loss_error": 0.11852965503931046, "val/loss_total": 0.13919226825237274 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.16050730645656586, "val/loss_error": 0.17315874993801117, "val/loss_total": 0.17062847316265106 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.10759793967008591, "val/loss_error": 0.2902950346469879, "val/loss_total": 0.2537556290626526 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.10233234614133835, "val/loss_error": 0.4002545475959778, "val/loss_total": 0.340670108795166 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.3810737729072571, "val/loss_error": 0.2799699008464813, "val/loss_total": 0.30019068717956543 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.3352818489074707, "val/loss_error": 0.2870366871356964, "val/loss_total": 0.29668572545051575 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.32673051953315735, "val/loss_error": 0.3387492001056671, "val/loss_total": 0.33634549379348755 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.12056879699230194, "val/loss_error": 0.13153424859046936, "val/loss_total": 0.12934115529060364 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.16448494791984558, "val/loss_error": 0.25852736830711365, "val/loss_total": 0.23971889913082123 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.783676266670227, "val/loss_error": 0.37573567032814026, "val/loss_total": 0.4573237895965576 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.817157506942749, "val/loss_error": 0.5455952882766724, "val/loss_total": 0.5999077558517456 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5621130466461182, "val/loss_error": 0.39866819977760315, "val/loss_total": 0.43135717511177063 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.459769606590271, "val/loss_error": 0.6367101669311523, "val/loss_total": 0.6013220548629761 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.08612766116857529, "val/loss_error": 0.2272319197654724, "val/loss_total": 0.19901107251644135 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.1788029819726944, "val/loss_error": 0.26904767751693726, "val/loss_total": 0.25099873542785645 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.10082969814538956, "val/loss_error": 0.25828781723976135, "val/loss_total": 0.2267961949110031 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5428022742271423, "val/loss_error": 0.42256027460098267, "val/loss_total": 0.44660866260528564 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.7910698652267456, "val/loss_error": 0.46177101135253906, "val/loss_total": 0.5276308059692383 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.2542940676212311, "val/loss_error": 0.5453252196311951, "val/loss_total": 0.48711901903152466 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.26327964663505554, "val/loss_error": 0.3631434738636017, "val/loss_total": 0.34317073225975037 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.22938291728496552, "val/loss_error": 0.24099691212177277, "val/loss_total": 0.2386741042137146 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.3399481177330017, "val/loss_error": 0.5737252831459045, "val/loss_total": 0.526969850063324 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5065654516220093, "val/loss_error": 0.5171324610710144, "val/loss_total": 0.5150190591812134 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.28908151388168335, "val/loss_error": 0.38589048385620117, "val/loss_total": 0.3665286898612976 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.14812591671943665, "val/loss_error": 0.221714049577713, "val/loss_total": 0.20699642598628998 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.017371706664562225, "val/loss_error": 0.07427627593278885, "val/loss_total": 0.06289536505937576 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.27256524562835693, "val/loss_error": 0.5943254232406616, "val/loss_total": 0.5299733877182007 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.08794596791267395, "val/loss_error": 0.17853567004203796, "val/loss_total": 0.16041773557662964 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.12909747660160065, "val/loss_error": 0.2503993511199951, "val/loss_total": 0.22613897919654846 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.09511659294366837, "val/loss_error": 0.13481752574443817, "val/loss_total": 0.1268773376941681 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5954804420471191, "val/loss_error": 0.6389543414115906, "val/loss_total": 0.6302595734596252 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.06816008687019348, "val/loss_error": 0.3057107627391815, "val/loss_total": 0.25820064544677734 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.21778355538845062, "val/loss_error": 0.36022138595581055, "val/loss_total": 0.3317338228225708 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.425350546836853, "val/loss_error": 0.5873180627822876, "val/loss_total": 0.5549245476722717 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.337068110704422, "val/loss_error": 0.40594589710235596, "val/loss_total": 0.39217033982276917 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.22391797602176666, "val/loss_error": 0.47947028279304504, "val/loss_total": 0.42835983633995056 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.6075563430786133, "val/loss_error": 0.694672167301178, "val/loss_total": 0.677249014377594 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.25145667791366577, "val/loss_error": 0.22845903038978577, "val/loss_total": 0.23305857181549072 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.25449836254119873, "val/loss_error": 0.2604964077472687, "val/loss_total": 0.25929680466651917 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5990200638771057, "val/loss_error": 0.935547411441803, "val/loss_total": 0.8682419657707214 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.4872681200504303, "val/loss_error": 0.5742895603179932, "val/loss_total": 0.5568852424621582 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.49751538038253784, "val/loss_error": 0.33356907963752747, "val/loss_total": 0.36635833978652954 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.3657495677471161, "val/loss_error": 0.6477700471878052, "val/loss_total": 0.5913659930229187 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.6473710536956787, "val/loss_error": 0.4657976031303406, "val/loss_total": 0.5021122694015503 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.3286575973033905, "val/loss_error": 0.525371789932251, "val/loss_total": 0.4860289692878723 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.29871299862861633, "val/loss_error": 0.2858392596244812, "val/loss_total": 0.28841400146484375 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.6415817141532898, "val/loss_error": 0.7970724701881409, "val/loss_total": 0.7659743428230286 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.23522622883319855, "val/loss_error": 0.7398200035095215, "val/loss_total": 0.6389012336730957 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.32973751425743103, "val/loss_error": 0.3189559280872345, "val/loss_total": 0.3211122453212738 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.4181536138057709, "val/loss_error": 0.3463008999824524, "val/loss_total": 0.3606714606285095 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5706080794334412, "val/loss_error": 0.8729390501976013, "val/loss_total": 0.8124728798866272 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.3328649699687958, "val/loss_error": 0.3061009347438812, "val/loss_total": 0.31145375967025757 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.22904954850673676, "val/loss_error": 0.5012235045433044, "val/loss_total": 0.4467887282371521 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.115186408162117, "val/loss_error": 0.21977739036083221, "val/loss_total": 0.19885919988155365 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.49662312865257263, "val/loss_error": 0.26763471961021423, "val/loss_total": 0.31343239545822144 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5834164023399353, "val/loss_error": 0.5653486251831055, "val/loss_total": 0.5689622163772583 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.08349885046482086, "val/loss_error": 0.12427686154842377, "val/loss_total": 0.11612126231193542 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.8775995969772339, "val/loss_error": 0.5807669758796692, "val/loss_total": 0.6401335000991821 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.678492546081543, "val/loss_error": 0.37283793091773987, "val/loss_total": 0.43396884202957153 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.082896389067173, "val/loss_error": 0.09558231383562088, "val/loss_total": 0.09304513037204742 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.1851939857006073, "val/loss_error": 0.23193804919719696, "val/loss_total": 0.22258923947811127 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.11538302153348923, "val/loss_error": 0.2195618599653244, "val/loss_total": 0.1987261027097702 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.2769984304904938, "val/loss_error": 0.42746400833129883, "val/loss_total": 0.3973709046840668 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.3472467362880707, "val/loss_error": 0.3077479600906372, "val/loss_total": 0.3156477212905884 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5227027535438538, "val/loss_error": 0.5034463405609131, "val/loss_total": 0.5072976350784302 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.42645421624183655, "val/loss_error": 0.46556323766708374, "val/loss_total": 0.4577414393424988 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5643413662910461, "val/loss_error": 0.49268028140068054, "val/loss_total": 0.5070124864578247 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.20443099737167358, "val/loss_error": 0.23611056804656982, "val/loss_total": 0.22977465391159058 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5466470718383789, "val/loss_error": 0.5907017588615417, "val/loss_total": 0.5818908214569092 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.47607746720314026, "val/loss_error": 0.6952195763587952, "val/loss_total": 0.6513911485671997 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.19576337933540344, "val/loss_error": 0.37341558933258057, "val/loss_total": 0.33788517117500305 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.2807046175003052, "val/loss_error": 0.31910815834999084, "val/loss_total": 0.3114274740219116 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.03618351370096207, "val/loss_error": 0.24603570997714996, "val/loss_total": 0.20406527817249298 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.770343542098999, "val/loss_error": 0.47062790393829346, "val/loss_total": 0.5305710434913635 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.7740697264671326, "val/loss_error": 0.37369805574417114, "val/loss_total": 0.4537723958492279 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5320001244544983, "val/loss_error": 0.3172321021556854, "val/loss_total": 0.36018574237823486 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.0030581075698137283, "val/loss_error": 0.3594382107257843, "val/loss_total": 0.2881622016429901 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.521463930606842, "val/loss_error": 0.2853222191333771, "val/loss_total": 0.3325505554676056 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.024880163371562958, "val/loss_error": 0.06749839335680008, "val/loss_total": 0.058974750339984894 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.16855624318122864, "val/loss_error": 0.2332584410905838, "val/loss_total": 0.220318004488945 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.49516287446022034, "val/loss_error": 0.3798414468765259, "val/loss_total": 0.40290573239326477 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.4216235280036926, "val/loss_error": 0.5802978873252869, "val/loss_total": 0.5485630035400391 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.491010844707489, "val/loss_error": 0.5725066661834717, "val/loss_total": 0.556207537651062 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.04079688340425491, "val/loss_error": 0.09641529619693756, "val/loss_total": 0.08529161661863327 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.37958472967147827, "val/loss_error": 0.3840288519859314, "val/loss_total": 0.38314002752304077 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.28022244572639465, "val/loss_error": 0.23536589741706848, "val/loss_total": 0.24433721601963043 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.3465307950973511, "val/loss_error": 0.4597403407096863, "val/loss_total": 0.4370984435081482 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.1983662247657776, "val/loss_error": 0.45736974477767944, "val/loss_total": 0.40556904673576355 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.9072129726409912, "val/loss_error": 0.41780781745910645, "val/loss_total": 0.5156888365745544 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.24553990364074707, "val/loss_error": 0.05860064551234245, "val/loss_total": 0.09598849713802338 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.40267956256866455, "val/loss_error": 0.3209279775619507, "val/loss_total": 0.3372783064842224 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5865243673324585, "val/loss_error": 0.539273202419281, "val/loss_total": 0.5487234592437744 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.15305311977863312, "val/loss_error": 0.19854766130447388, "val/loss_total": 0.1894487589597702 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.12265346199274063, "val/loss_error": 0.4449571669101715, "val/loss_total": 0.38049641251564026 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.6631127595901489, "val/loss_error": 0.552442729473114, "val/loss_total": 0.574576735496521 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.11918668448925018, "val/loss_error": 0.292680948972702, "val/loss_total": 0.2579821050167084 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.1475834995508194, "val/loss_error": 0.24509884417057037, "val/loss_total": 0.22559577226638794 }, { "epoch": 3.0, "eval_eval/f1_0": 0.5239294767379761, "eval_eval/f1_1": 0.9333568215370178, "eval_eval/precision_0": 0.584269642829895, "eval_eval/precision_1": 0.920055627822876, "eval_eval/recall_0": 0.47488585114479065, "eval_eval/recall_1": 0.9470483064651489, "eval_eval/wer": 0.1535340271836624, "eval_loss": 0.3741190433502197, "eval_runtime": 7.3493, "eval_samples_per_second": 13.607, "eval_steps_per_second": 13.607, "step": 1155 }, { "epoch": 3.0, "step": 1155, "train/loss_ctc": 0.4397742748260498, "train/loss_error": 0.27782750129699707, "train/loss_total": 0.31021687388420105 }, { "epoch": 3.0025974025974027, "step": 1156, "train/loss_ctc": 0.5669206976890564, "train/loss_error": 0.36870047450065613, "train/loss_total": 0.4083445072174072 }, { "epoch": 3.005194805194805, "step": 1157, "train/loss_ctc": 0.5759349465370178, "train/loss_error": 0.3529115617275238, "train/loss_total": 0.39751625061035156 }, { "epoch": 3.0077922077922077, "step": 1158, "train/loss_ctc": 0.5897786617279053, "train/loss_error": 0.30120518803596497, "train/loss_total": 0.3589198887348175 }, { "epoch": 3.0103896103896104, "step": 1159, "train/loss_ctc": 0.33148765563964844, "train/loss_error": 0.35147181153297424, "train/loss_total": 0.34747499227523804 }, { "epoch": 3.012987012987013, "grad_norm": 12935.6953125, "learning_rate": 1.1968831168831169e-05, "loss": 0.3659, "step": 1160 }, { "epoch": 3.012987012987013, "step": 1160, "train/loss_ctc": 0.28932613134384155, "train/loss_error": 0.2479289472103119, "train/loss_total": 0.2562083899974823 }, { "epoch": 3.0155844155844154, "step": 1161, "train/loss_ctc": 0.46804356575012207, "train/loss_error": 0.362758606672287, "train/loss_total": 0.38381558656692505 }, { "epoch": 3.018181818181818, "step": 1162, "train/loss_ctc": 0.4847143292427063, "train/loss_error": 0.3878646492958069, "train/loss_total": 0.4072346091270447 }, { "epoch": 3.020779220779221, "step": 1163, "train/loss_ctc": 0.44275039434432983, "train/loss_error": 0.33488860726356506, "train/loss_total": 0.3564609885215759 }, { "epoch": 3.0233766233766235, "step": 1164, "train/loss_ctc": 0.3997976779937744, "train/loss_error": 0.32416003942489624, "train/loss_total": 0.33928757905960083 }, { "epoch": 3.0259740259740258, "step": 1165, "train/loss_ctc": 0.3804735541343689, "train/loss_error": 0.2968178689479828, "train/loss_total": 0.3135490119457245 }, { "epoch": 3.0285714285714285, "step": 1166, "train/loss_ctc": 0.35522282123565674, "train/loss_error": 0.3426780700683594, "train/loss_total": 0.3451870381832123 }, { "epoch": 3.031168831168831, "step": 1167, "train/loss_ctc": 0.3018759489059448, "train/loss_error": 0.3742286264896393, "train/loss_total": 0.35975807905197144 }, { "epoch": 3.033766233766234, "step": 1168, "train/loss_ctc": 0.2822358310222626, "train/loss_error": 0.2996844947338104, "train/loss_total": 0.29619476199150085 }, { "epoch": 3.036363636363636, "step": 1169, "train/loss_ctc": 0.359291136264801, "train/loss_error": 0.36741334199905396, "train/loss_total": 0.36578890681266785 }, { "epoch": 3.038961038961039, "grad_norm": 15868.6708984375, "learning_rate": 1.1812987012987014e-05, "loss": 0.3423, "step": 1170 }, { "epoch": 3.038961038961039, "step": 1170, "train/loss_ctc": 0.3482920825481415, "train/loss_error": 0.35160544514656067, "train/loss_total": 0.35094279050827026 }, { "epoch": 3.0415584415584416, "step": 1171, "train/loss_ctc": 0.6288483738899231, "train/loss_error": 0.33399519324302673, "train/loss_total": 0.39296582341194153 }, { "epoch": 3.0441558441558443, "step": 1172, "train/loss_ctc": 0.3256325423717499, "train/loss_error": 0.3292348384857178, "train/loss_total": 0.3285143971443176 }, { "epoch": 3.0467532467532465, "step": 1173, "train/loss_ctc": 0.30756092071533203, "train/loss_error": 0.29037177562713623, "train/loss_total": 0.29380959272384644 }, { "epoch": 3.0493506493506493, "step": 1174, "train/loss_ctc": 0.4061161279678345, "train/loss_error": 0.385626882314682, "train/loss_total": 0.3897247314453125 }, { "epoch": 3.051948051948052, "step": 1175, "train/loss_ctc": 0.3376118838787079, "train/loss_error": 0.3124948740005493, "train/loss_total": 0.31751829385757446 }, { "epoch": 3.0545454545454547, "step": 1176, "train/loss_ctc": 0.48716890811920166, "train/loss_error": 0.3863060176372528, "train/loss_total": 0.4064785838127136 }, { "epoch": 3.057142857142857, "step": 1177, "train/loss_ctc": 0.32568925619125366, "train/loss_error": 0.3500104546546936, "train/loss_total": 0.3451462388038635 }, { "epoch": 3.0597402597402596, "step": 1178, "train/loss_ctc": 0.696115255355835, "train/loss_error": 0.4068076014518738, "train/loss_total": 0.4646691679954529 }, { "epoch": 3.0623376623376624, "step": 1179, "train/loss_ctc": 0.3937501907348633, "train/loss_error": 0.35003119707107544, "train/loss_total": 0.3587750196456909 }, { "epoch": 3.064935064935065, "grad_norm": 10853.2646484375, "learning_rate": 1.1657142857142857e-05, "loss": 0.3649, "step": 1180 }, { "epoch": 3.064935064935065, "step": 1180, "train/loss_ctc": 0.5451711416244507, "train/loss_error": 0.3427681028842926, "train/loss_total": 0.3832487165927887 }, { "epoch": 3.0675324675324673, "step": 1181, "train/loss_ctc": 0.42449983954429626, "train/loss_error": 0.3623098134994507, "train/loss_total": 0.3747478127479553 }, { "epoch": 3.07012987012987, "step": 1182, "train/loss_ctc": 0.4070984721183777, "train/loss_error": 0.3481074273586273, "train/loss_total": 0.3599056601524353 }, { "epoch": 3.0727272727272728, "step": 1183, "train/loss_ctc": 0.4963136315345764, "train/loss_error": 0.3442727327346802, "train/loss_total": 0.37468093633651733 }, { "epoch": 3.0753246753246755, "step": 1184, "train/loss_ctc": 0.42733219265937805, "train/loss_error": 0.36642783880233765, "train/loss_total": 0.37860873341560364 }, { "epoch": 3.0779220779220777, "step": 1185, "train/loss_ctc": 0.31578487157821655, "train/loss_error": 0.2602558434009552, "train/loss_total": 0.27136164903640747 }, { "epoch": 3.0805194805194804, "step": 1186, "train/loss_ctc": 0.4501042068004608, "train/loss_error": 0.4336707890033722, "train/loss_total": 0.4369574785232544 }, { "epoch": 3.083116883116883, "step": 1187, "train/loss_ctc": 0.4803403615951538, "train/loss_error": 0.2934953570365906, "train/loss_total": 0.3308643698692322 }, { "epoch": 3.085714285714286, "step": 1188, "train/loss_ctc": 0.4368302822113037, "train/loss_error": 0.33392199873924255, "train/loss_total": 0.35450369119644165 }, { "epoch": 3.088311688311688, "step": 1189, "train/loss_ctc": 0.317857950925827, "train/loss_error": 0.3734382390975952, "train/loss_total": 0.36232221126556396 }, { "epoch": 3.090909090909091, "grad_norm": 9622.572265625, "learning_rate": 1.1501298701298702e-05, "loss": 0.3627, "step": 1190 }, { "epoch": 3.090909090909091, "step": 1190, "train/loss_ctc": 0.3360961079597473, "train/loss_error": 0.3325328528881073, "train/loss_total": 0.33324551582336426 }, { "epoch": 3.0935064935064935, "step": 1191, "train/loss_ctc": 0.39358943700790405, "train/loss_error": 0.32812538743019104, "train/loss_total": 0.3412182033061981 }, { "epoch": 3.0961038961038962, "step": 1192, "train/loss_ctc": 0.6193915605545044, "train/loss_error": 0.4333972930908203, "train/loss_total": 0.4705961346626282 }, { "epoch": 3.0987012987012985, "step": 1193, "train/loss_ctc": 0.3781549334526062, "train/loss_error": 0.3403969705104828, "train/loss_total": 0.3479485809803009 }, { "epoch": 3.101298701298701, "step": 1194, "train/loss_ctc": 0.4638751149177551, "train/loss_error": 0.3703920841217041, "train/loss_total": 0.3890886902809143 }, { "epoch": 3.103896103896104, "step": 1195, "train/loss_ctc": 0.3532310724258423, "train/loss_error": 0.3138834536075592, "train/loss_total": 0.32175299525260925 }, { "epoch": 3.1064935064935066, "step": 1196, "train/loss_ctc": 0.8244426250457764, "train/loss_error": 0.3931836187839508, "train/loss_total": 0.47943544387817383 }, { "epoch": 3.109090909090909, "step": 1197, "train/loss_ctc": 0.47915709018707275, "train/loss_error": 0.4059717357158661, "train/loss_total": 0.4206088185310364 }, { "epoch": 3.1116883116883116, "step": 1198, "train/loss_ctc": 0.46139460802078247, "train/loss_error": 0.37894657254219055, "train/loss_total": 0.39543619751930237 }, { "epoch": 3.1142857142857143, "step": 1199, "train/loss_ctc": 0.36999639868736267, "train/loss_error": 0.24738460779190063, "train/loss_total": 0.2719069719314575 }, { "epoch": 3.116883116883117, "grad_norm": 10875.6279296875, "learning_rate": 1.1345454545454545e-05, "loss": 0.3771, "step": 1200 }, { "epoch": 3.116883116883117, "step": 1200, "train/loss_ctc": 0.26115837693214417, "train/loss_error": 0.30233174562454224, "train/loss_total": 0.29409706592559814 }, { "epoch": 3.1194805194805193, "step": 1201, "train/loss_ctc": 0.8637335896492004, "train/loss_error": 0.36132800579071045, "train/loss_total": 0.4618091285228729 }, { "epoch": 3.122077922077922, "step": 1202, "train/loss_ctc": 0.5260671377182007, "train/loss_error": 0.3910788595676422, "train/loss_total": 0.4180765151977539 }, { "epoch": 3.1246753246753247, "step": 1203, "train/loss_ctc": 0.35172799229621887, "train/loss_error": 0.41346094012260437, "train/loss_total": 0.4011143743991852 }, { "epoch": 3.1272727272727274, "step": 1204, "train/loss_ctc": 0.42212486267089844, "train/loss_error": 0.30513671040534973, "train/loss_total": 0.3285343647003174 }, { "epoch": 3.1298701298701297, "step": 1205, "train/loss_ctc": 0.2565247416496277, "train/loss_error": 0.2728341817855835, "train/loss_total": 0.26957228779792786 }, { "epoch": 3.1324675324675324, "step": 1206, "train/loss_ctc": 0.34294572472572327, "train/loss_error": 0.3435770869255066, "train/loss_total": 0.3434508144855499 }, { "epoch": 3.135064935064935, "step": 1207, "train/loss_ctc": 0.819007158279419, "train/loss_error": 0.3763025104999542, "train/loss_total": 0.4648434519767761 }, { "epoch": 3.137662337662338, "step": 1208, "train/loss_ctc": 0.39219027757644653, "train/loss_error": 0.32012301683425903, "train/loss_total": 0.33453649282455444 }, { "epoch": 3.14025974025974, "step": 1209, "train/loss_ctc": 0.47766607999801636, "train/loss_error": 0.33785948157310486, "train/loss_total": 0.36582082509994507 }, { "epoch": 3.142857142857143, "grad_norm": 11634.66796875, "learning_rate": 1.118961038961039e-05, "loss": 0.3682, "step": 1210 }, { "epoch": 3.142857142857143, "step": 1210, "train/loss_ctc": 0.3962758779525757, "train/loss_error": 0.3343026041984558, "train/loss_total": 0.34669727087020874 }, { "epoch": 3.1454545454545455, "step": 1211, "train/loss_ctc": 0.6759684085845947, "train/loss_error": 0.3495502769947052, "train/loss_total": 0.4148339033126831 }, { "epoch": 3.148051948051948, "step": 1212, "train/loss_ctc": 0.4594862461090088, "train/loss_error": 0.37689927220344543, "train/loss_total": 0.3934166729450226 }, { "epoch": 3.1506493506493505, "step": 1213, "train/loss_ctc": 0.4471725821495056, "train/loss_error": 0.3817751109600067, "train/loss_total": 0.3948546051979065 }, { "epoch": 3.153246753246753, "step": 1214, "train/loss_ctc": 0.37306690216064453, "train/loss_error": 0.41578489542007446, "train/loss_total": 0.4072413146495819 }, { "epoch": 3.155844155844156, "step": 1215, "train/loss_ctc": 0.5379006862640381, "train/loss_error": 0.27313926815986633, "train/loss_total": 0.32609155774116516 }, { "epoch": 3.1584415584415586, "step": 1216, "train/loss_ctc": 0.46564748883247375, "train/loss_error": 0.362135112285614, "train/loss_total": 0.38283759355545044 }, { "epoch": 3.161038961038961, "step": 1217, "train/loss_ctc": 0.6394009590148926, "train/loss_error": 0.4125988185405731, "train/loss_total": 0.45795923471450806 }, { "epoch": 3.1636363636363636, "step": 1218, "train/loss_ctc": 0.42375579476356506, "train/loss_error": 0.2826063632965088, "train/loss_total": 0.3108362555503845 }, { "epoch": 3.1662337662337663, "step": 1219, "train/loss_ctc": 0.40625840425491333, "train/loss_error": 0.3353709578514099, "train/loss_total": 0.34954845905303955 }, { "epoch": 3.168831168831169, "grad_norm": 12465.869140625, "learning_rate": 1.1033766233766234e-05, "loss": 0.3784, "step": 1220 }, { "epoch": 3.168831168831169, "step": 1220, "train/loss_ctc": 0.596881091594696, "train/loss_error": 0.307923823595047, "train/loss_total": 0.36571529507637024 }, { "epoch": 3.1714285714285713, "step": 1221, "train/loss_ctc": 0.3655451834201813, "train/loss_error": 0.3281305730342865, "train/loss_total": 0.335613489151001 }, { "epoch": 3.174025974025974, "step": 1222, "train/loss_ctc": 0.7114700675010681, "train/loss_error": 0.3680289089679718, "train/loss_total": 0.43671715259552 }, { "epoch": 3.1766233766233767, "step": 1223, "train/loss_ctc": 0.43775802850723267, "train/loss_error": 0.36380308866500854, "train/loss_total": 0.3785941004753113 }, { "epoch": 3.1792207792207794, "step": 1224, "train/loss_ctc": 0.344638466835022, "train/loss_error": 0.3300721347332001, "train/loss_total": 0.33298543095588684 }, { "epoch": 3.1818181818181817, "step": 1225, "train/loss_ctc": 0.4378778338432312, "train/loss_error": 0.3595426380634308, "train/loss_total": 0.3752096891403198 }, { "epoch": 3.1844155844155844, "step": 1226, "train/loss_ctc": 0.4560542702674866, "train/loss_error": 0.28412777185440063, "train/loss_total": 0.31851309537887573 }, { "epoch": 3.187012987012987, "step": 1227, "train/loss_ctc": 0.35361725091934204, "train/loss_error": 0.3716007173061371, "train/loss_total": 0.3680040240287781 }, { "epoch": 3.18961038961039, "step": 1228, "train/loss_ctc": 0.3804856836795807, "train/loss_error": 0.3047177493572235, "train/loss_total": 0.31987133622169495 }, { "epoch": 3.192207792207792, "step": 1229, "train/loss_ctc": 0.4000454545021057, "train/loss_error": 0.41123706102371216, "train/loss_total": 0.4089987576007843 }, { "epoch": 3.1948051948051948, "grad_norm": 11588.3173828125, "learning_rate": 1.0877922077922078e-05, "loss": 0.364, "step": 1230 }, { "epoch": 3.1948051948051948, "step": 1230, "train/loss_ctc": 0.4925099313259125, "train/loss_error": 0.3377663195133209, "train/loss_total": 0.3687150478363037 }, { "epoch": 3.1974025974025975, "step": 1231, "train/loss_ctc": 0.5415129065513611, "train/loss_error": 0.3762654662132263, "train/loss_total": 0.40931496024131775 }, { "epoch": 3.2, "step": 1232, "train/loss_ctc": 0.4445893168449402, "train/loss_error": 0.3152197003364563, "train/loss_total": 0.34109365940093994 }, { "epoch": 3.2025974025974024, "step": 1233, "train/loss_ctc": 0.45145875215530396, "train/loss_error": 0.29433882236480713, "train/loss_total": 0.3257628083229065 }, { "epoch": 3.205194805194805, "step": 1234, "train/loss_ctc": 0.4873988926410675, "train/loss_error": 0.27386629581451416, "train/loss_total": 0.31657281517982483 }, { "epoch": 3.207792207792208, "step": 1235, "train/loss_ctc": 0.6991333365440369, "train/loss_error": 0.41676080226898193, "train/loss_total": 0.4732353091239929 }, { "epoch": 3.2103896103896106, "step": 1236, "train/loss_ctc": 0.3062251806259155, "train/loss_error": 0.34749650955200195, "train/loss_total": 0.33924224972724915 }, { "epoch": 3.212987012987013, "step": 1237, "train/loss_ctc": 0.34912222623825073, "train/loss_error": 0.35477522015571594, "train/loss_total": 0.35364463925361633 }, { "epoch": 3.2155844155844155, "step": 1238, "train/loss_ctc": 0.3602435886859894, "train/loss_error": 0.35267046093940735, "train/loss_total": 0.3541851043701172 }, { "epoch": 3.2181818181818183, "step": 1239, "train/loss_ctc": 0.42662322521209717, "train/loss_error": 0.33253830671310425, "train/loss_total": 0.35135528445243835 }, { "epoch": 3.220779220779221, "grad_norm": 14231.515625, "learning_rate": 1.0722077922077922e-05, "loss": 0.3633, "step": 1240 }, { "epoch": 3.220779220779221, "step": 1240, "train/loss_ctc": 0.44394880533218384, "train/loss_error": 0.3080214858055115, "train/loss_total": 0.3352069556713104 }, { "epoch": 3.2233766233766232, "step": 1241, "train/loss_ctc": 0.4386140704154968, "train/loss_error": 0.33388882875442505, "train/loss_total": 0.3548338711261749 }, { "epoch": 3.225974025974026, "step": 1242, "train/loss_ctc": 0.3589886724948883, "train/loss_error": 0.3422466814517975, "train/loss_total": 0.3455950915813446 }, { "epoch": 3.2285714285714286, "step": 1243, "train/loss_ctc": 0.34103068709373474, "train/loss_error": 0.3849130868911743, "train/loss_total": 0.3761366009712219 }, { "epoch": 3.2311688311688314, "step": 1244, "train/loss_ctc": 0.4570619463920593, "train/loss_error": 0.3402842879295349, "train/loss_total": 0.36363983154296875 }, { "epoch": 3.2337662337662336, "step": 1245, "train/loss_ctc": 0.7452801465988159, "train/loss_error": 0.3534730076789856, "train/loss_total": 0.43183445930480957 }, { "epoch": 3.2363636363636363, "step": 1246, "train/loss_ctc": 0.5501331090927124, "train/loss_error": 0.3781147599220276, "train/loss_total": 0.4125184416770935 }, { "epoch": 3.238961038961039, "step": 1247, "train/loss_ctc": 0.5509181022644043, "train/loss_error": 0.4232659339904785, "train/loss_total": 0.4487963914871216 }, { "epoch": 3.2415584415584417, "step": 1248, "train/loss_ctc": 0.35209691524505615, "train/loss_error": 0.3306615948677063, "train/loss_total": 0.33494865894317627 }, { "epoch": 3.244155844155844, "step": 1249, "train/loss_ctc": 0.41594547033309937, "train/loss_error": 0.39510658383369446, "train/loss_total": 0.39927437901496887 }, { "epoch": 3.2467532467532467, "grad_norm": 13675.443359375, "learning_rate": 1.0566233766233765e-05, "loss": 0.3803, "step": 1250 }, { "epoch": 3.2467532467532467, "step": 1250, "train/loss_ctc": 0.32970893383026123, "train/loss_error": 0.39341336488723755, "train/loss_total": 0.38067248463630676 }, { "epoch": 3.2493506493506494, "step": 1251, "train/loss_ctc": 0.532305896282196, "train/loss_error": 0.30491676926612854, "train/loss_total": 0.350394606590271 }, { "epoch": 3.2519480519480517, "step": 1252, "train/loss_ctc": 0.31443917751312256, "train/loss_error": 0.2789790630340576, "train/loss_total": 0.2860710918903351 }, { "epoch": 3.2545454545454544, "step": 1253, "train/loss_ctc": 0.3605686128139496, "train/loss_error": 0.3280058801174164, "train/loss_total": 0.3345184326171875 }, { "epoch": 3.257142857142857, "step": 1254, "train/loss_ctc": 0.3572877049446106, "train/loss_error": 0.24150145053863525, "train/loss_total": 0.26465871930122375 }, { "epoch": 3.25974025974026, "step": 1255, "train/loss_ctc": 0.4140282869338989, "train/loss_error": 0.3124629855155945, "train/loss_total": 0.3327760398387909 }, { "epoch": 3.2623376623376625, "step": 1256, "train/loss_ctc": 0.3728312849998474, "train/loss_error": 0.40455129742622375, "train/loss_total": 0.39820730686187744 }, { "epoch": 3.264935064935065, "step": 1257, "train/loss_ctc": 0.704093337059021, "train/loss_error": 0.4177088737487793, "train/loss_total": 0.4749857783317566 }, { "epoch": 3.2675324675324675, "step": 1258, "train/loss_ctc": 0.5516400337219238, "train/loss_error": 0.33752450346946716, "train/loss_total": 0.3803476393222809 }, { "epoch": 3.27012987012987, "step": 1259, "train/loss_ctc": 0.5973231792449951, "train/loss_error": 0.3536814749240875, "train/loss_total": 0.4024098217487335 }, { "epoch": 3.2727272727272725, "grad_norm": 15630.0390625, "learning_rate": 1.041038961038961e-05, "loss": 0.3605, "step": 1260 }, { "epoch": 3.2727272727272725, "step": 1260, "train/loss_ctc": 0.6549766063690186, "train/loss_error": 0.3872382342815399, "train/loss_total": 0.44078588485717773 }, { "epoch": 3.275324675324675, "step": 1261, "train/loss_ctc": 0.4375436305999756, "train/loss_error": 0.34060049057006836, "train/loss_total": 0.35998913645744324 }, { "epoch": 3.277922077922078, "step": 1262, "train/loss_ctc": 0.37741518020629883, "train/loss_error": 0.3219893276691437, "train/loss_total": 0.33307451009750366 }, { "epoch": 3.2805194805194806, "step": 1263, "train/loss_ctc": 0.27448219060897827, "train/loss_error": 0.2879736125469208, "train/loss_total": 0.28527534008026123 }, { "epoch": 3.2831168831168833, "step": 1264, "train/loss_ctc": 0.4724734127521515, "train/loss_error": 0.4613330364227295, "train/loss_total": 0.46356111764907837 }, { "epoch": 3.2857142857142856, "step": 1265, "train/loss_ctc": 0.3388649523258209, "train/loss_error": 0.24602021276950836, "train/loss_total": 0.2645891606807709 }, { "epoch": 3.2883116883116883, "step": 1266, "train/loss_ctc": 0.45191043615341187, "train/loss_error": 0.2592332363128662, "train/loss_total": 0.2977686822414398 }, { "epoch": 3.290909090909091, "step": 1267, "train/loss_ctc": 0.7397627234458923, "train/loss_error": 0.4446910321712494, "train/loss_total": 0.5037053823471069 }, { "epoch": 3.2935064935064933, "step": 1268, "train/loss_ctc": 0.5008466243743896, "train/loss_error": 0.29887786507606506, "train/loss_total": 0.339271605014801 }, { "epoch": 3.296103896103896, "step": 1269, "train/loss_ctc": 0.40484344959259033, "train/loss_error": 0.3397311270236969, "train/loss_total": 0.35275357961654663 }, { "epoch": 3.2987012987012987, "grad_norm": 13049.8974609375, "learning_rate": 1.0254545454545453e-05, "loss": 0.3641, "step": 1270 }, { "epoch": 3.2987012987012987, "step": 1270, "train/loss_ctc": 0.38112393021583557, "train/loss_error": 0.2287888377904892, "train/loss_total": 0.25925585627555847 }, { "epoch": 3.3012987012987014, "step": 1271, "train/loss_ctc": 0.41412192583084106, "train/loss_error": 0.3932550847530365, "train/loss_total": 0.3974284529685974 }, { "epoch": 3.303896103896104, "step": 1272, "train/loss_ctc": 0.31921014189720154, "train/loss_error": 0.3739437460899353, "train/loss_total": 0.36299702525138855 }, { "epoch": 3.3064935064935064, "step": 1273, "train/loss_ctc": 0.4534679651260376, "train/loss_error": 0.329856812953949, "train/loss_total": 0.35457906126976013 }, { "epoch": 3.309090909090909, "step": 1274, "train/loss_ctc": 0.3314293324947357, "train/loss_error": 0.2961139380931854, "train/loss_total": 0.30317702889442444 }, { "epoch": 3.311688311688312, "step": 1275, "train/loss_ctc": 0.32028210163116455, "train/loss_error": 0.26281067728996277, "train/loss_total": 0.27430495619773865 }, { "epoch": 3.314285714285714, "step": 1276, "train/loss_ctc": 0.5126175880432129, "train/loss_error": 0.346371054649353, "train/loss_total": 0.37962037324905396 }, { "epoch": 3.3168831168831168, "step": 1277, "train/loss_ctc": 0.3876386284828186, "train/loss_error": 0.3065507113933563, "train/loss_total": 0.32276830077171326 }, { "epoch": 3.3194805194805195, "step": 1278, "train/loss_ctc": 0.4646502435207367, "train/loss_error": 0.2699057161808014, "train/loss_total": 0.3088546395301819 }, { "epoch": 3.322077922077922, "step": 1279, "train/loss_ctc": 0.20248889923095703, "train/loss_error": 0.24932289123535156, "train/loss_total": 0.2399560958147049 }, { "epoch": 3.324675324675325, "grad_norm": 8396.03125, "learning_rate": 1.0098701298701298e-05, "loss": 0.3203, "step": 1280 }, { "epoch": 3.324675324675325, "step": 1280, "train/loss_ctc": 0.4562304615974426, "train/loss_error": 0.31158313155174255, "train/loss_total": 0.34051260352134705 }, { "epoch": 3.327272727272727, "step": 1281, "train/loss_ctc": 0.3842810392379761, "train/loss_error": 0.36834606528282166, "train/loss_total": 0.3715330958366394 }, { "epoch": 3.32987012987013, "step": 1282, "train/loss_ctc": 0.3624483644962311, "train/loss_error": 0.2988964915275574, "train/loss_total": 0.31160685420036316 }, { "epoch": 3.3324675324675326, "step": 1283, "train/loss_ctc": 0.4819134473800659, "train/loss_error": 0.3838018774986267, "train/loss_total": 0.4034242033958435 }, { "epoch": 3.335064935064935, "step": 1284, "train/loss_ctc": 0.29670751094818115, "train/loss_error": 0.28833359479904175, "train/loss_total": 0.2900083661079407 }, { "epoch": 3.3376623376623376, "step": 1285, "train/loss_ctc": 0.3689732849597931, "train/loss_error": 0.32138389348983765, "train/loss_total": 0.33090177178382874 }, { "epoch": 3.3402597402597403, "step": 1286, "train/loss_ctc": 0.4374220669269562, "train/loss_error": 0.322104275226593, "train/loss_total": 0.3451678454875946 }, { "epoch": 3.342857142857143, "step": 1287, "train/loss_ctc": 0.5982828140258789, "train/loss_error": 0.2840671241283417, "train/loss_total": 0.3469102680683136 }, { "epoch": 3.3454545454545457, "step": 1288, "train/loss_ctc": 0.3665076792240143, "train/loss_error": 0.4057043790817261, "train/loss_total": 0.39786505699157715 }, { "epoch": 3.348051948051948, "step": 1289, "train/loss_ctc": 0.28013715147972107, "train/loss_error": 0.34180107712745667, "train/loss_total": 0.329468309879303 }, { "epoch": 3.3506493506493507, "grad_norm": 10303.1708984375, "learning_rate": 9.942857142857143e-06, "loss": 0.3467, "step": 1290 }, { "epoch": 3.3506493506493507, "step": 1290, "train/loss_ctc": 0.4082898199558258, "train/loss_error": 0.32392680644989014, "train/loss_total": 0.3407994210720062 }, { "epoch": 3.3532467532467534, "step": 1291, "train/loss_ctc": 0.6599149703979492, "train/loss_error": 0.39717617630958557, "train/loss_total": 0.4497239589691162 }, { "epoch": 3.3558441558441556, "step": 1292, "train/loss_ctc": 0.48946940898895264, "train/loss_error": 0.2592952847480774, "train/loss_total": 0.3053301274776459 }, { "epoch": 3.3584415584415583, "step": 1293, "train/loss_ctc": 0.38963326811790466, "train/loss_error": 0.36937400698661804, "train/loss_total": 0.3734258711338043 }, { "epoch": 3.361038961038961, "step": 1294, "train/loss_ctc": 0.22540068626403809, "train/loss_error": 0.21992890536785126, "train/loss_total": 0.22102326154708862 }, { "epoch": 3.3636363636363638, "step": 1295, "train/loss_ctc": 0.438028484582901, "train/loss_error": 0.31849604845046997, "train/loss_total": 0.34240251779556274 }, { "epoch": 3.3662337662337665, "step": 1296, "train/loss_ctc": 0.42364537715911865, "train/loss_error": 0.35590413212776184, "train/loss_total": 0.3694523870944977 }, { "epoch": 3.3688311688311687, "step": 1297, "train/loss_ctc": 0.6189830303192139, "train/loss_error": 0.3637867867946625, "train/loss_total": 0.41482603549957275 }, { "epoch": 3.3714285714285714, "step": 1298, "train/loss_ctc": 0.5334379076957703, "train/loss_error": 0.395584374666214, "train/loss_total": 0.4231550693511963 }, { "epoch": 3.374025974025974, "step": 1299, "train/loss_ctc": 0.4304255247116089, "train/loss_error": 0.3250105679035187, "train/loss_total": 0.3460935652256012 }, { "epoch": 3.3766233766233764, "grad_norm": 11150.6044921875, "learning_rate": 9.787012987012986e-06, "loss": 0.3586, "step": 1300 }, { "epoch": 3.3766233766233764, "step": 1300, "train/loss_ctc": 0.514143168926239, "train/loss_error": 0.3422201871871948, "train/loss_total": 0.3766047954559326 }, { "epoch": 3.379220779220779, "step": 1301, "train/loss_ctc": 0.5265023112297058, "train/loss_error": 0.32408323884010315, "train/loss_total": 0.3645670413970947 }, { "epoch": 3.381818181818182, "step": 1302, "train/loss_ctc": 0.410329669713974, "train/loss_error": 0.2329607903957367, "train/loss_total": 0.2684345543384552 }, { "epoch": 3.3844155844155845, "step": 1303, "train/loss_ctc": 0.661552369594574, "train/loss_error": 0.45211556553840637, "train/loss_total": 0.49400293827056885 }, { "epoch": 3.3870129870129873, "step": 1304, "train/loss_ctc": 0.4779919981956482, "train/loss_error": 0.3367976248264313, "train/loss_total": 0.3650365173816681 }, { "epoch": 3.3896103896103895, "step": 1305, "train/loss_ctc": 0.41216254234313965, "train/loss_error": 0.2962033450603485, "train/loss_total": 0.31939518451690674 }, { "epoch": 3.3922077922077922, "step": 1306, "train/loss_ctc": 0.6470104455947876, "train/loss_error": 0.3285446763038635, "train/loss_total": 0.3922378420829773 }, { "epoch": 3.394805194805195, "step": 1307, "train/loss_ctc": 0.4599168300628662, "train/loss_error": 0.3926595449447632, "train/loss_total": 0.4061110317707062 }, { "epoch": 3.397402597402597, "step": 1308, "train/loss_ctc": 0.581486701965332, "train/loss_error": 0.4503476023674011, "train/loss_total": 0.47657543420791626 }, { "epoch": 3.4, "step": 1309, "train/loss_ctc": 0.49871763586997986, "train/loss_error": 0.3041122853755951, "train/loss_total": 0.3430333733558655 }, { "epoch": 3.4025974025974026, "grad_norm": 11353.9384765625, "learning_rate": 9.631168831168831e-06, "loss": 0.3806, "step": 1310 }, { "epoch": 3.4025974025974026, "step": 1310, "train/loss_ctc": 0.3550449013710022, "train/loss_error": 0.32505983114242554, "train/loss_total": 0.3310568630695343 }, { "epoch": 3.4051948051948053, "step": 1311, "train/loss_ctc": 0.4309844970703125, "train/loss_error": 0.3979933261871338, "train/loss_total": 0.40459156036376953 }, { "epoch": 3.407792207792208, "step": 1312, "train/loss_ctc": 0.6339913606643677, "train/loss_error": 0.3458809554576874, "train/loss_total": 0.40350303053855896 }, { "epoch": 3.4103896103896103, "step": 1313, "train/loss_ctc": 0.5988548398017883, "train/loss_error": 0.32759931683540344, "train/loss_total": 0.3818504214286804 }, { "epoch": 3.412987012987013, "step": 1314, "train/loss_ctc": 0.4511449933052063, "train/loss_error": 0.3646090030670166, "train/loss_total": 0.38191622495651245 }, { "epoch": 3.4155844155844157, "step": 1315, "train/loss_ctc": 0.39888596534729004, "train/loss_error": 0.30327892303466797, "train/loss_total": 0.3224003314971924 }, { "epoch": 3.418181818181818, "step": 1316, "train/loss_ctc": 0.4410955011844635, "train/loss_error": 0.2859165072441101, "train/loss_total": 0.31695231795310974 }, { "epoch": 3.4207792207792207, "step": 1317, "train/loss_ctc": 0.3824216425418854, "train/loss_error": 0.38219887018203735, "train/loss_total": 0.38224342465400696 }, { "epoch": 3.4233766233766234, "step": 1318, "train/loss_ctc": 0.3544047474861145, "train/loss_error": 0.32633525133132935, "train/loss_total": 0.3319491446018219 }, { "epoch": 3.425974025974026, "step": 1319, "train/loss_ctc": 0.5538939237594604, "train/loss_error": 0.3576466143131256, "train/loss_total": 0.3968960642814636 }, { "epoch": 3.4285714285714284, "grad_norm": 14867.00390625, "learning_rate": 9.475324675324676e-06, "loss": 0.3653, "step": 1320 }, { "epoch": 3.4285714285714284, "step": 1320, "train/loss_ctc": 0.2898489832878113, "train/loss_error": 0.35098543763160706, "train/loss_total": 0.3387581706047058 }, { "epoch": 3.431168831168831, "step": 1321, "train/loss_ctc": 0.45564132928848267, "train/loss_error": 0.29948413372039795, "train/loss_total": 0.3307155668735504 }, { "epoch": 3.433766233766234, "step": 1322, "train/loss_ctc": 0.517824649810791, "train/loss_error": 0.4024183452129364, "train/loss_total": 0.4254996180534363 }, { "epoch": 3.4363636363636365, "step": 1323, "train/loss_ctc": 0.2872617542743683, "train/loss_error": 0.37620750069618225, "train/loss_total": 0.358418345451355 }, { "epoch": 3.4389610389610388, "step": 1324, "train/loss_ctc": 0.3040758967399597, "train/loss_error": 0.3229592740535736, "train/loss_total": 0.3191826045513153 }, { "epoch": 3.4415584415584415, "step": 1325, "train/loss_ctc": 0.2971556782722473, "train/loss_error": 0.32819923758506775, "train/loss_total": 0.3219905197620392 }, { "epoch": 3.444155844155844, "step": 1326, "train/loss_ctc": 0.29557132720947266, "train/loss_error": 0.31760263442993164, "train/loss_total": 0.3131963908672333 }, { "epoch": 3.446753246753247, "step": 1327, "train/loss_ctc": 0.5190359354019165, "train/loss_error": 0.3389643132686615, "train/loss_total": 0.374978631734848 }, { "epoch": 3.449350649350649, "step": 1328, "train/loss_ctc": 0.32662397623062134, "train/loss_error": 0.26614007353782654, "train/loss_total": 0.27823686599731445 }, { "epoch": 3.451948051948052, "step": 1329, "train/loss_ctc": 0.5457139015197754, "train/loss_error": 0.44037795066833496, "train/loss_total": 0.461445152759552 }, { "epoch": 3.4545454545454546, "grad_norm": 14471.279296875, "learning_rate": 9.31948051948052e-06, "loss": 0.3522, "step": 1330 }, { "epoch": 3.4545454545454546, "step": 1330, "train/loss_ctc": 0.5039575099945068, "train/loss_error": 0.28224730491638184, "train/loss_total": 0.32658934593200684 }, { "epoch": 3.4571428571428573, "step": 1331, "train/loss_ctc": 0.5165609121322632, "train/loss_error": 0.43593695759773254, "train/loss_total": 0.4520617723464966 }, { "epoch": 3.4597402597402596, "step": 1332, "train/loss_ctc": 0.41131827235221863, "train/loss_error": 0.32747843861579895, "train/loss_total": 0.34424641728401184 }, { "epoch": 3.4623376623376623, "step": 1333, "train/loss_ctc": 0.2938710153102875, "train/loss_error": 0.31242337822914124, "train/loss_total": 0.308712899684906 }, { "epoch": 3.464935064935065, "step": 1334, "train/loss_ctc": 0.4634096622467041, "train/loss_error": 0.2989174723625183, "train/loss_total": 0.3318159282207489 }, { "epoch": 3.4675324675324677, "step": 1335, "train/loss_ctc": 0.4479408860206604, "train/loss_error": 0.3628423810005188, "train/loss_total": 0.37986207008361816 }, { "epoch": 3.47012987012987, "step": 1336, "train/loss_ctc": 0.5343083739280701, "train/loss_error": 0.42029985785484314, "train/loss_total": 0.44310158491134644 }, { "epoch": 3.4727272727272727, "step": 1337, "train/loss_ctc": 0.465764582157135, "train/loss_error": 0.32402220368385315, "train/loss_total": 0.3523706793785095 }, { "epoch": 3.4753246753246754, "step": 1338, "train/loss_ctc": 0.6229563355445862, "train/loss_error": 0.44070520997047424, "train/loss_total": 0.4771554470062256 }, { "epoch": 3.477922077922078, "step": 1339, "train/loss_ctc": 0.3492184579372406, "train/loss_error": 0.3302060067653656, "train/loss_total": 0.33400851488113403 }, { "epoch": 3.4805194805194803, "grad_norm": 14626.9912109375, "learning_rate": 9.163636363636365e-06, "loss": 0.375, "step": 1340 }, { "epoch": 3.4805194805194803, "step": 1340, "train/loss_ctc": 0.5979167222976685, "train/loss_error": 0.4619074761867523, "train/loss_total": 0.4891093373298645 }, { "epoch": 3.483116883116883, "step": 1341, "train/loss_ctc": 0.3404043912887573, "train/loss_error": 0.30995306372642517, "train/loss_total": 0.31604334712028503 }, { "epoch": 3.4857142857142858, "step": 1342, "train/loss_ctc": 0.7340444922447205, "train/loss_error": 0.3535098135471344, "train/loss_total": 0.4296167492866516 }, { "epoch": 3.4883116883116885, "step": 1343, "train/loss_ctc": 0.3144181966781616, "train/loss_error": 0.29773467779159546, "train/loss_total": 0.3010713756084442 }, { "epoch": 3.4909090909090907, "step": 1344, "train/loss_ctc": 0.45081961154937744, "train/loss_error": 0.38194820284843445, "train/loss_total": 0.39572247862815857 }, { "epoch": 3.4935064935064934, "step": 1345, "train/loss_ctc": 0.2958543300628662, "train/loss_error": 0.23557013273239136, "train/loss_total": 0.24762697517871857 }, { "epoch": 3.496103896103896, "step": 1346, "train/loss_ctc": 0.5599008798599243, "train/loss_error": 0.32388895750045776, "train/loss_total": 0.3710913360118866 }, { "epoch": 3.498701298701299, "step": 1347, "train/loss_ctc": 0.5601643919944763, "train/loss_error": 0.33107423782348633, "train/loss_total": 0.3768922984600067 }, { "epoch": 3.501298701298701, "step": 1348, "train/loss_ctc": 0.41778331995010376, "train/loss_error": 0.30677613615989685, "train/loss_total": 0.3289775848388672 }, { "epoch": 3.503896103896104, "step": 1349, "train/loss_ctc": 0.4550985097885132, "train/loss_error": 0.3300922214984894, "train/loss_total": 0.35509347915649414 }, { "epoch": 3.5064935064935066, "grad_norm": 10517.96875, "learning_rate": 9.007792207792208e-06, "loss": 0.3611, "step": 1350 }, { "epoch": 3.5064935064935066, "step": 1350, "train/loss_ctc": 0.5088934898376465, "train/loss_error": 0.42451196908950806, "train/loss_total": 0.4413883090019226 }, { "epoch": 3.509090909090909, "step": 1351, "train/loss_ctc": 0.33816444873809814, "train/loss_error": 0.33869051933288574, "train/loss_total": 0.3385853171348572 }, { "epoch": 3.5116883116883115, "step": 1352, "train/loss_ctc": 0.39386576414108276, "train/loss_error": 0.36407899856567383, "train/loss_total": 0.37003636360168457 }, { "epoch": 3.5142857142857142, "step": 1353, "train/loss_ctc": 0.5527002811431885, "train/loss_error": 0.4300989806652069, "train/loss_total": 0.45461925864219666 }, { "epoch": 3.516883116883117, "step": 1354, "train/loss_ctc": 0.3074311316013336, "train/loss_error": 0.2723015546798706, "train/loss_total": 0.27932748198509216 }, { "epoch": 3.5194805194805197, "step": 1355, "train/loss_ctc": 0.7684006690979004, "train/loss_error": 0.35931333899497986, "train/loss_total": 0.4411308169364929 }, { "epoch": 3.522077922077922, "step": 1356, "train/loss_ctc": 0.4545864462852478, "train/loss_error": 0.32880493998527527, "train/loss_total": 0.3539612591266632 }, { "epoch": 3.5246753246753246, "step": 1357, "train/loss_ctc": 0.36480867862701416, "train/loss_error": 0.3341652452945709, "train/loss_total": 0.3402939438819885 }, { "epoch": 3.5272727272727273, "step": 1358, "train/loss_ctc": 0.3869844973087311, "train/loss_error": 0.2929374873638153, "train/loss_total": 0.31174689531326294 }, { "epoch": 3.5298701298701296, "step": 1359, "train/loss_ctc": 0.5242797136306763, "train/loss_error": 0.3470913767814636, "train/loss_total": 0.38252905011177063 }, { "epoch": 3.5324675324675323, "grad_norm": 15440.7529296875, "learning_rate": 8.851948051948053e-06, "loss": 0.3714, "step": 1360 }, { "epoch": 3.5324675324675323, "step": 1360, "train/loss_ctc": 0.5410693883895874, "train/loss_error": 0.3928721249103546, "train/loss_total": 0.42251157760620117 }, { "epoch": 3.535064935064935, "step": 1361, "train/loss_ctc": 0.3548257350921631, "train/loss_error": 0.26886090636253357, "train/loss_total": 0.286053866147995 }, { "epoch": 3.5376623376623377, "step": 1362, "train/loss_ctc": 0.4085543155670166, "train/loss_error": 0.4051879644393921, "train/loss_total": 0.4058612585067749 }, { "epoch": 3.5402597402597404, "step": 1363, "train/loss_ctc": 0.3722110390663147, "train/loss_error": 0.3286648988723755, "train/loss_total": 0.33737412095069885 }, { "epoch": 3.5428571428571427, "step": 1364, "train/loss_ctc": 0.5361945629119873, "train/loss_error": 0.39139747619628906, "train/loss_total": 0.4203568994998932 }, { "epoch": 3.5454545454545454, "step": 1365, "train/loss_ctc": 0.5001022219657898, "train/loss_error": 0.39582720398902893, "train/loss_total": 0.4166822135448456 }, { "epoch": 3.548051948051948, "step": 1366, "train/loss_ctc": 0.5441643595695496, "train/loss_error": 0.3653770685195923, "train/loss_total": 0.40113452076911926 }, { "epoch": 3.5506493506493504, "step": 1367, "train/loss_ctc": 0.3996058702468872, "train/loss_error": 0.30869919061660767, "train/loss_total": 0.326880544424057 }, { "epoch": 3.553246753246753, "step": 1368, "train/loss_ctc": 0.49148136377334595, "train/loss_error": 0.4213947653770447, "train/loss_total": 0.43541210889816284 }, { "epoch": 3.555844155844156, "step": 1369, "train/loss_ctc": 0.5436382293701172, "train/loss_error": 0.4091818928718567, "train/loss_total": 0.4360731840133667 }, { "epoch": 3.5584415584415585, "grad_norm": 12365.3984375, "learning_rate": 8.696103896103896e-06, "loss": 0.3888, "step": 1370 }, { "epoch": 3.5584415584415585, "step": 1370, "train/loss_ctc": 0.773661732673645, "train/loss_error": 0.3946281373500824, "train/loss_total": 0.47043487429618835 }, { "epoch": 3.5610389610389612, "step": 1371, "train/loss_ctc": 0.3554152250289917, "train/loss_error": 0.25802552700042725, "train/loss_total": 0.27750346064567566 }, { "epoch": 3.5636363636363635, "step": 1372, "train/loss_ctc": 0.3332677483558655, "train/loss_error": 0.250071257352829, "train/loss_total": 0.2667105793952942 }, { "epoch": 3.566233766233766, "step": 1373, "train/loss_ctc": 0.2845669686794281, "train/loss_error": 0.3322899639606476, "train/loss_total": 0.3227453827857971 }, { "epoch": 3.568831168831169, "step": 1374, "train/loss_ctc": 0.46997907757759094, "train/loss_error": 0.33922839164733887, "train/loss_total": 0.3653785288333893 }, { "epoch": 3.571428571428571, "step": 1375, "train/loss_ctc": 0.38540375232696533, "train/loss_error": 0.3535282015800476, "train/loss_total": 0.35990333557128906 }, { "epoch": 3.574025974025974, "step": 1376, "train/loss_ctc": 0.315027117729187, "train/loss_error": 0.2675262987613678, "train/loss_total": 0.2770264744758606 }, { "epoch": 3.5766233766233766, "step": 1377, "train/loss_ctc": 0.3559117913246155, "train/loss_error": 0.2733721435070038, "train/loss_total": 0.28988006711006165 }, { "epoch": 3.5792207792207793, "step": 1378, "train/loss_ctc": 0.6491421461105347, "train/loss_error": 0.36240383982658386, "train/loss_total": 0.41975152492523193 }, { "epoch": 3.581818181818182, "step": 1379, "train/loss_ctc": 0.5143790245056152, "train/loss_error": 0.3391295373439789, "train/loss_total": 0.3741794228553772 }, { "epoch": 3.5844155844155843, "grad_norm": 13025.5419921875, "learning_rate": 8.540259740259741e-06, "loss": 0.3424, "step": 1380 }, { "epoch": 3.5844155844155843, "step": 1380, "train/loss_ctc": 0.32897692918777466, "train/loss_error": 0.3223218619823456, "train/loss_total": 0.32365289330482483 }, { "epoch": 3.587012987012987, "step": 1381, "train/loss_ctc": 0.3706561326980591, "train/loss_error": 0.22999383509159088, "train/loss_total": 0.25812628865242004 }, { "epoch": 3.5896103896103897, "step": 1382, "train/loss_ctc": 0.38041964173316956, "train/loss_error": 0.3183835446834564, "train/loss_total": 0.33079075813293457 }, { "epoch": 3.592207792207792, "step": 1383, "train/loss_ctc": 0.4153446853160858, "train/loss_error": 0.3327285051345825, "train/loss_total": 0.34925174713134766 }, { "epoch": 3.5948051948051947, "step": 1384, "train/loss_ctc": 0.3104130029678345, "train/loss_error": 0.3301762640476227, "train/loss_total": 0.32622361183166504 }, { "epoch": 3.5974025974025974, "step": 1385, "train/loss_ctc": 0.38106024265289307, "train/loss_error": 0.30407845973968506, "train/loss_total": 0.31947481632232666 }, { "epoch": 3.6, "step": 1386, "train/loss_ctc": 0.8951590657234192, "train/loss_error": 0.3428584039211273, "train/loss_total": 0.4533185362815857 }, { "epoch": 3.602597402597403, "step": 1387, "train/loss_ctc": 0.29202279448509216, "train/loss_error": 0.2729814350605011, "train/loss_total": 0.27678969502449036 }, { "epoch": 3.605194805194805, "step": 1388, "train/loss_ctc": 0.25107520818710327, "train/loss_error": 0.2567267119884491, "train/loss_total": 0.25559642910957336 }, { "epoch": 3.6077922077922078, "step": 1389, "train/loss_ctc": 0.8334529995918274, "train/loss_error": 0.45303699374198914, "train/loss_total": 0.5291202068328857 }, { "epoch": 3.6103896103896105, "grad_norm": 20132.375, "learning_rate": 8.384415584415584e-06, "loss": 0.3422, "step": 1390 }, { "epoch": 3.6103896103896105, "step": 1390, "train/loss_ctc": 0.4839712679386139, "train/loss_error": 0.39600443840026855, "train/loss_total": 0.41359779238700867 }, { "epoch": 3.6129870129870127, "step": 1391, "train/loss_ctc": 0.7371808290481567, "train/loss_error": 0.3014082908630371, "train/loss_total": 0.38856279850006104 }, { "epoch": 3.6155844155844155, "step": 1392, "train/loss_ctc": 0.36944788694381714, "train/loss_error": 0.27415141463279724, "train/loss_total": 0.2932107150554657 }, { "epoch": 3.618181818181818, "step": 1393, "train/loss_ctc": 0.3477505147457123, "train/loss_error": 0.3201863169670105, "train/loss_total": 0.3256991505622864 }, { "epoch": 3.620779220779221, "step": 1394, "train/loss_ctc": 0.29272741079330444, "train/loss_error": 0.330539345741272, "train/loss_total": 0.3229769468307495 }, { "epoch": 3.6233766233766236, "step": 1395, "train/loss_ctc": 0.6942616701126099, "train/loss_error": 0.3187722861766815, "train/loss_total": 0.39387017488479614 }, { "epoch": 3.625974025974026, "step": 1396, "train/loss_ctc": 0.3256967067718506, "train/loss_error": 0.3070835769176483, "train/loss_total": 0.3108062148094177 }, { "epoch": 3.6285714285714286, "step": 1397, "train/loss_ctc": 0.4358832836151123, "train/loss_error": 0.4004419445991516, "train/loss_total": 0.4075302183628082 }, { "epoch": 3.6311688311688313, "step": 1398, "train/loss_ctc": 0.5073525309562683, "train/loss_error": 0.31691333651542664, "train/loss_total": 0.35500118136405945 }, { "epoch": 3.6337662337662335, "step": 1399, "train/loss_ctc": 0.5554956197738647, "train/loss_error": 0.4984595477581024, "train/loss_total": 0.5098667740821838 }, { "epoch": 3.6363636363636362, "grad_norm": 17591.69140625, "learning_rate": 8.22857142857143e-06, "loss": 0.3721, "step": 1400 }, { "epoch": 3.6363636363636362, "step": 1400, "train/loss_ctc": 0.39583492279052734, "train/loss_error": 0.3959212303161621, "train/loss_total": 0.39590397477149963 }, { "epoch": 3.638961038961039, "step": 1401, "train/loss_ctc": 0.4141707420349121, "train/loss_error": 0.3385099768638611, "train/loss_total": 0.35364213585853577 }, { "epoch": 3.6415584415584417, "step": 1402, "train/loss_ctc": 0.41654425859451294, "train/loss_error": 0.3608722984790802, "train/loss_total": 0.37200668454170227 }, { "epoch": 3.6441558441558444, "step": 1403, "train/loss_ctc": 0.8988666534423828, "train/loss_error": 0.36544597148895264, "train/loss_total": 0.4721301198005676 }, { "epoch": 3.6467532467532466, "step": 1404, "train/loss_ctc": 0.5366384983062744, "train/loss_error": 0.2797139883041382, "train/loss_total": 0.33109891414642334 }, { "epoch": 3.6493506493506493, "step": 1405, "train/loss_ctc": 0.3183499872684479, "train/loss_error": 0.24339935183525085, "train/loss_total": 0.2583894729614258 }, { "epoch": 3.651948051948052, "step": 1406, "train/loss_ctc": 0.5271725654602051, "train/loss_error": 0.3530624210834503, "train/loss_total": 0.3878844380378723 }, { "epoch": 3.6545454545454543, "step": 1407, "train/loss_ctc": 0.5914377570152283, "train/loss_error": 0.3443754315376282, "train/loss_total": 0.3937879204750061 }, { "epoch": 3.657142857142857, "step": 1408, "train/loss_ctc": 0.5290626883506775, "train/loss_error": 0.3751153349876404, "train/loss_total": 0.4059048295021057 }, { "epoch": 3.6597402597402597, "step": 1409, "train/loss_ctc": 0.41411876678466797, "train/loss_error": 0.3394218683242798, "train/loss_total": 0.35436126589775085 }, { "epoch": 3.6623376623376624, "grad_norm": 11692.6591796875, "learning_rate": 8.072727272727274e-06, "loss": 0.3725, "step": 1410 }, { "epoch": 3.6623376623376624, "step": 1410, "train/loss_ctc": 0.3702080249786377, "train/loss_error": 0.30094191431999207, "train/loss_total": 0.3147951364517212 }, { "epoch": 3.664935064935065, "step": 1411, "train/loss_ctc": 0.5009851455688477, "train/loss_error": 0.3473285138607025, "train/loss_total": 0.37805986404418945 }, { "epoch": 3.6675324675324674, "step": 1412, "train/loss_ctc": 0.34243592619895935, "train/loss_error": 0.32530948519706726, "train/loss_total": 0.32873478531837463 }, { "epoch": 3.67012987012987, "step": 1413, "train/loss_ctc": 0.32050198316574097, "train/loss_error": 0.3395712077617645, "train/loss_total": 0.33575737476348877 }, { "epoch": 3.672727272727273, "step": 1414, "train/loss_ctc": 0.37191861867904663, "train/loss_error": 0.2921336889266968, "train/loss_total": 0.3080906867980957 }, { "epoch": 3.675324675324675, "step": 1415, "train/loss_ctc": 0.5129961967468262, "train/loss_error": 0.40548470616340637, "train/loss_total": 0.4269869923591614 }, { "epoch": 3.677922077922078, "step": 1416, "train/loss_ctc": 0.320453941822052, "train/loss_error": 0.3261982798576355, "train/loss_total": 0.32504943013191223 }, { "epoch": 3.6805194805194805, "step": 1417, "train/loss_ctc": 0.4014742374420166, "train/loss_error": 0.3480602204799652, "train/loss_total": 0.3587430417537689 }, { "epoch": 3.6831168831168832, "step": 1418, "train/loss_ctc": 0.5173248648643494, "train/loss_error": 0.42566144466400146, "train/loss_total": 0.4439941644668579 }, { "epoch": 3.685714285714286, "step": 1419, "train/loss_ctc": 0.5521442890167236, "train/loss_error": 0.41572943329811096, "train/loss_total": 0.44301241636276245 }, { "epoch": 3.688311688311688, "grad_norm": 13105.7412109375, "learning_rate": 7.916883116883117e-06, "loss": 0.3663, "step": 1420 }, { "epoch": 3.688311688311688, "step": 1420, "train/loss_ctc": 0.4125576913356781, "train/loss_error": 0.4421195983886719, "train/loss_total": 0.43620723485946655 }, { "epoch": 3.690909090909091, "step": 1421, "train/loss_ctc": 0.49284371733665466, "train/loss_error": 0.33784717321395874, "train/loss_total": 0.36884647607803345 }, { "epoch": 3.6935064935064936, "step": 1422, "train/loss_ctc": 0.4510738253593445, "train/loss_error": 0.3471478521823883, "train/loss_total": 0.3679330348968506 }, { "epoch": 3.696103896103896, "step": 1423, "train/loss_ctc": 0.4019952416419983, "train/loss_error": 0.32737720012664795, "train/loss_total": 0.3423008322715759 }, { "epoch": 3.6987012987012986, "step": 1424, "train/loss_ctc": 0.6520242691040039, "train/loss_error": 0.3170813322067261, "train/loss_total": 0.38406991958618164 }, { "epoch": 3.7012987012987013, "step": 1425, "train/loss_ctc": 0.42040032148361206, "train/loss_error": 0.38720378279685974, "train/loss_total": 0.3938431143760681 }, { "epoch": 3.703896103896104, "step": 1426, "train/loss_ctc": 0.5021884441375732, "train/loss_error": 0.40049222111701965, "train/loss_total": 0.42083147168159485 }, { "epoch": 3.7064935064935067, "step": 1427, "train/loss_ctc": 0.45229387283325195, "train/loss_error": 0.388574481010437, "train/loss_total": 0.40131837129592896 }, { "epoch": 3.709090909090909, "step": 1428, "train/loss_ctc": 0.4091341197490692, "train/loss_error": 0.29750561714172363, "train/loss_total": 0.31983131170272827 }, { "epoch": 3.7116883116883117, "step": 1429, "train/loss_ctc": 0.40034574270248413, "train/loss_error": 0.3381359577178955, "train/loss_total": 0.3505779206752777 }, { "epoch": 3.7142857142857144, "grad_norm": 10501.53515625, "learning_rate": 7.761038961038962e-06, "loss": 0.3786, "step": 1430 }, { "epoch": 3.7142857142857144, "step": 1430, "train/loss_ctc": 0.4266519546508789, "train/loss_error": 0.4349673092365265, "train/loss_total": 0.4333042502403259 }, { "epoch": 3.7168831168831167, "step": 1431, "train/loss_ctc": 0.36882930994033813, "train/loss_error": 0.40603354573249817, "train/loss_total": 0.3985927104949951 }, { "epoch": 3.7194805194805194, "step": 1432, "train/loss_ctc": 0.43754321336746216, "train/loss_error": 0.20371362566947937, "train/loss_total": 0.2504795491695404 }, { "epoch": 3.722077922077922, "step": 1433, "train/loss_ctc": 0.4686324894428253, "train/loss_error": 0.3483988046646118, "train/loss_total": 0.3724455237388611 }, { "epoch": 3.724675324675325, "step": 1434, "train/loss_ctc": 0.5185533165931702, "train/loss_error": 0.44974157214164734, "train/loss_total": 0.46350395679473877 }, { "epoch": 3.7272727272727275, "step": 1435, "train/loss_ctc": 0.5094580054283142, "train/loss_error": 0.3983379006385803, "train/loss_total": 0.42056193947792053 }, { "epoch": 3.72987012987013, "step": 1436, "train/loss_ctc": 0.4176807403564453, "train/loss_error": 0.3596614897251129, "train/loss_total": 0.37126535177230835 }, { "epoch": 3.7324675324675325, "step": 1437, "train/loss_ctc": 0.3371213674545288, "train/loss_error": 0.29531291127204895, "train/loss_total": 0.3036746084690094 }, { "epoch": 3.735064935064935, "step": 1438, "train/loss_ctc": 0.5080126523971558, "train/loss_error": 0.42723676562309265, "train/loss_total": 0.44339194893836975 }, { "epoch": 3.7376623376623375, "step": 1439, "train/loss_ctc": 0.6276224851608276, "train/loss_error": 0.3428923487663269, "train/loss_total": 0.399838387966156 }, { "epoch": 3.74025974025974, "grad_norm": 15894.34765625, "learning_rate": 7.605194805194806e-06, "loss": 0.3857, "step": 1440 }, { "epoch": 3.74025974025974, "step": 1440, "train/loss_ctc": 0.6940932273864746, "train/loss_error": 0.30315983295440674, "train/loss_total": 0.38134652376174927 }, { "epoch": 3.742857142857143, "step": 1441, "train/loss_ctc": 0.5134188532829285, "train/loss_error": 0.32649335265159607, "train/loss_total": 0.363878458738327 }, { "epoch": 3.7454545454545456, "step": 1442, "train/loss_ctc": 0.7413984537124634, "train/loss_error": 0.41994214057922363, "train/loss_total": 0.48423340916633606 }, { "epoch": 3.7480519480519483, "step": 1443, "train/loss_ctc": 0.5446304678916931, "train/loss_error": 0.4618017375469208, "train/loss_total": 0.47836747765541077 }, { "epoch": 3.7506493506493506, "step": 1444, "train/loss_ctc": 0.4852025806903839, "train/loss_error": 0.3980903923511505, "train/loss_total": 0.4155128598213196 }, { "epoch": 3.7532467532467533, "step": 1445, "train/loss_ctc": 0.6152263879776001, "train/loss_error": 0.4057171940803528, "train/loss_total": 0.4476190209388733 }, { "epoch": 3.755844155844156, "step": 1446, "train/loss_ctc": 0.32245755195617676, "train/loss_error": 0.34771302342414856, "train/loss_total": 0.34266194701194763 }, { "epoch": 3.7584415584415583, "step": 1447, "train/loss_ctc": 0.4950721263885498, "train/loss_error": 0.37055978178977966, "train/loss_total": 0.3954622745513916 }, { "epoch": 3.761038961038961, "step": 1448, "train/loss_ctc": 0.388033390045166, "train/loss_error": 0.3097262680530548, "train/loss_total": 0.32538771629333496 }, { "epoch": 3.7636363636363637, "step": 1449, "train/loss_ctc": 0.4734245538711548, "train/loss_error": 0.3517723083496094, "train/loss_total": 0.3761027455329895 }, { "epoch": 3.7662337662337664, "grad_norm": 12027.9423828125, "learning_rate": 7.44935064935065e-06, "loss": 0.4011, "step": 1450 }, { "epoch": 3.7662337662337664, "step": 1450, "train/loss_ctc": 0.46091926097869873, "train/loss_error": 0.3528485894203186, "train/loss_total": 0.37446272373199463 }, { "epoch": 3.768831168831169, "step": 1451, "train/loss_ctc": 0.49971669912338257, "train/loss_error": 0.37265345454216003, "train/loss_total": 0.39806610345840454 }, { "epoch": 3.7714285714285714, "step": 1452, "train/loss_ctc": 0.7731975317001343, "train/loss_error": 0.38974010944366455, "train/loss_total": 0.4664316177368164 }, { "epoch": 3.774025974025974, "step": 1453, "train/loss_ctc": 0.3955535888671875, "train/loss_error": 0.3606290817260742, "train/loss_total": 0.3676139712333679 }, { "epoch": 3.7766233766233768, "step": 1454, "train/loss_ctc": 0.5112488269805908, "train/loss_error": 0.3940652906894684, "train/loss_total": 0.4175020158290863 }, { "epoch": 3.779220779220779, "step": 1455, "train/loss_ctc": 0.6329798102378845, "train/loss_error": 0.39883747696876526, "train/loss_total": 0.44566595554351807 }, { "epoch": 3.7818181818181817, "step": 1456, "train/loss_ctc": 0.473196804523468, "train/loss_error": 0.4246805012226105, "train/loss_total": 0.4343837797641754 }, { "epoch": 3.7844155844155845, "step": 1457, "train/loss_ctc": 0.3135078549385071, "train/loss_error": 0.36507511138916016, "train/loss_total": 0.35476168990135193 }, { "epoch": 3.787012987012987, "step": 1458, "train/loss_ctc": 0.3702933192253113, "train/loss_error": 0.33499404788017273, "train/loss_total": 0.3420538902282715 }, { "epoch": 3.78961038961039, "step": 1459, "train/loss_ctc": 0.4556751251220703, "train/loss_error": 0.33145084977149963, "train/loss_total": 0.35629570484161377 }, { "epoch": 3.792207792207792, "grad_norm": 13308.8564453125, "learning_rate": 7.293506493506494e-06, "loss": 0.3957, "step": 1460 }, { "epoch": 3.792207792207792, "step": 1460, "train/loss_ctc": 0.7887868881225586, "train/loss_error": 0.5034217834472656, "train/loss_total": 0.5604948401451111 }, { "epoch": 3.794805194805195, "step": 1461, "train/loss_ctc": 0.4987885355949402, "train/loss_error": 0.4077115058898926, "train/loss_total": 0.42592692375183105 }, { "epoch": 3.7974025974025976, "step": 1462, "train/loss_ctc": 0.625525712966919, "train/loss_error": 0.3523944318294525, "train/loss_total": 0.4070206880569458 }, { "epoch": 3.8, "step": 1463, "train/loss_ctc": 0.7822836637496948, "train/loss_error": 0.48568880558013916, "train/loss_total": 0.5450078248977661 }, { "epoch": 3.8025974025974025, "step": 1464, "train/loss_ctc": 0.40784013271331787, "train/loss_error": 0.37505844235420227, "train/loss_total": 0.3816148042678833 }, { "epoch": 3.8051948051948052, "step": 1465, "train/loss_ctc": 0.46929121017456055, "train/loss_error": 0.33048421144485474, "train/loss_total": 0.3582456111907959 }, { "epoch": 3.807792207792208, "step": 1466, "train/loss_ctc": 0.3277917802333832, "train/loss_error": 0.4125578701496124, "train/loss_total": 0.39560467004776 }, { "epoch": 3.8103896103896107, "step": 1467, "train/loss_ctc": 0.43451881408691406, "train/loss_error": 0.3296988606452942, "train/loss_total": 0.35066288709640503 }, { "epoch": 3.812987012987013, "step": 1468, "train/loss_ctc": 0.4589850604534149, "train/loss_error": 0.4227255582809448, "train/loss_total": 0.4299774765968323 }, { "epoch": 3.8155844155844156, "step": 1469, "train/loss_ctc": 0.7034968137741089, "train/loss_error": 0.3798121511936188, "train/loss_total": 0.4445490837097168 }, { "epoch": 3.8181818181818183, "grad_norm": 24086.958984375, "learning_rate": 7.137662337662338e-06, "loss": 0.4299, "step": 1470 }, { "epoch": 3.8181818181818183, "step": 1470, "train/loss_ctc": 0.31765514612197876, "train/loss_error": 0.3004254698753357, "train/loss_total": 0.30387142300605774 }, { "epoch": 3.8207792207792206, "step": 1471, "train/loss_ctc": 0.3929668664932251, "train/loss_error": 0.3972758948802948, "train/loss_total": 0.3964141011238098 }, { "epoch": 3.8233766233766233, "step": 1472, "train/loss_ctc": 0.586449384689331, "train/loss_error": 0.3582829535007477, "train/loss_total": 0.40391623973846436 }, { "epoch": 3.825974025974026, "step": 1473, "train/loss_ctc": 0.4025249779224396, "train/loss_error": 0.27496856451034546, "train/loss_total": 0.30047985911369324 }, { "epoch": 3.8285714285714287, "step": 1474, "train/loss_ctc": 0.35919302701950073, "train/loss_error": 0.296994686126709, "train/loss_total": 0.30943435430526733 }, { "epoch": 3.8311688311688314, "step": 1475, "train/loss_ctc": 0.4335325360298157, "train/loss_error": 0.4384324550628662, "train/loss_total": 0.437452495098114 }, { "epoch": 3.8337662337662337, "step": 1476, "train/loss_ctc": 0.41178032755851746, "train/loss_error": 0.3218232989311218, "train/loss_total": 0.3398147225379944 }, { "epoch": 3.8363636363636364, "step": 1477, "train/loss_ctc": 0.355867862701416, "train/loss_error": 0.3206872344017029, "train/loss_total": 0.3277233839035034 }, { "epoch": 3.838961038961039, "step": 1478, "train/loss_ctc": 0.5583286285400391, "train/loss_error": 0.46687641739845276, "train/loss_total": 0.48516687750816345 }, { "epoch": 3.8415584415584414, "step": 1479, "train/loss_ctc": 0.7158451080322266, "train/loss_error": 0.38713952898979187, "train/loss_total": 0.4528806805610657 }, { "epoch": 3.844155844155844, "grad_norm": 13700.2021484375, "learning_rate": 6.981818181818182e-06, "loss": 0.3757, "step": 1480 }, { "epoch": 3.844155844155844, "step": 1480, "train/loss_ctc": 0.3053557276725769, "train/loss_error": 0.249030202627182, "train/loss_total": 0.2602953016757965 }, { "epoch": 3.846753246753247, "step": 1481, "train/loss_ctc": 0.4217373728752136, "train/loss_error": 0.3005586564540863, "train/loss_total": 0.3247944116592407 }, { "epoch": 3.849350649350649, "step": 1482, "train/loss_ctc": 0.7324960231781006, "train/loss_error": 0.3461085855960846, "train/loss_total": 0.4233860969543457 }, { "epoch": 3.851948051948052, "step": 1483, "train/loss_ctc": 0.4454890787601471, "train/loss_error": 0.41190651059150696, "train/loss_total": 0.41862303018569946 }, { "epoch": 3.8545454545454545, "step": 1484, "train/loss_ctc": 0.3805028796195984, "train/loss_error": 0.3283134996891022, "train/loss_total": 0.3387514054775238 }, { "epoch": 3.857142857142857, "step": 1485, "train/loss_ctc": 0.6451544761657715, "train/loss_error": 0.3120235800743103, "train/loss_total": 0.3786497712135315 }, { "epoch": 3.85974025974026, "step": 1486, "train/loss_ctc": 0.412222683429718, "train/loss_error": 0.37651029229164124, "train/loss_total": 0.38365277647972107 }, { "epoch": 3.862337662337662, "step": 1487, "train/loss_ctc": 0.38672134280204773, "train/loss_error": 0.38503867387771606, "train/loss_total": 0.3853752017021179 }, { "epoch": 3.864935064935065, "step": 1488, "train/loss_ctc": 0.6292425394058228, "train/loss_error": 0.39384397864341736, "train/loss_total": 0.44092369079589844 }, { "epoch": 3.8675324675324676, "step": 1489, "train/loss_ctc": 0.3278552293777466, "train/loss_error": 0.31948021054267883, "train/loss_total": 0.32115522027015686 }, { "epoch": 3.87012987012987, "grad_norm": 13252.8740234375, "learning_rate": 6.825974025974026e-06, "loss": 0.3676, "step": 1490 }, { "epoch": 3.87012987012987, "step": 1490, "train/loss_ctc": 0.36369723081588745, "train/loss_error": 0.30322206020355225, "train/loss_total": 0.3153170943260193 }, { "epoch": 3.8727272727272726, "step": 1491, "train/loss_ctc": 0.33073312044143677, "train/loss_error": 0.3105018436908722, "train/loss_total": 0.3145481050014496 }, { "epoch": 3.8753246753246753, "step": 1492, "train/loss_ctc": 0.4112263321876526, "train/loss_error": 0.38548117876052856, "train/loss_total": 0.39063021540641785 }, { "epoch": 3.877922077922078, "step": 1493, "train/loss_ctc": 0.5394669771194458, "train/loss_error": 0.36048319935798645, "train/loss_total": 0.3962799608707428 }, { "epoch": 3.8805194805194807, "step": 1494, "train/loss_ctc": 0.2905554473400116, "train/loss_error": 0.3226025104522705, "train/loss_total": 0.3161931037902832 }, { "epoch": 3.883116883116883, "step": 1495, "train/loss_ctc": 0.3144266605377197, "train/loss_error": 0.32654818892478943, "train/loss_total": 0.32412388920783997 }, { "epoch": 3.8857142857142857, "step": 1496, "train/loss_ctc": 0.5075337886810303, "train/loss_error": 0.40185263752937317, "train/loss_total": 0.4229888916015625 }, { "epoch": 3.8883116883116884, "step": 1497, "train/loss_ctc": 0.6998605728149414, "train/loss_error": 0.41292205452919006, "train/loss_total": 0.4703097641468048 }, { "epoch": 3.8909090909090907, "step": 1498, "train/loss_ctc": 0.5120557546615601, "train/loss_error": 0.40642979741096497, "train/loss_total": 0.42755499482154846 }, { "epoch": 3.8935064935064934, "step": 1499, "train/loss_ctc": 0.7703448534011841, "train/loss_error": 0.41143307089805603, "train/loss_total": 0.48321545124053955 }, { "epoch": 3.896103896103896, "grad_norm": 15836.34765625, "learning_rate": 6.6701298701298705e-06, "loss": 0.3861, "step": 1500 }, { "epoch": 3.896103896103896, "step": 1500, "train/loss_ctc": 0.40604376792907715, "train/loss_error": 0.2811366617679596, "train/loss_total": 0.30611810088157654 }, { "epoch": 3.898701298701299, "step": 1501, "train/loss_ctc": 0.48116594552993774, "train/loss_error": 0.3448968827724457, "train/loss_total": 0.3721506893634796 }, { "epoch": 3.9012987012987015, "step": 1502, "train/loss_ctc": 0.23245540261268616, "train/loss_error": 0.2391543686389923, "train/loss_total": 0.23781459033489227 }, { "epoch": 3.9038961038961038, "step": 1503, "train/loss_ctc": 0.4150792360305786, "train/loss_error": 0.33155199885368347, "train/loss_total": 0.348257452249527 }, { "epoch": 3.9064935064935065, "step": 1504, "train/loss_ctc": 0.31187570095062256, "train/loss_error": 0.2815415859222412, "train/loss_total": 0.28760841488838196 }, { "epoch": 3.909090909090909, "step": 1505, "train/loss_ctc": 0.5637625455856323, "train/loss_error": 0.3575051724910736, "train/loss_total": 0.3987566828727722 }, { "epoch": 3.9116883116883114, "step": 1506, "train/loss_ctc": 0.4043261408805847, "train/loss_error": 0.2912542223930359, "train/loss_total": 0.31386861205101013 }, { "epoch": 3.914285714285714, "step": 1507, "train/loss_ctc": 0.4985427260398865, "train/loss_error": 0.3512415289878845, "train/loss_total": 0.38070178031921387 }, { "epoch": 3.916883116883117, "step": 1508, "train/loss_ctc": 0.4694189727306366, "train/loss_error": 0.2897146940231323, "train/loss_total": 0.3256555497646332 }, { "epoch": 3.9194805194805196, "step": 1509, "train/loss_ctc": 0.49730369448661804, "train/loss_error": 0.3932269513607025, "train/loss_total": 0.41404232382774353 }, { "epoch": 3.9220779220779223, "grad_norm": 15293.6416015625, "learning_rate": 6.5142857142857145e-06, "loss": 0.3385, "step": 1510 }, { "epoch": 3.9220779220779223, "step": 1510, "train/loss_ctc": 0.36778688430786133, "train/loss_error": 0.3311052918434143, "train/loss_total": 0.3384416103363037 }, { "epoch": 3.9246753246753245, "step": 1511, "train/loss_ctc": 0.5291942954063416, "train/loss_error": 0.38773253560066223, "train/loss_total": 0.4160248935222626 }, { "epoch": 3.9272727272727272, "step": 1512, "train/loss_ctc": 0.5383105874061584, "train/loss_error": 0.3494854271411896, "train/loss_total": 0.38725045323371887 }, { "epoch": 3.92987012987013, "step": 1513, "train/loss_ctc": 0.545945942401886, "train/loss_error": 0.3116112947463989, "train/loss_total": 0.35847821831703186 }, { "epoch": 3.9324675324675322, "step": 1514, "train/loss_ctc": 0.5873615741729736, "train/loss_error": 0.3217124938964844, "train/loss_total": 0.3748423159122467 }, { "epoch": 3.935064935064935, "step": 1515, "train/loss_ctc": 0.8443213701248169, "train/loss_error": 0.35265544056892395, "train/loss_total": 0.45098865032196045 }, { "epoch": 3.9376623376623376, "step": 1516, "train/loss_ctc": 0.5022940039634705, "train/loss_error": 0.39593279361724854, "train/loss_total": 0.4172050356864929 }, { "epoch": 3.9402597402597404, "step": 1517, "train/loss_ctc": 0.4005122482776642, "train/loss_error": 0.4009411633014679, "train/loss_total": 0.4008553922176361 }, { "epoch": 3.942857142857143, "step": 1518, "train/loss_ctc": 0.29430219531059265, "train/loss_error": 0.31023499369621277, "train/loss_total": 0.3070484399795532 }, { "epoch": 3.9454545454545453, "step": 1519, "train/loss_ctc": 0.9838787317276001, "train/loss_error": 0.481351375579834, "train/loss_total": 0.5818568468093872 }, { "epoch": 3.948051948051948, "grad_norm": 20797.728515625, "learning_rate": 6.3584415584415584e-06, "loss": 0.4033, "step": 1520 }, { "epoch": 3.948051948051948, "step": 1520, "train/loss_ctc": 0.6636033654212952, "train/loss_error": 0.40850889682769775, "train/loss_total": 0.45952779054641724 }, { "epoch": 3.9506493506493507, "step": 1521, "train/loss_ctc": 0.45566701889038086, "train/loss_error": 0.41974982619285583, "train/loss_total": 0.42693328857421875 }, { "epoch": 3.953246753246753, "step": 1522, "train/loss_ctc": 0.4562848210334778, "train/loss_error": 0.4184722304344177, "train/loss_total": 0.4260347783565521 }, { "epoch": 3.9558441558441557, "step": 1523, "train/loss_ctc": 0.44102197885513306, "train/loss_error": 0.37539565563201904, "train/loss_total": 0.3885209560394287 }, { "epoch": 3.9584415584415584, "step": 1524, "train/loss_ctc": 0.33495795726776123, "train/loss_error": 0.3482505977153778, "train/loss_total": 0.34559208154678345 }, { "epoch": 3.961038961038961, "step": 1525, "train/loss_ctc": 0.37007439136505127, "train/loss_error": 0.2633034586906433, "train/loss_total": 0.28465765714645386 }, { "epoch": 3.963636363636364, "step": 1526, "train/loss_ctc": 0.8026880621910095, "train/loss_error": 0.4319985508918762, "train/loss_total": 0.5061364769935608 }, { "epoch": 3.966233766233766, "step": 1527, "train/loss_ctc": 0.42587387561798096, "train/loss_error": 0.346813440322876, "train/loss_total": 0.3626255393028259 }, { "epoch": 3.968831168831169, "step": 1528, "train/loss_ctc": 0.6428526639938354, "train/loss_error": 0.34223151206970215, "train/loss_total": 0.40235573053359985 }, { "epoch": 3.9714285714285715, "step": 1529, "train/loss_ctc": 0.6524847745895386, "train/loss_error": 0.3855424225330353, "train/loss_total": 0.4389309287071228 }, { "epoch": 3.974025974025974, "grad_norm": 13274.2119140625, "learning_rate": 6.202597402597402e-06, "loss": 0.4041, "step": 1530 }, { "epoch": 3.974025974025974, "step": 1530, "train/loss_ctc": 0.6809784173965454, "train/loss_error": 0.4365241229534149, "train/loss_total": 0.485414981842041 }, { "epoch": 3.9766233766233765, "step": 1531, "train/loss_ctc": 0.37594494223594666, "train/loss_error": 0.3362012803554535, "train/loss_total": 0.34415003657341003 }, { "epoch": 3.979220779220779, "step": 1532, "train/loss_ctc": 0.5438770651817322, "train/loss_error": 0.37553080916404724, "train/loss_total": 0.4092000722885132 }, { "epoch": 3.981818181818182, "step": 1533, "train/loss_ctc": 0.9810503721237183, "train/loss_error": 0.418466180562973, "train/loss_total": 0.530983030796051 }, { "epoch": 3.9844155844155846, "step": 1534, "train/loss_ctc": 0.37198731303215027, "train/loss_error": 0.31900185346603394, "train/loss_total": 0.32959896326065063 }, { "epoch": 3.987012987012987, "step": 1535, "train/loss_ctc": 0.44666486978530884, "train/loss_error": 0.39952945709228516, "train/loss_total": 0.40895652770996094 }, { "epoch": 3.9896103896103896, "step": 1536, "train/loss_ctc": 0.490900456905365, "train/loss_error": 0.3829994201660156, "train/loss_total": 0.40457963943481445 }, { "epoch": 3.9922077922077923, "step": 1537, "train/loss_ctc": 0.3941459655761719, "train/loss_error": 0.30954602360725403, "train/loss_total": 0.32646602392196655 }, { "epoch": 3.9948051948051946, "step": 1538, "train/loss_ctc": 0.4590134918689728, "train/loss_error": 0.346615731716156, "train/loss_total": 0.3690952658653259 }, { "epoch": 3.9974025974025973, "step": 1539, "train/loss_ctc": 0.7721630334854126, "train/loss_error": 0.37261801958084106, "train/loss_total": 0.4525270462036133 }, { "epoch": 4.0, "grad_norm": 19543.771484375, "learning_rate": 6.046753246753246e-06, "loss": 0.4061, "step": 1540 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.6108070015907288, "val/loss_error": 0.14871269464492798, "val/loss_total": 0.24113155901432037 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.6304459571838379, "val/loss_error": 0.43613263964653015, "val/loss_total": 0.47499531507492065 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.07339660078287125, "val/loss_error": 0.22115260362625122, "val/loss_total": 0.19160139560699463 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.21931859850883484, "val/loss_error": 0.08003588020801544, "val/loss_total": 0.10789242386817932 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.14087937772274017, "val/loss_error": 0.1982002705335617, "val/loss_total": 0.1867360919713974 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.10735104978084564, "val/loss_error": 0.30469760298728943, "val/loss_total": 0.2652283012866974 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.17059682309627533, "val/loss_error": 0.4620661735534668, "val/loss_total": 0.4037723243236542 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.4136902093887329, "val/loss_error": 0.3216513991355896, "val/loss_total": 0.34005916118621826 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.3321163058280945, "val/loss_error": 0.28636789321899414, "val/loss_total": 0.29551756381988525 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.33511942625045776, "val/loss_error": 0.2744376063346863, "val/loss_total": 0.28657397627830505 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.12142495810985565, "val/loss_error": 0.1624734103679657, "val/loss_total": 0.1542637199163437 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.22045807540416718, "val/loss_error": 0.2672308385372162, "val/loss_total": 0.25787630677223206 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.766738772392273, "val/loss_error": 0.37119144201278687, "val/loss_total": 0.450300931930542 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.8668340444564819, "val/loss_error": 0.5975932478904724, "val/loss_total": 0.6514414548873901 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.46355918049812317, "val/loss_error": 0.3012222349643707, "val/loss_total": 0.3336896300315857 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.532599925994873, "val/loss_error": 0.5406889915466309, "val/loss_total": 0.5390712022781372 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.07411230355501175, "val/loss_error": 0.25623196363449097, "val/loss_total": 0.21980804204940796 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.1575203537940979, "val/loss_error": 0.2461078017950058, "val/loss_total": 0.22839030623435974 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.11782239377498627, "val/loss_error": 0.27414074540138245, "val/loss_total": 0.2428770661354065 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.6358099579811096, "val/loss_error": 0.3838818669319153, "val/loss_total": 0.434267520904541 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.803467869758606, "val/loss_error": 0.38785818219184875, "val/loss_total": 0.47098010778427124 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.2570599615573883, "val/loss_error": 0.4744274616241455, "val/loss_total": 0.4309539496898651 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.2765573561191559, "val/loss_error": 0.42313647270202637, "val/loss_total": 0.3938206434249878 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.18463627994060516, "val/loss_error": 0.2175787091255188, "val/loss_total": 0.21099022030830383 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.2887722849845886, "val/loss_error": 0.5382173657417297, "val/loss_total": 0.48832836747169495 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.47813013195991516, "val/loss_error": 0.5754708647727966, "val/loss_total": 0.5560027360916138 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.30527690052986145, "val/loss_error": 0.3186843693256378, "val/loss_total": 0.31600290536880493 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.1342298537492752, "val/loss_error": 0.2153022438287735, "val/loss_total": 0.19908776879310608 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.020544525235891342, "val/loss_error": 0.09530594199895859, "val/loss_total": 0.08035366237163544 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.24546842277050018, "val/loss_error": 0.5575332045555115, "val/loss_total": 0.49512025713920593 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.10096792131662369, "val/loss_error": 0.14918723702430725, "val/loss_total": 0.13954338431358337 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.10141830891370773, "val/loss_error": 0.24463163316249847, "val/loss_total": 0.21598896384239197 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.08457089215517044, "val/loss_error": 0.1602264940738678, "val/loss_total": 0.14509537816047668 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.6499009132385254, "val/loss_error": 0.6310842633247375, "val/loss_total": 0.6348476409912109 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.06885518878698349, "val/loss_error": 0.26689907908439636, "val/loss_total": 0.2272903025150299 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.19799332320690155, "val/loss_error": 0.38842350244522095, "val/loss_total": 0.3503374755382538 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5247838497161865, "val/loss_error": 0.5945099592208862, "val/loss_total": 0.5805647373199463 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.36618655920028687, "val/loss_error": 0.38403788208961487, "val/loss_total": 0.38046765327453613 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.23428642749786377, "val/loss_error": 0.534977376461029, "val/loss_total": 0.4748392105102539 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5560307502746582, "val/loss_error": 0.6148540377616882, "val/loss_total": 0.6030893921852112 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.22014737129211426, "val/loss_error": 0.2197331041097641, "val/loss_total": 0.2198159545660019 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.2783929705619812, "val/loss_error": 0.3718022108078003, "val/loss_total": 0.3531203866004944 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.6315423846244812, "val/loss_error": 0.9037126898765564, "val/loss_total": 0.8492786884307861 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.47924691438674927, "val/loss_error": 0.4793737828731537, "val/loss_total": 0.47934842109680176 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5101687908172607, "val/loss_error": 0.28011554479599, "val/loss_total": 0.32612618803977966 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.39479824900627136, "val/loss_error": 0.6547173857688904, "val/loss_total": 0.6027335524559021 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.6575723886489868, "val/loss_error": 0.4114125669002533, "val/loss_total": 0.46064454317092896 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.332327663898468, "val/loss_error": 0.5283367037773132, "val/loss_total": 0.48913490772247314 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.44579508900642395, "val/loss_error": 0.256684809923172, "val/loss_total": 0.29450687766075134 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.6330226063728333, "val/loss_error": 0.7777295708656311, "val/loss_total": 0.7487882375717163 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.34749820828437805, "val/loss_error": 0.7892658710479736, "val/loss_total": 0.700912356376648 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.3601461350917816, "val/loss_error": 0.4984378516674042, "val/loss_total": 0.47077950835227966 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.40985366702079773, "val/loss_error": 0.3886144459247589, "val/loss_total": 0.39286231994628906 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5873360633850098, "val/loss_error": 0.9101449847221375, "val/loss_total": 0.8455832004547119 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.43860575556755066, "val/loss_error": 0.41337457299232483, "val/loss_total": 0.41842079162597656 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.19593305885791779, "val/loss_error": 0.4246886074542999, "val/loss_total": 0.3789374828338623 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.14830052852630615, "val/loss_error": 0.2811950147151947, "val/loss_total": 0.2546161413192749 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.4909258186817169, "val/loss_error": 0.30233660340309143, "val/loss_total": 0.340054452419281 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5829991698265076, "val/loss_error": 0.5194646716117859, "val/loss_total": 0.5321716070175171 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.07523853331804276, "val/loss_error": 0.12336920201778412, "val/loss_total": 0.11374306678771973 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.8595620393753052, "val/loss_error": 0.6210249662399292, "val/loss_total": 0.6687324047088623 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.7251005172729492, "val/loss_error": 0.36806583404541016, "val/loss_total": 0.4394727945327759 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.06404977291822433, "val/loss_error": 0.08934058248996735, "val/loss_total": 0.08428242057561874 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.20801356434822083, "val/loss_error": 0.2587154805660248, "val/loss_total": 0.2485751062631607 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.11829302459955215, "val/loss_error": 0.2261824905872345, "val/loss_total": 0.2046045958995819 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.21489769220352173, "val/loss_error": 0.3536774516105652, "val/loss_total": 0.32592150568962097 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.2814479470252991, "val/loss_error": 0.27579978108406067, "val/loss_total": 0.27692940831184387 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5593850612640381, "val/loss_error": 0.4926133453845978, "val/loss_total": 0.5059676766395569 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.4777630865573883, "val/loss_error": 0.5205134749412537, "val/loss_total": 0.5119633674621582 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.44655975699424744, "val/loss_error": 0.5119443535804749, "val/loss_total": 0.4988674521446228 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.20198439061641693, "val/loss_error": 0.2210645228624344, "val/loss_total": 0.21724849939346313 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5832992196083069, "val/loss_error": 0.5031632781028748, "val/loss_total": 0.5191904902458191 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5089853405952454, "val/loss_error": 0.7163639664649963, "val/loss_total": 0.6748882532119751 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.21554940938949585, "val/loss_error": 0.3810400664806366, "val/loss_total": 0.34794196486473083 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.22545650601387024, "val/loss_error": 0.267609179019928, "val/loss_total": 0.25917863845825195 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.034917376935482025, "val/loss_error": 0.2527502477169037, "val/loss_total": 0.2091836780309677 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.7252546548843384, "val/loss_error": 0.5537205934524536, "val/loss_total": 0.5880274176597595 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.7432459592819214, "val/loss_error": 0.37653112411499023, "val/loss_total": 0.4498741030693054 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.6498732566833496, "val/loss_error": 0.35102856159210205, "val/loss_total": 0.41079747676849365 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.0034804970491677523, "val/loss_error": 0.2219546139240265, "val/loss_total": 0.17825978994369507 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.4816752076148987, "val/loss_error": 0.29911163449287415, "val/loss_total": 0.3356243669986725 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.028292985633015633, "val/loss_error": 0.10941319167613983, "val/loss_total": 0.09318915009498596 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.16454926133155823, "val/loss_error": 0.2413404881954193, "val/loss_total": 0.22598224878311157 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.43364620208740234, "val/loss_error": 0.3848989009857178, "val/loss_total": 0.39464837312698364 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.4053923785686493, "val/loss_error": 0.6592007279396057, "val/loss_total": 0.6084390878677368 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.4212489128112793, "val/loss_error": 0.5245116353034973, "val/loss_total": 0.5038591027259827 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.06878316402435303, "val/loss_error": 0.10925033688545227, "val/loss_total": 0.10115690529346466 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.3936697542667389, "val/loss_error": 0.43978366255760193, "val/loss_total": 0.4305608868598938 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.2247607260942459, "val/loss_error": 0.25728240609169006, "val/loss_total": 0.25077807903289795 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.31632959842681885, "val/loss_error": 0.38340604305267334, "val/loss_total": 0.3699907660484314 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.2880697250366211, "val/loss_error": 0.5174074172973633, "val/loss_total": 0.4715398848056793 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.9974081516265869, "val/loss_error": 0.45620226860046387, "val/loss_total": 0.5644434690475464 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.181112140417099, "val/loss_error": 0.08174696564674377, "val/loss_total": 0.10162000358104706 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.38187018036842346, "val/loss_error": 0.281594842672348, "val/loss_total": 0.30164992809295654 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5802484750747681, "val/loss_error": 0.5893639326095581, "val/loss_total": 0.587540864944458 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.2006026655435562, "val/loss_error": 0.1904543787240982, "val/loss_total": 0.192484050989151 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.09616649150848389, "val/loss_error": 0.38857582211494446, "val/loss_total": 0.33009395003318787 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.584559977054596, "val/loss_error": 0.5048948526382446, "val/loss_total": 0.5208278894424438 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.09512966126203537, "val/loss_error": 0.2999906837940216, "val/loss_total": 0.2590184807777405 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.19178394973278046, "val/loss_error": 0.22726638615131378, "val/loss_total": 0.22016990184783936 }, { "epoch": 4.0, "eval_eval/f1_0": 0.5270935893058777, "eval_eval/f1_1": 0.9320834875106812, "eval_eval/precision_0": 0.5721924901008606, "eval_eval/precision_1": 0.9216509461402893, "eval_eval/recall_0": 0.4885844886302948, "eval_eval/recall_1": 0.9427549242973328, "eval_eval/wer": 0.15060198073635145, "eval_loss": 0.3745430111885071, "eval_runtime": 7.3749, "eval_samples_per_second": 13.56, "eval_steps_per_second": 13.56, "step": 1540 }, { "epoch": 4.0, "step": 1540, "train/loss_ctc": 0.5264116525650024, "train/loss_error": 0.3790316581726074, "train/loss_total": 0.40850767493247986 }, { "epoch": 4.002597402597402, "step": 1541, "train/loss_ctc": 0.8496545553207397, "train/loss_error": 0.4332323670387268, "train/loss_total": 0.5165168046951294 }, { "epoch": 4.005194805194805, "step": 1542, "train/loss_ctc": 0.3010812997817993, "train/loss_error": 0.35738474130630493, "train/loss_total": 0.3461240530014038 }, { "epoch": 4.007792207792208, "step": 1543, "train/loss_ctc": 0.6685442328453064, "train/loss_error": 0.3650417625904083, "train/loss_total": 0.4257422685623169 }, { "epoch": 4.01038961038961, "step": 1544, "train/loss_ctc": 1.0764628648757935, "train/loss_error": 0.46245861053466797, "train/loss_total": 0.5852594375610352 }, { "epoch": 4.012987012987013, "step": 1545, "train/loss_ctc": 0.39370977878570557, "train/loss_error": 0.34373700618743896, "train/loss_total": 0.35373157262802124 }, { "epoch": 4.015584415584415, "step": 1546, "train/loss_ctc": 0.7669479846954346, "train/loss_error": 0.33116933703422546, "train/loss_total": 0.4183250665664673 }, { "epoch": 4.0181818181818185, "step": 1547, "train/loss_ctc": 0.5275654792785645, "train/loss_error": 0.32862573862075806, "train/loss_total": 0.36841368675231934 }, { "epoch": 4.020779220779221, "step": 1548, "train/loss_ctc": 0.42177411913871765, "train/loss_error": 0.3104780316352844, "train/loss_total": 0.3327372670173645 }, { "epoch": 4.023376623376623, "step": 1549, "train/loss_ctc": 0.44175195693969727, "train/loss_error": 0.40093231201171875, "train/loss_total": 0.40909624099731445 }, { "epoch": 4.025974025974026, "grad_norm": 12191.939453125, "learning_rate": 5.89090909090909e-06, "loss": 0.4164, "step": 1550 }, { "epoch": 4.025974025974026, "step": 1550, "train/loss_ctc": 0.4330451190471649, "train/loss_error": 0.37264230847358704, "train/loss_total": 0.38472288846969604 }, { "epoch": 4.0285714285714285, "step": 1551, "train/loss_ctc": 0.4034774601459503, "train/loss_error": 0.29916253685951233, "train/loss_total": 0.3200255334377289 }, { "epoch": 4.031168831168831, "step": 1552, "train/loss_ctc": 0.365505576133728, "train/loss_error": 0.3411548435688019, "train/loss_total": 0.3460249900817871 }, { "epoch": 4.033766233766234, "step": 1553, "train/loss_ctc": 0.804436206817627, "train/loss_error": 0.391533762216568, "train/loss_total": 0.4741142690181732 }, { "epoch": 4.036363636363636, "step": 1554, "train/loss_ctc": 0.3774441182613373, "train/loss_error": 0.3031335771083832, "train/loss_total": 0.31799569725990295 }, { "epoch": 4.038961038961039, "step": 1555, "train/loss_ctc": 0.31228676438331604, "train/loss_error": 0.2756184935569763, "train/loss_total": 0.28295212984085083 }, { "epoch": 4.041558441558442, "step": 1556, "train/loss_ctc": 0.43824303150177, "train/loss_error": 0.40715521574020386, "train/loss_total": 0.41337278485298157 }, { "epoch": 4.044155844155844, "step": 1557, "train/loss_ctc": 0.6517347097396851, "train/loss_error": 0.3374082148075104, "train/loss_total": 0.40027350187301636 }, { "epoch": 4.046753246753247, "step": 1558, "train/loss_ctc": 0.37888649106025696, "train/loss_error": 0.3402774930000305, "train/loss_total": 0.34799930453300476 }, { "epoch": 4.049350649350649, "step": 1559, "train/loss_ctc": 0.4138813614845276, "train/loss_error": 0.34713059663772583, "train/loss_total": 0.36048075556755066 }, { "epoch": 4.0519480519480515, "grad_norm": 13410.6552734375, "learning_rate": 5.735064935064936e-06, "loss": 0.3648, "step": 1560 }, { "epoch": 4.0519480519480515, "step": 1560, "train/loss_ctc": 0.6982173919677734, "train/loss_error": 0.40375858545303345, "train/loss_total": 0.4626503586769104 }, { "epoch": 4.054545454545455, "step": 1561, "train/loss_ctc": 0.6835360527038574, "train/loss_error": 0.27636489272117615, "train/loss_total": 0.35779911279678345 }, { "epoch": 4.057142857142857, "step": 1562, "train/loss_ctc": 0.2616802752017975, "train/loss_error": 0.2851397395133972, "train/loss_total": 0.2804478406906128 }, { "epoch": 4.05974025974026, "step": 1563, "train/loss_ctc": 0.5241789817810059, "train/loss_error": 0.34971970319747925, "train/loss_total": 0.384611576795578 }, { "epoch": 4.062337662337662, "step": 1564, "train/loss_ctc": 0.35564449429512024, "train/loss_error": 0.2911365330219269, "train/loss_total": 0.3040381371974945 }, { "epoch": 4.064935064935065, "step": 1565, "train/loss_ctc": 0.6520709991455078, "train/loss_error": 0.3702917993068695, "train/loss_total": 0.4266476631164551 }, { "epoch": 4.067532467532468, "step": 1566, "train/loss_ctc": 0.4215899109840393, "train/loss_error": 0.2592049241065979, "train/loss_total": 0.2916819453239441 }, { "epoch": 4.07012987012987, "step": 1567, "train/loss_ctc": 0.8647135496139526, "train/loss_error": 0.43110576272010803, "train/loss_total": 0.5178273320198059 }, { "epoch": 4.072727272727272, "step": 1568, "train/loss_ctc": 0.6347424983978271, "train/loss_error": 0.42567479610443115, "train/loss_total": 0.4674883484840393 }, { "epoch": 4.0753246753246755, "step": 1569, "train/loss_ctc": 0.4969777464866638, "train/loss_error": 0.3263380229473114, "train/loss_total": 0.36046597361564636 }, { "epoch": 4.077922077922078, "grad_norm": 11759.240234375, "learning_rate": 5.57922077922078e-06, "loss": 0.3854, "step": 1570 }, { "epoch": 4.077922077922078, "step": 1570, "train/loss_ctc": 0.27779242396354675, "train/loss_error": 0.3166942596435547, "train/loss_total": 0.3089138865470886 }, { "epoch": 4.080519480519481, "step": 1571, "train/loss_ctc": 0.4174672067165375, "train/loss_error": 0.35843122005462646, "train/loss_total": 0.37023842334747314 }, { "epoch": 4.083116883116883, "step": 1572, "train/loss_ctc": 0.5142124891281128, "train/loss_error": 0.42366868257522583, "train/loss_total": 0.44177746772766113 }, { "epoch": 4.085714285714285, "step": 1573, "train/loss_ctc": 0.5210083723068237, "train/loss_error": 0.32214322686195374, "train/loss_total": 0.36191627383232117 }, { "epoch": 4.088311688311689, "step": 1574, "train/loss_ctc": 0.5422226786613464, "train/loss_error": 0.40313756465911865, "train/loss_total": 0.43095460534095764 }, { "epoch": 4.090909090909091, "step": 1575, "train/loss_ctc": 0.38352975249290466, "train/loss_error": 0.30512943863868713, "train/loss_total": 0.3208095133304596 }, { "epoch": 4.093506493506493, "step": 1576, "train/loss_ctc": 0.6513584852218628, "train/loss_error": 0.535308301448822, "train/loss_total": 0.5585183501243591 }, { "epoch": 4.096103896103896, "step": 1577, "train/loss_ctc": 0.39958950877189636, "train/loss_error": 0.36148127913475037, "train/loss_total": 0.36910292506217957 }, { "epoch": 4.0987012987012985, "step": 1578, "train/loss_ctc": 0.5518328547477722, "train/loss_error": 0.38444119691848755, "train/loss_total": 0.4179195463657379 }, { "epoch": 4.101298701298702, "step": 1579, "train/loss_ctc": 0.4763992726802826, "train/loss_error": 0.33084380626678467, "train/loss_total": 0.3599548935890198 }, { "epoch": 4.103896103896104, "grad_norm": 12387.634765625, "learning_rate": 5.423376623376624e-06, "loss": 0.394, "step": 1580 }, { "epoch": 4.103896103896104, "step": 1580, "train/loss_ctc": 0.34248286485671997, "train/loss_error": 0.29313841462135315, "train/loss_total": 0.3030073046684265 }, { "epoch": 4.106493506493506, "step": 1581, "train/loss_ctc": 0.3981354832649231, "train/loss_error": 0.3421558737754822, "train/loss_total": 0.35335180163383484 }, { "epoch": 4.109090909090909, "step": 1582, "train/loss_ctc": 0.32601699233055115, "train/loss_error": 0.3044971227645874, "train/loss_total": 0.3088011145591736 }, { "epoch": 4.111688311688312, "step": 1583, "train/loss_ctc": 0.39661574363708496, "train/loss_error": 0.32460838556289673, "train/loss_total": 0.3390098512172699 }, { "epoch": 4.114285714285714, "step": 1584, "train/loss_ctc": 0.39051660895347595, "train/loss_error": 0.3939443528652191, "train/loss_total": 0.39325881004333496 }, { "epoch": 4.116883116883117, "step": 1585, "train/loss_ctc": 0.3028825521469116, "train/loss_error": 0.2905023992061615, "train/loss_total": 0.292978435754776 }, { "epoch": 4.119480519480519, "step": 1586, "train/loss_ctc": 0.37304410338401794, "train/loss_error": 0.431125283241272, "train/loss_total": 0.41950905323028564 }, { "epoch": 4.1220779220779225, "step": 1587, "train/loss_ctc": 0.5917024612426758, "train/loss_error": 0.3279516100883484, "train/loss_total": 0.38070178031921387 }, { "epoch": 4.124675324675325, "step": 1588, "train/loss_ctc": 0.7012611627578735, "train/loss_error": 0.3760777711868286, "train/loss_total": 0.4411144554615021 }, { "epoch": 4.127272727272727, "step": 1589, "train/loss_ctc": 0.4965001940727234, "train/loss_error": 0.363945871591568, "train/loss_total": 0.3904567360877991 }, { "epoch": 4.12987012987013, "grad_norm": 11930.712890625, "learning_rate": 5.267532467532468e-06, "loss": 0.3622, "step": 1590 }, { "epoch": 4.12987012987013, "step": 1590, "train/loss_ctc": 0.4209625720977783, "train/loss_error": 0.27337387204170227, "train/loss_total": 0.3028916120529175 }, { "epoch": 4.132467532467532, "step": 1591, "train/loss_ctc": 0.3032821714878082, "train/loss_error": 0.30014410614967346, "train/loss_total": 0.30077171325683594 }, { "epoch": 4.135064935064935, "step": 1592, "train/loss_ctc": 0.47258636355400085, "train/loss_error": 0.35897043347358704, "train/loss_total": 0.38169360160827637 }, { "epoch": 4.137662337662338, "step": 1593, "train/loss_ctc": 0.5470438003540039, "train/loss_error": 0.3217613101005554, "train/loss_total": 0.366817831993103 }, { "epoch": 4.14025974025974, "step": 1594, "train/loss_ctc": 0.7102273106575012, "train/loss_error": 0.37182241678237915, "train/loss_total": 0.43950340151786804 }, { "epoch": 4.142857142857143, "step": 1595, "train/loss_ctc": 0.3124656677246094, "train/loss_error": 0.35313868522644043, "train/loss_total": 0.3450041115283966 }, { "epoch": 4.1454545454545455, "step": 1596, "train/loss_ctc": 0.3335476815700531, "train/loss_error": 0.3207409381866455, "train/loss_total": 0.323302298784256 }, { "epoch": 4.148051948051948, "step": 1597, "train/loss_ctc": 0.29372385144233704, "train/loss_error": 0.2904520034790039, "train/loss_total": 0.29110637307167053 }, { "epoch": 4.150649350649351, "step": 1598, "train/loss_ctc": 0.9558556079864502, "train/loss_error": 0.4545430541038513, "train/loss_total": 0.5548055768013 }, { "epoch": 4.153246753246753, "step": 1599, "train/loss_ctc": 0.570469081401825, "train/loss_error": 0.4065849184989929, "train/loss_total": 0.4393617510795593 }, { "epoch": 4.1558441558441555, "grad_norm": 13494.0517578125, "learning_rate": 5.111688311688312e-06, "loss": 0.3745, "step": 1600 }, { "epoch": 4.1558441558441555, "step": 1600, "train/loss_ctc": 0.4756833612918854, "train/loss_error": 0.3171537518501282, "train/loss_total": 0.34885966777801514 }, { "epoch": 4.158441558441559, "step": 1601, "train/loss_ctc": 0.3852808475494385, "train/loss_error": 0.2928667366504669, "train/loss_total": 0.3113495707511902 }, { "epoch": 4.161038961038961, "step": 1602, "train/loss_ctc": 0.4596272110939026, "train/loss_error": 0.35588210821151733, "train/loss_total": 0.37663114070892334 }, { "epoch": 4.163636363636364, "step": 1603, "train/loss_ctc": 0.4754331409931183, "train/loss_error": 0.3875206708908081, "train/loss_total": 0.4051031768321991 }, { "epoch": 4.166233766233766, "step": 1604, "train/loss_ctc": 0.5487741231918335, "train/loss_error": 0.36193540692329407, "train/loss_total": 0.3993031680583954 }, { "epoch": 4.1688311688311686, "step": 1605, "train/loss_ctc": 0.3939465582370758, "train/loss_error": 0.35271018743515015, "train/loss_total": 0.36095747351646423 }, { "epoch": 4.171428571428572, "step": 1606, "train/loss_ctc": 0.455871045589447, "train/loss_error": 0.36183395981788635, "train/loss_total": 0.3806414008140564 }, { "epoch": 4.174025974025974, "step": 1607, "train/loss_ctc": 0.4292304515838623, "train/loss_error": 0.3297986686229706, "train/loss_total": 0.34968504309654236 }, { "epoch": 4.176623376623376, "step": 1608, "train/loss_ctc": 0.47049206495285034, "train/loss_error": 0.32182013988494873, "train/loss_total": 0.3515545427799225 }, { "epoch": 4.179220779220779, "step": 1609, "train/loss_ctc": 0.39160841703414917, "train/loss_error": 0.3553597033023834, "train/loss_total": 0.36260947585105896 }, { "epoch": 4.181818181818182, "grad_norm": 9538.6376953125, "learning_rate": 4.955844155844156e-06, "loss": 0.3647, "step": 1610 }, { "epoch": 4.181818181818182, "step": 1610, "train/loss_ctc": 0.4787585437297821, "train/loss_error": 0.41502541303634644, "train/loss_total": 0.42777204513549805 }, { "epoch": 4.184415584415585, "step": 1611, "train/loss_ctc": 0.3798374533653259, "train/loss_error": 0.340362548828125, "train/loss_total": 0.34825754165649414 }, { "epoch": 4.187012987012987, "step": 1612, "train/loss_ctc": 0.5441655516624451, "train/loss_error": 0.34791940450668335, "train/loss_total": 0.38716864585876465 }, { "epoch": 4.189610389610389, "step": 1613, "train/loss_ctc": 0.47925472259521484, "train/loss_error": 0.34141114354133606, "train/loss_total": 0.36897987127304077 }, { "epoch": 4.1922077922077925, "step": 1614, "train/loss_ctc": 0.3732912838459015, "train/loss_error": 0.35907062888145447, "train/loss_total": 0.3619147539138794 }, { "epoch": 4.194805194805195, "step": 1615, "train/loss_ctc": 0.43665140867233276, "train/loss_error": 0.3315894603729248, "train/loss_total": 0.3526018559932709 }, { "epoch": 4.197402597402597, "step": 1616, "train/loss_ctc": 0.35279178619384766, "train/loss_error": 0.2909136414527893, "train/loss_total": 0.3032892644405365 }, { "epoch": 4.2, "step": 1617, "train/loss_ctc": 0.4780438542366028, "train/loss_error": 0.36772894859313965, "train/loss_total": 0.38979193568229675 }, { "epoch": 4.202597402597402, "step": 1618, "train/loss_ctc": 0.40623414516448975, "train/loss_error": 0.33413955569267273, "train/loss_total": 0.3485584855079651 }, { "epoch": 4.205194805194806, "step": 1619, "train/loss_ctc": 0.4488881230354309, "train/loss_error": 0.41439589858055115, "train/loss_total": 0.42129433155059814 }, { "epoch": 4.207792207792208, "grad_norm": 14666.697265625, "learning_rate": 4.800000000000001e-06, "loss": 0.371, "step": 1620 }, { "epoch": 4.207792207792208, "step": 1620, "train/loss_ctc": 0.5709483623504639, "train/loss_error": 0.37926483154296875, "train/loss_total": 0.4176015555858612 }, { "epoch": 4.21038961038961, "step": 1621, "train/loss_ctc": 0.32375431060791016, "train/loss_error": 0.3087718188762665, "train/loss_total": 0.3117683231830597 }, { "epoch": 4.212987012987013, "step": 1622, "train/loss_ctc": 0.5602771639823914, "train/loss_error": 0.4525972306728363, "train/loss_total": 0.4741332530975342 }, { "epoch": 4.2155844155844155, "step": 1623, "train/loss_ctc": 0.3525269627571106, "train/loss_error": 0.31350645422935486, "train/loss_total": 0.3213105797767639 }, { "epoch": 4.218181818181818, "step": 1624, "train/loss_ctc": 0.3512880504131317, "train/loss_error": 0.2660752832889557, "train/loss_total": 0.2831178307533264 }, { "epoch": 4.220779220779221, "step": 1625, "train/loss_ctc": 0.35133057832717896, "train/loss_error": 0.2412075698375702, "train/loss_total": 0.26323217153549194 }, { "epoch": 4.223376623376623, "step": 1626, "train/loss_ctc": 0.379815936088562, "train/loss_error": 0.29141294956207275, "train/loss_total": 0.30909356474876404 }, { "epoch": 4.225974025974026, "step": 1627, "train/loss_ctc": 0.5138907432556152, "train/loss_error": 0.29184070229530334, "train/loss_total": 0.3362507224082947 }, { "epoch": 4.228571428571429, "step": 1628, "train/loss_ctc": 0.2395220398902893, "train/loss_error": 0.3028130531311035, "train/loss_total": 0.2901548445224762 }, { "epoch": 4.231168831168831, "step": 1629, "train/loss_ctc": 0.6293154954910278, "train/loss_error": 0.40838387608528137, "train/loss_total": 0.45257019996643066 }, { "epoch": 4.233766233766234, "grad_norm": 14089.9091796875, "learning_rate": 4.644155844155845e-06, "loss": 0.3459, "step": 1630 }, { "epoch": 4.233766233766234, "step": 1630, "train/loss_ctc": 0.49126094579696655, "train/loss_error": 0.3775891661643982, "train/loss_total": 0.4003235101699829 }, { "epoch": 4.236363636363636, "step": 1631, "train/loss_ctc": 0.5054614543914795, "train/loss_error": 0.34961333870887756, "train/loss_total": 0.38078296184539795 }, { "epoch": 4.238961038961039, "step": 1632, "train/loss_ctc": 0.5265476703643799, "train/loss_error": 0.29666370153427124, "train/loss_total": 0.3426404893398285 }, { "epoch": 4.241558441558442, "step": 1633, "train/loss_ctc": 0.5380909442901611, "train/loss_error": 0.428893506526947, "train/loss_total": 0.4507330060005188 }, { "epoch": 4.244155844155844, "step": 1634, "train/loss_ctc": 0.30550217628479004, "train/loss_error": 0.2489900141954422, "train/loss_total": 0.2602924406528473 }, { "epoch": 4.246753246753247, "step": 1635, "train/loss_ctc": 0.4321257770061493, "train/loss_error": 0.323903888463974, "train/loss_total": 0.34554827213287354 }, { "epoch": 4.249350649350649, "step": 1636, "train/loss_ctc": 0.43458086252212524, "train/loss_error": 0.31434154510498047, "train/loss_total": 0.33838942646980286 }, { "epoch": 4.251948051948052, "step": 1637, "train/loss_ctc": 0.3694554269313812, "train/loss_error": 0.3704592287540436, "train/loss_total": 0.3702584505081177 }, { "epoch": 4.254545454545455, "step": 1638, "train/loss_ctc": 0.29513299465179443, "train/loss_error": 0.3504282534122467, "train/loss_total": 0.3393692076206207 }, { "epoch": 4.257142857142857, "step": 1639, "train/loss_ctc": 0.6319746971130371, "train/loss_error": 0.4384007751941681, "train/loss_total": 0.47711557149887085 }, { "epoch": 4.259740259740259, "grad_norm": 17536.341796875, "learning_rate": 4.4883116883116886e-06, "loss": 0.3705, "step": 1640 }, { "epoch": 4.259740259740259, "step": 1640, "train/loss_ctc": 0.45284610986709595, "train/loss_error": 0.2556131184101105, "train/loss_total": 0.2950597107410431 }, { "epoch": 4.2623376623376625, "step": 1641, "train/loss_ctc": 0.4732794165611267, "train/loss_error": 0.26467689871788025, "train/loss_total": 0.306397408246994 }, { "epoch": 4.264935064935065, "step": 1642, "train/loss_ctc": 0.6055819988250732, "train/loss_error": 0.33647873997688293, "train/loss_total": 0.39029940962791443 }, { "epoch": 4.267532467532467, "step": 1643, "train/loss_ctc": 0.48036259412765503, "train/loss_error": 0.462152361869812, "train/loss_total": 0.4657944142818451 }, { "epoch": 4.27012987012987, "step": 1644, "train/loss_ctc": 0.6047447919845581, "train/loss_error": 0.32606908679008484, "train/loss_total": 0.3818042576313019 }, { "epoch": 4.2727272727272725, "step": 1645, "train/loss_ctc": 0.4210677146911621, "train/loss_error": 0.25769349932670593, "train/loss_total": 0.29036834836006165 }, { "epoch": 4.275324675324676, "step": 1646, "train/loss_ctc": 0.532503604888916, "train/loss_error": 0.3422383666038513, "train/loss_total": 0.3802914023399353 }, { "epoch": 4.277922077922078, "step": 1647, "train/loss_ctc": 0.45890897512435913, "train/loss_error": 0.34729018807411194, "train/loss_total": 0.3696139454841614 }, { "epoch": 4.28051948051948, "step": 1648, "train/loss_ctc": 0.9367424249649048, "train/loss_error": 0.39222612977027893, "train/loss_total": 0.5011293888092041 }, { "epoch": 4.283116883116883, "step": 1649, "train/loss_ctc": 0.5019376873970032, "train/loss_error": 0.34725067019462585, "train/loss_total": 0.3781880736351013 }, { "epoch": 4.285714285714286, "grad_norm": 13324.130859375, "learning_rate": 4.3324675324675325e-06, "loss": 0.3759, "step": 1650 }, { "epoch": 4.285714285714286, "step": 1650, "train/loss_ctc": 0.42465126514434814, "train/loss_error": 0.3145655691623688, "train/loss_total": 0.3365827202796936 }, { "epoch": 4.288311688311689, "step": 1651, "train/loss_ctc": 0.4685027301311493, "train/loss_error": 0.4172636568546295, "train/loss_total": 0.4275114834308624 }, { "epoch": 4.290909090909091, "step": 1652, "train/loss_ctc": 0.23072019219398499, "train/loss_error": 0.24859318137168884, "train/loss_total": 0.2450185865163803 }, { "epoch": 4.293506493506493, "step": 1653, "train/loss_ctc": 0.40620818734169006, "train/loss_error": 0.34023770689964294, "train/loss_total": 0.3534318208694458 }, { "epoch": 4.296103896103896, "step": 1654, "train/loss_ctc": 0.49162954092025757, "train/loss_error": 0.43040433526039124, "train/loss_total": 0.44264939427375793 }, { "epoch": 4.298701298701299, "step": 1655, "train/loss_ctc": 0.4743306636810303, "train/loss_error": 0.3576766550540924, "train/loss_total": 0.38100746273994446 }, { "epoch": 4.301298701298701, "step": 1656, "train/loss_ctc": 0.39657682180404663, "train/loss_error": 0.4367539584636688, "train/loss_total": 0.42871853709220886 }, { "epoch": 4.303896103896104, "step": 1657, "train/loss_ctc": 0.3222450613975525, "train/loss_error": 0.3585856258869171, "train/loss_total": 0.35131752490997314 }, { "epoch": 4.306493506493506, "step": 1658, "train/loss_ctc": 0.5161315202713013, "train/loss_error": 0.3589870035648346, "train/loss_total": 0.39041590690612793 }, { "epoch": 4.309090909090909, "step": 1659, "train/loss_ctc": 0.49662190675735474, "train/loss_error": 0.35311734676361084, "train/loss_total": 0.3818182647228241 }, { "epoch": 4.311688311688312, "grad_norm": 15740.8212890625, "learning_rate": 4.1766233766233765e-06, "loss": 0.3738, "step": 1660 }, { "epoch": 4.311688311688312, "step": 1660, "train/loss_ctc": 0.37987852096557617, "train/loss_error": 0.31847646832466125, "train/loss_total": 0.33075690269470215 }, { "epoch": 4.314285714285714, "step": 1661, "train/loss_ctc": 0.28958848118782043, "train/loss_error": 0.2978302836418152, "train/loss_total": 0.29618191719055176 }, { "epoch": 4.316883116883117, "step": 1662, "train/loss_ctc": 0.40770822763442993, "train/loss_error": 0.3628056049346924, "train/loss_total": 0.3717861473560333 }, { "epoch": 4.3194805194805195, "step": 1663, "train/loss_ctc": 0.49408069252967834, "train/loss_error": 0.40470585227012634, "train/loss_total": 0.4225808382034302 }, { "epoch": 4.322077922077922, "step": 1664, "train/loss_ctc": 0.6601415872573853, "train/loss_error": 0.35441020131111145, "train/loss_total": 0.41555649042129517 }, { "epoch": 4.324675324675325, "step": 1665, "train/loss_ctc": 0.3125847280025482, "train/loss_error": 0.27166104316711426, "train/loss_total": 0.2798457741737366 }, { "epoch": 4.327272727272727, "step": 1666, "train/loss_ctc": 0.3023897707462311, "train/loss_error": 0.24427403509616852, "train/loss_total": 0.25589719414711 }, { "epoch": 4.32987012987013, "step": 1667, "train/loss_ctc": 0.35400816798210144, "train/loss_error": 0.3789285123348236, "train/loss_total": 0.3739444613456726 }, { "epoch": 4.332467532467533, "step": 1668, "train/loss_ctc": 0.5681357383728027, "train/loss_error": 0.4631968140602112, "train/loss_total": 0.4841846227645874 }, { "epoch": 4.335064935064935, "step": 1669, "train/loss_ctc": 0.4985438585281372, "train/loss_error": 0.3611109256744385, "train/loss_total": 0.3885975182056427 }, { "epoch": 4.337662337662338, "grad_norm": 12865.810546875, "learning_rate": 4.0207792207792205e-06, "loss": 0.3619, "step": 1670 }, { "epoch": 4.337662337662338, "step": 1670, "train/loss_ctc": 0.5315379500389099, "train/loss_error": 0.3935863673686981, "train/loss_total": 0.4211767017841339 }, { "epoch": 4.34025974025974, "step": 1671, "train/loss_ctc": 0.5945537090301514, "train/loss_error": 0.267568975687027, "train/loss_total": 0.3329659104347229 }, { "epoch": 4.3428571428571425, "step": 1672, "train/loss_ctc": 0.42929568886756897, "train/loss_error": 0.3266124725341797, "train/loss_total": 0.347149133682251 }, { "epoch": 4.345454545454546, "step": 1673, "train/loss_ctc": 0.46957430243492126, "train/loss_error": 0.3555486500263214, "train/loss_total": 0.3783538043498993 }, { "epoch": 4.348051948051948, "step": 1674, "train/loss_ctc": 0.6251452565193176, "train/loss_error": 0.41969865560531616, "train/loss_total": 0.46078798174858093 }, { "epoch": 4.35064935064935, "step": 1675, "train/loss_ctc": 0.3062671720981598, "train/loss_error": 0.2602836787700653, "train/loss_total": 0.2694803774356842 }, { "epoch": 4.353246753246753, "step": 1676, "train/loss_ctc": 0.30381906032562256, "train/loss_error": 0.29092150926589966, "train/loss_total": 0.29350101947784424 }, { "epoch": 4.355844155844156, "step": 1677, "train/loss_ctc": 0.5041653513908386, "train/loss_error": 0.3596532642841339, "train/loss_total": 0.38855570554733276 }, { "epoch": 4.358441558441559, "step": 1678, "train/loss_ctc": 0.41688716411590576, "train/loss_error": 0.3936496675014496, "train/loss_total": 0.39829719066619873 }, { "epoch": 4.361038961038961, "step": 1679, "train/loss_ctc": 0.5179513096809387, "train/loss_error": 0.383622407913208, "train/loss_total": 0.41048818826675415 }, { "epoch": 4.363636363636363, "grad_norm": 16602.138671875, "learning_rate": 3.864935064935064e-06, "loss": 0.3701, "step": 1680 }, { "epoch": 4.363636363636363, "step": 1680, "train/loss_ctc": 0.42874059081077576, "train/loss_error": 0.34174299240112305, "train/loss_total": 0.359142541885376 }, { "epoch": 4.3662337662337665, "step": 1681, "train/loss_ctc": 0.3851499855518341, "train/loss_error": 0.32311320304870605, "train/loss_total": 0.33552056550979614 }, { "epoch": 4.368831168831169, "step": 1682, "train/loss_ctc": 0.5499356389045715, "train/loss_error": 0.38616567850112915, "train/loss_total": 0.4189196825027466 }, { "epoch": 4.371428571428572, "step": 1683, "train/loss_ctc": 0.4936828911304474, "train/loss_error": 0.33149683475494385, "train/loss_total": 0.36393406987190247 }, { "epoch": 4.374025974025974, "step": 1684, "train/loss_ctc": 0.38149750232696533, "train/loss_error": 0.339474081993103, "train/loss_total": 0.34787875413894653 }, { "epoch": 4.376623376623376, "step": 1685, "train/loss_ctc": 0.743409276008606, "train/loss_error": 0.39384737610816956, "train/loss_total": 0.46375977993011475 }, { "epoch": 4.37922077922078, "step": 1686, "train/loss_ctc": 0.3370553255081177, "train/loss_error": 0.3576493561267853, "train/loss_total": 0.35353055596351624 }, { "epoch": 4.381818181818182, "step": 1687, "train/loss_ctc": 0.3648395538330078, "train/loss_error": 0.3598279058933258, "train/loss_total": 0.36083024740219116 }, { "epoch": 4.384415584415584, "step": 1688, "train/loss_ctc": 0.43782472610473633, "train/loss_error": 0.36073195934295654, "train/loss_total": 0.376150518655777 }, { "epoch": 4.387012987012987, "step": 1689, "train/loss_ctc": 0.5410850644111633, "train/loss_error": 0.291546106338501, "train/loss_total": 0.3414539098739624 }, { "epoch": 4.3896103896103895, "grad_norm": 14940.705078125, "learning_rate": 3.7090909090909092e-06, "loss": 0.3721, "step": 1690 }, { "epoch": 4.3896103896103895, "step": 1690, "train/loss_ctc": 0.5598846673965454, "train/loss_error": 0.3421550393104553, "train/loss_total": 0.3857010006904602 }, { "epoch": 4.392207792207792, "step": 1691, "train/loss_ctc": 0.3896540403366089, "train/loss_error": 0.26741278171539307, "train/loss_total": 0.29186105728149414 }, { "epoch": 4.394805194805195, "step": 1692, "train/loss_ctc": 0.5233289003372192, "train/loss_error": 0.3684619069099426, "train/loss_total": 0.3994353115558624 }, { "epoch": 4.397402597402597, "step": 1693, "train/loss_ctc": 0.7378149032592773, "train/loss_error": 0.37941715121269226, "train/loss_total": 0.45109671354293823 }, { "epoch": 4.4, "step": 1694, "train/loss_ctc": 0.4028298556804657, "train/loss_error": 0.2986256182193756, "train/loss_total": 0.3194664716720581 }, { "epoch": 4.402597402597403, "step": 1695, "train/loss_ctc": 0.9188803434371948, "train/loss_error": 0.31597045063972473, "train/loss_total": 0.43655240535736084 }, { "epoch": 4.405194805194805, "step": 1696, "train/loss_ctc": 0.3551790714263916, "train/loss_error": 0.39629191160202026, "train/loss_total": 0.3880693316459656 }, { "epoch": 4.407792207792208, "step": 1697, "train/loss_ctc": 0.4213399291038513, "train/loss_error": 0.3580942749977112, "train/loss_total": 0.37074339389801025 }, { "epoch": 4.41038961038961, "step": 1698, "train/loss_ctc": 0.9106787443161011, "train/loss_error": 0.3479875326156616, "train/loss_total": 0.4605257511138916 }, { "epoch": 4.4129870129870135, "step": 1699, "train/loss_ctc": 0.45316919684410095, "train/loss_error": 0.29265332221984863, "train/loss_total": 0.3247565031051636 }, { "epoch": 4.415584415584416, "grad_norm": 16052.501953125, "learning_rate": 3.5532467532467536e-06, "loss": 0.3828, "step": 1700 }, { "epoch": 4.415584415584416, "step": 1700, "train/loss_ctc": 0.42437243461608887, "train/loss_error": 0.2740904986858368, "train/loss_total": 0.3041468858718872 }, { "epoch": 4.418181818181818, "step": 1701, "train/loss_ctc": 0.5242879390716553, "train/loss_error": 0.42396080493927, "train/loss_total": 0.44402623176574707 }, { "epoch": 4.420779220779221, "step": 1702, "train/loss_ctc": 0.594342052936554, "train/loss_error": 0.4178844094276428, "train/loss_total": 0.45317593216896057 }, { "epoch": 4.423376623376623, "step": 1703, "train/loss_ctc": 1.145328164100647, "train/loss_error": 0.5046966671943665, "train/loss_total": 0.6328229904174805 }, { "epoch": 4.425974025974026, "step": 1704, "train/loss_ctc": 0.33207058906555176, "train/loss_error": 0.29532086849212646, "train/loss_total": 0.30267083644866943 }, { "epoch": 4.428571428571429, "step": 1705, "train/loss_ctc": 0.3973536789417267, "train/loss_error": 0.3779587149620056, "train/loss_total": 0.38183772563934326 }, { "epoch": 4.431168831168831, "step": 1706, "train/loss_ctc": 0.5416828989982605, "train/loss_error": 0.3317987620830536, "train/loss_total": 0.3737756013870239 }, { "epoch": 4.433766233766233, "step": 1707, "train/loss_ctc": 0.538821816444397, "train/loss_error": 0.3725827932357788, "train/loss_total": 0.40583059191703796 }, { "epoch": 4.4363636363636365, "step": 1708, "train/loss_ctc": 0.5069718360900879, "train/loss_error": 0.29369407892227173, "train/loss_total": 0.33634963631629944 }, { "epoch": 4.438961038961039, "step": 1709, "train/loss_ctc": 0.45312657952308655, "train/loss_error": 0.37184831500053406, "train/loss_total": 0.3881039619445801 }, { "epoch": 4.441558441558442, "grad_norm": 12463.794921875, "learning_rate": 3.3974025974025976e-06, "loss": 0.4023, "step": 1710 }, { "epoch": 4.441558441558442, "step": 1710, "train/loss_ctc": 0.5458440780639648, "train/loss_error": 0.3103291094303131, "train/loss_total": 0.357432097196579 }, { "epoch": 4.444155844155844, "step": 1711, "train/loss_ctc": 0.5477659702301025, "train/loss_error": 0.3330341875553131, "train/loss_total": 0.37598055601119995 }, { "epoch": 4.4467532467532465, "step": 1712, "train/loss_ctc": 0.35589319467544556, "train/loss_error": 0.28009697794914246, "train/loss_total": 0.29525622725486755 }, { "epoch": 4.44935064935065, "step": 1713, "train/loss_ctc": 0.4920697510242462, "train/loss_error": 0.3591461777687073, "train/loss_total": 0.38573089241981506 }, { "epoch": 4.451948051948052, "step": 1714, "train/loss_ctc": 0.3708862364292145, "train/loss_error": 0.3190045952796936, "train/loss_total": 0.32938092947006226 }, { "epoch": 4.454545454545454, "step": 1715, "train/loss_ctc": 0.499997079372406, "train/loss_error": 0.4267738461494446, "train/loss_total": 0.44141849875450134 }, { "epoch": 4.457142857142857, "step": 1716, "train/loss_ctc": 0.47499406337738037, "train/loss_error": 0.31000080704689026, "train/loss_total": 0.3429994583129883 }, { "epoch": 4.45974025974026, "step": 1717, "train/loss_ctc": 0.4800914227962494, "train/loss_error": 0.3483985364437103, "train/loss_total": 0.37473711371421814 }, { "epoch": 4.462337662337663, "step": 1718, "train/loss_ctc": 0.37107646465301514, "train/loss_error": 0.3787488639354706, "train/loss_total": 0.3772144019603729 }, { "epoch": 4.464935064935065, "step": 1719, "train/loss_ctc": 0.6211974620819092, "train/loss_error": 0.3507193624973297, "train/loss_total": 0.4048149883747101 }, { "epoch": 4.467532467532467, "grad_norm": 15421.5634765625, "learning_rate": 3.2415584415584415e-06, "loss": 0.3685, "step": 1720 }, { "epoch": 4.467532467532467, "step": 1720, "train/loss_ctc": 0.4869767129421234, "train/loss_error": 0.3577222228050232, "train/loss_total": 0.38357311487197876 }, { "epoch": 4.47012987012987, "step": 1721, "train/loss_ctc": 0.3023887276649475, "train/loss_error": 0.3048669695854187, "train/loss_total": 0.30437132716178894 }, { "epoch": 4.472727272727273, "step": 1722, "train/loss_ctc": 0.3937382698059082, "train/loss_error": 0.32767730951309204, "train/loss_total": 0.34088951349258423 }, { "epoch": 4.475324675324675, "step": 1723, "train/loss_ctc": 0.4289250671863556, "train/loss_error": 0.399455726146698, "train/loss_total": 0.40534961223602295 }, { "epoch": 4.477922077922078, "step": 1724, "train/loss_ctc": 0.5902477502822876, "train/loss_error": 0.3577337861061096, "train/loss_total": 0.4042365849018097 }, { "epoch": 4.48051948051948, "step": 1725, "train/loss_ctc": 0.5163820385932922, "train/loss_error": 0.4213932454586029, "train/loss_total": 0.44039100408554077 }, { "epoch": 4.4831168831168835, "step": 1726, "train/loss_ctc": 0.3524417281150818, "train/loss_error": 0.29736313223838806, "train/loss_total": 0.3083788752555847 }, { "epoch": 4.485714285714286, "step": 1727, "train/loss_ctc": 0.6638855934143066, "train/loss_error": 0.36460110545158386, "train/loss_total": 0.42445802688598633 }, { "epoch": 4.488311688311688, "step": 1728, "train/loss_ctc": 0.39325451850891113, "train/loss_error": 0.38027411699295044, "train/loss_total": 0.3828701972961426 }, { "epoch": 4.490909090909091, "step": 1729, "train/loss_ctc": 0.34681087732315063, "train/loss_error": 0.3159412741661072, "train/loss_total": 0.3221151828765869 }, { "epoch": 4.4935064935064934, "grad_norm": 11852.9365234375, "learning_rate": 3.085714285714286e-06, "loss": 0.3717, "step": 1730 }, { "epoch": 4.4935064935064934, "step": 1730, "train/loss_ctc": 0.5058676600456238, "train/loss_error": 0.38601696491241455, "train/loss_total": 0.40998709201812744 }, { "epoch": 4.496103896103896, "step": 1731, "train/loss_ctc": 0.5603839159011841, "train/loss_error": 0.30128881335258484, "train/loss_total": 0.35310783982276917 }, { "epoch": 4.498701298701299, "step": 1732, "train/loss_ctc": 0.7278800010681152, "train/loss_error": 0.43332669138908386, "train/loss_total": 0.4922373592853546 }, { "epoch": 4.501298701298701, "step": 1733, "train/loss_ctc": 0.5021476745605469, "train/loss_error": 0.3411104381084442, "train/loss_total": 0.3733178973197937 }, { "epoch": 4.503896103896103, "step": 1734, "train/loss_ctc": 0.5637474060058594, "train/loss_error": 0.3580198287963867, "train/loss_total": 0.3991653621196747 }, { "epoch": 4.5064935064935066, "step": 1735, "train/loss_ctc": 0.426108181476593, "train/loss_error": 0.3431994318962097, "train/loss_total": 0.3597812056541443 }, { "epoch": 4.509090909090909, "step": 1736, "train/loss_ctc": 0.35551732778549194, "train/loss_error": 0.3668413460254669, "train/loss_total": 0.3645765781402588 }, { "epoch": 4.511688311688312, "step": 1737, "train/loss_ctc": 0.3025667667388916, "train/loss_error": 0.23978374898433685, "train/loss_total": 0.2523403465747833 }, { "epoch": 4.514285714285714, "step": 1738, "train/loss_ctc": 0.45826590061187744, "train/loss_error": 0.3463607132434845, "train/loss_total": 0.3687417507171631 }, { "epoch": 4.5168831168831165, "step": 1739, "train/loss_ctc": 0.5437764525413513, "train/loss_error": 0.41358327865600586, "train/loss_total": 0.4396219253540039 }, { "epoch": 4.51948051948052, "grad_norm": 17568.966796875, "learning_rate": 2.92987012987013e-06, "loss": 0.3813, "step": 1740 }, { "epoch": 4.51948051948052, "step": 1740, "train/loss_ctc": 0.4329705238342285, "train/loss_error": 0.3056623041629791, "train/loss_total": 0.331123948097229 }, { "epoch": 4.522077922077922, "step": 1741, "train/loss_ctc": 0.4802708625793457, "train/loss_error": 0.4095848500728607, "train/loss_total": 0.4237220585346222 }, { "epoch": 4.524675324675325, "step": 1742, "train/loss_ctc": 0.3546503186225891, "train/loss_error": 0.32786405086517334, "train/loss_total": 0.33322131633758545 }, { "epoch": 4.527272727272727, "step": 1743, "train/loss_ctc": 0.3753076195716858, "train/loss_error": 0.2707216143608093, "train/loss_total": 0.2916388213634491 }, { "epoch": 4.52987012987013, "step": 1744, "train/loss_ctc": 0.3392906188964844, "train/loss_error": 0.3382156193256378, "train/loss_total": 0.33843064308166504 }, { "epoch": 4.532467532467533, "step": 1745, "train/loss_ctc": 0.37307238578796387, "train/loss_error": 0.27696263790130615, "train/loss_total": 0.29618459939956665 }, { "epoch": 4.535064935064935, "step": 1746, "train/loss_ctc": 0.3933819532394409, "train/loss_error": 0.4114111661911011, "train/loss_total": 0.40780535340309143 }, { "epoch": 4.537662337662337, "step": 1747, "train/loss_ctc": 0.5074543356895447, "train/loss_error": 0.35744860768318176, "train/loss_total": 0.3874497413635254 }, { "epoch": 4.54025974025974, "step": 1748, "train/loss_ctc": 0.4424342215061188, "train/loss_error": 0.29121461510658264, "train/loss_total": 0.3214585483074188 }, { "epoch": 4.542857142857143, "step": 1749, "train/loss_ctc": 0.43179985880851746, "train/loss_error": 0.2764589190483093, "train/loss_total": 0.307527095079422 }, { "epoch": 4.545454545454545, "grad_norm": 10908.6025390625, "learning_rate": 2.774025974025974e-06, "loss": 0.3439, "step": 1750 }, { "epoch": 4.545454545454545, "step": 1750, "train/loss_ctc": 0.24357888102531433, "train/loss_error": 0.23552954196929932, "train/loss_total": 0.23713940382003784 }, { "epoch": 4.548051948051948, "step": 1751, "train/loss_ctc": 0.3619012236595154, "train/loss_error": 0.29829835891723633, "train/loss_total": 0.3110189437866211 }, { "epoch": 4.55064935064935, "step": 1752, "train/loss_ctc": 0.4413699507713318, "train/loss_error": 0.3353140950202942, "train/loss_total": 0.3565252721309662 }, { "epoch": 4.5532467532467535, "step": 1753, "train/loss_ctc": 0.4700813889503479, "train/loss_error": 0.4381551742553711, "train/loss_total": 0.44454044103622437 }, { "epoch": 4.555844155844156, "step": 1754, "train/loss_ctc": 0.47573328018188477, "train/loss_error": 0.2832722067832947, "train/loss_total": 0.32176440954208374 }, { "epoch": 4.558441558441558, "step": 1755, "train/loss_ctc": 0.2714657187461853, "train/loss_error": 0.3009456396102905, "train/loss_total": 0.29504966735839844 }, { "epoch": 4.561038961038961, "step": 1756, "train/loss_ctc": 0.3978571593761444, "train/loss_error": 0.39432600140571594, "train/loss_total": 0.39503222703933716 }, { "epoch": 4.5636363636363635, "step": 1757, "train/loss_ctc": 0.32027268409729004, "train/loss_error": 0.29664698243141174, "train/loss_total": 0.30137214064598083 }, { "epoch": 4.566233766233767, "step": 1758, "train/loss_ctc": 0.34265682101249695, "train/loss_error": 0.3397325873374939, "train/loss_total": 0.34031742811203003 }, { "epoch": 4.568831168831169, "step": 1759, "train/loss_ctc": 0.33753353357315063, "train/loss_error": 0.3828791379928589, "train/loss_total": 0.3738100230693817 }, { "epoch": 4.571428571428571, "grad_norm": 10084.359375, "learning_rate": 2.6181818181818183e-06, "loss": 0.3377, "step": 1760 }, { "epoch": 4.571428571428571, "step": 1760, "train/loss_ctc": 0.46888816356658936, "train/loss_error": 0.30403029918670654, "train/loss_total": 0.33700188994407654 }, { "epoch": 4.574025974025974, "step": 1761, "train/loss_ctc": 0.4099580943584442, "train/loss_error": 0.38381779193878174, "train/loss_total": 0.3890458643436432 }, { "epoch": 4.576623376623377, "step": 1762, "train/loss_ctc": 0.3546020984649658, "train/loss_error": 0.3516969382762909, "train/loss_total": 0.3522779941558838 }, { "epoch": 4.579220779220779, "step": 1763, "train/loss_ctc": 0.4285764694213867, "train/loss_error": 0.3146323263645172, "train/loss_total": 0.33742114901542664 }, { "epoch": 4.581818181818182, "step": 1764, "train/loss_ctc": 0.43740880489349365, "train/loss_error": 0.44369420409202576, "train/loss_total": 0.44243714213371277 }, { "epoch": 4.584415584415584, "step": 1765, "train/loss_ctc": 0.5207366943359375, "train/loss_error": 0.3788785934448242, "train/loss_total": 0.40725022554397583 }, { "epoch": 4.5870129870129865, "step": 1766, "train/loss_ctc": 0.5240669250488281, "train/loss_error": 0.33918219804763794, "train/loss_total": 0.3761591613292694 }, { "epoch": 4.58961038961039, "step": 1767, "train/loss_ctc": 0.5074079632759094, "train/loss_error": 0.3286322355270386, "train/loss_total": 0.3643873929977417 }, { "epoch": 4.592207792207792, "step": 1768, "train/loss_ctc": 0.4397282004356384, "train/loss_error": 0.31839460134506226, "train/loss_total": 0.3426613211631775 }, { "epoch": 4.594805194805195, "step": 1769, "train/loss_ctc": 0.3451780378818512, "train/loss_error": 0.3766482472419739, "train/loss_total": 0.37035423517227173 }, { "epoch": 4.597402597402597, "grad_norm": 10342.181640625, "learning_rate": 2.4623376623376626e-06, "loss": 0.3719, "step": 1770 }, { "epoch": 4.597402597402597, "step": 1770, "train/loss_ctc": 0.673514187335968, "train/loss_error": 0.38749051094055176, "train/loss_total": 0.44469523429870605 }, { "epoch": 4.6, "step": 1771, "train/loss_ctc": 0.6683873534202576, "train/loss_error": 0.40755757689476013, "train/loss_total": 0.4597235321998596 }, { "epoch": 4.602597402597403, "step": 1772, "train/loss_ctc": 0.49145668745040894, "train/loss_error": 0.28446564078330994, "train/loss_total": 0.3258638381958008 }, { "epoch": 4.605194805194805, "step": 1773, "train/loss_ctc": 0.3203176259994507, "train/loss_error": 0.2908592224121094, "train/loss_total": 0.29675090312957764 }, { "epoch": 4.607792207792208, "step": 1774, "train/loss_ctc": 0.4440547525882721, "train/loss_error": 0.3256712853908539, "train/loss_total": 0.3493479788303375 }, { "epoch": 4.6103896103896105, "step": 1775, "train/loss_ctc": 0.40282300114631653, "train/loss_error": 0.3407711386680603, "train/loss_total": 0.35318154096603394 }, { "epoch": 4.612987012987013, "step": 1776, "train/loss_ctc": 0.38879185914993286, "train/loss_error": 0.33087417483329773, "train/loss_total": 0.34245771169662476 }, { "epoch": 4.615584415584416, "step": 1777, "train/loss_ctc": 0.584128737449646, "train/loss_error": 0.404556006193161, "train/loss_total": 0.4404705762863159 }, { "epoch": 4.618181818181818, "step": 1778, "train/loss_ctc": 0.30566272139549255, "train/loss_error": 0.3770371675491333, "train/loss_total": 0.36276230216026306 }, { "epoch": 4.62077922077922, "step": 1779, "train/loss_ctc": 0.4886375665664673, "train/loss_error": 0.33638468384742737, "train/loss_total": 0.36683526635169983 }, { "epoch": 4.623376623376624, "grad_norm": 16350.5458984375, "learning_rate": 2.3064935064935066e-06, "loss": 0.3742, "step": 1780 }, { "epoch": 4.623376623376624, "step": 1780, "train/loss_ctc": 0.4744813144207001, "train/loss_error": 0.2912651300430298, "train/loss_total": 0.32790836691856384 }, { "epoch": 4.625974025974026, "step": 1781, "train/loss_ctc": 0.5109178423881531, "train/loss_error": 0.2632324993610382, "train/loss_total": 0.3127695620059967 }, { "epoch": 4.628571428571428, "step": 1782, "train/loss_ctc": 0.3371947407722473, "train/loss_error": 0.3769443929195404, "train/loss_total": 0.36899447441101074 }, { "epoch": 4.631168831168831, "step": 1783, "train/loss_ctc": 0.3866609036922455, "train/loss_error": 0.26576778292655945, "train/loss_total": 0.28994640707969666 }, { "epoch": 4.6337662337662335, "step": 1784, "train/loss_ctc": 0.6346707344055176, "train/loss_error": 0.3868533670902252, "train/loss_total": 0.4364168643951416 }, { "epoch": 4.636363636363637, "step": 1785, "train/loss_ctc": 0.6873793005943298, "train/loss_error": 0.4181675612926483, "train/loss_total": 0.47200989723205566 }, { "epoch": 4.638961038961039, "step": 1786, "train/loss_ctc": 0.3712672293186188, "train/loss_error": 0.317092627286911, "train/loss_total": 0.3279275596141815 }, { "epoch": 4.641558441558441, "step": 1787, "train/loss_ctc": 0.4143245220184326, "train/loss_error": 0.35408711433410645, "train/loss_total": 0.3661346137523651 }, { "epoch": 4.644155844155844, "step": 1788, "train/loss_ctc": 0.6292366981506348, "train/loss_error": 0.43357405066490173, "train/loss_total": 0.4727065861225128 }, { "epoch": 4.646753246753247, "step": 1789, "train/loss_ctc": 0.38727420568466187, "train/loss_error": 0.3136875629425049, "train/loss_total": 0.32840490341186523 }, { "epoch": 4.64935064935065, "grad_norm": 12492.59375, "learning_rate": 2.1506493506493506e-06, "loss": 0.3703, "step": 1790 }, { "epoch": 4.64935064935065, "step": 1790, "train/loss_ctc": 0.5636283159255981, "train/loss_error": 0.34562352299690247, "train/loss_total": 0.38922449946403503 }, { "epoch": 4.651948051948052, "step": 1791, "train/loss_ctc": 0.970160186290741, "train/loss_error": 0.35068440437316895, "train/loss_total": 0.4745795726776123 }, { "epoch": 4.654545454545454, "step": 1792, "train/loss_ctc": 0.36136531829833984, "train/loss_error": 0.37409356236457825, "train/loss_total": 0.3715479373931885 }, { "epoch": 4.6571428571428575, "step": 1793, "train/loss_ctc": 0.3328041434288025, "train/loss_error": 0.2981174886226654, "train/loss_total": 0.30505481362342834 }, { "epoch": 4.65974025974026, "step": 1794, "train/loss_ctc": 0.3704034090042114, "train/loss_error": 0.3230755627155304, "train/loss_total": 0.3325411379337311 }, { "epoch": 4.662337662337662, "step": 1795, "train/loss_ctc": 0.38251861929893494, "train/loss_error": 0.36316153407096863, "train/loss_total": 0.3670329451560974 }, { "epoch": 4.664935064935065, "step": 1796, "train/loss_ctc": 0.4644274413585663, "train/loss_error": 0.28933340311050415, "train/loss_total": 0.3243522047996521 }, { "epoch": 4.667532467532467, "step": 1797, "train/loss_ctc": 0.5524426698684692, "train/loss_error": 0.3873124420642853, "train/loss_total": 0.4203384816646576 }, { "epoch": 4.67012987012987, "step": 1798, "train/loss_ctc": 0.4549449682235718, "train/loss_error": 0.37716588377952576, "train/loss_total": 0.3927217125892639 }, { "epoch": 4.672727272727273, "step": 1799, "train/loss_ctc": 0.3782256245613098, "train/loss_error": 0.3102821409702301, "train/loss_total": 0.32387083768844604 }, { "epoch": 4.675324675324675, "grad_norm": 13225.0048828125, "learning_rate": 1.994805194805195e-06, "loss": 0.3701, "step": 1800 }, { "epoch": 4.675324675324675, "step": 1800, "train/loss_ctc": 0.47959810495376587, "train/loss_error": 0.35687175393104553, "train/loss_total": 0.38141703605651855 }, { "epoch": 4.677922077922078, "step": 1801, "train/loss_ctc": 0.3131231665611267, "train/loss_error": 0.25351259112358093, "train/loss_total": 0.26543471217155457 }, { "epoch": 4.6805194805194805, "step": 1802, "train/loss_ctc": 0.4576428532600403, "train/loss_error": 0.35153728723526, "train/loss_total": 0.3727583885192871 }, { "epoch": 4.683116883116883, "step": 1803, "train/loss_ctc": 0.6561508178710938, "train/loss_error": 0.44983744621276855, "train/loss_total": 0.49110013246536255 }, { "epoch": 4.685714285714286, "step": 1804, "train/loss_ctc": 0.35031309723854065, "train/loss_error": 0.3000495731830597, "train/loss_total": 0.31010228395462036 }, { "epoch": 4.688311688311688, "step": 1805, "train/loss_ctc": 0.4390621483325958, "train/loss_error": 0.3608601987361908, "train/loss_total": 0.37650057673454285 }, { "epoch": 4.690909090909091, "step": 1806, "train/loss_ctc": 0.27258604764938354, "train/loss_error": 0.23176135122776031, "train/loss_total": 0.2399262934923172 }, { "epoch": 4.693506493506494, "step": 1807, "train/loss_ctc": 0.3712272047996521, "train/loss_error": 0.31319350004196167, "train/loss_total": 0.3248002529144287 }, { "epoch": 4.696103896103896, "step": 1808, "train/loss_ctc": 0.6460719704627991, "train/loss_error": 0.35946333408355713, "train/loss_total": 0.4167850613594055 }, { "epoch": 4.698701298701299, "step": 1809, "train/loss_ctc": 0.4688573181629181, "train/loss_error": 0.3562087416648865, "train/loss_total": 0.3787384629249573 }, { "epoch": 4.701298701298701, "grad_norm": 13082.6435546875, "learning_rate": 1.838961038961039e-06, "loss": 0.3558, "step": 1810 }, { "epoch": 4.701298701298701, "step": 1810, "train/loss_ctc": 0.6960006952285767, "train/loss_error": 0.41322067379951477, "train/loss_total": 0.4697766900062561 }, { "epoch": 4.703896103896104, "step": 1811, "train/loss_ctc": 0.4545247554779053, "train/loss_error": 0.325456827878952, "train/loss_total": 0.3512704074382782 }, { "epoch": 4.706493506493507, "step": 1812, "train/loss_ctc": 0.39432039856910706, "train/loss_error": 0.2943155765533447, "train/loss_total": 0.3143165409564972 }, { "epoch": 4.709090909090909, "step": 1813, "train/loss_ctc": 0.8917479515075684, "train/loss_error": 0.3700442612171173, "train/loss_total": 0.47438502311706543 }, { "epoch": 4.711688311688311, "step": 1814, "train/loss_ctc": 0.3949338495731354, "train/loss_error": 0.3457533121109009, "train/loss_total": 0.3555894196033478 }, { "epoch": 4.714285714285714, "step": 1815, "train/loss_ctc": 0.5646682381629944, "train/loss_error": 0.42779237031936646, "train/loss_total": 0.4551675319671631 }, { "epoch": 4.716883116883117, "step": 1816, "train/loss_ctc": 0.4838142395019531, "train/loss_error": 0.4133470356464386, "train/loss_total": 0.42744046449661255 }, { "epoch": 4.71948051948052, "step": 1817, "train/loss_ctc": 0.33877140283584595, "train/loss_error": 0.3361877202987671, "train/loss_total": 0.3367044925689697 }, { "epoch": 4.722077922077922, "step": 1818, "train/loss_ctc": 0.2493864893913269, "train/loss_error": 0.2873873710632324, "train/loss_total": 0.27978718280792236 }, { "epoch": 4.724675324675324, "step": 1819, "train/loss_ctc": 0.4646071195602417, "train/loss_error": 0.3669871687889099, "train/loss_total": 0.3865111768245697 }, { "epoch": 4.7272727272727275, "grad_norm": 12543.01953125, "learning_rate": 1.6831168831168833e-06, "loss": 0.3851, "step": 1820 }, { "epoch": 4.7272727272727275, "step": 1820, "train/loss_ctc": 0.41936570405960083, "train/loss_error": 0.28285282850265503, "train/loss_total": 0.3101554214954376 }, { "epoch": 4.72987012987013, "step": 1821, "train/loss_ctc": 0.4886781573295593, "train/loss_error": 0.3913860023021698, "train/loss_total": 0.41084444522857666 }, { "epoch": 4.732467532467533, "step": 1822, "train/loss_ctc": 0.5288512110710144, "train/loss_error": 0.4440126121044159, "train/loss_total": 0.4609803557395935 }, { "epoch": 4.735064935064935, "step": 1823, "train/loss_ctc": 0.7336521148681641, "train/loss_error": 0.35244935750961304, "train/loss_total": 0.4286899268627167 }, { "epoch": 4.7376623376623375, "step": 1824, "train/loss_ctc": 0.5283362865447998, "train/loss_error": 0.37355688214302063, "train/loss_total": 0.40451276302337646 }, { "epoch": 4.740259740259741, "step": 1825, "train/loss_ctc": 0.5098491907119751, "train/loss_error": 0.3497977554798126, "train/loss_total": 0.3818080425262451 }, { "epoch": 4.742857142857143, "step": 1826, "train/loss_ctc": 0.45632317662239075, "train/loss_error": 0.25675997138023376, "train/loss_total": 0.29667261242866516 }, { "epoch": 4.745454545454545, "step": 1827, "train/loss_ctc": 0.3179604411125183, "train/loss_error": 0.3292199373245239, "train/loss_total": 0.32696807384490967 }, { "epoch": 4.748051948051948, "step": 1828, "train/loss_ctc": 0.49485719203948975, "train/loss_error": 0.27222689986228943, "train/loss_total": 0.31675297021865845 }, { "epoch": 4.750649350649351, "step": 1829, "train/loss_ctc": 0.355282723903656, "train/loss_error": 0.3676517605781555, "train/loss_total": 0.3651779592037201 }, { "epoch": 4.753246753246753, "grad_norm": 14959.96875, "learning_rate": 1.5272727272727273e-06, "loss": 0.3703, "step": 1830 }, { "epoch": 4.753246753246753, "step": 1830, "train/loss_ctc": 0.4238358736038208, "train/loss_error": 0.34040161967277527, "train/loss_total": 0.35708850622177124 }, { "epoch": 4.755844155844156, "step": 1831, "train/loss_ctc": 0.601475715637207, "train/loss_error": 0.37714967131614685, "train/loss_total": 0.42201489210128784 }, { "epoch": 4.758441558441558, "step": 1832, "train/loss_ctc": 0.3603931963443756, "train/loss_error": 0.311896413564682, "train/loss_total": 0.3215957581996918 }, { "epoch": 4.761038961038961, "step": 1833, "train/loss_ctc": 0.5201610326766968, "train/loss_error": 0.32547527551651, "train/loss_total": 0.36441245675086975 }, { "epoch": 4.763636363636364, "step": 1834, "train/loss_ctc": 0.2968393862247467, "train/loss_error": 0.3488273620605469, "train/loss_total": 0.3384297490119934 }, { "epoch": 4.766233766233766, "step": 1835, "train/loss_ctc": 0.3980433940887451, "train/loss_error": 0.3636539578437805, "train/loss_total": 0.3705318570137024 }, { "epoch": 4.768831168831169, "step": 1836, "train/loss_ctc": 0.5545719861984253, "train/loss_error": 0.331576943397522, "train/loss_total": 0.37617596983909607 }, { "epoch": 4.771428571428571, "step": 1837, "train/loss_ctc": 0.5008302927017212, "train/loss_error": 0.322034627199173, "train/loss_total": 0.35779374837875366 }, { "epoch": 4.7740259740259745, "step": 1838, "train/loss_ctc": 0.3416571021080017, "train/loss_error": 0.3109133243560791, "train/loss_total": 0.3170620799064636 }, { "epoch": 4.776623376623377, "step": 1839, "train/loss_ctc": 0.3800255060195923, "train/loss_error": 0.28502848744392395, "train/loss_total": 0.3040279150009155 }, { "epoch": 4.779220779220779, "grad_norm": 10728.0517578125, "learning_rate": 1.3714285714285715e-06, "loss": 0.3529, "step": 1840 }, { "epoch": 4.779220779220779, "step": 1840, "train/loss_ctc": 0.5966446399688721, "train/loss_error": 0.47774359583854675, "train/loss_total": 0.5015237927436829 }, { "epoch": 4.781818181818182, "step": 1841, "train/loss_ctc": 0.43040281534194946, "train/loss_error": 0.37627556920051575, "train/loss_total": 0.38710105419158936 }, { "epoch": 4.7844155844155845, "step": 1842, "train/loss_ctc": 0.7101888060569763, "train/loss_error": 0.45192280411720276, "train/loss_total": 0.5035760402679443 }, { "epoch": 4.787012987012987, "step": 1843, "train/loss_ctc": 0.3689204752445221, "train/loss_error": 0.3614874482154846, "train/loss_total": 0.36297404766082764 }, { "epoch": 4.78961038961039, "step": 1844, "train/loss_ctc": 0.4710047245025635, "train/loss_error": 0.2906482219696045, "train/loss_total": 0.3267195224761963 }, { "epoch": 4.792207792207792, "step": 1845, "train/loss_ctc": 0.36959952116012573, "train/loss_error": 0.34711378812789917, "train/loss_total": 0.3516109585762024 }, { "epoch": 4.794805194805194, "step": 1846, "train/loss_ctc": 0.316233366727829, "train/loss_error": 0.3169024586677551, "train/loss_total": 0.3167686462402344 }, { "epoch": 4.797402597402598, "step": 1847, "train/loss_ctc": 0.37338942289352417, "train/loss_error": 0.3446548879146576, "train/loss_total": 0.35040178894996643 }, { "epoch": 4.8, "step": 1848, "train/loss_ctc": 0.4559589624404907, "train/loss_error": 0.35508421063423157, "train/loss_total": 0.3752591609954834 }, { "epoch": 4.802597402597403, "step": 1849, "train/loss_ctc": 0.3255149722099304, "train/loss_error": 0.2946803867816925, "train/loss_total": 0.30084729194641113 }, { "epoch": 4.805194805194805, "grad_norm": 9720.9326171875, "learning_rate": 1.2155844155844156e-06, "loss": 0.3777, "step": 1850 }, { "epoch": 4.805194805194805, "step": 1850, "train/loss_ctc": 0.46330443024635315, "train/loss_error": 0.30658769607543945, "train/loss_total": 0.3379310369491577 }, { "epoch": 4.8077922077922075, "step": 1851, "train/loss_ctc": 0.3450726270675659, "train/loss_error": 0.3236660957336426, "train/loss_total": 0.3279474079608917 }, { "epoch": 4.810389610389611, "step": 1852, "train/loss_ctc": 0.470742404460907, "train/loss_error": 0.2909294366836548, "train/loss_total": 0.32689201831817627 }, { "epoch": 4.812987012987013, "step": 1853, "train/loss_ctc": 0.5352873802185059, "train/loss_error": 0.3674354553222656, "train/loss_total": 0.4010058641433716 }, { "epoch": 4.815584415584416, "step": 1854, "train/loss_ctc": 0.4371885657310486, "train/loss_error": 0.31205418705940247, "train/loss_total": 0.33708107471466064 }, { "epoch": 4.818181818181818, "step": 1855, "train/loss_ctc": 0.5538568496704102, "train/loss_error": 0.34018468856811523, "train/loss_total": 0.3829191327095032 }, { "epoch": 4.820779220779221, "step": 1856, "train/loss_ctc": 0.3701854348182678, "train/loss_error": 0.24751055240631104, "train/loss_total": 0.2720455527305603 }, { "epoch": 4.823376623376624, "step": 1857, "train/loss_ctc": 0.4201434850692749, "train/loss_error": 0.33263546228408813, "train/loss_total": 0.35013705492019653 }, { "epoch": 4.825974025974026, "step": 1858, "train/loss_ctc": 0.43791699409484863, "train/loss_error": 0.32099032402038574, "train/loss_total": 0.3443756699562073 }, { "epoch": 4.828571428571428, "step": 1859, "train/loss_ctc": 0.2582935392856598, "train/loss_error": 0.28357240557670593, "train/loss_total": 0.27851665019989014 }, { "epoch": 4.8311688311688314, "grad_norm": 7614.73388671875, "learning_rate": 1.0597402597402598e-06, "loss": 0.3359, "step": 1860 }, { "epoch": 4.8311688311688314, "step": 1860, "train/loss_ctc": 0.4183503985404968, "train/loss_error": 0.2742226719856262, "train/loss_total": 0.3030482232570648 }, { "epoch": 4.833766233766234, "step": 1861, "train/loss_ctc": 0.38347747921943665, "train/loss_error": 0.3771800994873047, "train/loss_total": 0.3784395754337311 }, { "epoch": 4.836363636363636, "step": 1862, "train/loss_ctc": 0.34480684995651245, "train/loss_error": 0.3416669964790344, "train/loss_total": 0.34229499101638794 }, { "epoch": 4.838961038961039, "step": 1863, "train/loss_ctc": 0.38262224197387695, "train/loss_error": 0.37277647852897644, "train/loss_total": 0.3747456669807434 }, { "epoch": 4.841558441558441, "step": 1864, "train/loss_ctc": 0.5593939423561096, "train/loss_error": 0.35967469215393066, "train/loss_total": 0.399618536233902 }, { "epoch": 4.8441558441558445, "step": 1865, "train/loss_ctc": 0.38433724641799927, "train/loss_error": 0.2791198194026947, "train/loss_total": 0.30016329884529114 }, { "epoch": 4.846753246753247, "step": 1866, "train/loss_ctc": 0.4356418550014496, "train/loss_error": 0.4203330874443054, "train/loss_total": 0.4233948588371277 }, { "epoch": 4.849350649350649, "step": 1867, "train/loss_ctc": 1.0716524124145508, "train/loss_error": 0.44017741084098816, "train/loss_total": 0.5664724111557007 }, { "epoch": 4.851948051948052, "step": 1868, "train/loss_ctc": 0.37710875272750854, "train/loss_error": 0.2862227261066437, "train/loss_total": 0.30439993739128113 }, { "epoch": 4.8545454545454545, "step": 1869, "train/loss_ctc": 0.5114738941192627, "train/loss_error": 0.34174850583076477, "train/loss_total": 0.37569358944892883 }, { "epoch": 4.857142857142857, "grad_norm": 14555.029296875, "learning_rate": 9.03896103896104e-07, "loss": 0.3768, "step": 1870 }, { "epoch": 4.857142857142857, "step": 1870, "train/loss_ctc": 0.38248175382614136, "train/loss_error": 0.3664298355579376, "train/loss_total": 0.3696402311325073 }, { "epoch": 4.85974025974026, "step": 1871, "train/loss_ctc": 0.48088836669921875, "train/loss_error": 0.31895557045936584, "train/loss_total": 0.3513421416282654 }, { "epoch": 4.862337662337662, "step": 1872, "train/loss_ctc": 0.4754698872566223, "train/loss_error": 0.29211118817329407, "train/loss_total": 0.32878291606903076 }, { "epoch": 4.8649350649350644, "step": 1873, "train/loss_ctc": 0.5782651901245117, "train/loss_error": 0.289130836725235, "train/loss_total": 0.3469577133655548 }, { "epoch": 4.867532467532468, "step": 1874, "train/loss_ctc": 0.48591020703315735, "train/loss_error": 0.3343304693698883, "train/loss_total": 0.36464640498161316 }, { "epoch": 4.87012987012987, "step": 1875, "train/loss_ctc": 0.43416982889175415, "train/loss_error": 0.3256148397922516, "train/loss_total": 0.34732586145401 }, { "epoch": 4.872727272727273, "step": 1876, "train/loss_ctc": 0.5808994174003601, "train/loss_error": 0.39004024863243103, "train/loss_total": 0.42821207642555237 }, { "epoch": 4.875324675324675, "step": 1877, "train/loss_ctc": 0.45052024722099304, "train/loss_error": 0.2762475609779358, "train/loss_total": 0.31110209226608276 }, { "epoch": 4.8779220779220775, "step": 1878, "train/loss_ctc": 0.4011489748954773, "train/loss_error": 0.4125778079032898, "train/loss_total": 0.41029202938079834 }, { "epoch": 4.880519480519481, "step": 1879, "train/loss_ctc": 0.6957187652587891, "train/loss_error": 0.2913946211338043, "train/loss_total": 0.3722594380378723 }, { "epoch": 4.883116883116883, "grad_norm": 18816.955078125, "learning_rate": 7.480519480519481e-07, "loss": 0.3631, "step": 1880 }, { "epoch": 4.883116883116883, "step": 1880, "train/loss_ctc": 0.5458420515060425, "train/loss_error": 0.4077455699443817, "train/loss_total": 0.43536487221717834 }, { "epoch": 4.885714285714286, "step": 1881, "train/loss_ctc": 0.7939369678497314, "train/loss_error": 0.49305984377861023, "train/loss_total": 0.5532352924346924 }, { "epoch": 4.888311688311688, "step": 1882, "train/loss_ctc": 0.6827054619789124, "train/loss_error": 0.3138599097728729, "train/loss_total": 0.38762903213500977 }, { "epoch": 4.890909090909091, "step": 1883, "train/loss_ctc": 0.7771693468093872, "train/loss_error": 0.4301982820034027, "train/loss_total": 0.49959248304367065 }, { "epoch": 4.893506493506494, "step": 1884, "train/loss_ctc": 0.5469666123390198, "train/loss_error": 0.318338543176651, "train/loss_total": 0.36406415700912476 }, { "epoch": 4.896103896103896, "step": 1885, "train/loss_ctc": 0.3599787950515747, "train/loss_error": 0.3030228614807129, "train/loss_total": 0.31441405415534973 }, { "epoch": 4.898701298701298, "step": 1886, "train/loss_ctc": 0.3956599533557892, "train/loss_error": 0.34428948163986206, "train/loss_total": 0.3545635938644409 }, { "epoch": 4.9012987012987015, "step": 1887, "train/loss_ctc": 0.5000158548355103, "train/loss_error": 0.36408334970474243, "train/loss_total": 0.39126986265182495 }, { "epoch": 4.903896103896104, "step": 1888, "train/loss_ctc": 0.36489999294281006, "train/loss_error": 0.283273309469223, "train/loss_total": 0.2995986342430115 }, { "epoch": 4.906493506493506, "step": 1889, "train/loss_ctc": 0.3194197714328766, "train/loss_error": 0.3814532160758972, "train/loss_total": 0.36904653906822205 }, { "epoch": 4.909090909090909, "grad_norm": 9077.45703125, "learning_rate": 5.922077922077922e-07, "loss": 0.3969, "step": 1890 }, { "epoch": 4.909090909090909, "step": 1890, "train/loss_ctc": 0.3140220642089844, "train/loss_error": 0.3173488676548004, "train/loss_total": 0.3166835308074951 }, { "epoch": 4.911688311688311, "step": 1891, "train/loss_ctc": 0.3103610873222351, "train/loss_error": 0.3652729094028473, "train/loss_total": 0.35429054498672485 }, { "epoch": 4.914285714285715, "step": 1892, "train/loss_ctc": 0.45631951093673706, "train/loss_error": 0.3312831223011017, "train/loss_total": 0.35629040002822876 }, { "epoch": 4.916883116883117, "step": 1893, "train/loss_ctc": 0.5245206356048584, "train/loss_error": 0.4289000630378723, "train/loss_total": 0.4480242133140564 }, { "epoch": 4.919480519480519, "step": 1894, "train/loss_ctc": 0.4294789433479309, "train/loss_error": 0.38720330595970154, "train/loss_total": 0.3956584334373474 }, { "epoch": 4.922077922077922, "step": 1895, "train/loss_ctc": 0.47275733947753906, "train/loss_error": 0.3921484351158142, "train/loss_total": 0.4082702398300171 }, { "epoch": 4.9246753246753245, "step": 1896, "train/loss_ctc": 0.3770599663257599, "train/loss_error": 0.3100910484790802, "train/loss_total": 0.3234848380088806 }, { "epoch": 4.927272727272728, "step": 1897, "train/loss_ctc": 0.3133673071861267, "train/loss_error": 0.29202526807785034, "train/loss_total": 0.2962936758995056 }, { "epoch": 4.92987012987013, "step": 1898, "train/loss_ctc": 0.2902228534221649, "train/loss_error": 0.2504384517669678, "train/loss_total": 0.25839534401893616 }, { "epoch": 4.932467532467532, "step": 1899, "train/loss_ctc": 0.415492445230484, "train/loss_error": 0.2923637330532074, "train/loss_total": 0.3169894814491272 }, { "epoch": 4.935064935064935, "grad_norm": 12105.5166015625, "learning_rate": 4.3636363636363636e-07, "loss": 0.3474, "step": 1900 }, { "epoch": 4.935064935064935, "step": 1900, "train/loss_ctc": 0.3321300745010376, "train/loss_error": 0.3065105676651001, "train/loss_total": 0.31163448095321655 }, { "epoch": 4.937662337662338, "step": 1901, "train/loss_ctc": 0.8193004727363586, "train/loss_error": 0.36720648407936096, "train/loss_total": 0.45762526988983154 }, { "epoch": 4.94025974025974, "step": 1902, "train/loss_ctc": 0.5020947456359863, "train/loss_error": 0.4121650159358978, "train/loss_total": 0.43015098571777344 }, { "epoch": 4.942857142857143, "step": 1903, "train/loss_ctc": 0.4351004660129547, "train/loss_error": 0.35012638568878174, "train/loss_total": 0.36712121963500977 }, { "epoch": 4.945454545454545, "step": 1904, "train/loss_ctc": 0.5637869834899902, "train/loss_error": 0.3682195842266083, "train/loss_total": 0.4073330760002136 }, { "epoch": 4.948051948051948, "step": 1905, "train/loss_ctc": 0.6072202920913696, "train/loss_error": 0.3800641894340515, "train/loss_total": 0.425495445728302 }, { "epoch": 4.950649350649351, "step": 1906, "train/loss_ctc": 0.3188381493091583, "train/loss_error": 0.27601268887519836, "train/loss_total": 0.28457778692245483 }, { "epoch": 4.953246753246753, "step": 1907, "train/loss_ctc": 0.33819007873535156, "train/loss_error": 0.37995773553848267, "train/loss_total": 0.37160420417785645 }, { "epoch": 4.955844155844156, "step": 1908, "train/loss_ctc": 0.42900699377059937, "train/loss_error": 0.32494598627090454, "train/loss_total": 0.34575819969177246 }, { "epoch": 4.958441558441558, "step": 1909, "train/loss_ctc": 0.46481841802597046, "train/loss_error": 0.3727739751338959, "train/loss_total": 0.39118286967277527 }, { "epoch": 4.961038961038961, "grad_norm": 12544.232421875, "learning_rate": 2.8051948051948053e-07, "loss": 0.3792, "step": 1910 }, { "epoch": 4.961038961038961, "step": 1910, "train/loss_ctc": 0.4348629117012024, "train/loss_error": 0.3794926702976227, "train/loss_total": 0.39056673645973206 }, { "epoch": 4.963636363636364, "step": 1911, "train/loss_ctc": 0.5774006843566895, "train/loss_error": 0.3893633484840393, "train/loss_total": 0.42697083950042725 }, { "epoch": 4.966233766233766, "step": 1912, "train/loss_ctc": 0.4103315472602844, "train/loss_error": 0.3724864721298218, "train/loss_total": 0.3800554871559143 }, { "epoch": 4.968831168831169, "step": 1913, "train/loss_ctc": 1.0385472774505615, "train/loss_error": 0.40944886207580566, "train/loss_total": 0.5352685451507568 }, { "epoch": 4.9714285714285715, "step": 1914, "train/loss_ctc": 0.4049646258354187, "train/loss_error": 0.2847217321395874, "train/loss_total": 0.3087703287601471 }, { "epoch": 4.974025974025974, "step": 1915, "train/loss_ctc": 0.7677231431007385, "train/loss_error": 0.4134467840194702, "train/loss_total": 0.4843020737171173 }, { "epoch": 4.976623376623377, "step": 1916, "train/loss_ctc": 0.4006579518318176, "train/loss_error": 0.29329171776771545, "train/loss_total": 0.31476497650146484 }, { "epoch": 4.979220779220779, "step": 1917, "train/loss_ctc": 0.7014670372009277, "train/loss_error": 0.40446484088897705, "train/loss_total": 0.4638652801513672 }, { "epoch": 4.9818181818181815, "step": 1918, "train/loss_ctc": 0.3717190623283386, "train/loss_error": 0.3761925995349884, "train/loss_total": 0.3752979040145874 }, { "epoch": 4.984415584415585, "step": 1919, "train/loss_ctc": 0.5896443128585815, "train/loss_error": 0.3431680500507355, "train/loss_total": 0.3924632966518402 }, { "epoch": 4.987012987012987, "grad_norm": 15636.3916015625, "learning_rate": 1.2467532467532469e-07, "loss": 0.4072, "step": 1920 }, { "epoch": 4.987012987012987, "step": 1920, "train/loss_ctc": 0.3978821039199829, "train/loss_error": 0.32593125104904175, "train/loss_total": 0.34032145142555237 }, { "epoch": 4.989610389610389, "step": 1921, "train/loss_ctc": 0.46665728092193604, "train/loss_error": 0.3350077271461487, "train/loss_total": 0.3613376319408417 }, { "epoch": 4.992207792207792, "step": 1922, "train/loss_ctc": 0.3633562922477722, "train/loss_error": 0.2835646867752075, "train/loss_total": 0.2995229959487915 }, { "epoch": 4.994805194805195, "step": 1923, "train/loss_ctc": 0.476561963558197, "train/loss_error": 0.30369073152542114, "train/loss_total": 0.33826497197151184 }, { "epoch": 4.997402597402598, "step": 1924, "train/loss_ctc": 0.3097357749938965, "train/loss_error": 0.2602827250957489, "train/loss_total": 0.2701733410358429 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.6246699690818787, "val/loss_error": 0.15397024154663086, "val/loss_total": 0.24811019003391266 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.607525646686554, "val/loss_error": 0.47989800572395325, "val/loss_total": 0.5054235458374023 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.07125519961118698, "val/loss_error": 0.24175792932510376, "val/loss_total": 0.20765738189220428 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.24455730617046356, "val/loss_error": 0.10023341327905655, "val/loss_total": 0.12909819185733795 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.12588806450366974, "val/loss_error": 0.16558758914470673, "val/loss_total": 0.15764768421649933 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.11731574684381485, "val/loss_error": 0.2914775013923645, "val/loss_total": 0.256645143032074 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.1777358502149582, "val/loss_error": 0.49354052543640137, "val/loss_total": 0.43037959933280945 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.38250893354415894, "val/loss_error": 0.24908500909805298, "val/loss_total": 0.27576979994773865 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.3566465675830841, "val/loss_error": 0.30396732687950134, "val/loss_total": 0.31450319290161133 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.31596797704696655, "val/loss_error": 0.3075321912765503, "val/loss_total": 0.3092193603515625 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.10968615114688873, "val/loss_error": 0.12754297256469727, "val/loss_total": 0.1239716112613678 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.18461894989013672, "val/loss_error": 0.2666967213153839, "val/loss_total": 0.2502811551094055 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.7553161382675171, "val/loss_error": 0.35318684577941895, "val/loss_total": 0.4336127042770386 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.7990849018096924, "val/loss_error": 0.5796869397163391, "val/loss_total": 0.6235665082931519 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5408079028129578, "val/loss_error": 0.3084508180618286, "val/loss_total": 0.35492223501205444 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5449475049972534, "val/loss_error": 0.5876696705818176, "val/loss_total": 0.5791252255439758 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.07044278085231781, "val/loss_error": 0.2334357500076294, "val/loss_total": 0.2008371651172638 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.15581205487251282, "val/loss_error": 0.2229645997285843, "val/loss_total": 0.20953409373760223 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.10505729913711548, "val/loss_error": 0.25985509157180786, "val/loss_total": 0.22889553010463715 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.6028918623924255, "val/loss_error": 0.3683081269264221, "val/loss_total": 0.4152248799800873 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.8084850907325745, "val/loss_error": 0.42098307609558105, "val/loss_total": 0.49848347902297974 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.2494509369134903, "val/loss_error": 0.39444443583488464, "val/loss_total": 0.36544573307037354 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.29084324836730957, "val/loss_error": 0.4021511673927307, "val/loss_total": 0.379889577627182 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.20026829838752747, "val/loss_error": 0.17662009596824646, "val/loss_total": 0.1813497394323349 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.3147900104522705, "val/loss_error": 0.5160038471221924, "val/loss_total": 0.4757610857486725 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.44883406162261963, "val/loss_error": 0.5703834295272827, "val/loss_total": 0.5460735559463501 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.31248924136161804, "val/loss_error": 0.3362652659416199, "val/loss_total": 0.331510066986084 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.12958107888698578, "val/loss_error": 0.20100289583206177, "val/loss_total": 0.18671853840351105 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.022003373131155968, "val/loss_error": 0.06520331650972366, "val/loss_total": 0.056563328951597214 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.25627633929252625, "val/loss_error": 0.5287688970565796, "val/loss_total": 0.47427037358283997 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.10070011764764786, "val/loss_error": 0.11953214555978775, "val/loss_total": 0.11576574295759201 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.11117617785930634, "val/loss_error": 0.23909318447113037, "val/loss_total": 0.21350978314876556 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.08980347961187363, "val/loss_error": 0.14025340974330902, "val/loss_total": 0.13016341626644135 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.6035685539245605, "val/loss_error": 0.6489872336387634, "val/loss_total": 0.6399034857749939 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.08684325218200684, "val/loss_error": 0.2829248309135437, "val/loss_total": 0.2437085062265396 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.1746019870042801, "val/loss_error": 0.3701065480709076, "val/loss_total": 0.33100563287734985 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.47330304980278015, "val/loss_error": 0.5752632021903992, "val/loss_total": 0.5548712015151978 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.37739622592926025, "val/loss_error": 0.37572625279426575, "val/loss_total": 0.37606024742126465 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.22429095208644867, "val/loss_error": 0.49482566118240356, "val/loss_total": 0.44071871042251587 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5595707297325134, "val/loss_error": 0.6221680045127869, "val/loss_total": 0.6096485257148743 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.2374076396226883, "val/loss_error": 0.22656159102916718, "val/loss_total": 0.22873079776763916 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.28289762139320374, "val/loss_error": 0.39486968517303467, "val/loss_total": 0.3724752962589264 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.6165946125984192, "val/loss_error": 0.9331052303314209, "val/loss_total": 0.8698031306266785 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5059840679168701, "val/loss_error": 0.616759181022644, "val/loss_total": 0.5946041345596313 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.49061229825019836, "val/loss_error": 0.3190176784992218, "val/loss_total": 0.3533366322517395 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.39841872453689575, "val/loss_error": 0.6992455720901489, "val/loss_total": 0.6390801668167114 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.6413567662239075, "val/loss_error": 0.389420747756958, "val/loss_total": 0.4398079514503479 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.3157205879688263, "val/loss_error": 0.49929603934288025, "val/loss_total": 0.46258094906806946 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.36816784739494324, "val/loss_error": 0.2405136525630951, "val/loss_total": 0.2660444974899292 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.6190939545631409, "val/loss_error": 0.8094415068626404, "val/loss_total": 0.7713720202445984 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.3211616575717926, "val/loss_error": 0.7925539612770081, "val/loss_total": 0.6982755064964294 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.3093317449092865, "val/loss_error": 0.4245705008506775, "val/loss_total": 0.40152275562286377 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.3481307029724121, "val/loss_error": 0.3831369876861572, "val/loss_total": 0.3761357367038727 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5585353970527649, "val/loss_error": 0.8331480026245117, "val/loss_total": 0.7782254815101624 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.36772119998931885, "val/loss_error": 0.36340615153312683, "val/loss_total": 0.3642691671848297 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.1718088686466217, "val/loss_error": 0.43088147044181824, "val/loss_total": 0.37906697392463684 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.14035199582576752, "val/loss_error": 0.2618764042854309, "val/loss_total": 0.23757153749465942 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.4814637303352356, "val/loss_error": 0.3269810974597931, "val/loss_total": 0.35787761211395264 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5914303660392761, "val/loss_error": 0.5789550542831421, "val/loss_total": 0.5814501047134399 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.06065122038125992, "val/loss_error": 0.12235617637634277, "val/loss_total": 0.11001519113779068 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.8275797963142395, "val/loss_error": 0.5802265405654907, "val/loss_total": 0.6296972036361694 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.6482733488082886, "val/loss_error": 0.358978271484375, "val/loss_total": 0.41683730483055115 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.05473272502422333, "val/loss_error": 0.06687375158071518, "val/loss_total": 0.06444554775953293 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.19380243122577667, "val/loss_error": 0.25727716088294983, "val/loss_total": 0.24458222091197968 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.1336812824010849, "val/loss_error": 0.2176511436700821, "val/loss_total": 0.20085717737674713 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.207158163189888, "val/loss_error": 0.36678585410118103, "val/loss_total": 0.33486032485961914 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.27584996819496155, "val/loss_error": 0.25906720757484436, "val/loss_total": 0.2624237537384033 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5595126152038574, "val/loss_error": 0.5580713748931885, "val/loss_total": 0.5583596229553223 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.4218448996543884, "val/loss_error": 0.539309561252594, "val/loss_total": 0.5158166289329529 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.4675092101097107, "val/loss_error": 0.4811660051345825, "val/loss_total": 0.47843465209007263 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.2053380310535431, "val/loss_error": 0.23864498734474182, "val/loss_total": 0.23198360204696655 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5808184146881104, "val/loss_error": 0.5502936244010925, "val/loss_total": 0.5563985705375671 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5218967795372009, "val/loss_error": 0.6645674705505371, "val/loss_total": 0.6360333561897278 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.201784148812294, "val/loss_error": 0.35592785477638245, "val/loss_total": 0.3250991106033325 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.2276170700788498, "val/loss_error": 0.2693210244178772, "val/loss_total": 0.2609802484512329 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.03736936300992966, "val/loss_error": 0.22808407247066498, "val/loss_total": 0.1899411380290985 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.7240115404129028, "val/loss_error": 0.5327463150024414, "val/loss_total": 0.5709993839263916 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.6928349733352661, "val/loss_error": 0.3579254746437073, "val/loss_total": 0.424907386302948 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.6073930263519287, "val/loss_error": 0.3472973108291626, "val/loss_total": 0.3993164598941803 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.00463070860132575, "val/loss_error": 0.2237877994775772, "val/loss_total": 0.1799563765525818 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.49578362703323364, "val/loss_error": 0.27387312054634094, "val/loss_total": 0.318255215883255 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.03500795736908913, "val/loss_error": 0.0888034999370575, "val/loss_total": 0.07804439216852188 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.1805315613746643, "val/loss_error": 0.2515946626663208, "val/loss_total": 0.23738205432891846 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.41977444291114807, "val/loss_error": 0.40446117520332336, "val/loss_total": 0.40752384066581726 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.3650522530078888, "val/loss_error": 0.6925578713417053, "val/loss_total": 0.6270567774772644 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.41919535398483276, "val/loss_error": 0.5377658605575562, "val/loss_total": 0.5140517950057983 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.06399211287498474, "val/loss_error": 0.1026756539940834, "val/loss_total": 0.09493894875049591 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.3961379826068878, "val/loss_error": 0.41785141825675964, "val/loss_total": 0.41350874304771423 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.258306622505188, "val/loss_error": 0.25161927938461304, "val/loss_total": 0.252956748008728 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.3502500057220459, "val/loss_error": 0.36165475845336914, "val/loss_total": 0.3593738079071045 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.2665698528289795, "val/loss_error": 0.5255231261253357, "val/loss_total": 0.47373247146606445 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.9101464748382568, "val/loss_error": 0.48031097650527954, "val/loss_total": 0.5662781000137329 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.20106622576713562, "val/loss_error": 0.06461511552333832, "val/loss_total": 0.09190534055233002 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.3802759051322937, "val/loss_error": 0.28711503744125366, "val/loss_total": 0.30574721097946167 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5905624628067017, "val/loss_error": 0.5207822918891907, "val/loss_total": 0.534738302230835 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.1903264969587326, "val/loss_error": 0.1887292116880417, "val/loss_total": 0.1890486776828766 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.09403572976589203, "val/loss_error": 0.24251359701156616, "val/loss_total": 0.21281802654266357 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5647215843200684, "val/loss_error": 0.5006731152534485, "val/loss_total": 0.5134828090667725 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.09947756677865982, "val/loss_error": 0.3041309416294098, "val/loss_total": 0.2632002830505371 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.19930794835090637, "val/loss_error": 0.22306972742080688, "val/loss_total": 0.21831737458705902 }, { "epoch": 5.0, "eval_eval/f1_0": 0.5425400733947754, "eval_eval/f1_1": 0.934394359588623, "eval_eval/precision_0": 0.5898123383522034, "eval_eval/precision_1": 0.9237762093544006, "eval_eval/recall_0": 0.5022830963134766, "eval_eval/recall_1": 0.9452593922615051, "eval_eval/wer": 0.1495095855136229, "eval_loss": 0.36842384934425354, "eval_runtime": 7.3892, "eval_samples_per_second": 13.533, "eval_steps_per_second": 13.533, "step": 1925 } ], "logging_steps": 10, "max_steps": 1925, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }