{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 1925, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0, "step": 0, "train/loss_ctc": 0.694847822189331, "train/loss_error": 0.4741579592227936, "train/loss_total": 0.51829594373703 }, { "epoch": 0.0025974025974025974, "step": 1, "train/loss_ctc": 0.3723033666610718, "train/loss_error": 0.37305036187171936, "train/loss_total": 0.37290099263191223 }, { "epoch": 0.005194805194805195, "step": 2, "train/loss_ctc": 0.9261850714683533, "train/loss_error": 0.47874051332473755, "train/loss_total": 0.5682294368743896 }, { "epoch": 0.007792207792207792, "step": 3, "train/loss_ctc": 0.26508471369743347, "train/loss_error": 0.40710076689720154, "train/loss_total": 0.37869754433631897 }, { "epoch": 0.01038961038961039, "step": 4, "train/loss_ctc": 0.4772428870201111, "train/loss_error": 0.4527019262313843, "train/loss_total": 0.4576101303100586 }, { "epoch": 0.012987012987012988, "step": 5, "train/loss_ctc": 1.1765501499176025, "train/loss_error": 0.4540162682533264, "train/loss_total": 0.5985230803489685 }, { "epoch": 0.015584415584415584, "step": 6, "train/loss_ctc": 0.5156809091567993, "train/loss_error": 0.3797151446342468, "train/loss_total": 0.4069083034992218 }, { "epoch": 0.01818181818181818, "step": 7, "train/loss_ctc": 0.43020400404930115, "train/loss_error": 0.37856730818748474, "train/loss_total": 0.388894647359848 }, { "epoch": 0.02077922077922078, "step": 8, "train/loss_ctc": 0.44236594438552856, "train/loss_error": 0.44689488410949707, "train/loss_total": 0.44598910212516785 }, { "epoch": 0.023376623376623377, "step": 9, "train/loss_ctc": 0.41499727964401245, "train/loss_error": 0.3972800374031067, "train/loss_total": 0.4008235037326813 }, { "epoch": 0.025974025974025976, "grad_norm": 12016.8369140625, "learning_rate": 2.985974025974026e-05, "loss": 0.4537, "step": 10 }, { "epoch": 0.025974025974025976, "step": 10, "train/loss_ctc": 0.5102351903915405, "train/loss_error": 0.39993810653686523, "train/loss_total": 0.4219975471496582 }, { "epoch": 0.02857142857142857, "step": 11, "train/loss_ctc": 0.6526217460632324, "train/loss_error": 0.5021234154701233, "train/loss_total": 0.532223105430603 }, { "epoch": 0.03116883116883117, "step": 12, "train/loss_ctc": 0.49539780616760254, "train/loss_error": 0.40407660603523254, "train/loss_total": 0.42234086990356445 }, { "epoch": 0.033766233766233764, "step": 13, "train/loss_ctc": 0.5866650342941284, "train/loss_error": 0.43195992708206177, "train/loss_total": 0.46290093660354614 }, { "epoch": 0.03636363636363636, "step": 14, "train/loss_ctc": 0.34654372930526733, "train/loss_error": 0.3847479522228241, "train/loss_total": 0.3771071135997772 }, { "epoch": 0.03896103896103896, "step": 15, "train/loss_ctc": 0.42292290925979614, "train/loss_error": 0.40521734952926636, "train/loss_total": 0.4087584912776947 }, { "epoch": 0.04155844155844156, "step": 16, "train/loss_ctc": 0.9357887506484985, "train/loss_error": 0.4207207262516022, "train/loss_total": 0.5237343311309814 }, { "epoch": 0.04415584415584416, "step": 17, "train/loss_ctc": 0.5965677499771118, "train/loss_error": 0.49248793721199036, "train/loss_total": 0.5133039355278015 }, { "epoch": 0.046753246753246755, "step": 18, "train/loss_ctc": 0.6182276010513306, "train/loss_error": 0.5362619757652283, "train/loss_total": 0.5526551008224487 }, { "epoch": 0.04935064935064935, "step": 19, "train/loss_ctc": 0.8662735223770142, "train/loss_error": 0.4674307405948639, "train/loss_total": 0.5471993088722229 }, { "epoch": 0.05194805194805195, "grad_norm": 14429.919921875, "learning_rate": 2.9703896103896104e-05, "loss": 0.4762, "step": 20 }, { "epoch": 0.05194805194805195, "step": 20, "train/loss_ctc": 0.542626142501831, "train/loss_error": 0.4929944574832916, "train/loss_total": 0.5029208064079285 }, { "epoch": 0.05454545454545454, "step": 21, "train/loss_ctc": 0.6155562400817871, "train/loss_error": 0.4884151220321655, "train/loss_total": 0.5138433575630188 }, { "epoch": 0.05714285714285714, "step": 22, "train/loss_ctc": 0.7565929889678955, "train/loss_error": 0.4126456677913666, "train/loss_total": 0.4814351201057434 }, { "epoch": 0.05974025974025974, "step": 23, "train/loss_ctc": 0.23116707801818848, "train/loss_error": 0.39073434472084045, "train/loss_total": 0.3588208854198456 }, { "epoch": 0.06233766233766234, "step": 24, "train/loss_ctc": 0.5480751991271973, "train/loss_error": 0.44282689690589905, "train/loss_total": 0.46387654542922974 }, { "epoch": 0.06493506493506493, "step": 25, "train/loss_ctc": 0.431088924407959, "train/loss_error": 0.4524277150630951, "train/loss_total": 0.44815996289253235 }, { "epoch": 0.06753246753246753, "step": 26, "train/loss_ctc": 0.36495620012283325, "train/loss_error": 0.4761621356010437, "train/loss_total": 0.45392096042633057 }, { "epoch": 0.07012987012987013, "step": 27, "train/loss_ctc": 0.45267802476882935, "train/loss_error": 0.3130171000957489, "train/loss_total": 0.34094929695129395 }, { "epoch": 0.07272727272727272, "step": 28, "train/loss_ctc": 0.5199288725852966, "train/loss_error": 0.4530007243156433, "train/loss_total": 0.4663863778114319 }, { "epoch": 0.07532467532467532, "step": 29, "train/loss_ctc": 0.4434507191181183, "train/loss_error": 0.45997747778892517, "train/loss_total": 0.45667213201522827 }, { "epoch": 0.07792207792207792, "grad_norm": 12181.98046875, "learning_rate": 2.954805194805195e-05, "loss": 0.4487, "step": 30 }, { "epoch": 0.07792207792207792, "step": 30, "train/loss_ctc": 0.8396193981170654, "train/loss_error": 0.4296696186065674, "train/loss_total": 0.511659562587738 }, { "epoch": 0.08051948051948052, "step": 31, "train/loss_ctc": 0.3793363571166992, "train/loss_error": 0.4265888035297394, "train/loss_total": 0.41713830828666687 }, { "epoch": 0.08311688311688312, "step": 32, "train/loss_ctc": 0.6027369499206543, "train/loss_error": 0.4996735155582428, "train/loss_total": 0.5202862024307251 }, { "epoch": 0.08571428571428572, "step": 33, "train/loss_ctc": 0.43069902062416077, "train/loss_error": 0.3591897189617157, "train/loss_total": 0.3734915852546692 }, { "epoch": 0.08831168831168831, "step": 34, "train/loss_ctc": 0.6149111986160278, "train/loss_error": 0.45989498496055603, "train/loss_total": 0.4908982217311859 }, { "epoch": 0.09090909090909091, "step": 35, "train/loss_ctc": 0.5249230861663818, "train/loss_error": 0.4087159335613251, "train/loss_total": 0.4319573640823364 }, { "epoch": 0.09350649350649351, "step": 36, "train/loss_ctc": 0.6643285751342773, "train/loss_error": 0.49235257506370544, "train/loss_total": 0.5267477631568909 }, { "epoch": 0.09610389610389611, "step": 37, "train/loss_ctc": 0.8580278158187866, "train/loss_error": 0.4466114640235901, "train/loss_total": 0.5288947224617004 }, { "epoch": 0.0987012987012987, "step": 38, "train/loss_ctc": 0.5038363337516785, "train/loss_error": 0.38499900698661804, "train/loss_total": 0.408766508102417 }, { "epoch": 0.1012987012987013, "step": 39, "train/loss_ctc": 0.38870590925216675, "train/loss_error": 0.3637288510799408, "train/loss_total": 0.3687242567539215 }, { "epoch": 0.1038961038961039, "grad_norm": 11977.173828125, "learning_rate": 2.939220779220779e-05, "loss": 0.4579, "step": 40 }, { "epoch": 0.1038961038961039, "step": 40, "train/loss_ctc": 0.6793447732925415, "train/loss_error": 0.4505559802055359, "train/loss_total": 0.49631375074386597 }, { "epoch": 0.10649350649350649, "step": 41, "train/loss_ctc": 0.7093327045440674, "train/loss_error": 0.47992387413978577, "train/loss_total": 0.525805652141571 }, { "epoch": 0.10909090909090909, "step": 42, "train/loss_ctc": 0.5849858522415161, "train/loss_error": 0.4291219413280487, "train/loss_total": 0.4602947533130646 }, { "epoch": 0.11168831168831168, "step": 43, "train/loss_ctc": 0.4660418927669525, "train/loss_error": 0.43575045466423035, "train/loss_total": 0.4418087303638458 }, { "epoch": 0.11428571428571428, "step": 44, "train/loss_ctc": 0.4985731840133667, "train/loss_error": 0.5299038887023926, "train/loss_total": 0.5236377716064453 }, { "epoch": 0.11688311688311688, "step": 45, "train/loss_ctc": 0.41148003935813904, "train/loss_error": 0.412060409784317, "train/loss_total": 0.41194435954093933 }, { "epoch": 0.11948051948051948, "step": 46, "train/loss_ctc": 0.5926523804664612, "train/loss_error": 0.44150787591934204, "train/loss_total": 0.4717367887496948 }, { "epoch": 0.12207792207792208, "step": 47, "train/loss_ctc": 0.3527876138687134, "train/loss_error": 0.4003596901893616, "train/loss_total": 0.39084529876708984 }, { "epoch": 0.12467532467532468, "step": 48, "train/loss_ctc": 0.46948474645614624, "train/loss_error": 0.49810272455215454, "train/loss_total": 0.4923791289329529 }, { "epoch": 0.12727272727272726, "step": 49, "train/loss_ctc": 0.5447032451629639, "train/loss_error": 0.4804885685443878, "train/loss_total": 0.49333152174949646 }, { "epoch": 0.12987012987012986, "grad_norm": 12921.349609375, "learning_rate": 2.9236363636363635e-05, "loss": 0.4708, "step": 50 }, { "epoch": 0.12987012987012986, "step": 50, "train/loss_ctc": 0.4550197124481201, "train/loss_error": 0.418063759803772, "train/loss_total": 0.4254549741744995 }, { "epoch": 0.13246753246753246, "step": 51, "train/loss_ctc": 0.564122200012207, "train/loss_error": 0.3327409625053406, "train/loss_total": 0.3790172040462494 }, { "epoch": 0.13506493506493505, "step": 52, "train/loss_ctc": 0.617279052734375, "train/loss_error": 0.4624791741371155, "train/loss_total": 0.4934391677379608 }, { "epoch": 0.13766233766233765, "step": 53, "train/loss_ctc": 0.6506682634353638, "train/loss_error": 0.4075768291950226, "train/loss_total": 0.4561951160430908 }, { "epoch": 0.14025974025974025, "step": 54, "train/loss_ctc": 0.47502148151397705, "train/loss_error": 0.41296595335006714, "train/loss_total": 0.4253770709037781 }, { "epoch": 0.14285714285714285, "step": 55, "train/loss_ctc": 0.42169317603111267, "train/loss_error": 0.46754148602485657, "train/loss_total": 0.4583718180656433 }, { "epoch": 0.14545454545454545, "step": 56, "train/loss_ctc": 0.42899608612060547, "train/loss_error": 0.4293688237667084, "train/loss_total": 0.42929428815841675 }, { "epoch": 0.14805194805194805, "step": 57, "train/loss_ctc": 0.9978979229927063, "train/loss_error": 0.4699254035949707, "train/loss_total": 0.5755199193954468 }, { "epoch": 0.15064935064935064, "step": 58, "train/loss_ctc": 0.6721453070640564, "train/loss_error": 0.4724445641040802, "train/loss_total": 0.5123847126960754 }, { "epoch": 0.15324675324675324, "step": 59, "train/loss_ctc": 0.39731431007385254, "train/loss_error": 0.5176360011100769, "train/loss_total": 0.4935716688632965 }, { "epoch": 0.15584415584415584, "grad_norm": 10906.537109375, "learning_rate": 2.908051948051948e-05, "loss": 0.4649, "step": 60 }, { "epoch": 0.15584415584415584, "step": 60, "train/loss_ctc": 0.7905935049057007, "train/loss_error": 0.4196770489215851, "train/loss_total": 0.4938603639602661 }, { "epoch": 0.15844155844155844, "step": 61, "train/loss_ctc": 0.466178297996521, "train/loss_error": 0.4129438102245331, "train/loss_total": 0.4235907196998596 }, { "epoch": 0.16103896103896104, "step": 62, "train/loss_ctc": 0.4900413453578949, "train/loss_error": 0.3796130418777466, "train/loss_total": 0.4016987085342407 }, { "epoch": 0.16363636363636364, "step": 63, "train/loss_ctc": 0.47806528210639954, "train/loss_error": 0.4747025966644287, "train/loss_total": 0.47537514567375183 }, { "epoch": 0.16623376623376623, "step": 64, "train/loss_ctc": 0.60530024766922, "train/loss_error": 0.4331114590167999, "train/loss_total": 0.467549204826355 }, { "epoch": 0.16883116883116883, "step": 65, "train/loss_ctc": 0.6666626334190369, "train/loss_error": 0.3927488327026367, "train/loss_total": 0.4475315809249878 }, { "epoch": 0.17142857142857143, "step": 66, "train/loss_ctc": 0.4167201519012451, "train/loss_error": 0.4001433849334717, "train/loss_total": 0.40345874428749084 }, { "epoch": 0.17402597402597403, "step": 67, "train/loss_ctc": 0.741320788860321, "train/loss_error": 0.4323989152908325, "train/loss_total": 0.4941833019256592 }, { "epoch": 0.17662337662337663, "step": 68, "train/loss_ctc": 0.5626553297042847, "train/loss_error": 0.39185377955436707, "train/loss_total": 0.42601409554481506 }, { "epoch": 0.17922077922077922, "step": 69, "train/loss_ctc": 0.41166380047798157, "train/loss_error": 0.43042322993278503, "train/loss_total": 0.4266713559627533 }, { "epoch": 0.18181818181818182, "grad_norm": 14124.3154296875, "learning_rate": 2.8924675324675328e-05, "loss": 0.446, "step": 70 }, { "epoch": 0.18181818181818182, "step": 70, "train/loss_ctc": 0.8468095660209656, "train/loss_error": 0.49531567096710205, "train/loss_total": 0.5656144618988037 }, { "epoch": 0.18441558441558442, "step": 71, "train/loss_ctc": 0.456368088722229, "train/loss_error": 0.42593738436698914, "train/loss_total": 0.4320235252380371 }, { "epoch": 0.18701298701298702, "step": 72, "train/loss_ctc": 0.3759976029396057, "train/loss_error": 0.4089236259460449, "train/loss_total": 0.4023384153842926 }, { "epoch": 0.18961038961038962, "step": 73, "train/loss_ctc": 0.4734193682670593, "train/loss_error": 0.4078376889228821, "train/loss_total": 0.42095404863357544 }, { "epoch": 0.19220779220779222, "step": 74, "train/loss_ctc": 0.5773196220397949, "train/loss_error": 0.598096489906311, "train/loss_total": 0.5939411520957947 }, { "epoch": 0.19480519480519481, "step": 75, "train/loss_ctc": 0.479409784078598, "train/loss_error": 0.4564332365989685, "train/loss_total": 0.4610285758972168 }, { "epoch": 0.1974025974025974, "step": 76, "train/loss_ctc": 0.7672194838523865, "train/loss_error": 0.39408984780311584, "train/loss_total": 0.4687157869338989 }, { "epoch": 0.2, "step": 77, "train/loss_ctc": 0.49705979228019714, "train/loss_error": 0.3994363248348236, "train/loss_total": 0.4189610183238983 }, { "epoch": 0.2025974025974026, "step": 78, "train/loss_ctc": 0.46368837356567383, "train/loss_error": 0.422127366065979, "train/loss_total": 0.4304395616054535 }, { "epoch": 0.2051948051948052, "step": 79, "train/loss_ctc": 0.5929354429244995, "train/loss_error": 0.5014660358428955, "train/loss_total": 0.5197599530220032 }, { "epoch": 0.2077922077922078, "grad_norm": 11533.623046875, "learning_rate": 2.8768831168831172e-05, "loss": 0.4714, "step": 80 }, { "epoch": 0.2077922077922078, "step": 80, "train/loss_ctc": 0.5618027448654175, "train/loss_error": 0.5087939500808716, "train/loss_total": 0.5193957090377808 }, { "epoch": 0.21038961038961038, "step": 81, "train/loss_ctc": 0.47679510712623596, "train/loss_error": 0.40070971846580505, "train/loss_total": 0.41592681407928467 }, { "epoch": 0.21298701298701297, "step": 82, "train/loss_ctc": 0.5186178684234619, "train/loss_error": 0.447969526052475, "train/loss_total": 0.46209922432899475 }, { "epoch": 0.21558441558441557, "step": 83, "train/loss_ctc": 0.8178589344024658, "train/loss_error": 0.5058044791221619, "train/loss_total": 0.5682153701782227 }, { "epoch": 0.21818181818181817, "step": 84, "train/loss_ctc": 0.554175078868866, "train/loss_error": 0.3967260718345642, "train/loss_total": 0.428215891122818 }, { "epoch": 0.22077922077922077, "step": 85, "train/loss_ctc": 0.4375346899032593, "train/loss_error": 0.4162735939025879, "train/loss_total": 0.42052581906318665 }, { "epoch": 0.22337662337662337, "step": 86, "train/loss_ctc": 0.6339104771614075, "train/loss_error": 0.45103728771209717, "train/loss_total": 0.48761194944381714 }, { "epoch": 0.22597402597402597, "step": 87, "train/loss_ctc": 0.7311000823974609, "train/loss_error": 0.40043121576309204, "train/loss_total": 0.46656501293182373 }, { "epoch": 0.22857142857142856, "step": 88, "train/loss_ctc": 0.460066020488739, "train/loss_error": 0.3895653784275055, "train/loss_total": 0.4036655128002167 }, { "epoch": 0.23116883116883116, "step": 89, "train/loss_ctc": 0.5910898447036743, "train/loss_error": 0.46283045411109924, "train/loss_total": 0.48848235607147217 }, { "epoch": 0.23376623376623376, "grad_norm": 13310.0966796875, "learning_rate": 2.8612987012987014e-05, "loss": 0.4661, "step": 90 }, { "epoch": 0.23376623376623376, "step": 90, "train/loss_ctc": 0.450400710105896, "train/loss_error": 0.44167450070381165, "train/loss_total": 0.44341975450515747 }, { "epoch": 0.23636363636363636, "step": 91, "train/loss_ctc": 0.36392512917518616, "train/loss_error": 0.3669217824935913, "train/loss_total": 0.36632245779037476 }, { "epoch": 0.23896103896103896, "step": 92, "train/loss_ctc": 0.4017769992351532, "train/loss_error": 0.38956573605537415, "train/loss_total": 0.3920080065727234 }, { "epoch": 0.24155844155844156, "step": 93, "train/loss_ctc": 0.5900647640228271, "train/loss_error": 0.5102027058601379, "train/loss_total": 0.5261751413345337 }, { "epoch": 0.24415584415584415, "step": 94, "train/loss_ctc": 0.7423102259635925, "train/loss_error": 0.40996813774108887, "train/loss_total": 0.4764365553855896 }, { "epoch": 0.24675324675324675, "step": 95, "train/loss_ctc": 0.6362175941467285, "train/loss_error": 0.40239080786705017, "train/loss_total": 0.44915616512298584 }, { "epoch": 0.24935064935064935, "step": 96, "train/loss_ctc": 0.5263576507568359, "train/loss_error": 0.41567742824554443, "train/loss_total": 0.4378134608268738 }, { "epoch": 0.2519480519480519, "step": 97, "train/loss_ctc": 0.5835067629814148, "train/loss_error": 0.4252483546733856, "train/loss_total": 0.45690006017684937 }, { "epoch": 0.2545454545454545, "step": 98, "train/loss_ctc": 0.5541237592697144, "train/loss_error": 0.45291438698768616, "train/loss_total": 0.47315627336502075 }, { "epoch": 0.2571428571428571, "step": 99, "train/loss_ctc": 0.5311332941055298, "train/loss_error": 0.5581430792808533, "train/loss_total": 0.5527411103248596 }, { "epoch": 0.2597402597402597, "grad_norm": 43387.609375, "learning_rate": 2.845714285714286e-05, "loss": 0.4574, "step": 100 }, { "epoch": 0.2597402597402597, "step": 100, "train/loss_ctc": 0.4485066533088684, "train/loss_error": 0.4215376079082489, "train/loss_total": 0.4269314110279083 }, { "epoch": 0.2623376623376623, "step": 101, "train/loss_ctc": 0.39266133308410645, "train/loss_error": 0.3870480954647064, "train/loss_total": 0.3881707489490509 }, { "epoch": 0.2649350649350649, "step": 102, "train/loss_ctc": 0.7396215200424194, "train/loss_error": 0.4740568995475769, "train/loss_total": 0.5271698236465454 }, { "epoch": 0.2675324675324675, "step": 103, "train/loss_ctc": 0.515573263168335, "train/loss_error": 0.46192696690559387, "train/loss_total": 0.47265625 }, { "epoch": 0.2701298701298701, "step": 104, "train/loss_ctc": 0.5795592069625854, "train/loss_error": 0.47372958064079285, "train/loss_total": 0.4948955178260803 }, { "epoch": 0.2727272727272727, "step": 105, "train/loss_ctc": 0.6253448724746704, "train/loss_error": 0.38724929094314575, "train/loss_total": 0.43486839532852173 }, { "epoch": 0.2753246753246753, "step": 106, "train/loss_ctc": 0.4871644973754883, "train/loss_error": 0.36779728531837463, "train/loss_total": 0.39167073369026184 }, { "epoch": 0.2779220779220779, "step": 107, "train/loss_ctc": 0.8678815364837646, "train/loss_error": 0.4788866937160492, "train/loss_total": 0.5566856861114502 }, { "epoch": 0.2805194805194805, "step": 108, "train/loss_ctc": 1.0229829549789429, "train/loss_error": 0.45859473943710327, "train/loss_total": 0.5714724063873291 }, { "epoch": 0.2831168831168831, "step": 109, "train/loss_ctc": 0.5975233316421509, "train/loss_error": 0.3950938582420349, "train/loss_total": 0.43557974696159363 }, { "epoch": 0.2857142857142857, "grad_norm": 26670.6171875, "learning_rate": 2.8301298701298703e-05, "loss": 0.47, "step": 110 }, { "epoch": 0.2857142857142857, "step": 110, "train/loss_ctc": 0.5073396563529968, "train/loss_error": 0.45853665471076965, "train/loss_total": 0.4682972729206085 }, { "epoch": 0.2883116883116883, "step": 111, "train/loss_ctc": 0.511534571647644, "train/loss_error": 0.4300502836704254, "train/loss_total": 0.446347177028656 }, { "epoch": 0.2909090909090909, "step": 112, "train/loss_ctc": 0.56026291847229, "train/loss_error": 0.36755481362342834, "train/loss_total": 0.4060964584350586 }, { "epoch": 0.2935064935064935, "step": 113, "train/loss_ctc": 0.494110643863678, "train/loss_error": 0.482489675283432, "train/loss_total": 0.4848138689994812 }, { "epoch": 0.2961038961038961, "step": 114, "train/loss_ctc": 0.3575074374675751, "train/loss_error": 0.4084242582321167, "train/loss_total": 0.3982408940792084 }, { "epoch": 0.2987012987012987, "step": 115, "train/loss_ctc": 0.48857998847961426, "train/loss_error": 0.3596517741680145, "train/loss_total": 0.3854374289512634 }, { "epoch": 0.3012987012987013, "step": 116, "train/loss_ctc": 0.6169968247413635, "train/loss_error": 0.4649410545825958, "train/loss_total": 0.49535223841667175 }, { "epoch": 0.3038961038961039, "step": 117, "train/loss_ctc": 0.4906640946865082, "train/loss_error": 0.4488445520401001, "train/loss_total": 0.45720845460891724 }, { "epoch": 0.3064935064935065, "step": 118, "train/loss_ctc": 0.3822084963321686, "train/loss_error": 0.34759724140167236, "train/loss_total": 0.3545195162296295 }, { "epoch": 0.3090909090909091, "step": 119, "train/loss_ctc": 0.5509300231933594, "train/loss_error": 0.5224983096122742, "train/loss_total": 0.5281846523284912 }, { "epoch": 0.3116883116883117, "grad_norm": 44933.5546875, "learning_rate": 2.8145454545454548e-05, "loss": 0.4424, "step": 120 }, { "epoch": 0.3116883116883117, "step": 120, "train/loss_ctc": 0.3936852216720581, "train/loss_error": 0.4130989611148834, "train/loss_total": 0.4092162251472473 }, { "epoch": 0.3142857142857143, "step": 121, "train/loss_ctc": 0.3246385455131531, "train/loss_error": 0.35658320784568787, "train/loss_total": 0.3501942753791809 }, { "epoch": 0.3168831168831169, "step": 122, "train/loss_ctc": 0.418913334608078, "train/loss_error": 0.4482111632823944, "train/loss_total": 0.4423516094684601 }, { "epoch": 0.3194805194805195, "step": 123, "train/loss_ctc": 0.602005660533905, "train/loss_error": 0.5229321718215942, "train/loss_total": 0.5387468934059143 }, { "epoch": 0.3220779220779221, "step": 124, "train/loss_ctc": 0.4243882894515991, "train/loss_error": 0.3559877872467041, "train/loss_total": 0.3696679174900055 }, { "epoch": 0.3246753246753247, "step": 125, "train/loss_ctc": 0.7160326242446899, "train/loss_error": 0.5305467247962952, "train/loss_total": 0.5676438808441162 }, { "epoch": 0.32727272727272727, "step": 126, "train/loss_ctc": 0.38261646032333374, "train/loss_error": 0.42523232102394104, "train/loss_total": 0.41670915484428406 }, { "epoch": 0.32987012987012987, "step": 127, "train/loss_ctc": 0.5586296916007996, "train/loss_error": 0.4024519622325897, "train/loss_total": 0.4336875081062317 }, { "epoch": 0.33246753246753247, "step": 128, "train/loss_ctc": 0.6108089089393616, "train/loss_error": 0.4284844398498535, "train/loss_total": 0.4649493396282196 }, { "epoch": 0.33506493506493507, "step": 129, "train/loss_ctc": 0.47188377380371094, "train/loss_error": 0.38435113430023193, "train/loss_total": 0.4018576741218567 }, { "epoch": 0.33766233766233766, "grad_norm": 24695.896484375, "learning_rate": 2.7989610389610393e-05, "loss": 0.4395, "step": 130 }, { "epoch": 0.33766233766233766, "step": 130, "train/loss_ctc": 0.36422258615493774, "train/loss_error": 0.48705172538757324, "train/loss_total": 0.4624859094619751 }, { "epoch": 0.34025974025974026, "step": 131, "train/loss_ctc": 0.36398395895957947, "train/loss_error": 0.379377156496048, "train/loss_total": 0.3762985169887543 }, { "epoch": 0.34285714285714286, "step": 132, "train/loss_ctc": 0.4718188941478729, "train/loss_error": 0.4176892042160034, "train/loss_total": 0.42851513624191284 }, { "epoch": 0.34545454545454546, "step": 133, "train/loss_ctc": 0.314931184053421, "train/loss_error": 0.41632378101348877, "train/loss_total": 0.3960452675819397 }, { "epoch": 0.34805194805194806, "step": 134, "train/loss_ctc": 0.42645639181137085, "train/loss_error": 0.36802390217781067, "train/loss_total": 0.37971043586730957 }, { "epoch": 0.35064935064935066, "step": 135, "train/loss_ctc": 0.4179978668689728, "train/loss_error": 0.43115234375, "train/loss_total": 0.42852145433425903 }, { "epoch": 0.35324675324675325, "step": 136, "train/loss_ctc": 1.2257776260375977, "train/loss_error": 0.5022720694541931, "train/loss_total": 0.646973192691803 }, { "epoch": 0.35584415584415585, "step": 137, "train/loss_ctc": 0.6215097904205322, "train/loss_error": 0.40348541736602783, "train/loss_total": 0.4470902979373932 }, { "epoch": 0.35844155844155845, "step": 138, "train/loss_ctc": 0.5693950057029724, "train/loss_error": 0.43772202730178833, "train/loss_total": 0.4640566110610962 }, { "epoch": 0.36103896103896105, "step": 139, "train/loss_ctc": 0.3710927367210388, "train/loss_error": 0.41996240615844727, "train/loss_total": 0.4101884663105011 }, { "epoch": 0.36363636363636365, "grad_norm": 18522.671875, "learning_rate": 2.7833766233766234e-05, "loss": 0.444, "step": 140 }, { "epoch": 0.36363636363636365, "step": 140, "train/loss_ctc": 0.6660314798355103, "train/loss_error": 0.4004495143890381, "train/loss_total": 0.45356589555740356 }, { "epoch": 0.36623376623376624, "step": 141, "train/loss_ctc": 0.7704581022262573, "train/loss_error": 0.4437367618083954, "train/loss_total": 0.5090810060501099 }, { "epoch": 0.36883116883116884, "step": 142, "train/loss_ctc": 0.883712887763977, "train/loss_error": 0.4850730001926422, "train/loss_total": 0.5648009777069092 }, { "epoch": 0.37142857142857144, "step": 143, "train/loss_ctc": 0.9189484715461731, "train/loss_error": 0.5185495018959045, "train/loss_total": 0.5986292958259583 }, { "epoch": 0.37402597402597404, "step": 144, "train/loss_ctc": 0.4688904583454132, "train/loss_error": 0.4677773118019104, "train/loss_total": 0.46799996495246887 }, { "epoch": 0.37662337662337664, "step": 145, "train/loss_ctc": 0.7310811877250671, "train/loss_error": 0.4646577537059784, "train/loss_total": 0.5179424285888672 }, { "epoch": 0.37922077922077924, "step": 146, "train/loss_ctc": 0.481586217880249, "train/loss_error": 0.45260176062583923, "train/loss_total": 0.45839864015579224 }, { "epoch": 0.38181818181818183, "step": 147, "train/loss_ctc": 0.36366763710975647, "train/loss_error": 0.4541378915309906, "train/loss_total": 0.4360438287258148 }, { "epoch": 0.38441558441558443, "step": 148, "train/loss_ctc": 0.4388599991798401, "train/loss_error": 0.4470514953136444, "train/loss_total": 0.445413202047348 }, { "epoch": 0.38701298701298703, "step": 149, "train/loss_ctc": 0.4058709144592285, "train/loss_error": 0.3584604263305664, "train/loss_total": 0.36794254183769226 }, { "epoch": 0.38961038961038963, "grad_norm": 22015.357421875, "learning_rate": 2.767792207792208e-05, "loss": 0.482, "step": 150 }, { "epoch": 0.38961038961038963, "step": 150, "train/loss_ctc": 0.571031928062439, "train/loss_error": 0.4360535740852356, "train/loss_total": 0.4630492329597473 }, { "epoch": 0.3922077922077922, "step": 151, "train/loss_ctc": 0.45183104276657104, "train/loss_error": 0.4048255681991577, "train/loss_total": 0.4142266809940338 }, { "epoch": 0.3948051948051948, "step": 152, "train/loss_ctc": 0.708226203918457, "train/loss_error": 0.47719481587409973, "train/loss_total": 0.5234010815620422 }, { "epoch": 0.3974025974025974, "step": 153, "train/loss_ctc": 0.398032009601593, "train/loss_error": 0.35495662689208984, "train/loss_total": 0.36357173323631287 }, { "epoch": 0.4, "step": 154, "train/loss_ctc": 0.4809565544128418, "train/loss_error": 0.41559720039367676, "train/loss_total": 0.4286690950393677 }, { "epoch": 0.4025974025974026, "step": 155, "train/loss_ctc": 0.5673643350601196, "train/loss_error": 0.4595203697681427, "train/loss_total": 0.48108917474746704 }, { "epoch": 0.4051948051948052, "step": 156, "train/loss_ctc": 0.46091583371162415, "train/loss_error": 0.3687645196914673, "train/loss_total": 0.38719478249549866 }, { "epoch": 0.4077922077922078, "step": 157, "train/loss_ctc": 0.5808408260345459, "train/loss_error": 0.3480362296104431, "train/loss_total": 0.3945971727371216 }, { "epoch": 0.4103896103896104, "step": 158, "train/loss_ctc": 0.4464465081691742, "train/loss_error": 0.36486366391181946, "train/loss_total": 0.3811802566051483 }, { "epoch": 0.412987012987013, "step": 159, "train/loss_ctc": 0.6190758943557739, "train/loss_error": 0.36768338084220886, "train/loss_total": 0.41796189546585083 }, { "epoch": 0.4155844155844156, "grad_norm": 33683.46484375, "learning_rate": 2.7522077922077924e-05, "loss": 0.4255, "step": 160 }, { "epoch": 0.4155844155844156, "step": 160, "train/loss_ctc": 0.7037824988365173, "train/loss_error": 0.3916843831539154, "train/loss_total": 0.4541040062904358 }, { "epoch": 0.41818181818181815, "step": 161, "train/loss_ctc": 0.6537892818450928, "train/loss_error": 0.47114044427871704, "train/loss_total": 0.5076702237129211 }, { "epoch": 0.42077922077922075, "step": 162, "train/loss_ctc": 0.42495718598365784, "train/loss_error": 0.41306865215301514, "train/loss_total": 0.41544634103775024 }, { "epoch": 0.42337662337662335, "step": 163, "train/loss_ctc": 0.3958876430988312, "train/loss_error": 0.4012632369995117, "train/loss_total": 0.4001881182193756 }, { "epoch": 0.42597402597402595, "step": 164, "train/loss_ctc": 0.3755267262458801, "train/loss_error": 0.34935861825942993, "train/loss_total": 0.3545922338962555 }, { "epoch": 0.42857142857142855, "step": 165, "train/loss_ctc": 0.4458908140659332, "train/loss_error": 0.38131725788116455, "train/loss_total": 0.39423197507858276 }, { "epoch": 0.43116883116883115, "step": 166, "train/loss_ctc": 0.3181658387184143, "train/loss_error": 0.3449333906173706, "train/loss_total": 0.33957988023757935 }, { "epoch": 0.43376623376623374, "step": 167, "train/loss_ctc": 0.371226042509079, "train/loss_error": 0.36941075325012207, "train/loss_total": 0.36977383494377136 }, { "epoch": 0.43636363636363634, "step": 168, "train/loss_ctc": 0.3162132799625397, "train/loss_error": 0.40878045558929443, "train/loss_total": 0.390267014503479 }, { "epoch": 0.43896103896103894, "step": 169, "train/loss_ctc": 0.41068559885025024, "train/loss_error": 0.31856998801231384, "train/loss_total": 0.33699309825897217 }, { "epoch": 0.44155844155844154, "grad_norm": 21341.10546875, "learning_rate": 2.736623376623377e-05, "loss": 0.3963, "step": 170 }, { "epoch": 0.44155844155844154, "step": 170, "train/loss_ctc": 0.5195636749267578, "train/loss_error": 0.38710781931877136, "train/loss_total": 0.41359901428222656 }, { "epoch": 0.44415584415584414, "step": 171, "train/loss_ctc": 0.5002082586288452, "train/loss_error": 0.40380340814590454, "train/loss_total": 0.4230843782424927 }, { "epoch": 0.44675324675324674, "step": 172, "train/loss_ctc": 0.5325273275375366, "train/loss_error": 0.6496865749359131, "train/loss_total": 0.6262547373771667 }, { "epoch": 0.44935064935064933, "step": 173, "train/loss_ctc": 0.5061541795730591, "train/loss_error": 0.4385828375816345, "train/loss_total": 0.4520971179008484 }, { "epoch": 0.45194805194805193, "step": 174, "train/loss_ctc": 0.6845451593399048, "train/loss_error": 0.3592882454395294, "train/loss_total": 0.4243396520614624 }, { "epoch": 0.45454545454545453, "step": 175, "train/loss_ctc": 0.6034549474716187, "train/loss_error": 0.4488508999347687, "train/loss_total": 0.4797717332839966 }, { "epoch": 0.45714285714285713, "step": 176, "train/loss_ctc": 0.38590097427368164, "train/loss_error": 0.3211212754249573, "train/loss_total": 0.33407723903656006 }, { "epoch": 0.4597402597402597, "step": 177, "train/loss_ctc": 0.5472739934921265, "train/loss_error": 0.3941081464290619, "train/loss_total": 0.42474132776260376 }, { "epoch": 0.4623376623376623, "step": 178, "train/loss_ctc": 0.5276693105697632, "train/loss_error": 0.42501595616340637, "train/loss_total": 0.44554662704467773 }, { "epoch": 0.4649350649350649, "step": 179, "train/loss_ctc": 0.4784081280231476, "train/loss_error": 0.3514203429222107, "train/loss_total": 0.376817911863327 }, { "epoch": 0.4675324675324675, "grad_norm": 14542.3525390625, "learning_rate": 2.7225974025974027e-05, "loss": 0.44, "step": 180 }, { "epoch": 0.4675324675324675, "step": 180, "train/loss_ctc": 0.48748940229415894, "train/loss_error": 0.3759630024433136, "train/loss_total": 0.39826828241348267 }, { "epoch": 0.4701298701298701, "step": 181, "train/loss_ctc": 0.39702194929122925, "train/loss_error": 0.32495781779289246, "train/loss_total": 0.33937063813209534 }, { "epoch": 0.4727272727272727, "step": 182, "train/loss_ctc": 0.4024379551410675, "train/loss_error": 0.3362906873226166, "train/loss_total": 0.34952014684677124 }, { "epoch": 0.4753246753246753, "step": 183, "train/loss_ctc": 0.5769008994102478, "train/loss_error": 0.3889283537864685, "train/loss_total": 0.4265228509902954 }, { "epoch": 0.4779220779220779, "step": 184, "train/loss_ctc": 0.42895057797431946, "train/loss_error": 0.37379464507102966, "train/loss_total": 0.38482585549354553 }, { "epoch": 0.4805194805194805, "step": 185, "train/loss_ctc": 0.49320125579833984, "train/loss_error": 0.4561750888824463, "train/loss_total": 0.46358034014701843 }, { "epoch": 0.4831168831168831, "step": 186, "train/loss_ctc": 0.4256981611251831, "train/loss_error": 0.4000067710876465, "train/loss_total": 0.4051450490951538 }, { "epoch": 0.4857142857142857, "step": 187, "train/loss_ctc": 0.5347601771354675, "train/loss_error": 0.4745819866657257, "train/loss_total": 0.4866176247596741 }, { "epoch": 0.4883116883116883, "step": 188, "train/loss_ctc": 0.3705996870994568, "train/loss_error": 0.410880446434021, "train/loss_total": 0.4028242826461792 }, { "epoch": 0.4909090909090909, "step": 189, "train/loss_ctc": 0.5106054544448853, "train/loss_error": 0.3481922447681427, "train/loss_total": 0.38067489862442017 }, { "epoch": 0.4935064935064935, "grad_norm": 12283.8564453125, "learning_rate": 2.7070129870129872e-05, "loss": 0.4037, "step": 190 }, { "epoch": 0.4935064935064935, "step": 190, "train/loss_ctc": 0.6678175926208496, "train/loss_error": 0.4818575978279114, "train/loss_total": 0.5190496444702148 }, { "epoch": 0.4961038961038961, "step": 191, "train/loss_ctc": 0.48850974440574646, "train/loss_error": 0.43228405714035034, "train/loss_total": 0.4435291886329651 }, { "epoch": 0.4987012987012987, "step": 192, "train/loss_ctc": 0.42701205611228943, "train/loss_error": 0.41948139667510986, "train/loss_total": 0.4209875464439392 }, { "epoch": 0.5012987012987012, "step": 193, "train/loss_ctc": 0.5930408239364624, "train/loss_error": 0.40519317984580994, "train/loss_total": 0.44276273250579834 }, { "epoch": 0.5038961038961038, "step": 194, "train/loss_ctc": 0.4181671142578125, "train/loss_error": 0.4605180025100708, "train/loss_total": 0.45204782485961914 }, { "epoch": 0.5064935064935064, "step": 195, "train/loss_ctc": 0.4362983703613281, "train/loss_error": 0.4468342959880829, "train/loss_total": 0.4447271227836609 }, { "epoch": 0.509090909090909, "step": 196, "train/loss_ctc": 0.32031014561653137, "train/loss_error": 0.3704223334789276, "train/loss_total": 0.36039990186691284 }, { "epoch": 0.5116883116883116, "step": 197, "train/loss_ctc": 0.27914196252822876, "train/loss_error": 0.3769702911376953, "train/loss_total": 0.3574046194553375 }, { "epoch": 0.5142857142857142, "step": 198, "train/loss_ctc": 0.43773049116134644, "train/loss_error": 0.426435261964798, "train/loss_total": 0.42869433760643005 }, { "epoch": 0.5168831168831168, "step": 199, "train/loss_ctc": 0.565542459487915, "train/loss_error": 0.3974165916442871, "train/loss_total": 0.43104177713394165 }, { "epoch": 0.5194805194805194, "grad_norm": 13083.265625, "learning_rate": 2.6914285714285713e-05, "loss": 0.4301, "step": 200 }, { "epoch": 0.5194805194805194, "step": 200, "train/loss_ctc": 0.636461615562439, "train/loss_error": 0.564273476600647, "train/loss_total": 0.5787110924720764 }, { "epoch": 0.522077922077922, "step": 201, "train/loss_ctc": 0.3893153965473175, "train/loss_error": 0.4210117757320404, "train/loss_total": 0.41467249393463135 }, { "epoch": 0.5246753246753246, "step": 202, "train/loss_ctc": 0.5203227996826172, "train/loss_error": 0.35457298159599304, "train/loss_total": 0.3877229392528534 }, { "epoch": 0.5272727272727272, "step": 203, "train/loss_ctc": 0.48496514558792114, "train/loss_error": 0.3458348214626312, "train/loss_total": 0.3736608922481537 }, { "epoch": 0.5298701298701298, "step": 204, "train/loss_ctc": 0.51988685131073, "train/loss_error": 0.4417828619480133, "train/loss_total": 0.45740368962287903 }, { "epoch": 0.5324675324675324, "step": 205, "train/loss_ctc": 0.4620157480239868, "train/loss_error": 0.49517908692359924, "train/loss_total": 0.4885464310646057 }, { "epoch": 0.535064935064935, "step": 206, "train/loss_ctc": 0.663906455039978, "train/loss_error": 0.4551914632320404, "train/loss_total": 0.4969344735145569 }, { "epoch": 0.5376623376623376, "step": 207, "train/loss_ctc": 0.47060027718544006, "train/loss_error": 0.41259995102882385, "train/loss_total": 0.42420002818107605 }, { "epoch": 0.5402597402597402, "step": 208, "train/loss_ctc": 0.6171635389328003, "train/loss_error": 0.4134629964828491, "train/loss_total": 0.45420312881469727 }, { "epoch": 0.5428571428571428, "step": 209, "train/loss_ctc": 0.5674647092819214, "train/loss_error": 0.5152274370193481, "train/loss_total": 0.5256748795509338 }, { "epoch": 0.5454545454545454, "grad_norm": 12028.9619140625, "learning_rate": 2.6758441558441558e-05, "loss": 0.4602, "step": 210 }, { "epoch": 0.5454545454545454, "step": 210, "train/loss_ctc": 0.453008234500885, "train/loss_error": 0.4163069427013397, "train/loss_total": 0.4236472249031067 }, { "epoch": 0.548051948051948, "step": 211, "train/loss_ctc": 0.339884489774704, "train/loss_error": 0.3831951916217804, "train/loss_total": 0.3745330572128296 }, { "epoch": 0.5506493506493506, "step": 212, "train/loss_ctc": 0.7014938592910767, "train/loss_error": 0.4342828691005707, "train/loss_total": 0.48772507905960083 }, { "epoch": 0.5532467532467532, "step": 213, "train/loss_ctc": 0.5570043325424194, "train/loss_error": 0.4410848319530487, "train/loss_total": 0.4642687439918518 }, { "epoch": 0.5558441558441558, "step": 214, "train/loss_ctc": 0.4670715928077698, "train/loss_error": 0.4773990213871002, "train/loss_total": 0.475333571434021 }, { "epoch": 0.5584415584415584, "step": 215, "train/loss_ctc": 0.4445573389530182, "train/loss_error": 0.45350319147109985, "train/loss_total": 0.45171403884887695 }, { "epoch": 0.561038961038961, "step": 216, "train/loss_ctc": 0.6298832893371582, "train/loss_error": 0.4286919832229614, "train/loss_total": 0.4689302444458008 }, { "epoch": 0.5636363636363636, "step": 217, "train/loss_ctc": 0.43114539980888367, "train/loss_error": 0.44985586404800415, "train/loss_total": 0.44611379504203796 }, { "epoch": 0.5662337662337662, "step": 218, "train/loss_ctc": 0.48047664761543274, "train/loss_error": 0.3855012357234955, "train/loss_total": 0.40449631214141846 }, { "epoch": 0.5688311688311688, "step": 219, "train/loss_ctc": 0.678550124168396, "train/loss_error": 0.5056227445602417, "train/loss_total": 0.5402082204818726 }, { "epoch": 0.5714285714285714, "grad_norm": 13737.333984375, "learning_rate": 2.6602597402597403e-05, "loss": 0.4537, "step": 220 }, { "epoch": 0.5714285714285714, "step": 220, "train/loss_ctc": 0.49419650435447693, "train/loss_error": 0.42504972219467163, "train/loss_total": 0.4388791024684906 }, { "epoch": 0.574025974025974, "step": 221, "train/loss_ctc": 0.4729907214641571, "train/loss_error": 0.38991719484329224, "train/loss_total": 0.4065319001674652 }, { "epoch": 0.5766233766233766, "step": 222, "train/loss_ctc": 0.3768821954727173, "train/loss_error": 0.41616132855415344, "train/loss_total": 0.4083055257797241 }, { "epoch": 0.5792207792207792, "step": 223, "train/loss_ctc": 0.537695050239563, "train/loss_error": 0.524414598941803, "train/loss_total": 0.5270707011222839 }, { "epoch": 0.5818181818181818, "step": 224, "train/loss_ctc": 0.4202202558517456, "train/loss_error": 0.5184571743011475, "train/loss_total": 0.498809814453125 }, { "epoch": 0.5844155844155844, "step": 225, "train/loss_ctc": 0.6063609719276428, "train/loss_error": 0.45721229910850525, "train/loss_total": 0.48704203963279724 }, { "epoch": 0.587012987012987, "step": 226, "train/loss_ctc": 0.9125233888626099, "train/loss_error": 0.5300382971763611, "train/loss_total": 0.6065353155136108 }, { "epoch": 0.5896103896103896, "step": 227, "train/loss_ctc": 0.34471505880355835, "train/loss_error": 0.3318951725959778, "train/loss_total": 0.3344591557979584 }, { "epoch": 0.5922077922077922, "step": 228, "train/loss_ctc": 0.46127015352249146, "train/loss_error": 0.4096316397190094, "train/loss_total": 0.4199593663215637 }, { "epoch": 0.5948051948051948, "step": 229, "train/loss_ctc": 0.42463213205337524, "train/loss_error": 0.40696027874946594, "train/loss_total": 0.4104946553707123 }, { "epoch": 0.5974025974025974, "grad_norm": 14390.7626953125, "learning_rate": 2.6446753246753248e-05, "loss": 0.4538, "step": 230 }, { "epoch": 0.5974025974025974, "step": 230, "train/loss_ctc": 0.7257915735244751, "train/loss_error": 0.49610233306884766, "train/loss_total": 0.542040228843689 }, { "epoch": 0.6, "step": 231, "train/loss_ctc": 0.5684338808059692, "train/loss_error": 0.5185420513153076, "train/loss_total": 0.528520405292511 }, { "epoch": 0.6025974025974026, "step": 232, "train/loss_ctc": 0.45149388909339905, "train/loss_error": 0.4000440239906311, "train/loss_total": 0.4103339910507202 }, { "epoch": 0.6051948051948052, "step": 233, "train/loss_ctc": 0.4115580916404724, "train/loss_error": 0.4692164957523346, "train/loss_total": 0.45768484473228455 }, { "epoch": 0.6077922077922078, "step": 234, "train/loss_ctc": 0.44797223806381226, "train/loss_error": 0.4924132823944092, "train/loss_total": 0.4835250973701477 }, { "epoch": 0.6103896103896104, "step": 235, "train/loss_ctc": 0.40565401315689087, "train/loss_error": 0.40175703167915344, "train/loss_total": 0.40253642201423645 }, { "epoch": 0.612987012987013, "step": 236, "train/loss_ctc": 0.8054758310317993, "train/loss_error": 0.4911223351955414, "train/loss_total": 0.5539930462837219 }, { "epoch": 0.6155844155844156, "step": 237, "train/loss_ctc": 0.3872474730014801, "train/loss_error": 0.4035918414592743, "train/loss_total": 0.40032297372817993 }, { "epoch": 0.6181818181818182, "step": 238, "train/loss_ctc": 0.538615345954895, "train/loss_error": 0.4664310812950134, "train/loss_total": 0.4808679223060608 }, { "epoch": 0.6207792207792208, "step": 239, "train/loss_ctc": 0.38705024123191833, "train/loss_error": 0.4122221767902374, "train/loss_total": 0.4071877896785736 }, { "epoch": 0.6233766233766234, "grad_norm": 15926.8720703125, "learning_rate": 2.6290909090909092e-05, "loss": 0.4667, "step": 240 }, { "epoch": 0.6233766233766234, "step": 240, "train/loss_ctc": 0.41426828503608704, "train/loss_error": 0.4365774095058441, "train/loss_total": 0.4321156144142151 }, { "epoch": 0.625974025974026, "step": 241, "train/loss_ctc": 0.3699045479297638, "train/loss_error": 0.4457038342952728, "train/loss_total": 0.4305439591407776 }, { "epoch": 0.6285714285714286, "step": 242, "train/loss_ctc": 0.4492357671260834, "train/loss_error": 0.38027310371398926, "train/loss_total": 0.39406564831733704 }, { "epoch": 0.6311688311688312, "step": 243, "train/loss_ctc": 0.7188501954078674, "train/loss_error": 0.4636821746826172, "train/loss_total": 0.5147157907485962 }, { "epoch": 0.6337662337662338, "step": 244, "train/loss_ctc": 0.4499160349369049, "train/loss_error": 0.42798492312431335, "train/loss_total": 0.4323711395263672 }, { "epoch": 0.6363636363636364, "step": 245, "train/loss_ctc": 0.47223347425460815, "train/loss_error": 0.38021063804626465, "train/loss_total": 0.3986152112483978 }, { "epoch": 0.638961038961039, "step": 246, "train/loss_ctc": 0.4428882598876953, "train/loss_error": 0.38598814606666565, "train/loss_total": 0.3973681926727295 }, { "epoch": 0.6415584415584416, "step": 247, "train/loss_ctc": 0.5742753744125366, "train/loss_error": 0.5108848810195923, "train/loss_total": 0.5235629677772522 }, { "epoch": 0.6441558441558441, "step": 248, "train/loss_ctc": 0.6068830490112305, "train/loss_error": 0.4352036118507385, "train/loss_total": 0.46953949332237244 }, { "epoch": 0.6467532467532467, "step": 249, "train/loss_ctc": 0.436700701713562, "train/loss_error": 0.41439712047576904, "train/loss_total": 0.4188578426837921 }, { "epoch": 0.6493506493506493, "grad_norm": 10542.8017578125, "learning_rate": 2.6135064935064934e-05, "loss": 0.4412, "step": 250 }, { "epoch": 0.6493506493506493, "step": 250, "train/loss_ctc": 0.30312299728393555, "train/loss_error": 0.31911903619766235, "train/loss_total": 0.3159198462963104 }, { "epoch": 0.6519480519480519, "step": 251, "train/loss_ctc": 0.4237211048603058, "train/loss_error": 0.4296495318412781, "train/loss_total": 0.4284638464450836 }, { "epoch": 0.6545454545454545, "step": 252, "train/loss_ctc": 0.5786790251731873, "train/loss_error": 0.48712125420570374, "train/loss_total": 0.5054328441619873 }, { "epoch": 0.6571428571428571, "step": 253, "train/loss_ctc": 0.9540936350822449, "train/loss_error": 0.43748170137405396, "train/loss_total": 0.5408040881156921 }, { "epoch": 0.6597402597402597, "step": 254, "train/loss_ctc": 0.3593224883079529, "train/loss_error": 0.5246496200561523, "train/loss_total": 0.4915841817855835 }, { "epoch": 0.6623376623376623, "step": 255, "train/loss_ctc": 0.5179242491722107, "train/loss_error": 0.32054004073143005, "train/loss_total": 0.3600168824195862 }, { "epoch": 0.6649350649350649, "step": 256, "train/loss_ctc": 0.6017791032791138, "train/loss_error": 0.3417007625102997, "train/loss_total": 0.393716424703598 }, { "epoch": 0.6675324675324675, "step": 257, "train/loss_ctc": 0.5852051377296448, "train/loss_error": 0.45802414417266846, "train/loss_total": 0.48346033692359924 }, { "epoch": 0.6701298701298701, "step": 258, "train/loss_ctc": 0.6817496418952942, "train/loss_error": 0.3558428883552551, "train/loss_total": 0.42102426290512085 }, { "epoch": 0.6727272727272727, "step": 259, "train/loss_ctc": 0.9122869372367859, "train/loss_error": 0.446368008852005, "train/loss_total": 0.5395517945289612 }, { "epoch": 0.6753246753246753, "grad_norm": 19701.7109375, "learning_rate": 2.597922077922078e-05, "loss": 0.448, "step": 260 }, { "epoch": 0.6753246753246753, "step": 260, "train/loss_ctc": 0.4665682315826416, "train/loss_error": 0.4771418869495392, "train/loss_total": 0.4750271439552307 }, { "epoch": 0.6779220779220779, "step": 261, "train/loss_ctc": 0.4029620289802551, "train/loss_error": 0.40848374366760254, "train/loss_total": 0.4073793888092041 }, { "epoch": 0.6805194805194805, "step": 262, "train/loss_ctc": 0.3463667929172516, "train/loss_error": 0.4066176116466522, "train/loss_total": 0.39456745982170105 }, { "epoch": 0.6831168831168831, "step": 263, "train/loss_ctc": 0.5899643898010254, "train/loss_error": 0.48394784331321716, "train/loss_total": 0.5051511526107788 }, { "epoch": 0.6857142857142857, "step": 264, "train/loss_ctc": 0.5702892541885376, "train/loss_error": 0.5134250521659851, "train/loss_total": 0.5247979164123535 }, { "epoch": 0.6883116883116883, "step": 265, "train/loss_ctc": 0.5346920490264893, "train/loss_error": 0.4088478982448578, "train/loss_total": 0.43401673436164856 }, { "epoch": 0.6909090909090909, "step": 266, "train/loss_ctc": 0.4835936427116394, "train/loss_error": 0.405282586812973, "train/loss_total": 0.42094480991363525 }, { "epoch": 0.6935064935064935, "step": 267, "train/loss_ctc": 0.60472571849823, "train/loss_error": 0.4594457447528839, "train/loss_total": 0.48850175738334656 }, { "epoch": 0.6961038961038961, "step": 268, "train/loss_ctc": 0.533643364906311, "train/loss_error": 0.4977452754974365, "train/loss_total": 0.5049248933792114 }, { "epoch": 0.6987012987012987, "step": 269, "train/loss_ctc": 0.31537556648254395, "train/loss_error": 0.36827531456947327, "train/loss_total": 0.3576953709125519 }, { "epoch": 0.7012987012987013, "grad_norm": 9720.0087890625, "learning_rate": 2.5823376623376623e-05, "loss": 0.4513, "step": 270 }, { "epoch": 0.7012987012987013, "step": 270, "train/loss_ctc": 0.463398277759552, "train/loss_error": 0.458618700504303, "train/loss_total": 0.4595746099948883 }, { "epoch": 0.7038961038961039, "step": 271, "train/loss_ctc": 0.36472025513648987, "train/loss_error": 0.42337921261787415, "train/loss_total": 0.41164740920066833 }, { "epoch": 0.7064935064935065, "step": 272, "train/loss_ctc": 0.5855190753936768, "train/loss_error": 0.4171290993690491, "train/loss_total": 0.4508070945739746 }, { "epoch": 0.7090909090909091, "step": 273, "train/loss_ctc": 0.6197136044502258, "train/loss_error": 0.5339711904525757, "train/loss_total": 0.5511196851730347 }, { "epoch": 0.7116883116883117, "step": 274, "train/loss_ctc": 0.4075319468975067, "train/loss_error": 0.432346910238266, "train/loss_total": 0.4273839294910431 }, { "epoch": 0.7142857142857143, "step": 275, "train/loss_ctc": 0.43753695487976074, "train/loss_error": 0.4575241208076477, "train/loss_total": 0.45352670550346375 }, { "epoch": 0.7168831168831169, "step": 276, "train/loss_ctc": 0.425190806388855, "train/loss_error": 0.4822400212287903, "train/loss_total": 0.47083017230033875 }, { "epoch": 0.7194805194805195, "step": 277, "train/loss_ctc": 0.8450477123260498, "train/loss_error": 0.5041634440422058, "train/loss_total": 0.5723403096199036 }, { "epoch": 0.7220779220779221, "step": 278, "train/loss_ctc": 0.538542628288269, "train/loss_error": 0.38632532954216003, "train/loss_total": 0.41676878929138184 }, { "epoch": 0.7246753246753247, "step": 279, "train/loss_ctc": 0.6349853873252869, "train/loss_error": 0.4244873821735382, "train/loss_total": 0.46658700704574585 }, { "epoch": 0.7272727272727273, "grad_norm": 14758.689453125, "learning_rate": 2.5667532467532468e-05, "loss": 0.4681, "step": 280 }, { "epoch": 0.7272727272727273, "step": 280, "train/loss_ctc": 0.4618496894836426, "train/loss_error": 0.4753122329711914, "train/loss_total": 0.4726197421550751 }, { "epoch": 0.7298701298701299, "step": 281, "train/loss_ctc": 0.5127853155136108, "train/loss_error": 0.47361382842063904, "train/loss_total": 0.48144814372062683 }, { "epoch": 0.7324675324675325, "step": 282, "train/loss_ctc": 0.5186810493469238, "train/loss_error": 0.3540695309638977, "train/loss_total": 0.38699185848236084 }, { "epoch": 0.7350649350649351, "step": 283, "train/loss_ctc": 0.3553265929222107, "train/loss_error": 0.37179139256477356, "train/loss_total": 0.36849844455718994 }, { "epoch": 0.7376623376623377, "step": 284, "train/loss_ctc": 0.44730424880981445, "train/loss_error": 0.3714148700237274, "train/loss_total": 0.3865927457809448 }, { "epoch": 0.7402597402597403, "step": 285, "train/loss_ctc": 0.7690060138702393, "train/loss_error": 0.49174824357032776, "train/loss_total": 0.5471998453140259 }, { "epoch": 0.7428571428571429, "step": 286, "train/loss_ctc": 0.8050577044487, "train/loss_error": 0.4881535768508911, "train/loss_total": 0.5515344142913818 }, { "epoch": 0.7454545454545455, "step": 287, "train/loss_ctc": 0.5391800999641418, "train/loss_error": 0.36031076312065125, "train/loss_total": 0.39608466625213623 }, { "epoch": 0.7480519480519481, "step": 288, "train/loss_ctc": 0.563965380191803, "train/loss_error": 0.3809317350387573, "train/loss_total": 0.41753849387168884 }, { "epoch": 0.7506493506493507, "step": 289, "train/loss_ctc": 0.41731327772140503, "train/loss_error": 0.3997521698474884, "train/loss_total": 0.4032644033432007 }, { "epoch": 0.7532467532467533, "grad_norm": 12638.984375, "learning_rate": 2.5511688311688313e-05, "loss": 0.4412, "step": 290 }, { "epoch": 0.7532467532467533, "step": 290, "train/loss_ctc": 0.34575679898262024, "train/loss_error": 0.3878324031829834, "train/loss_total": 0.3794173002243042 }, { "epoch": 0.7558441558441559, "step": 291, "train/loss_ctc": 0.4165225028991699, "train/loss_error": 0.38631829619407654, "train/loss_total": 0.3923591375350952 }, { "epoch": 0.7584415584415585, "step": 292, "train/loss_ctc": 0.7680728435516357, "train/loss_error": 0.42085400223731995, "train/loss_total": 0.490297794342041 }, { "epoch": 0.7610389610389611, "step": 293, "train/loss_ctc": 0.3300358057022095, "train/loss_error": 0.368034303188324, "train/loss_total": 0.3604346215724945 }, { "epoch": 0.7636363636363637, "step": 294, "train/loss_ctc": 0.7681301832199097, "train/loss_error": 0.4499627649784088, "train/loss_total": 0.51359623670578 }, { "epoch": 0.7662337662337663, "step": 295, "train/loss_ctc": 0.43694934248924255, "train/loss_error": 0.4586755037307739, "train/loss_total": 0.4543302655220032 }, { "epoch": 0.7688311688311689, "step": 296, "train/loss_ctc": 0.5589066743850708, "train/loss_error": 0.5085680484771729, "train/loss_total": 0.5186358094215393 }, { "epoch": 0.7714285714285715, "step": 297, "train/loss_ctc": 0.597864031791687, "train/loss_error": 0.4301302433013916, "train/loss_total": 0.4636770188808441 }, { "epoch": 0.7740259740259741, "step": 298, "train/loss_ctc": 0.48409369587898254, "train/loss_error": 0.35313066840171814, "train/loss_total": 0.379323273897171 }, { "epoch": 0.7766233766233767, "step": 299, "train/loss_ctc": 0.4264247417449951, "train/loss_error": 0.3632739782333374, "train/loss_total": 0.3759041428565979 }, { "epoch": 0.7792207792207793, "grad_norm": 10845.140625, "learning_rate": 2.5355844155844154e-05, "loss": 0.4328, "step": 300 }, { "epoch": 0.7792207792207793, "step": 300, "train/loss_ctc": 0.406425416469574, "train/loss_error": 0.40227946639060974, "train/loss_total": 0.4031086564064026 }, { "epoch": 0.7818181818181819, "step": 301, "train/loss_ctc": 0.47284236550331116, "train/loss_error": 0.4230504333972931, "train/loss_total": 0.4330088496208191 }, { "epoch": 0.7844155844155845, "step": 302, "train/loss_ctc": 0.450747013092041, "train/loss_error": 0.46763136982917786, "train/loss_total": 0.4642544984817505 }, { "epoch": 0.787012987012987, "step": 303, "train/loss_ctc": 0.4047471284866333, "train/loss_error": 0.35155677795410156, "train/loss_total": 0.36219486594200134 }, { "epoch": 0.7896103896103897, "step": 304, "train/loss_ctc": 0.4897924065589905, "train/loss_error": 0.44595813751220703, "train/loss_total": 0.4547249972820282 }, { "epoch": 0.7922077922077922, "step": 305, "train/loss_ctc": 0.4190228283405304, "train/loss_error": 0.48409321904182434, "train/loss_total": 0.47107917070388794 }, { "epoch": 0.7948051948051948, "step": 306, "train/loss_ctc": 0.49051767587661743, "train/loss_error": 0.4499370753765106, "train/loss_total": 0.45805323123931885 }, { "epoch": 0.7974025974025974, "step": 307, "train/loss_ctc": 0.4805491864681244, "train/loss_error": 0.3931584358215332, "train/loss_total": 0.4106366038322449 }, { "epoch": 0.8, "step": 308, "train/loss_ctc": 0.6654008626937866, "train/loss_error": 0.4201053977012634, "train/loss_total": 0.46916449069976807 }, { "epoch": 0.8025974025974026, "step": 309, "train/loss_ctc": 0.5398128628730774, "train/loss_error": 0.426874041557312, "train/loss_total": 0.44946181774139404 }, { "epoch": 0.8051948051948052, "grad_norm": 14215.314453125, "learning_rate": 2.52e-05, "loss": 0.4376, "step": 310 }, { "epoch": 0.8051948051948052, "step": 310, "train/loss_ctc": 0.5723519325256348, "train/loss_error": 0.419887900352478, "train/loss_total": 0.45038071274757385 }, { "epoch": 0.8077922077922078, "step": 311, "train/loss_ctc": 0.5955041646957397, "train/loss_error": 0.47120341658592224, "train/loss_total": 0.49606359004974365 }, { "epoch": 0.8103896103896104, "step": 312, "train/loss_ctc": 0.8016881942749023, "train/loss_error": 0.4913681745529175, "train/loss_total": 0.5534321665763855 }, { "epoch": 0.812987012987013, "step": 313, "train/loss_ctc": 0.5920800566673279, "train/loss_error": 0.4914494752883911, "train/loss_total": 0.5115755796432495 }, { "epoch": 0.8155844155844156, "step": 314, "train/loss_ctc": 0.6778717041015625, "train/loss_error": 0.36184585094451904, "train/loss_total": 0.4250510334968567 }, { "epoch": 0.8181818181818182, "step": 315, "train/loss_ctc": 0.4037536382675171, "train/loss_error": 0.4391977787017822, "train/loss_total": 0.43210896849632263 }, { "epoch": 0.8207792207792208, "step": 316, "train/loss_ctc": 0.5494859218597412, "train/loss_error": 0.44129055738449097, "train/loss_total": 0.4629296362400055 }, { "epoch": 0.8233766233766234, "step": 317, "train/loss_ctc": 0.5633087158203125, "train/loss_error": 0.434218168258667, "train/loss_total": 0.4600362777709961 }, { "epoch": 0.825974025974026, "step": 318, "train/loss_ctc": 0.6497985124588013, "train/loss_error": 0.4523210823535919, "train/loss_total": 0.49181658029556274 }, { "epoch": 0.8285714285714286, "step": 319, "train/loss_ctc": 0.422196626663208, "train/loss_error": 0.3856998682022095, "train/loss_total": 0.39299923181533813 }, { "epoch": 0.8311688311688312, "grad_norm": 10593.2099609375, "learning_rate": 2.5044155844155844e-05, "loss": 0.4676, "step": 320 }, { "epoch": 0.8311688311688312, "step": 320, "train/loss_ctc": 0.495840847492218, "train/loss_error": 0.4014514982700348, "train/loss_total": 0.42032939195632935 }, { "epoch": 0.8337662337662337, "step": 321, "train/loss_ctc": 0.5657000541687012, "train/loss_error": 0.4909168481826782, "train/loss_total": 0.5058735013008118 }, { "epoch": 0.8363636363636363, "step": 322, "train/loss_ctc": 0.36684131622314453, "train/loss_error": 0.28804171085357666, "train/loss_total": 0.30380165576934814 }, { "epoch": 0.8389610389610389, "step": 323, "train/loss_ctc": 0.37701624631881714, "train/loss_error": 0.414108544588089, "train/loss_total": 0.4066900908946991 }, { "epoch": 0.8415584415584415, "step": 324, "train/loss_ctc": 0.512583315372467, "train/loss_error": 0.3851414918899536, "train/loss_total": 0.41062986850738525 }, { "epoch": 0.8441558441558441, "step": 325, "train/loss_ctc": 0.48052525520324707, "train/loss_error": 0.39575228095054626, "train/loss_total": 0.4127069115638733 }, { "epoch": 0.8467532467532467, "step": 326, "train/loss_ctc": 0.4500676393508911, "train/loss_error": 0.37638625502586365, "train/loss_total": 0.3911225497722626 }, { "epoch": 0.8493506493506493, "step": 327, "train/loss_ctc": 0.47372013330459595, "train/loss_error": 0.4833221137523651, "train/loss_total": 0.4814017117023468 }, { "epoch": 0.8519480519480519, "step": 328, "train/loss_ctc": 0.4860488474369049, "train/loss_error": 0.5253537893295288, "train/loss_total": 0.5174928307533264 }, { "epoch": 0.8545454545454545, "step": 329, "train/loss_ctc": 0.5093758702278137, "train/loss_error": 0.4163684546947479, "train/loss_total": 0.434969961643219 }, { "epoch": 0.8571428571428571, "grad_norm": 13013.47265625, "learning_rate": 2.488831168831169e-05, "loss": 0.4285, "step": 330 }, { "epoch": 0.8571428571428571, "step": 330, "train/loss_ctc": 0.5442733764648438, "train/loss_error": 0.48877957463264465, "train/loss_total": 0.4998783469200134 }, { "epoch": 0.8597402597402597, "step": 331, "train/loss_ctc": 0.45224183797836304, "train/loss_error": 0.48327121138572693, "train/loss_total": 0.4770653545856476 }, { "epoch": 0.8623376623376623, "step": 332, "train/loss_ctc": 0.5128874778747559, "train/loss_error": 0.43081575632095337, "train/loss_total": 0.44723013043403625 }, { "epoch": 0.8649350649350649, "step": 333, "train/loss_ctc": 0.38780996203422546, "train/loss_error": 0.43389105796813965, "train/loss_total": 0.4246748685836792 }, { "epoch": 0.8675324675324675, "step": 334, "train/loss_ctc": 0.7437102794647217, "train/loss_error": 0.44625651836395264, "train/loss_total": 0.5057472586631775 }, { "epoch": 0.8701298701298701, "step": 335, "train/loss_ctc": 0.4212670922279358, "train/loss_error": 0.5176345705986023, "train/loss_total": 0.49836108088493347 }, { "epoch": 0.8727272727272727, "step": 336, "train/loss_ctc": 0.47546958923339844, "train/loss_error": 0.42168936133384705, "train/loss_total": 0.4324454069137573 }, { "epoch": 0.8753246753246753, "step": 337, "train/loss_ctc": 0.5238399505615234, "train/loss_error": 0.30512845516204834, "train/loss_total": 0.34887075424194336 }, { "epoch": 0.8779220779220779, "step": 338, "train/loss_ctc": 0.73314368724823, "train/loss_error": 0.46629825234413147, "train/loss_total": 0.519667387008667 }, { "epoch": 0.8805194805194805, "step": 339, "train/loss_ctc": 0.6850738525390625, "train/loss_error": 0.4480302631855011, "train/loss_total": 0.49543899297714233 }, { "epoch": 0.8831168831168831, "grad_norm": 15793.3720703125, "learning_rate": 2.473246753246753e-05, "loss": 0.4649, "step": 340 }, { "epoch": 0.8831168831168831, "step": 340, "train/loss_ctc": 0.4764045476913452, "train/loss_error": 0.3963364064693451, "train/loss_total": 0.412350058555603 }, { "epoch": 0.8857142857142857, "step": 341, "train/loss_ctc": 0.54991614818573, "train/loss_error": 0.4962140619754791, "train/loss_total": 0.5069544911384583 }, { "epoch": 0.8883116883116883, "step": 342, "train/loss_ctc": 0.5448359251022339, "train/loss_error": 0.39246445894241333, "train/loss_total": 0.4229387640953064 }, { "epoch": 0.8909090909090909, "step": 343, "train/loss_ctc": 0.499101847410202, "train/loss_error": 0.38140690326690674, "train/loss_total": 0.40494590997695923 }, { "epoch": 0.8935064935064935, "step": 344, "train/loss_ctc": 0.44352850317955017, "train/loss_error": 0.43010228872299194, "train/loss_total": 0.43278753757476807 }, { "epoch": 0.8961038961038961, "step": 345, "train/loss_ctc": 0.7641052007675171, "train/loss_error": 0.40858760476112366, "train/loss_total": 0.47969114780426025 }, { "epoch": 0.8987012987012987, "step": 346, "train/loss_ctc": 0.3334445357322693, "train/loss_error": 0.38666942715644836, "train/loss_total": 0.3760244846343994 }, { "epoch": 0.9012987012987013, "step": 347, "train/loss_ctc": 0.39917176961898804, "train/loss_error": 0.4427673816680908, "train/loss_total": 0.43404829502105713 }, { "epoch": 0.9038961038961039, "step": 348, "train/loss_ctc": 0.42667752504348755, "train/loss_error": 0.4083763062953949, "train/loss_total": 0.4120365381240845 }, { "epoch": 0.9064935064935065, "step": 349, "train/loss_ctc": 0.42429712414741516, "train/loss_error": 0.31961530447006226, "train/loss_total": 0.3405516743659973 }, { "epoch": 0.9090909090909091, "grad_norm": 11329.1767578125, "learning_rate": 2.4576623376623375e-05, "loss": 0.4222, "step": 350 }, { "epoch": 0.9090909090909091, "step": 350, "train/loss_ctc": 0.8168613910675049, "train/loss_error": 0.5108219981193542, "train/loss_total": 0.5720298886299133 }, { "epoch": 0.9116883116883117, "step": 351, "train/loss_ctc": 0.674929141998291, "train/loss_error": 0.43722423911094666, "train/loss_total": 0.4847652316093445 }, { "epoch": 0.9142857142857143, "step": 352, "train/loss_ctc": 0.4130552411079407, "train/loss_error": 0.42760220170021057, "train/loss_total": 0.4246928095817566 }, { "epoch": 0.9168831168831169, "step": 353, "train/loss_ctc": 0.5909594297409058, "train/loss_error": 0.4452938139438629, "train/loss_total": 0.4744269549846649 }, { "epoch": 0.9194805194805195, "step": 354, "train/loss_ctc": 0.4876194894313812, "train/loss_error": 0.4155755639076233, "train/loss_total": 0.42998436093330383 }, { "epoch": 0.922077922077922, "step": 355, "train/loss_ctc": 0.7593616247177124, "train/loss_error": 0.5306193232536316, "train/loss_total": 0.5763677954673767 }, { "epoch": 0.9246753246753247, "step": 356, "train/loss_ctc": 0.5891261696815491, "train/loss_error": 0.4047496020793915, "train/loss_total": 0.4416249394416809 }, { "epoch": 0.9272727272727272, "step": 357, "train/loss_ctc": 0.5428358316421509, "train/loss_error": 0.369256854057312, "train/loss_total": 0.40397265553474426 }, { "epoch": 0.9298701298701298, "step": 358, "train/loss_ctc": 0.588638424873352, "train/loss_error": 0.4503944218158722, "train/loss_total": 0.47804322838783264 }, { "epoch": 0.9324675324675324, "step": 359, "train/loss_ctc": 0.40572020411491394, "train/loss_error": 0.4550376534461975, "train/loss_total": 0.4451741576194763 }, { "epoch": 0.935064935064935, "grad_norm": 13438.7197265625, "learning_rate": 2.442077922077922e-05, "loss": 0.4731, "step": 360 }, { "epoch": 0.935064935064935, "step": 360, "train/loss_ctc": 0.5036501884460449, "train/loss_error": 0.5259048938751221, "train/loss_total": 0.5214539766311646 }, { "epoch": 0.9376623376623376, "step": 361, "train/loss_ctc": 0.44653573632240295, "train/loss_error": 0.4333121180534363, "train/loss_total": 0.4359568655490875 }, { "epoch": 0.9402597402597402, "step": 362, "train/loss_ctc": 0.441516637802124, "train/loss_error": 0.3856276273727417, "train/loss_total": 0.39680543541908264 }, { "epoch": 0.9428571428571428, "step": 363, "train/loss_ctc": 0.442773699760437, "train/loss_error": 0.4009542763233185, "train/loss_total": 0.4093181788921356 }, { "epoch": 0.9454545454545454, "step": 364, "train/loss_ctc": 0.4823257327079773, "train/loss_error": 0.43804776668548584, "train/loss_total": 0.4469033479690552 }, { "epoch": 0.948051948051948, "step": 365, "train/loss_ctc": 0.5254629850387573, "train/loss_error": 0.3498455584049225, "train/loss_total": 0.3849690556526184 }, { "epoch": 0.9506493506493506, "step": 366, "train/loss_ctc": 0.4761037528514862, "train/loss_error": 0.4413817524909973, "train/loss_total": 0.44832614064216614 }, { "epoch": 0.9532467532467532, "step": 367, "train/loss_ctc": 0.49694541096687317, "train/loss_error": 0.40724343061447144, "train/loss_total": 0.42518383264541626 }, { "epoch": 0.9558441558441558, "step": 368, "train/loss_ctc": 0.592665433883667, "train/loss_error": 0.4636036157608032, "train/loss_total": 0.4894160032272339 }, { "epoch": 0.9584415584415584, "step": 369, "train/loss_ctc": 0.6820040941238403, "train/loss_error": 0.4353382885456085, "train/loss_total": 0.4846714437007904 }, { "epoch": 0.961038961038961, "grad_norm": 21791.73046875, "learning_rate": 2.4264935064935065e-05, "loss": 0.4443, "step": 370 }, { "epoch": 0.961038961038961, "step": 370, "train/loss_ctc": 0.29406315088272095, "train/loss_error": 0.4127374589443207, "train/loss_total": 0.38900259137153625 }, { "epoch": 0.9636363636363636, "step": 371, "train/loss_ctc": 0.5122777819633484, "train/loss_error": 0.43552371859550476, "train/loss_total": 0.45087453722953796 }, { "epoch": 0.9662337662337662, "step": 372, "train/loss_ctc": 0.5106490850448608, "train/loss_error": 0.39982226490974426, "train/loss_total": 0.4219876229763031 }, { "epoch": 0.9688311688311688, "step": 373, "train/loss_ctc": 0.6345514059066772, "train/loss_error": 0.4658709168434143, "train/loss_total": 0.49960702657699585 }, { "epoch": 0.9714285714285714, "step": 374, "train/loss_ctc": 0.5672173500061035, "train/loss_error": 0.41468313336372375, "train/loss_total": 0.4451899826526642 }, { "epoch": 0.974025974025974, "step": 375, "train/loss_ctc": 0.6495596766471863, "train/loss_error": 0.4442780613899231, "train/loss_total": 0.4853343963623047 }, { "epoch": 0.9766233766233766, "step": 376, "train/loss_ctc": 0.7115393280982971, "train/loss_error": 0.45198410749435425, "train/loss_total": 0.5038951635360718 }, { "epoch": 0.9792207792207792, "step": 377, "train/loss_ctc": 0.4175011217594147, "train/loss_error": 0.4126114547252655, "train/loss_total": 0.4135894179344177 }, { "epoch": 0.9818181818181818, "step": 378, "train/loss_ctc": 0.5740886330604553, "train/loss_error": 0.4767829477787018, "train/loss_total": 0.4962441027164459 }, { "epoch": 0.9844155844155844, "step": 379, "train/loss_ctc": 0.7323096990585327, "train/loss_error": 0.5038710832595825, "train/loss_total": 0.5495588183403015 }, { "epoch": 0.987012987012987, "grad_norm": 14975.7109375, "learning_rate": 2.410909090909091e-05, "loss": 0.4655, "step": 380 }, { "epoch": 0.987012987012987, "step": 380, "train/loss_ctc": 0.30461806058883667, "train/loss_error": 0.32928234338760376, "train/loss_total": 0.3243494927883148 }, { "epoch": 0.9896103896103896, "step": 381, "train/loss_ctc": 0.3463175296783447, "train/loss_error": 0.4094979166984558, "train/loss_total": 0.39686185121536255 }, { "epoch": 0.9922077922077922, "step": 382, "train/loss_ctc": 0.48353156447410583, "train/loss_error": 0.49711835384368896, "train/loss_total": 0.4944009780883789 }, { "epoch": 0.9948051948051948, "step": 383, "train/loss_ctc": 1.0355541706085205, "train/loss_error": 0.4756161570549011, "train/loss_total": 0.587603747844696 }, { "epoch": 0.9974025974025974, "step": 384, "train/loss_ctc": 0.5864169597625732, "train/loss_error": 0.337484747171402, "train/loss_total": 0.3872711956501007 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.5850890278816223, "val/loss_error": 0.15646322071552277, "val/loss_total": 0.24218837916851044 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.6685696840286255, "val/loss_error": 0.5376557111740112, "val/loss_total": 0.563838541507721 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.13284416496753693, "val/loss_error": 0.3624674081802368, "val/loss_total": 0.31654277443885803 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.13311293721199036, "val/loss_error": 0.1707940399646759, "val/loss_total": 0.16325782239437103 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.1470770388841629, "val/loss_error": 0.2709096372127533, "val/loss_total": 0.2461431324481964 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.18062393367290497, "val/loss_error": 0.30556830763816833, "val/loss_total": 0.28057944774627686 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.21850554645061493, "val/loss_error": 0.5827322602272034, "val/loss_total": 0.5098869204521179 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.356916606426239, "val/loss_error": 0.3648688793182373, "val/loss_total": 0.36327844858169556 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.4766559600830078, "val/loss_error": 0.3500139117240906, "val/loss_total": 0.37534233927726746 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.23494145274162292, "val/loss_error": 0.36613327264785767, "val/loss_total": 0.3398948907852173 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.1751912236213684, "val/loss_error": 0.16532136499881744, "val/loss_total": 0.16729533672332764 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.34874099493026733, "val/loss_error": 0.3987298607826233, "val/loss_total": 0.38873207569122314 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.8347690105438232, "val/loss_error": 0.4350520074367523, "val/loss_total": 0.5149954557418823 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.835744321346283, "val/loss_error": 0.6179299354553223, "val/loss_total": 0.6614928245544434 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.7594460248947144, "val/loss_error": 0.46163326501846313, "val/loss_total": 0.5211958289146423 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.5759822130203247, "val/loss_error": 0.5922759771347046, "val/loss_total": 0.5890172123908997 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.11256518214941025, "val/loss_error": 0.2676049470901489, "val/loss_total": 0.2365969866514206 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.2038545310497284, "val/loss_error": 0.3085962235927582, "val/loss_total": 0.28764790296554565 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.0906648263335228, "val/loss_error": 0.37408313155174255, "val/loss_total": 0.3173994719982147 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.5147131681442261, "val/loss_error": 0.554377019405365, "val/loss_total": 0.5464442372322083 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.8135906457901001, "val/loss_error": 0.6403398513793945, "val/loss_total": 0.6749899983406067 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.28343501687049866, "val/loss_error": 0.48009228706359863, "val/loss_total": 0.4407608211040497 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.28137001395225525, "val/loss_error": 0.5100228190422058, "val/loss_total": 0.4642922580242157 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.22347809374332428, "val/loss_error": 0.2700389623641968, "val/loss_total": 0.26072680950164795 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.3573582172393799, "val/loss_error": 0.5725762248039246, "val/loss_total": 0.5295326113700867 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.5307639241218567, "val/loss_error": 0.6621344685554504, "val/loss_total": 0.6358603835105896 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.4015779197216034, "val/loss_error": 0.439198762178421, "val/loss_total": 0.431674599647522 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.1297222077846527, "val/loss_error": 0.25667065382003784, "val/loss_total": 0.23128096759319305 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.04224799573421478, "val/loss_error": 0.10972410440444946, "val/loss_total": 0.09622888267040253 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.3080141842365265, "val/loss_error": 0.5552491545677185, "val/loss_total": 0.5058021545410156 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.1305800974369049, "val/loss_error": 0.15846461057662964, "val/loss_total": 0.15288770198822021 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.13080619275569916, "val/loss_error": 0.3178443908691406, "val/loss_total": 0.28043675422668457 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.06283526867628098, "val/loss_error": 0.1691150665283203, "val/loss_total": 0.1478591114282608 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.5783337354660034, "val/loss_error": 0.7891045212745667, "val/loss_total": 0.7469503879547119 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.08126572519540787, "val/loss_error": 0.39645543694496155, "val/loss_total": 0.33341750502586365 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.17227581143379211, "val/loss_error": 0.3936431109905243, "val/loss_total": 0.3493696451187134 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.45474380254745483, "val/loss_error": 0.6438060402870178, "val/loss_total": 0.6059936285018921 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.519850492477417, "val/loss_error": 0.41113969683647156, "val/loss_total": 0.4328818619251251 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.24529731273651123, "val/loss_error": 0.5152789950370789, "val/loss_total": 0.4612826406955719 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.6518040299415588, "val/loss_error": 0.6925783753395081, "val/loss_total": 0.684423565864563 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.3141807019710541, "val/loss_error": 0.32309815287590027, "val/loss_total": 0.32131466269493103 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.33116284012794495, "val/loss_error": 0.5101414918899536, "val/loss_total": 0.47434577345848083 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.4294654130935669, "val/loss_error": 1.0855071544647217, "val/loss_total": 0.9542988538742065 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.6538249254226685, "val/loss_error": 0.5530783534049988, "val/loss_total": 0.5732276439666748 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.580619752407074, "val/loss_error": 0.3730297088623047, "val/loss_total": 0.41454771161079407 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.3796667158603668, "val/loss_error": 0.6666717529296875, "val/loss_total": 0.6092707514762878 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.6612417101860046, "val/loss_error": 0.46050551533699036, "val/loss_total": 0.5006527900695801 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.4188068211078644, "val/loss_error": 0.5232993960380554, "val/loss_total": 0.5024008750915527 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.5089583396911621, "val/loss_error": 0.3149538040161133, "val/loss_total": 0.3537547290325165 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.7816178798675537, "val/loss_error": 0.9584941267967224, "val/loss_total": 0.9231189489364624 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.4131067097187042, "val/loss_error": 0.8739575743675232, "val/loss_total": 0.7817873954772949 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.3217580318450928, "val/loss_error": 0.46770167350769043, "val/loss_total": 0.4385129511356354 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.401536226272583, "val/loss_error": 0.43853759765625, "val/loss_total": 0.4311373233795166 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.4263943135738373, "val/loss_error": 1.045138955116272, "val/loss_total": 0.9213900566101074 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.16100935637950897, "val/loss_error": 0.3619976043701172, "val/loss_total": 0.3217999339103699 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.20381803810596466, "val/loss_error": 0.5157153606414795, "val/loss_total": 0.4533359110355377 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.10764407366514206, "val/loss_error": 0.35759326815605164, "val/loss_total": 0.3076034188270569 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.4498700201511383, "val/loss_error": 0.36949652433395386, "val/loss_total": 0.3855712413787842 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.7353336811065674, "val/loss_error": 0.7266898155212402, "val/loss_total": 0.7284185886383057 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.02823951281607151, "val/loss_error": 0.13013508915901184, "val/loss_total": 0.1097559779882431 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.7872844934463501, "val/loss_error": 0.5519052147865295, "val/loss_total": 0.5989810824394226 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.6617932915687561, "val/loss_error": 0.4546843469142914, "val/loss_total": 0.4961061477661133 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.0847245454788208, "val/loss_error": 0.11163110285997391, "val/loss_total": 0.10624979436397552 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.20844914019107819, "val/loss_error": 0.3706419765949249, "val/loss_total": 0.33820343017578125 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.13692471385002136, "val/loss_error": 0.22884461283683777, "val/loss_total": 0.2104606330394745 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.2373884618282318, "val/loss_error": 0.47003838419914246, "val/loss_total": 0.4235084056854248 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.29863592982292175, "val/loss_error": 0.2775871753692627, "val/loss_total": 0.281796932220459 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.5868434309959412, "val/loss_error": 0.6202366352081299, "val/loss_total": 0.6135579943656921 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.5083684325218201, "val/loss_error": 0.5082134008407593, "val/loss_total": 0.5082443952560425 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.4926762580871582, "val/loss_error": 0.49914610385894775, "val/loss_total": 0.4978521466255188 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.20757508277893066, "val/loss_error": 0.2950453460216522, "val/loss_total": 0.2775512933731079 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.4551757872104645, "val/loss_error": 0.5115163326263428, "val/loss_total": 0.5002481937408447 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.6905673146247864, "val/loss_error": 0.6658170819282532, "val/loss_total": 0.6707671880722046 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.2695978879928589, "val/loss_error": 0.3769007623195648, "val/loss_total": 0.3554401993751526 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.24921521544456482, "val/loss_error": 0.3325859308242798, "val/loss_total": 0.31591179966926575 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.03269151225686073, "val/loss_error": 0.30631715059280396, "val/loss_total": 0.2515920400619507 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.7920424938201904, "val/loss_error": 0.5815716981887817, "val/loss_total": 0.6236658692359924 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.545952558517456, "val/loss_error": 0.341249018907547, "val/loss_total": 0.3821897506713867 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.6585754156112671, "val/loss_error": 0.3991970419883728, "val/loss_total": 0.45107272267341614 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.0036876052618026733, "val/loss_error": 0.43654805421829224, "val/loss_total": 0.34997597336769104 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.3994354009628296, "val/loss_error": 0.29038527607917786, "val/loss_total": 0.3121953010559082 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.030672434717416763, "val/loss_error": 0.10334896296262741, "val/loss_total": 0.0888136625289917 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.26116716861724854, "val/loss_error": 0.3654705584049225, "val/loss_total": 0.3446098864078522 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.317179799079895, "val/loss_error": 0.34296759963035583, "val/loss_total": 0.33781006932258606 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.41330334544181824, "val/loss_error": 0.5662981271743774, "val/loss_total": 0.535699188709259 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.4912623167037964, "val/loss_error": 0.6049366593360901, "val/loss_total": 0.5822017788887024 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.02418038807809353, "val/loss_error": 0.15599827468395233, "val/loss_total": 0.12963469326496124 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.5373027324676514, "val/loss_error": 0.48361527919769287, "val/loss_total": 0.494352787733078 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.29020214080810547, "val/loss_error": 0.3223745822906494, "val/loss_total": 0.31594011187553406 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.40084123611450195, "val/loss_error": 0.40404513478279114, "val/loss_total": 0.4034043550491333 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.3198506236076355, "val/loss_error": 0.5083996057510376, "val/loss_total": 0.4706898033618927 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.7516082525253296, "val/loss_error": 0.4623936414718628, "val/loss_total": 0.5202365517616272 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.26654475927352905, "val/loss_error": 0.08989621698856354, "val/loss_total": 0.12522593140602112 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.23737557232379913, "val/loss_error": 0.3644457459449768, "val/loss_total": 0.3390316963195801 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.6498298048973083, "val/loss_error": 0.6004876494407654, "val/loss_total": 0.6103560924530029 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.2141427993774414, "val/loss_error": 0.21717660129070282, "val/loss_total": 0.21656984090805054 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.07836173474788666, "val/loss_error": 0.4467732906341553, "val/loss_total": 0.3730909824371338 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.6660016179084778, "val/loss_error": 0.5441977381706238, "val/loss_total": 0.5685585141181946 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.15959042310714722, "val/loss_error": 0.30841198563575745, "val/loss_total": 0.27864769101142883 }, { "epoch": 1.0, "step": 385, "val/loss_ctc": 0.20690491795539856, "val/loss_error": 0.26009178161621094, "val/loss_total": 0.24945440888404846 }, { "epoch": 1.0, "eval_eval/f1_0": 0.5247524976730347, "eval_eval/f1_1": 0.9321314692497253, "eval_eval/precision_0": 0.5729729533195496, "eval_eval/precision_1": 0.9210618138313293, "eval_eval/recall_0": 0.4840182662010193, "eval_eval/recall_1": 0.9434704780578613, "eval_eval/wer": 0.15265448639550036, "eval_loss": 0.42154571413993835, "eval_runtime": 7.6389, "eval_samples_per_second": 13.091, "eval_steps_per_second": 13.091, "step": 385 }, { "epoch": 1.0, "step": 385, "train/loss_ctc": 0.5850752592086792, "train/loss_error": 0.4817230701446533, "train/loss_total": 0.5023934841156006 }, { "epoch": 1.0025974025974025, "step": 386, "train/loss_ctc": 0.3563157916069031, "train/loss_error": 0.392827570438385, "train/loss_total": 0.3855252265930176 }, { "epoch": 1.0051948051948052, "step": 387, "train/loss_ctc": 0.8473033905029297, "train/loss_error": 0.4803182780742645, "train/loss_total": 0.5537153482437134 }, { "epoch": 1.0077922077922077, "step": 388, "train/loss_ctc": 0.7135196924209595, "train/loss_error": 0.49780765175819397, "train/loss_total": 0.5409500598907471 }, { "epoch": 1.0103896103896104, "step": 389, "train/loss_ctc": 0.38363614678382874, "train/loss_error": 0.33372756838798523, "train/loss_total": 0.3437092900276184 }, { "epoch": 1.0129870129870129, "grad_norm": 9952.89453125, "learning_rate": 2.395324675324675e-05, "loss": 0.4517, "step": 390 }, { "epoch": 1.0129870129870129, "step": 390, "train/loss_ctc": 0.48662590980529785, "train/loss_error": 0.45378348231315613, "train/loss_total": 0.46035197377204895 }, { "epoch": 1.0155844155844156, "step": 391, "train/loss_ctc": 0.4064636826515198, "train/loss_error": 0.3271266520023346, "train/loss_total": 0.3429940640926361 }, { "epoch": 1.018181818181818, "step": 392, "train/loss_ctc": 0.5792995095252991, "train/loss_error": 0.41508299112319946, "train/loss_total": 0.44792628288269043 }, { "epoch": 1.0207792207792208, "step": 393, "train/loss_ctc": 0.40563976764678955, "train/loss_error": 0.38682323694229126, "train/loss_total": 0.3905865550041199 }, { "epoch": 1.0233766233766233, "step": 394, "train/loss_ctc": 0.4558628797531128, "train/loss_error": 0.4072920083999634, "train/loss_total": 0.4170061945915222 }, { "epoch": 1.025974025974026, "step": 395, "train/loss_ctc": 0.311847984790802, "train/loss_error": 0.35666319727897644, "train/loss_total": 0.3477001488208771 }, { "epoch": 1.0285714285714285, "step": 396, "train/loss_ctc": 0.5324610471725464, "train/loss_error": 0.38994038105010986, "train/loss_total": 0.41844454407691956 }, { "epoch": 1.0311688311688312, "step": 397, "train/loss_ctc": 0.42905381321907043, "train/loss_error": 0.42365768551826477, "train/loss_total": 0.4247369170188904 }, { "epoch": 1.0337662337662337, "step": 398, "train/loss_ctc": 0.5141938924789429, "train/loss_error": 0.41634494066238403, "train/loss_total": 0.4359147548675537 }, { "epoch": 1.0363636363636364, "step": 399, "train/loss_ctc": 0.609045684337616, "train/loss_error": 0.43339407444000244, "train/loss_total": 0.46852439641952515 }, { "epoch": 1.0389610389610389, "grad_norm": 12613.462890625, "learning_rate": 2.3797402597402596e-05, "loss": 0.4154, "step": 400 }, { "epoch": 1.0389610389610389, "step": 400, "train/loss_ctc": 0.3778412938117981, "train/loss_error": 0.30264970660209656, "train/loss_total": 0.3176880478858948 }, { "epoch": 1.0415584415584416, "step": 401, "train/loss_ctc": 0.6499009728431702, "train/loss_error": 0.49926242232322693, "train/loss_total": 0.5293901562690735 }, { "epoch": 1.044155844155844, "step": 402, "train/loss_ctc": 0.5761820673942566, "train/loss_error": 0.39305049180984497, "train/loss_total": 0.4296768307685852 }, { "epoch": 1.0467532467532468, "step": 403, "train/loss_ctc": 0.7358410954475403, "train/loss_error": 0.4943608343601227, "train/loss_total": 0.5426568984985352 }, { "epoch": 1.0493506493506493, "step": 404, "train/loss_ctc": 0.3814470171928406, "train/loss_error": 0.3702351152896881, "train/loss_total": 0.3724775016307831 }, { "epoch": 1.051948051948052, "step": 405, "train/loss_ctc": 0.6227798461914062, "train/loss_error": 0.3675815165042877, "train/loss_total": 0.4186211824417114 }, { "epoch": 1.0545454545454545, "step": 406, "train/loss_ctc": 0.4686512351036072, "train/loss_error": 0.5046194195747375, "train/loss_total": 0.49742579460144043 }, { "epoch": 1.0571428571428572, "step": 407, "train/loss_ctc": 0.6435899138450623, "train/loss_error": 0.533774733543396, "train/loss_total": 0.5557377934455872 }, { "epoch": 1.0597402597402596, "step": 408, "train/loss_ctc": 0.5738219022750854, "train/loss_error": 0.4182100296020508, "train/loss_total": 0.44933241605758667 }, { "epoch": 1.0623376623376624, "step": 409, "train/loss_ctc": 0.5217333436012268, "train/loss_error": 0.3972095251083374, "train/loss_total": 0.4221142828464508 }, { "epoch": 1.0649350649350648, "grad_norm": 11580.0888671875, "learning_rate": 2.364155844155844e-05, "loss": 0.4535, "step": 410 }, { "epoch": 1.0649350649350648, "step": 410, "train/loss_ctc": 0.6285779476165771, "train/loss_error": 0.3911778926849365, "train/loss_total": 0.4386579394340515 }, { "epoch": 1.0675324675324676, "step": 411, "train/loss_ctc": 0.37928324937820435, "train/loss_error": 0.41111281514167786, "train/loss_total": 0.4047469198703766 }, { "epoch": 1.07012987012987, "step": 412, "train/loss_ctc": 0.4571111798286438, "train/loss_error": 0.4999127686023712, "train/loss_total": 0.4913524389266968 }, { "epoch": 1.0727272727272728, "step": 413, "train/loss_ctc": 0.6270039677619934, "train/loss_error": 0.4042035937309265, "train/loss_total": 0.4487636685371399 }, { "epoch": 1.0753246753246752, "step": 414, "train/loss_ctc": 0.5108661651611328, "train/loss_error": 0.35521191358566284, "train/loss_total": 0.38634276390075684 }, { "epoch": 1.077922077922078, "step": 415, "train/loss_ctc": 0.46381065249443054, "train/loss_error": 0.5136527419090271, "train/loss_total": 0.5036843419075012 }, { "epoch": 1.0805194805194804, "step": 416, "train/loss_ctc": 0.5383804440498352, "train/loss_error": 0.44796475768089294, "train/loss_total": 0.46604791283607483 }, { "epoch": 1.0831168831168831, "step": 417, "train/loss_ctc": 0.3633260130882263, "train/loss_error": 0.42254897952079773, "train/loss_total": 0.4107044041156769 }, { "epoch": 1.0857142857142856, "step": 418, "train/loss_ctc": 0.7021875381469727, "train/loss_error": 0.44592368602752686, "train/loss_total": 0.49717646837234497 }, { "epoch": 1.0883116883116883, "step": 419, "train/loss_ctc": 0.40766826272010803, "train/loss_error": 0.4724121391773224, "train/loss_total": 0.45946335792541504 }, { "epoch": 1.0909090909090908, "grad_norm": 11784.087890625, "learning_rate": 2.3485714285714285e-05, "loss": 0.4507, "step": 420 }, { "epoch": 1.0909090909090908, "step": 420, "train/loss_ctc": 0.5773164629936218, "train/loss_error": 0.5482038259506226, "train/loss_total": 0.5540263652801514 }, { "epoch": 1.0935064935064935, "step": 421, "train/loss_ctc": 0.43904563784599304, "train/loss_error": 0.4022524058818817, "train/loss_total": 0.4096110463142395 }, { "epoch": 1.096103896103896, "step": 422, "train/loss_ctc": 0.5040805339813232, "train/loss_error": 0.4167911410331726, "train/loss_total": 0.43424901366233826 }, { "epoch": 1.0987012987012987, "step": 423, "train/loss_ctc": 0.4365085959434509, "train/loss_error": 0.479126513004303, "train/loss_total": 0.47060295939445496 }, { "epoch": 1.1012987012987012, "step": 424, "train/loss_ctc": 0.5181259512901306, "train/loss_error": 0.5001798272132874, "train/loss_total": 0.503769040107727 }, { "epoch": 1.103896103896104, "step": 425, "train/loss_ctc": 0.3145093023777008, "train/loss_error": 0.3562110662460327, "train/loss_total": 0.34787070751190186 }, { "epoch": 1.1064935064935064, "step": 426, "train/loss_ctc": 0.43216249346733093, "train/loss_error": 0.3918277621269226, "train/loss_total": 0.39989471435546875 }, { "epoch": 1.1090909090909091, "step": 427, "train/loss_ctc": 0.544712483882904, "train/loss_error": 0.4414825439453125, "train/loss_total": 0.46212854981422424 }, { "epoch": 1.1116883116883116, "step": 428, "train/loss_ctc": 0.4181838035583496, "train/loss_error": 0.3882819414138794, "train/loss_total": 0.39426231384277344 }, { "epoch": 1.1142857142857143, "step": 429, "train/loss_ctc": 0.47637584805488586, "train/loss_error": 0.396330863237381, "train/loss_total": 0.41233986616134644 }, { "epoch": 1.1168831168831168, "grad_norm": 13364.294921875, "learning_rate": 2.332987012987013e-05, "loss": 0.4389, "step": 430 }, { "epoch": 1.1168831168831168, "step": 430, "train/loss_ctc": 0.4194212555885315, "train/loss_error": 0.44346168637275696, "train/loss_total": 0.4386535882949829 }, { "epoch": 1.1194805194805195, "step": 431, "train/loss_ctc": 0.5409796237945557, "train/loss_error": 0.47470971941947937, "train/loss_total": 0.4879637062549591 }, { "epoch": 1.122077922077922, "step": 432, "train/loss_ctc": 1.0257964134216309, "train/loss_error": 0.5053595900535583, "train/loss_total": 0.6094469428062439 }, { "epoch": 1.1246753246753247, "step": 433, "train/loss_ctc": 0.5376659631729126, "train/loss_error": 0.42248281836509705, "train/loss_total": 0.44551944732666016 }, { "epoch": 1.1272727272727272, "step": 434, "train/loss_ctc": 0.5304258465766907, "train/loss_error": 0.4499804675579071, "train/loss_total": 0.4660695493221283 }, { "epoch": 1.12987012987013, "step": 435, "train/loss_ctc": 0.39612317085266113, "train/loss_error": 0.3746173679828644, "train/loss_total": 0.3789185583591461 }, { "epoch": 1.1324675324675324, "step": 436, "train/loss_ctc": 0.4370388686656952, "train/loss_error": 0.4173071086406708, "train/loss_total": 0.4212534725666046 }, { "epoch": 1.135064935064935, "step": 437, "train/loss_ctc": 0.6178821921348572, "train/loss_error": 0.536853015422821, "train/loss_total": 0.5530588626861572 }, { "epoch": 1.1376623376623376, "step": 438, "train/loss_ctc": 0.43816256523132324, "train/loss_error": 0.43502214550971985, "train/loss_total": 0.4356502294540405 }, { "epoch": 1.1402597402597403, "step": 439, "train/loss_ctc": 0.46879735589027405, "train/loss_error": 0.5107969045639038, "train/loss_total": 0.5023970007896423 }, { "epoch": 1.1428571428571428, "grad_norm": 11373.58984375, "learning_rate": 2.317402597402597e-05, "loss": 0.4739, "step": 440 }, { "epoch": 1.1428571428571428, "step": 440, "train/loss_ctc": 0.3924143314361572, "train/loss_error": 0.3833776116371155, "train/loss_total": 0.38518497347831726 }, { "epoch": 1.1454545454545455, "step": 441, "train/loss_ctc": 0.4320562183856964, "train/loss_error": 0.427455872297287, "train/loss_total": 0.4283759295940399 }, { "epoch": 1.148051948051948, "step": 442, "train/loss_ctc": 0.3836585581302643, "train/loss_error": 0.37511131167411804, "train/loss_total": 0.37682077288627625 }, { "epoch": 1.1506493506493507, "step": 443, "train/loss_ctc": 0.4160490334033966, "train/loss_error": 0.3724731504917145, "train/loss_total": 0.3811883330345154 }, { "epoch": 1.1532467532467532, "step": 444, "train/loss_ctc": 0.4673062264919281, "train/loss_error": 0.3350518047809601, "train/loss_total": 0.3615027070045471 }, { "epoch": 1.155844155844156, "step": 445, "train/loss_ctc": 0.5486899614334106, "train/loss_error": 0.4397902190685272, "train/loss_total": 0.4615701735019684 }, { "epoch": 1.1584415584415584, "step": 446, "train/loss_ctc": 0.6298480033874512, "train/loss_error": 0.38568374514579773, "train/loss_total": 0.4345166087150574 }, { "epoch": 1.161038961038961, "step": 447, "train/loss_ctc": 0.37272214889526367, "train/loss_error": 0.38609036803245544, "train/loss_total": 0.3834167420864105 }, { "epoch": 1.1636363636363636, "step": 448, "train/loss_ctc": 0.49855557084083557, "train/loss_error": 0.39604446291923523, "train/loss_total": 0.41654670238494873 }, { "epoch": 1.1662337662337663, "step": 449, "train/loss_ctc": 0.5660229921340942, "train/loss_error": 0.4360395073890686, "train/loss_total": 0.46203622221946716 }, { "epoch": 1.1688311688311688, "grad_norm": 14255.4833984375, "learning_rate": 2.3018181818181816e-05, "loss": 0.4091, "step": 450 }, { "epoch": 1.1688311688311688, "step": 450, "train/loss_ctc": 0.3541427552700043, "train/loss_error": 0.4021000564098358, "train/loss_total": 0.3925085961818695 }, { "epoch": 1.1714285714285715, "step": 451, "train/loss_ctc": 0.5151469707489014, "train/loss_error": 0.4143333435058594, "train/loss_total": 0.43449607491493225 }, { "epoch": 1.174025974025974, "step": 452, "train/loss_ctc": 0.5866044759750366, "train/loss_error": 0.4510266184806824, "train/loss_total": 0.4781422019004822 }, { "epoch": 1.1766233766233767, "step": 453, "train/loss_ctc": 0.5455751419067383, "train/loss_error": 0.41385355591773987, "train/loss_total": 0.4401978850364685 }, { "epoch": 1.1792207792207792, "step": 454, "train/loss_ctc": 0.2847733497619629, "train/loss_error": 0.31129390001296997, "train/loss_total": 0.3059898018836975 }, { "epoch": 1.1818181818181819, "step": 455, "train/loss_ctc": 0.4697730839252472, "train/loss_error": 0.42589229345321655, "train/loss_total": 0.4346684515476227 }, { "epoch": 1.1844155844155844, "step": 456, "train/loss_ctc": 0.6028781533241272, "train/loss_error": 0.46088457107543945, "train/loss_total": 0.4892832934856415 }, { "epoch": 1.187012987012987, "step": 457, "train/loss_ctc": 0.3667582869529724, "train/loss_error": 0.3922768831253052, "train/loss_total": 0.3871731758117676 }, { "epoch": 1.1896103896103896, "step": 458, "train/loss_ctc": 0.45095095038414, "train/loss_error": 0.3594757318496704, "train/loss_total": 0.3777707815170288 }, { "epoch": 1.1922077922077923, "step": 459, "train/loss_ctc": 0.5079978704452515, "train/loss_error": 0.29953518509864807, "train/loss_total": 0.3412277102470398 }, { "epoch": 1.1948051948051948, "grad_norm": 12157.8056640625, "learning_rate": 2.2862337662337665e-05, "loss": 0.4081, "step": 460 }, { "epoch": 1.1948051948051948, "step": 460, "train/loss_ctc": 0.4698159992694855, "train/loss_error": 0.4825795888900757, "train/loss_total": 0.48002687096595764 }, { "epoch": 1.1974025974025975, "step": 461, "train/loss_ctc": 0.37948834896087646, "train/loss_error": 0.35641202330589294, "train/loss_total": 0.3610273003578186 }, { "epoch": 1.2, "step": 462, "train/loss_ctc": 0.3844801187515259, "train/loss_error": 0.5036032199859619, "train/loss_total": 0.47977858781814575 }, { "epoch": 1.2025974025974027, "step": 463, "train/loss_ctc": 0.42109113931655884, "train/loss_error": 0.3798549175262451, "train/loss_total": 0.3881021738052368 }, { "epoch": 1.2051948051948052, "step": 464, "train/loss_ctc": 0.664932370185852, "train/loss_error": 0.49868637323379517, "train/loss_total": 0.5319355726242065 }, { "epoch": 1.2077922077922079, "step": 465, "train/loss_ctc": 0.5306627750396729, "train/loss_error": 0.5481559038162231, "train/loss_total": 0.544657289981842 }, { "epoch": 1.2103896103896103, "step": 466, "train/loss_ctc": 0.4241098165512085, "train/loss_error": 0.4101329743862152, "train/loss_total": 0.41292834281921387 }, { "epoch": 1.212987012987013, "step": 467, "train/loss_ctc": 0.5555671453475952, "train/loss_error": 0.41930514574050903, "train/loss_total": 0.44655755162239075 }, { "epoch": 1.2155844155844155, "step": 468, "train/loss_ctc": 0.5906224250793457, "train/loss_error": 0.40295201539993286, "train/loss_total": 0.4404861032962799 }, { "epoch": 1.2181818181818183, "step": 469, "train/loss_ctc": 0.32890474796295166, "train/loss_error": 0.40111619234085083, "train/loss_total": 0.3866739273071289 }, { "epoch": 1.2207792207792207, "grad_norm": 12816.3955078125, "learning_rate": 2.270649350649351e-05, "loss": 0.4472, "step": 470 }, { "epoch": 1.2207792207792207, "step": 470, "train/loss_ctc": 0.28999656438827515, "train/loss_error": 0.32299524545669556, "train/loss_total": 0.31639552116394043 }, { "epoch": 1.2233766233766235, "step": 471, "train/loss_ctc": 0.5742143988609314, "train/loss_error": 0.5084536075592041, "train/loss_total": 0.5216057896614075 }, { "epoch": 1.225974025974026, "step": 472, "train/loss_ctc": 0.4847123920917511, "train/loss_error": 0.4719597101211548, "train/loss_total": 0.4745102524757385 }, { "epoch": 1.2285714285714286, "step": 473, "train/loss_ctc": 0.9444057941436768, "train/loss_error": 0.5527226328849792, "train/loss_total": 0.6310592889785767 }, { "epoch": 1.2311688311688311, "step": 474, "train/loss_ctc": 0.4046381711959839, "train/loss_error": 0.3468643128871918, "train/loss_total": 0.3584190905094147 }, { "epoch": 1.2337662337662338, "step": 475, "train/loss_ctc": 0.4736343026161194, "train/loss_error": 0.3477218747138977, "train/loss_total": 0.37290436029434204 }, { "epoch": 1.2363636363636363, "step": 476, "train/loss_ctc": 0.5357938408851624, "train/loss_error": 0.43466562032699585, "train/loss_total": 0.45489129424095154 }, { "epoch": 1.238961038961039, "step": 477, "train/loss_ctc": 0.4626949429512024, "train/loss_error": 0.3646146357059479, "train/loss_total": 0.38423070311546326 }, { "epoch": 1.2415584415584415, "step": 478, "train/loss_ctc": 0.6404685378074646, "train/loss_error": 0.42889928817749023, "train/loss_total": 0.471213161945343 }, { "epoch": 1.2441558441558442, "step": 479, "train/loss_ctc": 0.4832950830459595, "train/loss_error": 0.39515185356140137, "train/loss_total": 0.4127805233001709 }, { "epoch": 1.2467532467532467, "grad_norm": 18200.13671875, "learning_rate": 2.2550649350649354e-05, "loss": 0.4398, "step": 480 }, { "epoch": 1.2467532467532467, "step": 480, "train/loss_ctc": 0.4535917341709137, "train/loss_error": 0.49475035071372986, "train/loss_total": 0.48651865124702454 }, { "epoch": 1.2493506493506494, "step": 481, "train/loss_ctc": 0.5995355844497681, "train/loss_error": 0.4451226592063904, "train/loss_total": 0.4760052561759949 }, { "epoch": 1.251948051948052, "step": 482, "train/loss_ctc": 0.5357411503791809, "train/loss_error": 0.38769230246543884, "train/loss_total": 0.41730207204818726 }, { "epoch": 1.2545454545454544, "step": 483, "train/loss_ctc": 0.6708081960678101, "train/loss_error": 0.478177547454834, "train/loss_total": 0.5167036652565002 }, { "epoch": 1.2571428571428571, "step": 484, "train/loss_ctc": 0.5265322923660278, "train/loss_error": 0.4037075340747833, "train/loss_total": 0.4282724857330322 }, { "epoch": 1.2597402597402598, "step": 485, "train/loss_ctc": 0.28605660796165466, "train/loss_error": 0.44784459471702576, "train/loss_total": 0.41548699140548706 }, { "epoch": 1.2623376623376623, "step": 486, "train/loss_ctc": 0.3764401078224182, "train/loss_error": 0.4460128843784332, "train/loss_total": 0.4320983290672302 }, { "epoch": 1.2649350649350648, "step": 487, "train/loss_ctc": 0.6162067651748657, "train/loss_error": 0.33967551589012146, "train/loss_total": 0.3949817717075348 }, { "epoch": 1.2675324675324675, "step": 488, "train/loss_ctc": 0.5678637623786926, "train/loss_error": 0.4448745846748352, "train/loss_total": 0.46947240829467773 }, { "epoch": 1.2701298701298702, "step": 489, "train/loss_ctc": 0.36398088932037354, "train/loss_error": 0.39020606875419617, "train/loss_total": 0.3849610686302185 }, { "epoch": 1.2727272727272727, "grad_norm": 8082.6572265625, "learning_rate": 2.2394805194805196e-05, "loss": 0.4422, "step": 490 }, { "epoch": 1.2727272727272727, "step": 490, "train/loss_ctc": 0.6460669040679932, "train/loss_error": 0.45628198981285095, "train/loss_total": 0.4942389726638794 }, { "epoch": 1.2753246753246752, "step": 491, "train/loss_ctc": 0.7129509449005127, "train/loss_error": 0.4177573025226593, "train/loss_total": 0.47679603099823 }, { "epoch": 1.277922077922078, "step": 492, "train/loss_ctc": 0.42293834686279297, "train/loss_error": 0.4055692255496979, "train/loss_total": 0.4090430438518524 }, { "epoch": 1.2805194805194806, "step": 493, "train/loss_ctc": 0.5791569948196411, "train/loss_error": 0.4858042597770691, "train/loss_total": 0.5044748187065125 }, { "epoch": 1.283116883116883, "step": 494, "train/loss_ctc": 0.5799546837806702, "train/loss_error": 0.4812522232532501, "train/loss_total": 0.5009927153587341 }, { "epoch": 1.2857142857142856, "step": 495, "train/loss_ctc": 0.31917262077331543, "train/loss_error": 0.380170613527298, "train/loss_total": 0.3679710030555725 }, { "epoch": 1.2883116883116883, "step": 496, "train/loss_ctc": 0.4284976124763489, "train/loss_error": 0.3887322247028351, "train/loss_total": 0.39668530225753784 }, { "epoch": 1.290909090909091, "step": 497, "train/loss_ctc": 0.26924073696136475, "train/loss_error": 0.35040295124053955, "train/loss_total": 0.33417052030563354 }, { "epoch": 1.2935064935064935, "step": 498, "train/loss_ctc": 0.42598944902420044, "train/loss_error": 0.45800766348838806, "train/loss_total": 0.45160403847694397 }, { "epoch": 1.296103896103896, "step": 499, "train/loss_ctc": 0.31400248408317566, "train/loss_error": 0.3662009537220001, "train/loss_total": 0.35576125979423523 }, { "epoch": 1.2987012987012987, "grad_norm": 9664.9150390625, "learning_rate": 2.223896103896104e-05, "loss": 0.4292, "step": 500 }, { "epoch": 1.2987012987012987, "step": 500, "train/loss_ctc": 0.4991791248321533, "train/loss_error": 0.43600261211395264, "train/loss_total": 0.4486379027366638 }, { "epoch": 1.3012987012987014, "step": 501, "train/loss_ctc": 0.42969343066215515, "train/loss_error": 0.4648178517818451, "train/loss_total": 0.4577929675579071 }, { "epoch": 1.3038961038961039, "step": 502, "train/loss_ctc": 0.3509193956851959, "train/loss_error": 0.300504595041275, "train/loss_total": 0.3105875551700592 }, { "epoch": 1.3064935064935064, "step": 503, "train/loss_ctc": 0.7791563272476196, "train/loss_error": 0.47901299595832825, "train/loss_total": 0.5390416383743286 }, { "epoch": 1.309090909090909, "step": 504, "train/loss_ctc": 0.4831152856349945, "train/loss_error": 0.4949560761451721, "train/loss_total": 0.4925879240036011 }, { "epoch": 1.3116883116883118, "step": 505, "train/loss_ctc": 0.550849974155426, "train/loss_error": 0.4047818183898926, "train/loss_total": 0.43399545550346375 }, { "epoch": 1.3142857142857143, "step": 506, "train/loss_ctc": 0.5600680112838745, "train/loss_error": 0.4138337969779968, "train/loss_total": 0.4430806636810303 }, { "epoch": 1.3168831168831168, "step": 507, "train/loss_ctc": 0.5341476202011108, "train/loss_error": 0.3986845314502716, "train/loss_total": 0.4257771670818329 }, { "epoch": 1.3194805194805195, "step": 508, "train/loss_ctc": 0.30324792861938477, "train/loss_error": 0.3087727129459381, "train/loss_total": 0.3076677620410919 }, { "epoch": 1.3220779220779222, "step": 509, "train/loss_ctc": 0.4091183543205261, "train/loss_error": 0.37711960077285767, "train/loss_total": 0.38351935148239136 }, { "epoch": 1.3246753246753247, "grad_norm": 12718.951171875, "learning_rate": 2.2083116883116885e-05, "loss": 0.4243, "step": 510 }, { "epoch": 1.3246753246753247, "step": 510, "train/loss_ctc": 0.651812732219696, "train/loss_error": 0.4555288255214691, "train/loss_total": 0.4947856068611145 }, { "epoch": 1.3272727272727272, "step": 511, "train/loss_ctc": 0.6255940198898315, "train/loss_error": 0.4851270020008087, "train/loss_total": 0.5132204294204712 }, { "epoch": 1.3298701298701299, "step": 512, "train/loss_ctc": 0.6334453821182251, "train/loss_error": 0.35403624176979065, "train/loss_total": 0.40991806983947754 }, { "epoch": 1.3324675324675326, "step": 513, "train/loss_ctc": 0.536929726600647, "train/loss_error": 0.4631749391555786, "train/loss_total": 0.4779258966445923 }, { "epoch": 1.335064935064935, "step": 514, "train/loss_ctc": 0.7417004108428955, "train/loss_error": 0.4748733341693878, "train/loss_total": 0.5282387733459473 }, { "epoch": 1.3376623376623376, "step": 515, "train/loss_ctc": 0.566964864730835, "train/loss_error": 0.41022300720214844, "train/loss_total": 0.4415713846683502 }, { "epoch": 1.3402597402597403, "step": 516, "train/loss_ctc": 0.5274872779846191, "train/loss_error": 0.39678314328193665, "train/loss_total": 0.4229239821434021 }, { "epoch": 1.342857142857143, "step": 517, "train/loss_ctc": 0.6838077306747437, "train/loss_error": 0.4748755097389221, "train/loss_total": 0.5166620016098022 }, { "epoch": 1.3454545454545455, "step": 518, "train/loss_ctc": 0.3950461447238922, "train/loss_error": 0.33663538098335266, "train/loss_total": 0.34831753373146057 }, { "epoch": 1.348051948051948, "step": 519, "train/loss_ctc": 0.58773273229599, "train/loss_error": 0.36149510741233826, "train/loss_total": 0.406742662191391 }, { "epoch": 1.3506493506493507, "grad_norm": 17953.515625, "learning_rate": 2.192727272727273e-05, "loss": 0.456, "step": 520 }, { "epoch": 1.3506493506493507, "step": 520, "train/loss_ctc": 0.4458099603652954, "train/loss_error": 0.5203057527542114, "train/loss_total": 0.5054066181182861 }, { "epoch": 1.3532467532467534, "step": 521, "train/loss_ctc": 0.4802694320678711, "train/loss_error": 0.5162683129310608, "train/loss_total": 0.5090685486793518 }, { "epoch": 1.3558441558441559, "step": 522, "train/loss_ctc": 0.5591474771499634, "train/loss_error": 0.34346067905426025, "train/loss_total": 0.38659805059432983 }, { "epoch": 1.3584415584415583, "step": 523, "train/loss_ctc": 0.43695300817489624, "train/loss_error": 0.4012746512889862, "train/loss_total": 0.40841034054756165 }, { "epoch": 1.361038961038961, "step": 524, "train/loss_ctc": 0.4756784439086914, "train/loss_error": 0.3845369517803192, "train/loss_total": 0.4027652442455292 }, { "epoch": 1.3636363636363638, "step": 525, "train/loss_ctc": 0.4706064760684967, "train/loss_error": 0.3844946622848511, "train/loss_total": 0.40171703696250916 }, { "epoch": 1.3662337662337662, "step": 526, "train/loss_ctc": 0.5182939171791077, "train/loss_error": 0.36321377754211426, "train/loss_total": 0.39422982931137085 }, { "epoch": 1.3688311688311687, "step": 527, "train/loss_ctc": 0.5850343108177185, "train/loss_error": 0.36962783336639404, "train/loss_total": 0.41270914673805237 }, { "epoch": 1.3714285714285714, "step": 528, "train/loss_ctc": 0.44550126791000366, "train/loss_error": 0.3914671540260315, "train/loss_total": 0.4022740125656128 }, { "epoch": 1.3740259740259742, "step": 529, "train/loss_ctc": 0.38006895780563354, "train/loss_error": 0.5281417965888977, "train/loss_total": 0.49852725863456726 }, { "epoch": 1.3766233766233766, "grad_norm": 13377.36328125, "learning_rate": 2.177142857142857e-05, "loss": 0.4322, "step": 530 }, { "epoch": 1.3766233766233766, "step": 530, "train/loss_ctc": 0.5773483514785767, "train/loss_error": 0.45276376605033875, "train/loss_total": 0.47768068313598633 }, { "epoch": 1.3792207792207791, "step": 531, "train/loss_ctc": 0.44189512729644775, "train/loss_error": 0.42124128341674805, "train/loss_total": 0.42537206411361694 }, { "epoch": 1.3818181818181818, "step": 532, "train/loss_ctc": 0.44888484477996826, "train/loss_error": 0.35455530881881714, "train/loss_total": 0.37342122197151184 }, { "epoch": 1.3844155844155845, "step": 533, "train/loss_ctc": 0.5193815231323242, "train/loss_error": 0.3977604806423187, "train/loss_total": 0.4220846891403198 }, { "epoch": 1.387012987012987, "step": 534, "train/loss_ctc": 0.5399290323257446, "train/loss_error": 0.463765025138855, "train/loss_total": 0.4789978265762329 }, { "epoch": 1.3896103896103895, "step": 535, "train/loss_ctc": 0.5590641498565674, "train/loss_error": 0.4549209773540497, "train/loss_total": 0.4757496118545532 }, { "epoch": 1.3922077922077922, "step": 536, "train/loss_ctc": 0.4022553563117981, "train/loss_error": 0.4689479470252991, "train/loss_total": 0.45560944080352783 }, { "epoch": 1.394805194805195, "step": 537, "train/loss_ctc": 0.4969099164009094, "train/loss_error": 0.42457571625709534, "train/loss_total": 0.4390425682067871 }, { "epoch": 1.3974025974025974, "step": 538, "train/loss_ctc": 0.38078099489212036, "train/loss_error": 0.4578838646411896, "train/loss_total": 0.44246330857276917 }, { "epoch": 1.4, "step": 539, "train/loss_ctc": 0.34173059463500977, "train/loss_error": 0.4033532738685608, "train/loss_total": 0.3910287320613861 }, { "epoch": 1.4025974025974026, "grad_norm": 8956.73828125, "learning_rate": 2.1615584415584416e-05, "loss": 0.4381, "step": 540 }, { "epoch": 1.4025974025974026, "step": 540, "train/loss_ctc": 0.5944470167160034, "train/loss_error": 0.46106502413749695, "train/loss_total": 0.4877414107322693 }, { "epoch": 1.4051948051948053, "step": 541, "train/loss_ctc": 0.6270057559013367, "train/loss_error": 0.47306036949157715, "train/loss_total": 0.503849446773529 }, { "epoch": 1.4077922077922078, "step": 542, "train/loss_ctc": 0.7037888765335083, "train/loss_error": 0.5113322734832764, "train/loss_total": 0.5498235821723938 }, { "epoch": 1.4103896103896103, "step": 543, "train/loss_ctc": 0.4359579086303711, "train/loss_error": 0.45941632986068726, "train/loss_total": 0.45472463965415955 }, { "epoch": 1.412987012987013, "step": 544, "train/loss_ctc": 0.6012882590293884, "train/loss_error": 0.45094653964042664, "train/loss_total": 0.4810148775577545 }, { "epoch": 1.4155844155844157, "step": 545, "train/loss_ctc": 0.5158230066299438, "train/loss_error": 0.46630239486694336, "train/loss_total": 0.47620654106140137 }, { "epoch": 1.4181818181818182, "step": 546, "train/loss_ctc": 0.6660903692245483, "train/loss_error": 0.35119131207466125, "train/loss_total": 0.41417112946510315 }, { "epoch": 1.4207792207792207, "step": 547, "train/loss_ctc": 0.33324918150901794, "train/loss_error": 0.42275574803352356, "train/loss_total": 0.404854416847229 }, { "epoch": 1.4233766233766234, "step": 548, "train/loss_ctc": 0.35566580295562744, "train/loss_error": 0.37313729524612427, "train/loss_total": 0.3696430027484894 }, { "epoch": 1.425974025974026, "step": 549, "train/loss_ctc": 0.5921926498413086, "train/loss_error": 0.47182127833366394, "train/loss_total": 0.4958955645561218 }, { "epoch": 1.4285714285714286, "grad_norm": 13675.501953125, "learning_rate": 2.145974025974026e-05, "loss": 0.4638, "step": 550 }, { "epoch": 1.4285714285714286, "step": 550, "train/loss_ctc": 0.7926477193832397, "train/loss_error": 0.4374622404575348, "train/loss_total": 0.5084993839263916 }, { "epoch": 1.431168831168831, "step": 551, "train/loss_ctc": 0.44758331775665283, "train/loss_error": 0.4179244935512543, "train/loss_total": 0.423856258392334 }, { "epoch": 1.4337662337662338, "step": 552, "train/loss_ctc": 0.4906901717185974, "train/loss_error": 0.4150024354457855, "train/loss_total": 0.4301399886608124 }, { "epoch": 1.4363636363636363, "step": 553, "train/loss_ctc": 0.5552089810371399, "train/loss_error": 0.4061211347579956, "train/loss_total": 0.4359387159347534 }, { "epoch": 1.438961038961039, "step": 554, "train/loss_ctc": 0.6437962055206299, "train/loss_error": 0.4346274137496948, "train/loss_total": 0.47646117210388184 }, { "epoch": 1.4415584415584415, "step": 555, "train/loss_ctc": 0.4782348573207855, "train/loss_error": 0.379209041595459, "train/loss_total": 0.3990142047405243 }, { "epoch": 1.4441558441558442, "step": 556, "train/loss_ctc": 0.39623042941093445, "train/loss_error": 0.45357000827789307, "train/loss_total": 0.4421020746231079 }, { "epoch": 1.4467532467532467, "step": 557, "train/loss_ctc": 0.3138650953769684, "train/loss_error": 0.34415265917778015, "train/loss_total": 0.33809515833854675 }, { "epoch": 1.4493506493506494, "step": 558, "train/loss_ctc": 0.6850205659866333, "train/loss_error": 0.3908621370792389, "train/loss_total": 0.44969385862350464 }, { "epoch": 1.4519480519480519, "step": 559, "train/loss_ctc": 0.43321651220321655, "train/loss_error": 0.35769015550613403, "train/loss_total": 0.372795432806015 }, { "epoch": 1.4545454545454546, "grad_norm": 12022.9541015625, "learning_rate": 2.1303896103896106e-05, "loss": 0.4277, "step": 560 }, { "epoch": 1.4545454545454546, "step": 560, "train/loss_ctc": 0.4359273314476013, "train/loss_error": 0.3898954391479492, "train/loss_total": 0.3991018235683441 }, { "epoch": 1.457142857142857, "step": 561, "train/loss_ctc": 0.5215237736701965, "train/loss_error": 0.49614477157592773, "train/loss_total": 0.5012205839157104 }, { "epoch": 1.4597402597402598, "step": 562, "train/loss_ctc": 0.4089413583278656, "train/loss_error": 0.3430430591106415, "train/loss_total": 0.3562227189540863 }, { "epoch": 1.4623376623376623, "step": 563, "train/loss_ctc": 0.39418137073516846, "train/loss_error": 0.48825928568840027, "train/loss_total": 0.46944373846054077 }, { "epoch": 1.464935064935065, "step": 564, "train/loss_ctc": 0.6463356018066406, "train/loss_error": 0.37080761790275574, "train/loss_total": 0.4259132146835327 }, { "epoch": 1.4675324675324675, "step": 565, "train/loss_ctc": 0.46846312284469604, "train/loss_error": 0.456970751285553, "train/loss_total": 0.4592692255973816 }, { "epoch": 1.4701298701298702, "step": 566, "train/loss_ctc": 0.44413840770721436, "train/loss_error": 0.27477872371673584, "train/loss_total": 0.3086506724357605 }, { "epoch": 1.4727272727272727, "step": 567, "train/loss_ctc": 0.5424789190292358, "train/loss_error": 0.36597245931625366, "train/loss_total": 0.40127378702163696 }, { "epoch": 1.4753246753246754, "step": 568, "train/loss_ctc": 0.5202217698097229, "train/loss_error": 0.5066185593605042, "train/loss_total": 0.5093392133712769 }, { "epoch": 1.4779220779220779, "step": 569, "train/loss_ctc": 0.43378257751464844, "train/loss_error": 0.38787221908569336, "train/loss_total": 0.3970543146133423 }, { "epoch": 1.4805194805194806, "grad_norm": 16820.33203125, "learning_rate": 2.114805194805195e-05, "loss": 0.4227, "step": 570 }, { "epoch": 1.4805194805194806, "step": 570, "train/loss_ctc": 0.5181453227996826, "train/loss_error": 0.4394575357437134, "train/loss_total": 0.4551951289176941 }, { "epoch": 1.483116883116883, "step": 571, "train/loss_ctc": 0.3269159197807312, "train/loss_error": 0.3693723976612091, "train/loss_total": 0.36088111996650696 }, { "epoch": 1.4857142857142858, "step": 572, "train/loss_ctc": 0.3966122269630432, "train/loss_error": 0.3866136968135834, "train/loss_total": 0.38861343264579773 }, { "epoch": 1.4883116883116883, "step": 573, "train/loss_ctc": 0.48870059847831726, "train/loss_error": 0.3778630197048187, "train/loss_total": 0.4000305235385895 }, { "epoch": 1.490909090909091, "step": 574, "train/loss_ctc": 0.6348576545715332, "train/loss_error": 0.4927232265472412, "train/loss_total": 0.5211501121520996 }, { "epoch": 1.4935064935064934, "step": 575, "train/loss_ctc": 0.7885884046554565, "train/loss_error": 0.506908655166626, "train/loss_total": 0.563244640827179 }, { "epoch": 1.4961038961038962, "step": 576, "train/loss_ctc": 0.7677744626998901, "train/loss_error": 0.4691813588142395, "train/loss_total": 0.5288999676704407 }, { "epoch": 1.4987012987012986, "step": 577, "train/loss_ctc": 0.7404606938362122, "train/loss_error": 0.40165987610816956, "train/loss_total": 0.46942001581192017 }, { "epoch": 1.5012987012987011, "step": 578, "train/loss_ctc": 0.43727147579193115, "train/loss_error": 0.36014416813850403, "train/loss_total": 0.3755696415901184 }, { "epoch": 1.5038961038961038, "step": 579, "train/loss_ctc": 0.739143967628479, "train/loss_error": 0.5179652571678162, "train/loss_total": 0.5622010231018066 }, { "epoch": 1.5064935064935066, "grad_norm": 13588.2666015625, "learning_rate": 2.0992207792207792e-05, "loss": 0.4625, "step": 580 }, { "epoch": 1.5064935064935066, "step": 580, "train/loss_ctc": 0.6560373306274414, "train/loss_error": 0.45283645391464233, "train/loss_total": 0.49347662925720215 }, { "epoch": 1.509090909090909, "step": 581, "train/loss_ctc": 0.4712487459182739, "train/loss_error": 0.34655991196632385, "train/loss_total": 0.3714976906776428 }, { "epoch": 1.5116883116883115, "step": 582, "train/loss_ctc": 0.4738585650920868, "train/loss_error": 0.4671991169452667, "train/loss_total": 0.4685310125350952 }, { "epoch": 1.5142857142857142, "step": 583, "train/loss_ctc": 0.48708686232566833, "train/loss_error": 0.4571058750152588, "train/loss_total": 0.4631021022796631 }, { "epoch": 1.516883116883117, "step": 584, "train/loss_ctc": 0.6512737274169922, "train/loss_error": 0.4701080620288849, "train/loss_total": 0.5063412189483643 }, { "epoch": 1.5194805194805194, "step": 585, "train/loss_ctc": 0.36722850799560547, "train/loss_error": 0.41455334424972534, "train/loss_total": 0.4050883948802948 }, { "epoch": 1.522077922077922, "step": 586, "train/loss_ctc": 0.5089607834815979, "train/loss_error": 0.36582767963409424, "train/loss_total": 0.39445430040359497 }, { "epoch": 1.5246753246753246, "step": 587, "train/loss_ctc": 0.6822967529296875, "train/loss_error": 0.5626192688941956, "train/loss_total": 0.586554765701294 }, { "epoch": 1.5272727272727273, "step": 588, "train/loss_ctc": 0.9367376565933228, "train/loss_error": 0.48776575922966003, "train/loss_total": 0.5775601863861084 }, { "epoch": 1.5298701298701298, "step": 589, "train/loss_ctc": 0.46932703256607056, "train/loss_error": 0.49243617057800293, "train/loss_total": 0.48781436681747437 }, { "epoch": 1.5324675324675323, "grad_norm": 12909.8857421875, "learning_rate": 2.0836363636363637e-05, "loss": 0.4754, "step": 590 }, { "epoch": 1.5324675324675323, "step": 590, "train/loss_ctc": 0.319924920797348, "train/loss_error": 0.482850044965744, "train/loss_total": 0.4502650201320648 }, { "epoch": 1.535064935064935, "step": 591, "train/loss_ctc": 0.37456953525543213, "train/loss_error": 0.3806517720222473, "train/loss_total": 0.37943533062934875 }, { "epoch": 1.5376623376623377, "step": 592, "train/loss_ctc": 0.5184667110443115, "train/loss_error": 0.44780951738357544, "train/loss_total": 0.4619409441947937 }, { "epoch": 1.5402597402597402, "step": 593, "train/loss_ctc": 0.4502066373825073, "train/loss_error": 0.4840274453163147, "train/loss_total": 0.47726330161094666 }, { "epoch": 1.5428571428571427, "step": 594, "train/loss_ctc": 0.5933941006660461, "train/loss_error": 0.41438451409339905, "train/loss_total": 0.45018643140792847 }, { "epoch": 1.5454545454545454, "step": 595, "train/loss_ctc": 0.32038456201553345, "train/loss_error": 0.2996804714202881, "train/loss_total": 0.30382129549980164 }, { "epoch": 1.5480519480519481, "step": 596, "train/loss_ctc": 0.7219564914703369, "train/loss_error": 0.40262433886528015, "train/loss_total": 0.466490775346756 }, { "epoch": 1.5506493506493506, "step": 597, "train/loss_ctc": 0.6793842911720276, "train/loss_error": 0.4813334345817566, "train/loss_total": 0.5209436416625977 }, { "epoch": 1.553246753246753, "step": 598, "train/loss_ctc": 0.8013080358505249, "train/loss_error": 0.4284047484397888, "train/loss_total": 0.502985417842865 }, { "epoch": 1.5558441558441558, "step": 599, "train/loss_ctc": 0.35186290740966797, "train/loss_error": 0.44855672121047974, "train/loss_total": 0.42921796441078186 }, { "epoch": 1.5584415584415585, "grad_norm": 10922.8447265625, "learning_rate": 2.0680519480519482e-05, "loss": 0.4443, "step": 600 }, { "epoch": 1.5584415584415585, "step": 600, "train/loss_ctc": 0.5146725177764893, "train/loss_error": 0.4555650055408478, "train/loss_total": 0.46738651394844055 }, { "epoch": 1.561038961038961, "step": 601, "train/loss_ctc": 0.4669667184352875, "train/loss_error": 0.5245593786239624, "train/loss_total": 0.5130408406257629 }, { "epoch": 1.5636363636363635, "step": 602, "train/loss_ctc": 0.418283075094223, "train/loss_error": 0.33903366327285767, "train/loss_total": 0.3548835515975952 }, { "epoch": 1.5662337662337662, "step": 603, "train/loss_ctc": 0.5428087711334229, "train/loss_error": 0.4684072434902191, "train/loss_total": 0.4832875430583954 }, { "epoch": 1.568831168831169, "step": 604, "train/loss_ctc": 0.43852633237838745, "train/loss_error": 0.4223793148994446, "train/loss_total": 0.42560875415802 }, { "epoch": 1.5714285714285714, "step": 605, "train/loss_ctc": 0.6604779958724976, "train/loss_error": 0.46750548481941223, "train/loss_total": 0.5060999989509583 }, { "epoch": 1.5740259740259739, "step": 606, "train/loss_ctc": 0.48788490891456604, "train/loss_error": 0.41343674063682556, "train/loss_total": 0.4283263683319092 }, { "epoch": 1.5766233766233766, "step": 607, "train/loss_ctc": 0.5802578330039978, "train/loss_error": 0.5291554927825928, "train/loss_total": 0.5393759608268738 }, { "epoch": 1.5792207792207793, "step": 608, "train/loss_ctc": 0.6474225521087646, "train/loss_error": 0.4275383949279785, "train/loss_total": 0.4715152382850647 }, { "epoch": 1.5818181818181818, "step": 609, "train/loss_ctc": 0.5171098709106445, "train/loss_error": 0.45256876945495605, "train/loss_total": 0.46547701954841614 }, { "epoch": 1.5844155844155843, "grad_norm": 13419.373046875, "learning_rate": 2.0524675324675327e-05, "loss": 0.4655, "step": 610 }, { "epoch": 1.5844155844155843, "step": 610, "train/loss_ctc": 0.3520013689994812, "train/loss_error": 0.5073508024215698, "train/loss_total": 0.47628092765808105 }, { "epoch": 1.587012987012987, "step": 611, "train/loss_ctc": 0.5071874856948853, "train/loss_error": 0.5072435140609741, "train/loss_total": 0.5072323083877563 }, { "epoch": 1.5896103896103897, "step": 612, "train/loss_ctc": 0.4642595052719116, "train/loss_error": 0.3060194253921509, "train/loss_total": 0.33766743540763855 }, { "epoch": 1.5922077922077922, "step": 613, "train/loss_ctc": 0.46602874994277954, "train/loss_error": 0.42373254895210266, "train/loss_total": 0.43219178915023804 }, { "epoch": 1.5948051948051947, "step": 614, "train/loss_ctc": 0.3598201274871826, "train/loss_error": 0.394644558429718, "train/loss_total": 0.38767966628074646 }, { "epoch": 1.5974025974025974, "step": 615, "train/loss_ctc": 0.5193796753883362, "train/loss_error": 0.4526275098323822, "train/loss_total": 0.4659779369831085 }, { "epoch": 1.6, "step": 616, "train/loss_ctc": 0.397167444229126, "train/loss_error": 0.5245645642280579, "train/loss_total": 0.4990851581096649 }, { "epoch": 1.6025974025974026, "step": 617, "train/loss_ctc": 0.5788940191268921, "train/loss_error": 0.5046259164810181, "train/loss_total": 0.519479513168335 }, { "epoch": 1.605194805194805, "step": 618, "train/loss_ctc": 0.25758516788482666, "train/loss_error": 0.4356691539287567, "train/loss_total": 0.40005236864089966 }, { "epoch": 1.6077922077922078, "step": 619, "train/loss_ctc": 0.5940976142883301, "train/loss_error": 0.3906233608722687, "train/loss_total": 0.4313182234764099 }, { "epoch": 1.6103896103896105, "grad_norm": 14076.06640625, "learning_rate": 2.036883116883117e-05, "loss": 0.4457, "step": 620 }, { "epoch": 1.6103896103896105, "step": 620, "train/loss_ctc": 0.48789095878601074, "train/loss_error": 0.426087886095047, "train/loss_total": 0.4384485185146332 }, { "epoch": 1.612987012987013, "step": 621, "train/loss_ctc": 0.4056876003742218, "train/loss_error": 0.43816089630126953, "train/loss_total": 0.4316662549972534 }, { "epoch": 1.6155844155844155, "step": 622, "train/loss_ctc": 0.42751675844192505, "train/loss_error": 0.4824672341346741, "train/loss_total": 0.4714771509170532 }, { "epoch": 1.6181818181818182, "step": 623, "train/loss_ctc": 0.39822036027908325, "train/loss_error": 0.39270272850990295, "train/loss_total": 0.3938062787055969 }, { "epoch": 1.6207792207792209, "step": 624, "train/loss_ctc": 0.497310608625412, "train/loss_error": 0.4018491506576538, "train/loss_total": 0.42094144225120544 }, { "epoch": 1.6233766233766234, "step": 625, "train/loss_ctc": 0.5246763825416565, "train/loss_error": 0.36580848693847656, "train/loss_total": 0.397582083940506 }, { "epoch": 1.6259740259740258, "step": 626, "train/loss_ctc": 0.6265509128570557, "train/loss_error": 0.43994924426078796, "train/loss_total": 0.47726958990097046 }, { "epoch": 1.6285714285714286, "step": 627, "train/loss_ctc": 0.8528670072555542, "train/loss_error": 0.42943474650382996, "train/loss_total": 0.5141211748123169 }, { "epoch": 1.6311688311688313, "step": 628, "train/loss_ctc": 0.5296465754508972, "train/loss_error": 0.38120123744010925, "train/loss_total": 0.4108903110027313 }, { "epoch": 1.6337662337662338, "step": 629, "train/loss_ctc": 0.3892641067504883, "train/loss_error": 0.352253794670105, "train/loss_total": 0.35965585708618164 }, { "epoch": 1.6363636363636362, "grad_norm": 12153.212890625, "learning_rate": 2.0212987012987013e-05, "loss": 0.4316, "step": 630 }, { "epoch": 1.6363636363636362, "step": 630, "train/loss_ctc": 0.6218644380569458, "train/loss_error": 0.4183930456638336, "train/loss_total": 0.4590873420238495 }, { "epoch": 1.638961038961039, "step": 631, "train/loss_ctc": 0.43102359771728516, "train/loss_error": 0.36114200949668884, "train/loss_total": 0.37511831521987915 }, { "epoch": 1.6415584415584417, "step": 632, "train/loss_ctc": 0.6346060037612915, "train/loss_error": 0.5049745440483093, "train/loss_total": 0.5309008359909058 }, { "epoch": 1.6441558441558441, "step": 633, "train/loss_ctc": 0.3591400384902954, "train/loss_error": 0.4017753601074219, "train/loss_total": 0.3932482898235321 }, { "epoch": 1.6467532467532466, "step": 634, "train/loss_ctc": 0.41865018010139465, "train/loss_error": 0.3991761803627014, "train/loss_total": 0.40307098627090454 }, { "epoch": 1.6493506493506493, "step": 635, "train/loss_ctc": 0.39433300495147705, "train/loss_error": 0.3665936589241028, "train/loss_total": 0.3721415400505066 }, { "epoch": 1.651948051948052, "step": 636, "train/loss_ctc": 0.5417003631591797, "train/loss_error": 0.5199500322341919, "train/loss_total": 0.5243000984191895 }, { "epoch": 1.6545454545454545, "step": 637, "train/loss_ctc": 0.5378527641296387, "train/loss_error": 0.4268370568752289, "train/loss_total": 0.4490402042865753 }, { "epoch": 1.657142857142857, "step": 638, "train/loss_ctc": 0.6627585291862488, "train/loss_error": 0.45304134488105774, "train/loss_total": 0.49498480558395386 }, { "epoch": 1.6597402597402597, "step": 639, "train/loss_ctc": 0.5468976497650146, "train/loss_error": 0.40239691734313965, "train/loss_total": 0.43129706382751465 }, { "epoch": 1.6623376623376624, "grad_norm": 12131.65625, "learning_rate": 2.0057142857142858e-05, "loss": 0.4433, "step": 640 }, { "epoch": 1.6623376623376624, "step": 640, "train/loss_ctc": 0.4740447402000427, "train/loss_error": 0.4181217849254608, "train/loss_total": 0.42930638790130615 }, { "epoch": 1.664935064935065, "step": 641, "train/loss_ctc": 0.39896902441978455, "train/loss_error": 0.43873563408851624, "train/loss_total": 0.4307823181152344 }, { "epoch": 1.6675324675324674, "step": 642, "train/loss_ctc": 0.40717488527297974, "train/loss_error": 0.49874448776245117, "train/loss_total": 0.48043060302734375 }, { "epoch": 1.6701298701298701, "step": 643, "train/loss_ctc": 0.4540250599384308, "train/loss_error": 0.36412081122398376, "train/loss_total": 0.3821016848087311 }, { "epoch": 1.6727272727272728, "step": 644, "train/loss_ctc": 0.464885413646698, "train/loss_error": 0.48115968704223633, "train/loss_total": 0.4779048264026642 }, { "epoch": 1.6753246753246753, "step": 645, "train/loss_ctc": 0.3602196276187897, "train/loss_error": 0.42244186997413635, "train/loss_total": 0.40999743342399597 }, { "epoch": 1.6779220779220778, "step": 646, "train/loss_ctc": 1.180898904800415, "train/loss_error": 0.4854589104652405, "train/loss_total": 0.6245469450950623 }, { "epoch": 1.6805194805194805, "step": 647, "train/loss_ctc": 0.7672433853149414, "train/loss_error": 0.42233142256736755, "train/loss_total": 0.4913138151168823 }, { "epoch": 1.6831168831168832, "step": 648, "train/loss_ctc": 0.31238222122192383, "train/loss_error": 0.4029539227485657, "train/loss_total": 0.38483959436416626 }, { "epoch": 1.6857142857142857, "step": 649, "train/loss_ctc": 0.645094633102417, "train/loss_error": 0.5649255514144897, "train/loss_total": 0.5809593796730042 }, { "epoch": 1.6883116883116882, "grad_norm": 12145.6533203125, "learning_rate": 1.9901298701298703e-05, "loss": 0.4692, "step": 650 }, { "epoch": 1.6883116883116882, "step": 650, "train/loss_ctc": 0.7188942432403564, "train/loss_error": 0.3927106261253357, "train/loss_total": 0.45794737339019775 }, { "epoch": 1.690909090909091, "step": 651, "train/loss_ctc": 0.6205147504806519, "train/loss_error": 0.40333789587020874, "train/loss_total": 0.4467732608318329 }, { "epoch": 1.6935064935064936, "step": 652, "train/loss_ctc": 0.3766503930091858, "train/loss_error": 0.37028175592422485, "train/loss_total": 0.37155547738075256 }, { "epoch": 1.6961038961038961, "step": 653, "train/loss_ctc": 0.41249316930770874, "train/loss_error": 0.4037098288536072, "train/loss_total": 0.4054664969444275 }, { "epoch": 1.6987012987012986, "step": 654, "train/loss_ctc": 0.799534797668457, "train/loss_error": 0.48555609583854675, "train/loss_total": 0.5483518242835999 }, { "epoch": 1.7012987012987013, "step": 655, "train/loss_ctc": 0.42786577343940735, "train/loss_error": 0.3737027943134308, "train/loss_total": 0.38453540205955505 }, { "epoch": 1.703896103896104, "step": 656, "train/loss_ctc": 0.4111911356449127, "train/loss_error": 0.39247575402259827, "train/loss_total": 0.39621883630752563 }, { "epoch": 1.7064935064935065, "step": 657, "train/loss_ctc": 0.3404443860054016, "train/loss_error": 0.33500173687934875, "train/loss_total": 0.3360902965068817 }, { "epoch": 1.709090909090909, "step": 658, "train/loss_ctc": 0.44381871819496155, "train/loss_error": 0.4756125807762146, "train/loss_total": 0.469253808259964 }, { "epoch": 1.7116883116883117, "step": 659, "train/loss_ctc": 0.43956202268600464, "train/loss_error": 0.46334108710289, "train/loss_total": 0.45858529210090637 }, { "epoch": 1.7142857142857144, "grad_norm": 13509.1376953125, "learning_rate": 1.9745454545454547e-05, "loss": 0.4275, "step": 660 }, { "epoch": 1.7142857142857144, "step": 660, "train/loss_ctc": 0.4581942558288574, "train/loss_error": 0.45909932255744934, "train/loss_total": 0.45891833305358887 }, { "epoch": 1.716883116883117, "step": 661, "train/loss_ctc": 0.8773304224014282, "train/loss_error": 0.4531882405281067, "train/loss_total": 0.538016676902771 }, { "epoch": 1.7194805194805194, "step": 662, "train/loss_ctc": 1.0483920574188232, "train/loss_error": 0.48350194096565247, "train/loss_total": 0.5964800119400024 }, { "epoch": 1.722077922077922, "step": 663, "train/loss_ctc": 0.4054456353187561, "train/loss_error": 0.3332667946815491, "train/loss_total": 0.34770259261131287 }, { "epoch": 1.7246753246753248, "step": 664, "train/loss_ctc": 0.7266868948936462, "train/loss_error": 0.36465051770210266, "train/loss_total": 0.4370577931404114 }, { "epoch": 1.7272727272727273, "step": 665, "train/loss_ctc": 0.5294763445854187, "train/loss_error": 0.5488013625144958, "train/loss_total": 0.5449363589286804 }, { "epoch": 1.7298701298701298, "step": 666, "train/loss_ctc": 0.5331743955612183, "train/loss_error": 0.3734265863895416, "train/loss_total": 0.4053761661052704 }, { "epoch": 1.7324675324675325, "step": 667, "train/loss_ctc": 0.4170442819595337, "train/loss_error": 0.29198896884918213, "train/loss_total": 0.31700003147125244 }, { "epoch": 1.7350649350649352, "step": 668, "train/loss_ctc": 0.43445804715156555, "train/loss_error": 0.456394761800766, "train/loss_total": 0.4520074427127838 }, { "epoch": 1.7376623376623377, "step": 669, "train/loss_ctc": 0.6563950777053833, "train/loss_error": 0.315766304731369, "train/loss_total": 0.3838920593261719 }, { "epoch": 1.7402597402597402, "grad_norm": 20375.572265625, "learning_rate": 1.9589610389610392e-05, "loss": 0.4481, "step": 670 }, { "epoch": 1.7402597402597402, "step": 670, "train/loss_ctc": 0.3489413261413574, "train/loss_error": 0.3861934542655945, "train/loss_total": 0.378743052482605 }, { "epoch": 1.7428571428571429, "step": 671, "train/loss_ctc": 0.49945569038391113, "train/loss_error": 0.3700405955314636, "train/loss_total": 0.3959236145019531 }, { "epoch": 1.7454545454545456, "step": 672, "train/loss_ctc": 0.5632936954498291, "train/loss_error": 0.49658843874931335, "train/loss_total": 0.5099295377731323 }, { "epoch": 1.748051948051948, "step": 673, "train/loss_ctc": 0.5111880898475647, "train/loss_error": 0.44224631786346436, "train/loss_total": 0.45603466033935547 }, { "epoch": 1.7506493506493506, "step": 674, "train/loss_ctc": 0.513247013092041, "train/loss_error": 0.4258171319961548, "train/loss_total": 0.44330310821533203 }, { "epoch": 1.7532467532467533, "step": 675, "train/loss_ctc": 0.38028740882873535, "train/loss_error": 0.3963707685470581, "train/loss_total": 0.393154114484787 }, { "epoch": 1.755844155844156, "step": 676, "train/loss_ctc": 0.2963399291038513, "train/loss_error": 0.3633609712123871, "train/loss_total": 0.34995678067207336 }, { "epoch": 1.7584415584415585, "step": 677, "train/loss_ctc": 0.48295915126800537, "train/loss_error": 0.5003724098205566, "train/loss_total": 0.49688977003097534 }, { "epoch": 1.761038961038961, "step": 678, "train/loss_ctc": 0.3532203435897827, "train/loss_error": 0.4153042435646057, "train/loss_total": 0.4028874933719635 }, { "epoch": 1.7636363636363637, "step": 679, "train/loss_ctc": 0.5673993229866028, "train/loss_error": 0.3376362919807434, "train/loss_total": 0.38358891010284424 }, { "epoch": 1.7662337662337664, "grad_norm": 13589.376953125, "learning_rate": 1.9433766233766234e-05, "loss": 0.421, "step": 680 }, { "epoch": 1.7662337662337664, "step": 680, "train/loss_ctc": 0.33397021889686584, "train/loss_error": 0.4455782175064087, "train/loss_total": 0.42325660586357117 }, { "epoch": 1.7688311688311689, "step": 681, "train/loss_ctc": 0.5507590770721436, "train/loss_error": 0.4941459000110626, "train/loss_total": 0.5054685473442078 }, { "epoch": 1.7714285714285714, "step": 682, "train/loss_ctc": 0.4856727719306946, "train/loss_error": 0.3426738381385803, "train/loss_total": 0.37127363681793213 }, { "epoch": 1.774025974025974, "step": 683, "train/loss_ctc": 0.4975511431694031, "train/loss_error": 0.4880368709564209, "train/loss_total": 0.48993971943855286 }, { "epoch": 1.7766233766233768, "step": 684, "train/loss_ctc": 0.5242135524749756, "train/loss_error": 0.39967942237854004, "train/loss_total": 0.4245862662792206 }, { "epoch": 1.7792207792207793, "step": 685, "train/loss_ctc": 0.34245598316192627, "train/loss_error": 0.4655245542526245, "train/loss_total": 0.44091084599494934 }, { "epoch": 1.7818181818181817, "step": 686, "train/loss_ctc": 0.3015326261520386, "train/loss_error": 0.36143288016319275, "train/loss_total": 0.34945282340049744 }, { "epoch": 1.7844155844155845, "step": 687, "train/loss_ctc": 0.5621632933616638, "train/loss_error": 0.4163164496421814, "train/loss_total": 0.44548583030700684 }, { "epoch": 1.7870129870129872, "step": 688, "train/loss_ctc": 0.4489537477493286, "train/loss_error": 0.3594694137573242, "train/loss_total": 0.377366304397583 }, { "epoch": 1.7896103896103897, "step": 689, "train/loss_ctc": 0.5295306444168091, "train/loss_error": 0.4747498333454132, "train/loss_total": 0.48570600152015686 }, { "epoch": 1.7922077922077921, "grad_norm": 13454.7978515625, "learning_rate": 1.927792207792208e-05, "loss": 0.4313, "step": 690 }, { "epoch": 1.7922077922077921, "step": 690, "train/loss_ctc": 0.5634748935699463, "train/loss_error": 0.399019330739975, "train/loss_total": 0.4319104552268982 }, { "epoch": 1.7948051948051948, "step": 691, "train/loss_ctc": 0.3253403604030609, "train/loss_error": 0.35372307896614075, "train/loss_total": 0.34804654121398926 }, { "epoch": 1.7974025974025976, "step": 692, "train/loss_ctc": 0.5204123854637146, "train/loss_error": 0.48465579748153687, "train/loss_total": 0.49180710315704346 }, { "epoch": 1.8, "step": 693, "train/loss_ctc": 0.45956292748451233, "train/loss_error": 0.36194950342178345, "train/loss_total": 0.3814722001552582 }, { "epoch": 1.8025974025974025, "step": 694, "train/loss_ctc": 0.4777572751045227, "train/loss_error": 0.4468281865119934, "train/loss_total": 0.4530140161514282 }, { "epoch": 1.8051948051948052, "step": 695, "train/loss_ctc": 0.5269793272018433, "train/loss_error": 0.4458874762058258, "train/loss_total": 0.4621058702468872 }, { "epoch": 1.807792207792208, "step": 696, "train/loss_ctc": 0.6988600492477417, "train/loss_error": 0.4281385838985443, "train/loss_total": 0.4822828769683838 }, { "epoch": 1.8103896103896104, "step": 697, "train/loss_ctc": 0.3824452757835388, "train/loss_error": 0.3568798303604126, "train/loss_total": 0.3619929254055023 }, { "epoch": 1.812987012987013, "step": 698, "train/loss_ctc": 0.6222847104072571, "train/loss_error": 0.4820188283920288, "train/loss_total": 0.5100719928741455 }, { "epoch": 1.8155844155844156, "step": 699, "train/loss_ctc": 0.7759758830070496, "train/loss_error": 0.48644620180130005, "train/loss_total": 0.5443521738052368 }, { "epoch": 1.8181818181818183, "grad_norm": 16301.6259765625, "learning_rate": 1.9122077922077923e-05, "loss": 0.4467, "step": 700 }, { "epoch": 1.8181818181818183, "step": 700, "train/loss_ctc": 0.7943434715270996, "train/loss_error": 0.4096769392490387, "train/loss_total": 0.4866102635860443 }, { "epoch": 1.8207792207792208, "step": 701, "train/loss_ctc": 0.706234335899353, "train/loss_error": 0.5245950818061829, "train/loss_total": 0.5609229207038879 }, { "epoch": 1.8233766233766233, "step": 702, "train/loss_ctc": 0.6225611567497253, "train/loss_error": 0.45393913984298706, "train/loss_total": 0.48766353726387024 }, { "epoch": 1.825974025974026, "step": 703, "train/loss_ctc": 0.49108466506004333, "train/loss_error": 0.4439312219619751, "train/loss_total": 0.4533619284629822 }, { "epoch": 1.8285714285714287, "step": 704, "train/loss_ctc": 0.5411069393157959, "train/loss_error": 0.4816889464855194, "train/loss_total": 0.49357253313064575 }, { "epoch": 1.8311688311688312, "step": 705, "train/loss_ctc": 0.5772061347961426, "train/loss_error": 0.4883875250816345, "train/loss_total": 0.5061512589454651 }, { "epoch": 1.8337662337662337, "step": 706, "train/loss_ctc": 0.337287962436676, "train/loss_error": 0.4379150867462158, "train/loss_total": 0.41778966784477234 }, { "epoch": 1.8363636363636364, "step": 707, "train/loss_ctc": 0.6538031697273254, "train/loss_error": 0.36180350184440613, "train/loss_total": 0.42020344734191895 }, { "epoch": 1.838961038961039, "step": 708, "train/loss_ctc": 0.6534125804901123, "train/loss_error": 0.5282236337661743, "train/loss_total": 0.553261399269104 }, { "epoch": 1.8415584415584414, "step": 709, "train/loss_ctc": 0.40431147813796997, "train/loss_error": 0.377130389213562, "train/loss_total": 0.3825666308403015 }, { "epoch": 1.844155844155844, "grad_norm": 9582.0595703125, "learning_rate": 1.8966233766233768e-05, "loss": 0.4762, "step": 710 }, { "epoch": 1.844155844155844, "step": 710, "train/loss_ctc": 0.4519469141960144, "train/loss_error": 0.4065411686897278, "train/loss_total": 0.415622353553772 }, { "epoch": 1.8467532467532468, "step": 711, "train/loss_ctc": 0.5256520509719849, "train/loss_error": 0.38287079334259033, "train/loss_total": 0.4114270508289337 }, { "epoch": 1.8493506493506493, "step": 712, "train/loss_ctc": 0.4552084505558014, "train/loss_error": 0.3743671774864197, "train/loss_total": 0.390535444021225 }, { "epoch": 1.8519480519480518, "step": 713, "train/loss_ctc": 0.36671262979507446, "train/loss_error": 0.33411139249801636, "train/loss_total": 0.3406316637992859 }, { "epoch": 1.8545454545454545, "step": 714, "train/loss_ctc": 0.6996239423751831, "train/loss_error": 0.5391477346420288, "train/loss_total": 0.5712429881095886 }, { "epoch": 1.8571428571428572, "step": 715, "train/loss_ctc": 0.443052738904953, "train/loss_error": 0.33813050389289856, "train/loss_total": 0.35911497473716736 }, { "epoch": 1.8597402597402597, "step": 716, "train/loss_ctc": 0.7320840954780579, "train/loss_error": 0.4283214807510376, "train/loss_total": 0.4890739917755127 }, { "epoch": 1.8623376623376622, "step": 717, "train/loss_ctc": 0.43433845043182373, "train/loss_error": 0.3617010712623596, "train/loss_total": 0.37622854113578796 }, { "epoch": 1.864935064935065, "step": 718, "train/loss_ctc": 1.1098427772521973, "train/loss_error": 0.5009755492210388, "train/loss_total": 0.6227489709854126 }, { "epoch": 1.8675324675324676, "step": 719, "train/loss_ctc": 0.5731562972068787, "train/loss_error": 0.35603323578834534, "train/loss_total": 0.3994578719139099 }, { "epoch": 1.87012987012987, "grad_norm": 11722.2314453125, "learning_rate": 1.881038961038961e-05, "loss": 0.4376, "step": 720 }, { "epoch": 1.87012987012987, "step": 720, "train/loss_ctc": 0.7369223833084106, "train/loss_error": 0.4783039689064026, "train/loss_total": 0.5300276875495911 }, { "epoch": 1.8727272727272726, "step": 721, "train/loss_ctc": 0.39394283294677734, "train/loss_error": 0.38102972507476807, "train/loss_total": 0.38361236453056335 }, { "epoch": 1.8753246753246753, "step": 722, "train/loss_ctc": 0.48900341987609863, "train/loss_error": 0.4769112765789032, "train/loss_total": 0.4793297052383423 }, { "epoch": 1.877922077922078, "step": 723, "train/loss_ctc": 0.44373631477355957, "train/loss_error": 0.4342517554759979, "train/loss_total": 0.43614867329597473 }, { "epoch": 1.8805194805194805, "step": 724, "train/loss_ctc": 0.6731038093566895, "train/loss_error": 0.4763578772544861, "train/loss_total": 0.5157070755958557 }, { "epoch": 1.883116883116883, "step": 725, "train/loss_ctc": 0.4215391278266907, "train/loss_error": 0.3436170816421509, "train/loss_total": 0.35920149087905884 }, { "epoch": 1.8857142857142857, "step": 726, "train/loss_ctc": 0.5490015745162964, "train/loss_error": 0.46516716480255127, "train/loss_total": 0.4819340407848358 }, { "epoch": 1.8883116883116884, "step": 727, "train/loss_ctc": 0.4511847198009491, "train/loss_error": 0.3480672836303711, "train/loss_total": 0.3686907887458801 }, { "epoch": 1.8909090909090909, "step": 728, "train/loss_ctc": 0.5338492393493652, "train/loss_error": 0.44016778469085693, "train/loss_total": 0.45890408754348755 }, { "epoch": 1.8935064935064934, "step": 729, "train/loss_ctc": 0.47241559624671936, "train/loss_error": 0.4500066041946411, "train/loss_total": 0.4544883966445923 }, { "epoch": 1.896103896103896, "grad_norm": 10374.9833984375, "learning_rate": 1.8654545454545454e-05, "loss": 0.4468, "step": 730 }, { "epoch": 1.896103896103896, "step": 730, "train/loss_ctc": 0.4433000981807709, "train/loss_error": 0.4201076030731201, "train/loss_total": 0.4247461259365082 }, { "epoch": 1.8987012987012988, "step": 731, "train/loss_ctc": 0.35742270946502686, "train/loss_error": 0.47614946961402893, "train/loss_total": 0.45240411162376404 }, { "epoch": 1.9012987012987013, "step": 732, "train/loss_ctc": 0.6166063547134399, "train/loss_error": 0.4408595561981201, "train/loss_total": 0.47600892186164856 }, { "epoch": 1.9038961038961038, "step": 733, "train/loss_ctc": 0.45360952615737915, "train/loss_error": 0.4364159405231476, "train/loss_total": 0.4398546516895294 }, { "epoch": 1.9064935064935065, "step": 734, "train/loss_ctc": 0.4911096692085266, "train/loss_error": 0.434873104095459, "train/loss_total": 0.44612041115760803 }, { "epoch": 1.9090909090909092, "step": 735, "train/loss_ctc": 0.31540048122406006, "train/loss_error": 0.4390723705291748, "train/loss_total": 0.41433799266815186 }, { "epoch": 1.9116883116883117, "step": 736, "train/loss_ctc": 0.33879226446151733, "train/loss_error": 0.3797464668750763, "train/loss_total": 0.3715556263923645 }, { "epoch": 1.9142857142857141, "step": 737, "train/loss_ctc": 0.5072047710418701, "train/loss_error": 0.4927273988723755, "train/loss_total": 0.4956229031085968 }, { "epoch": 1.9168831168831169, "step": 738, "train/loss_ctc": 0.5609079599380493, "train/loss_error": 0.4063657224178314, "train/loss_total": 0.43727418780326843 }, { "epoch": 1.9194805194805196, "step": 739, "train/loss_ctc": 0.6507038474082947, "train/loss_error": 0.5107905864715576, "train/loss_total": 0.538773238658905 }, { "epoch": 1.922077922077922, "grad_norm": 13648.1826171875, "learning_rate": 1.84987012987013e-05, "loss": 0.4497, "step": 740 }, { "epoch": 1.922077922077922, "step": 740, "train/loss_ctc": 0.5177335739135742, "train/loss_error": 0.3683600425720215, "train/loss_total": 0.3982347548007965 }, { "epoch": 1.9246753246753245, "step": 741, "train/loss_ctc": 0.5925301909446716, "train/loss_error": 0.49510613083839417, "train/loss_total": 0.5145909190177917 }, { "epoch": 1.9272727272727272, "step": 742, "train/loss_ctc": 0.4657670855522156, "train/loss_error": 0.45718613266944885, "train/loss_total": 0.4589023292064667 }, { "epoch": 1.92987012987013, "step": 743, "train/loss_ctc": 0.3556225299835205, "train/loss_error": 0.4169783294200897, "train/loss_total": 0.4047071933746338 }, { "epoch": 1.9324675324675324, "step": 744, "train/loss_ctc": 0.6097930669784546, "train/loss_error": 0.5637280941009521, "train/loss_total": 0.5729410648345947 }, { "epoch": 1.935064935064935, "step": 745, "train/loss_ctc": 0.44805434346199036, "train/loss_error": 0.4124574661254883, "train/loss_total": 0.41957685351371765 }, { "epoch": 1.9376623376623376, "step": 746, "train/loss_ctc": 0.50528883934021, "train/loss_error": 0.4280138909816742, "train/loss_total": 0.4434688985347748 }, { "epoch": 1.9402597402597404, "step": 747, "train/loss_ctc": 0.5609554648399353, "train/loss_error": 0.42048537731170654, "train/loss_total": 0.44857943058013916 }, { "epoch": 1.9428571428571428, "step": 748, "train/loss_ctc": 0.5424987077713013, "train/loss_error": 0.4132038652896881, "train/loss_total": 0.43906283378601074 }, { "epoch": 1.9454545454545453, "step": 749, "train/loss_ctc": 0.47490352392196655, "train/loss_error": 0.55413818359375, "train/loss_total": 0.5382912755012512 }, { "epoch": 1.948051948051948, "grad_norm": 12288.205078125, "learning_rate": 1.8342857142857144e-05, "loss": 0.4638, "step": 750 }, { "epoch": 1.948051948051948, "step": 750, "train/loss_ctc": 0.5154086351394653, "train/loss_error": 0.4339587688446045, "train/loss_total": 0.45024874806404114 }, { "epoch": 1.9506493506493507, "step": 751, "train/loss_ctc": 0.5225628018379211, "train/loss_error": 0.49716344475746155, "train/loss_total": 0.5022433400154114 }, { "epoch": 1.9532467532467532, "step": 752, "train/loss_ctc": 0.5740833282470703, "train/loss_error": 0.43955862522125244, "train/loss_total": 0.466463565826416 }, { "epoch": 1.9558441558441557, "step": 753, "train/loss_ctc": 0.49688416719436646, "train/loss_error": 0.40511834621429443, "train/loss_total": 0.42347151041030884 }, { "epoch": 1.9584415584415584, "step": 754, "train/loss_ctc": 1.0787670612335205, "train/loss_error": 0.47241589426994324, "train/loss_total": 0.5936861634254456 }, { "epoch": 1.9610389610389611, "step": 755, "train/loss_ctc": 0.45252448320388794, "train/loss_error": 0.4030362665653229, "train/loss_total": 0.41293394565582275 }, { "epoch": 1.9636363636363636, "step": 756, "train/loss_ctc": 0.47963249683380127, "train/loss_error": 0.4237360656261444, "train/loss_total": 0.43491536378860474 }, { "epoch": 1.9662337662337661, "step": 757, "train/loss_ctc": 0.3660067915916443, "train/loss_error": 0.4249142110347748, "train/loss_total": 0.4131327271461487 }, { "epoch": 1.9688311688311688, "step": 758, "train/loss_ctc": 0.444174587726593, "train/loss_error": 0.48593947291374207, "train/loss_total": 0.4775865077972412 }, { "epoch": 1.9714285714285715, "step": 759, "train/loss_ctc": 0.35585498809814453, "train/loss_error": 0.36885809898376465, "train/loss_total": 0.3662574887275696 }, { "epoch": 1.974025974025974, "grad_norm": 11549.3046875, "learning_rate": 1.818701298701299e-05, "loss": 0.4541, "step": 760 }, { "epoch": 1.974025974025974, "step": 760, "train/loss_ctc": 0.5118741989135742, "train/loss_error": 0.48890480399131775, "train/loss_total": 0.49349871277809143 }, { "epoch": 1.9766233766233765, "step": 761, "train/loss_ctc": 0.3130386173725128, "train/loss_error": 0.32327479124069214, "train/loss_total": 0.3212275803089142 }, { "epoch": 1.9792207792207792, "step": 762, "train/loss_ctc": 0.2906043827533722, "train/loss_error": 0.3671187460422516, "train/loss_total": 0.3518158793449402 }, { "epoch": 1.981818181818182, "step": 763, "train/loss_ctc": 0.3576236069202423, "train/loss_error": 0.38594552874565125, "train/loss_total": 0.38028115034103394 }, { "epoch": 1.9844155844155844, "step": 764, "train/loss_ctc": 0.6682290434837341, "train/loss_error": 0.39614439010620117, "train/loss_total": 0.4505613446235657 }, { "epoch": 1.987012987012987, "step": 765, "train/loss_ctc": 0.5382173657417297, "train/loss_error": 0.4087863266468048, "train/loss_total": 0.4346725642681122 }, { "epoch": 1.9896103896103896, "step": 766, "train/loss_ctc": 0.4863837957382202, "train/loss_error": 0.4989496171474457, "train/loss_total": 0.4964364767074585 }, { "epoch": 1.9922077922077923, "step": 767, "train/loss_ctc": 0.49977394938468933, "train/loss_error": 0.4030526876449585, "train/loss_total": 0.4223969280719757 }, { "epoch": 1.9948051948051948, "step": 768, "train/loss_ctc": 0.6031382083892822, "train/loss_error": 0.5635022521018982, "train/loss_total": 0.5714294910430908 }, { "epoch": 1.9974025974025973, "step": 769, "train/loss_ctc": 0.36966729164123535, "train/loss_error": 0.41744130849838257, "train/loss_total": 0.4078865051269531 }, { "epoch": 2.0, "grad_norm": 17152.76171875, "learning_rate": 1.803116883116883e-05, "loss": 0.433, "step": 770 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.680915117263794, "val/loss_error": 0.17383477091789246, "val/loss_total": 0.2752508521080017 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.8060282468795776, "val/loss_error": 0.5437341928482056, "val/loss_total": 0.5961930155754089 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.10961653292179108, "val/loss_error": 0.30613142251968384, "val/loss_total": 0.2668284475803375 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.16724073886871338, "val/loss_error": 0.18268820643424988, "val/loss_total": 0.17959871888160706 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.18930652737617493, "val/loss_error": 0.2223411202430725, "val/loss_total": 0.21573419868946075 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.13767023384571075, "val/loss_error": 0.3320491313934326, "val/loss_total": 0.29317334294319153 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.2555467188358307, "val/loss_error": 0.4245340824127197, "val/loss_total": 0.3907366096973419 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.3978431224822998, "val/loss_error": 0.3270410895347595, "val/loss_total": 0.341201514005661 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.4549086093902588, "val/loss_error": 0.3640858232975006, "val/loss_total": 0.3822503983974457 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.22309783101081848, "val/loss_error": 0.36338868737220764, "val/loss_total": 0.3353305160999298 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.14883291721343994, "val/loss_error": 0.19182395935058594, "val/loss_total": 0.18322576582431793 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.24520044028759003, "val/loss_error": 0.36192604899406433, "val/loss_total": 0.3385809361934662 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.7559667229652405, "val/loss_error": 0.4997890293598175, "val/loss_total": 0.5510245561599731 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.7237899899482727, "val/loss_error": 0.49167799949645996, "val/loss_total": 0.5381004214286804 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.6082496643066406, "val/loss_error": 0.3421156704425812, "val/loss_total": 0.39534246921539307 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5476247072219849, "val/loss_error": 0.6928290724754333, "val/loss_total": 0.6637882590293884 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.10170631110668182, "val/loss_error": 0.25085440278053284, "val/loss_total": 0.2210247814655304 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.15542952716350555, "val/loss_error": 0.27319857478141785, "val/loss_total": 0.24964477121829987 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.16137950122356415, "val/loss_error": 0.466147243976593, "val/loss_total": 0.4051936864852905 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5402660369873047, "val/loss_error": 0.4846173822879791, "val/loss_total": 0.4957471191883087 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.6375261545181274, "val/loss_error": 0.6617396473884583, "val/loss_total": 0.6568969488143921 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.2998090088367462, "val/loss_error": 0.5731147527694702, "val/loss_total": 0.5184535980224609 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.2897116243839264, "val/loss_error": 0.47858917713165283, "val/loss_total": 0.44081369042396545 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.19411799311637878, "val/loss_error": 0.1625235229730606, "val/loss_total": 0.16884241998195648 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.31777605414390564, "val/loss_error": 0.5552201271057129, "val/loss_total": 0.5077313184738159 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.4462614953517914, "val/loss_error": 0.5801241993904114, "val/loss_total": 0.5533517003059387 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.29203441739082336, "val/loss_error": 0.3821226954460144, "val/loss_total": 0.3641050457954407 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.17092865705490112, "val/loss_error": 0.2892080247402191, "val/loss_total": 0.2655521631240845 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.024685628712177277, "val/loss_error": 0.1057719886302948, "val/loss_total": 0.08955471962690353 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.23853114247322083, "val/loss_error": 0.542203426361084, "val/loss_total": 0.48146897554397583 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.08882196247577667, "val/loss_error": 0.14753510057926178, "val/loss_total": 0.13579247891902924 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.10950889438390732, "val/loss_error": 0.3281564712524414, "val/loss_total": 0.2844269573688507 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.08957140892744064, "val/loss_error": 0.2394886016845703, "val/loss_total": 0.20950517058372498 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5829106569290161, "val/loss_error": 0.7909975051879883, "val/loss_total": 0.7493801712989807 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.07092205435037613, "val/loss_error": 0.46962419152259827, "val/loss_total": 0.38988378643989563 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.16700707376003265, "val/loss_error": 0.4935098886489868, "val/loss_total": 0.4282093048095703 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.46746471524238586, "val/loss_error": 0.7834643721580505, "val/loss_total": 0.7202644348144531 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.4690442979335785, "val/loss_error": 0.4496062099933624, "val/loss_total": 0.4534938335418701 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.21087072789669037, "val/loss_error": 0.5163319706916809, "val/loss_total": 0.4552397131919861 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5315051078796387, "val/loss_error": 0.5862923264503479, "val/loss_total": 0.575334906578064 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.2565003037452698, "val/loss_error": 0.23431779444217682, "val/loss_total": 0.2387543022632599 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.32829374074935913, "val/loss_error": 0.4355319142341614, "val/loss_total": 0.4140842854976654 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.48402899503707886, "val/loss_error": 0.9711059331893921, "val/loss_total": 0.8736905455589294 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5985138416290283, "val/loss_error": 0.5698925852775574, "val/loss_total": 0.5756168365478516 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5102601051330566, "val/loss_error": 0.34004977345466614, "val/loss_total": 0.37409186363220215 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.44496917724609375, "val/loss_error": 0.6575208306312561, "val/loss_total": 0.6150104999542236 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.684817373752594, "val/loss_error": 0.4152393341064453, "val/loss_total": 0.469154953956604 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.3982682526111603, "val/loss_error": 0.603018581867218, "val/loss_total": 0.562068521976471 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.48363208770751953, "val/loss_error": 0.23473531007766724, "val/loss_total": 0.2845146656036377 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.6565335988998413, "val/loss_error": 0.8413261771202087, "val/loss_total": 0.8043676614761353 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.4021139144897461, "val/loss_error": 0.8887939453125, "val/loss_total": 0.7914579510688782 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.37995460629463196, "val/loss_error": 0.41000574827194214, "val/loss_total": 0.4039955139160156 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.39248940348625183, "val/loss_error": 0.5010086297988892, "val/loss_total": 0.47930479049682617 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.34725192189216614, "val/loss_error": 0.7700244784355164, "val/loss_total": 0.6854699850082397 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.1673702895641327, "val/loss_error": 0.40413305163383484, "val/loss_total": 0.3567804992198944 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.20595411956310272, "val/loss_error": 0.5172392725944519, "val/loss_total": 0.4549822509288788 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.13778527081012726, "val/loss_error": 0.26690641045570374, "val/loss_total": 0.24108219146728516 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5651096701622009, "val/loss_error": 0.3527616858482361, "val/loss_total": 0.39523130655288696 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.6988420486450195, "val/loss_error": 0.7937723994255066, "val/loss_total": 0.7747863531112671 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.03742444887757301, "val/loss_error": 0.13006365299224854, "val/loss_total": 0.11153581738471985 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.7800800800323486, "val/loss_error": 0.5457596182823181, "val/loss_total": 0.5926237106323242 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.7032290101051331, "val/loss_error": 0.41950079798698425, "val/loss_total": 0.4762464463710785 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.08104585111141205, "val/loss_error": 0.10089635848999023, "val/loss_total": 0.0969262570142746 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.22762000560760498, "val/loss_error": 0.406170517206192, "val/loss_total": 0.3704604208469391 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.14449213445186615, "val/loss_error": 0.21165692806243896, "val/loss_total": 0.19822397828102112 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.21460691094398499, "val/loss_error": 0.46275702118873596, "val/loss_total": 0.41312700510025024 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.32429707050323486, "val/loss_error": 0.3062078356742859, "val/loss_total": 0.30982568860054016 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5677694082260132, "val/loss_error": 0.49876168370246887, "val/loss_total": 0.5125632286071777 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5457555055618286, "val/loss_error": 0.5316460728645325, "val/loss_total": 0.5344679951667786 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.3984563648700714, "val/loss_error": 0.42100703716278076, "val/loss_total": 0.4164969325065613 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.2824959456920624, "val/loss_error": 0.30511489510536194, "val/loss_total": 0.3005911111831665 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5554267764091492, "val/loss_error": 0.46245086193084717, "val/loss_total": 0.48104605078697205 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.6201465725898743, "val/loss_error": 0.6388517618179321, "val/loss_total": 0.6351107358932495 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.21783427894115448, "val/loss_error": 0.3391264081001282, "val/loss_total": 0.3148679733276367 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.2591453790664673, "val/loss_error": 0.3364295959472656, "val/loss_total": 0.320972740650177 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.04813365265727043, "val/loss_error": 0.31931862235069275, "val/loss_total": 0.26508161425590515 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.8048776388168335, "val/loss_error": 0.5582758784294128, "val/loss_total": 0.607596218585968 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5932003259658813, "val/loss_error": 0.40115928649902344, "val/loss_total": 0.439567506313324 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.6592594981193542, "val/loss_error": 0.5043666362762451, "val/loss_total": 0.535345196723938 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.0033660910557955503, "val/loss_error": 0.4622323215007782, "val/loss_total": 0.37045907974243164 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.3382970690727234, "val/loss_error": 0.31255364418029785, "val/loss_total": 0.3177023231983185 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.023993873968720436, "val/loss_error": 0.16450150310993195, "val/loss_total": 0.13639996945858002 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.18195828795433044, "val/loss_error": 0.3301336169242859, "val/loss_total": 0.3004985451698303 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.466138631105423, "val/loss_error": 0.37003329396247864, "val/loss_total": 0.3892543911933899 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5441433191299438, "val/loss_error": 0.6861721277236938, "val/loss_total": 0.6577664017677307 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5327460765838623, "val/loss_error": 0.6834729313850403, "val/loss_total": 0.6533275842666626 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.03153311461210251, "val/loss_error": 0.15580891072750092, "val/loss_total": 0.13095375895500183 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.4566832184791565, "val/loss_error": 0.49693813920021057, "val/loss_total": 0.48888716101646423 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.23437203466892242, "val/loss_error": 0.2771236002445221, "val/loss_total": 0.2685732841491699 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.3594389259815216, "val/loss_error": 0.45472103357315063, "val/loss_total": 0.4356646239757538 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.29800480604171753, "val/loss_error": 0.5825624465942383, "val/loss_total": 0.5256509184837341 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.8191030025482178, "val/loss_error": 0.35431498289108276, "val/loss_total": 0.4472725987434387 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.29211482405662537, "val/loss_error": 0.14701907336711884, "val/loss_total": 0.1760382205247879 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.26017898321151733, "val/loss_error": 0.3364645838737488, "val/loss_total": 0.3212074935436249 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.4907594919204712, "val/loss_error": 0.5256654024124146, "val/loss_total": 0.5186842083930969 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.22545066475868225, "val/loss_error": 0.2276301383972168, "val/loss_total": 0.22719424962997437 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.13214454054832458, "val/loss_error": 0.319469690322876, "val/loss_total": 0.2820046544075012 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.5661864280700684, "val/loss_error": 0.43962815403938293, "val/loss_total": 0.46493983268737793 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.1313091516494751, "val/loss_error": 0.27762553095817566, "val/loss_total": 0.24836225807666779 }, { "epoch": 2.0, "step": 770, "val/loss_ctc": 0.17518070340156555, "val/loss_error": 0.2925023138523102, "val/loss_total": 0.26903799176216125 }, { "epoch": 2.0, "eval_eval/f1_0": 0.5153664350509644, "eval_eval/f1_1": 0.9270462393760681, "eval_eval/precision_0": 0.5343137383460999, "eval_eval/precision_1": 0.922123908996582, "eval_eval/recall_0": 0.49771690368652344, "eval_eval/recall_1": 0.93202143907547, "eval_eval/wer": 0.1541646259149732, "eval_loss": 0.41123276948928833, "eval_runtime": 7.5304, "eval_samples_per_second": 13.279, "eval_steps_per_second": 13.279, "step": 770 }, { "epoch": 2.0, "step": 770, "train/loss_ctc": 0.41347262263298035, "train/loss_error": 0.39450815320014954, "train/loss_total": 0.39830106496810913 }, { "epoch": 2.0025974025974027, "step": 771, "train/loss_ctc": 0.39697128534317017, "train/loss_error": 0.4635641276836395, "train/loss_total": 0.45024555921554565 }, { "epoch": 2.005194805194805, "step": 772, "train/loss_ctc": 0.5271583795547485, "train/loss_error": 0.4320818781852722, "train/loss_total": 0.45109719038009644 }, { "epoch": 2.0077922077922077, "step": 773, "train/loss_ctc": 0.42960798740386963, "train/loss_error": 0.3448140323162079, "train/loss_total": 0.3617728352546692 }, { "epoch": 2.0103896103896104, "step": 774, "train/loss_ctc": 0.369924932718277, "train/loss_error": 0.40433061122894287, "train/loss_total": 0.39744946360588074 }, { "epoch": 2.012987012987013, "step": 775, "train/loss_ctc": 0.4533753991127014, "train/loss_error": 0.36876094341278076, "train/loss_total": 0.3856838345527649 }, { "epoch": 2.0155844155844154, "step": 776, "train/loss_ctc": 0.3096637725830078, "train/loss_error": 0.35335803031921387, "train/loss_total": 0.34461918473243713 }, { "epoch": 2.018181818181818, "step": 777, "train/loss_ctc": 0.4668458104133606, "train/loss_error": 0.3008449971675873, "train/loss_total": 0.3340451717376709 }, { "epoch": 2.020779220779221, "step": 778, "train/loss_ctc": 0.6917890310287476, "train/loss_error": 0.38959038257598877, "train/loss_total": 0.4500300884246826 }, { "epoch": 2.0233766233766235, "step": 779, "train/loss_ctc": 0.30596011877059937, "train/loss_error": 0.2883242070674896, "train/loss_total": 0.2918514013290405 }, { "epoch": 2.0259740259740258, "grad_norm": 9573.7578125, "learning_rate": 1.7875324675324675e-05, "loss": 0.3865, "step": 780 }, { "epoch": 2.0259740259740258, "step": 780, "train/loss_ctc": 0.3007905185222626, "train/loss_error": 0.38954198360443115, "train/loss_total": 0.37179169058799744 }, { "epoch": 2.0285714285714285, "step": 781, "train/loss_ctc": 0.505885124206543, "train/loss_error": 0.37434542179107666, "train/loss_total": 0.4006533920764923 }, { "epoch": 2.031168831168831, "step": 782, "train/loss_ctc": 0.4268304705619812, "train/loss_error": 0.5332185626029968, "train/loss_total": 0.5119409561157227 }, { "epoch": 2.033766233766234, "step": 783, "train/loss_ctc": 0.6379947662353516, "train/loss_error": 0.3595232367515564, "train/loss_total": 0.4152175784111023 }, { "epoch": 2.036363636363636, "step": 784, "train/loss_ctc": 0.3897671699523926, "train/loss_error": 0.36012378334999084, "train/loss_total": 0.36605244874954224 }, { "epoch": 2.038961038961039, "step": 785, "train/loss_ctc": 0.4012129306793213, "train/loss_error": 0.44366762042045593, "train/loss_total": 0.43517667055130005 }, { "epoch": 2.0415584415584416, "step": 786, "train/loss_ctc": 0.5858749151229858, "train/loss_error": 0.5218259692192078, "train/loss_total": 0.5346357822418213 }, { "epoch": 2.0441558441558443, "step": 787, "train/loss_ctc": 0.6771523952484131, "train/loss_error": 0.5487126708030701, "train/loss_total": 0.5744006633758545 }, { "epoch": 2.0467532467532465, "step": 788, "train/loss_ctc": 0.4529455304145813, "train/loss_error": 0.4266066253185272, "train/loss_total": 0.43187442421913147 }, { "epoch": 2.0493506493506493, "step": 789, "train/loss_ctc": 0.5032700300216675, "train/loss_error": 0.41126930713653564, "train/loss_total": 0.42966946959495544 }, { "epoch": 2.051948051948052, "grad_norm": 13240.4375, "learning_rate": 1.771948051948052e-05, "loss": 0.4471, "step": 790 }, { "epoch": 2.051948051948052, "step": 790, "train/loss_ctc": 0.46230748295783997, "train/loss_error": 0.3916381001472473, "train/loss_total": 0.40577197074890137 }, { "epoch": 2.0545454545454547, "step": 791, "train/loss_ctc": 0.6950356960296631, "train/loss_error": 0.5216543078422546, "train/loss_total": 0.5563305616378784 }, { "epoch": 2.057142857142857, "step": 792, "train/loss_ctc": 0.4351199269294739, "train/loss_error": 0.38637909293174744, "train/loss_total": 0.39612728357315063 }, { "epoch": 2.0597402597402596, "step": 793, "train/loss_ctc": 0.9195190668106079, "train/loss_error": 0.46983662247657776, "train/loss_total": 0.5597730875015259 }, { "epoch": 2.0623376623376624, "step": 794, "train/loss_ctc": 0.6844026446342468, "train/loss_error": 0.4721972644329071, "train/loss_total": 0.514638364315033 }, { "epoch": 2.064935064935065, "step": 795, "train/loss_ctc": 0.4423374533653259, "train/loss_error": 0.48171353340148926, "train/loss_total": 0.47383832931518555 }, { "epoch": 2.0675324675324673, "step": 796, "train/loss_ctc": 0.5774547457695007, "train/loss_error": 0.4047378897666931, "train/loss_total": 0.43928125500679016 }, { "epoch": 2.07012987012987, "step": 797, "train/loss_ctc": 0.6038207411766052, "train/loss_error": 0.3756146728992462, "train/loss_total": 0.421255886554718 }, { "epoch": 2.0727272727272728, "step": 798, "train/loss_ctc": 0.45140713453292847, "train/loss_error": 0.4139549136161804, "train/loss_total": 0.421445369720459 }, { "epoch": 2.0753246753246755, "step": 799, "train/loss_ctc": 0.4403112828731537, "train/loss_error": 0.3888787627220154, "train/loss_total": 0.3991652727127075 }, { "epoch": 2.0779220779220777, "grad_norm": 10369.375, "learning_rate": 1.7563636363636365e-05, "loss": 0.4588, "step": 800 }, { "epoch": 2.0779220779220777, "step": 800, "train/loss_ctc": 0.5085282325744629, "train/loss_error": 0.4031386375427246, "train/loss_total": 0.4242165684700012 }, { "epoch": 2.0805194805194804, "step": 801, "train/loss_ctc": 0.4089943468570709, "train/loss_error": 0.282321035861969, "train/loss_total": 0.3076556921005249 }, { "epoch": 2.083116883116883, "step": 802, "train/loss_ctc": 0.5095385313034058, "train/loss_error": 0.3591122329235077, "train/loss_total": 0.3891974985599518 }, { "epoch": 2.085714285714286, "step": 803, "train/loss_ctc": 0.4988602101802826, "train/loss_error": 0.45945557951927185, "train/loss_total": 0.467336505651474 }, { "epoch": 2.088311688311688, "step": 804, "train/loss_ctc": 0.6910096406936646, "train/loss_error": 0.35891079902648926, "train/loss_total": 0.42533057928085327 }, { "epoch": 2.090909090909091, "step": 805, "train/loss_ctc": 0.5617861747741699, "train/loss_error": 0.4695836901664734, "train/loss_total": 0.48802417516708374 }, { "epoch": 2.0935064935064935, "step": 806, "train/loss_ctc": 0.27749913930892944, "train/loss_error": 0.3666432499885559, "train/loss_total": 0.3488144278526306 }, { "epoch": 2.0961038961038962, "step": 807, "train/loss_ctc": 0.6592663526535034, "train/loss_error": 0.44512006640434265, "train/loss_total": 0.48794931173324585 }, { "epoch": 2.0987012987012985, "step": 808, "train/loss_ctc": 0.4478517472743988, "train/loss_error": 0.4273723363876343, "train/loss_total": 0.4314682185649872 }, { "epoch": 2.101298701298701, "step": 809, "train/loss_ctc": 0.37942540645599365, "train/loss_error": 0.4595605134963989, "train/loss_total": 0.4435335099697113 }, { "epoch": 2.103896103896104, "grad_norm": 9718.974609375, "learning_rate": 1.740779220779221e-05, "loss": 0.4214, "step": 810 }, { "epoch": 2.103896103896104, "step": 810, "train/loss_ctc": 0.5936996936798096, "train/loss_error": 0.4211881756782532, "train/loss_total": 0.4556904733181 }, { "epoch": 2.1064935064935066, "step": 811, "train/loss_ctc": 0.41449594497680664, "train/loss_error": 0.4304904341697693, "train/loss_total": 0.42729154229164124 }, { "epoch": 2.109090909090909, "step": 812, "train/loss_ctc": 0.3641122877597809, "train/loss_error": 0.3459959626197815, "train/loss_total": 0.3496192395687103 }, { "epoch": 2.1116883116883116, "step": 813, "train/loss_ctc": 0.35837745666503906, "train/loss_error": 0.3982096016407013, "train/loss_total": 0.39024317264556885 }, { "epoch": 2.1142857142857143, "step": 814, "train/loss_ctc": 0.4291490912437439, "train/loss_error": 0.44020071625709534, "train/loss_total": 0.4379903972148895 }, { "epoch": 2.116883116883117, "step": 815, "train/loss_ctc": 0.5382010340690613, "train/loss_error": 0.39055997133255005, "train/loss_total": 0.4200882017612457 }, { "epoch": 2.1194805194805193, "step": 816, "train/loss_ctc": 0.46529632806777954, "train/loss_error": 0.430375874042511, "train/loss_total": 0.4373599886894226 }, { "epoch": 2.122077922077922, "step": 817, "train/loss_ctc": 0.6236335039138794, "train/loss_error": 0.47409358620643616, "train/loss_total": 0.5040015578269958 }, { "epoch": 2.1246753246753247, "step": 818, "train/loss_ctc": 0.4060022234916687, "train/loss_error": 0.32274556159973145, "train/loss_total": 0.3393968939781189 }, { "epoch": 2.1272727272727274, "step": 819, "train/loss_ctc": 0.7216699123382568, "train/loss_error": 0.4981381893157959, "train/loss_total": 0.5428445339202881 }, { "epoch": 2.1298701298701297, "grad_norm": 11628.2421875, "learning_rate": 1.725194805194805e-05, "loss": 0.4305, "step": 820 }, { "epoch": 2.1298701298701297, "step": 820, "train/loss_ctc": 0.409427285194397, "train/loss_error": 0.36920514702796936, "train/loss_total": 0.3772495687007904 }, { "epoch": 2.1324675324675324, "step": 821, "train/loss_ctc": 0.5463451147079468, "train/loss_error": 0.4218142330646515, "train/loss_total": 0.4467204213142395 }, { "epoch": 2.135064935064935, "step": 822, "train/loss_ctc": 0.508476972579956, "train/loss_error": 0.435830295085907, "train/loss_total": 0.45035964250564575 }, { "epoch": 2.137662337662338, "step": 823, "train/loss_ctc": 0.6742210388183594, "train/loss_error": 0.3813769221305847, "train/loss_total": 0.4399457573890686 }, { "epoch": 2.14025974025974, "step": 824, "train/loss_ctc": 0.43271195888519287, "train/loss_error": 0.46758541464805603, "train/loss_total": 0.4606107473373413 }, { "epoch": 2.142857142857143, "step": 825, "train/loss_ctc": 0.5058780312538147, "train/loss_error": 0.45330989360809326, "train/loss_total": 0.463823527097702 }, { "epoch": 2.1454545454545455, "step": 826, "train/loss_ctc": 0.3335184156894684, "train/loss_error": 0.31776323914527893, "train/loss_total": 0.32091426849365234 }, { "epoch": 2.148051948051948, "step": 827, "train/loss_ctc": 0.5591492652893066, "train/loss_error": 0.4727970361709595, "train/loss_total": 0.4900674819946289 }, { "epoch": 2.1506493506493505, "step": 828, "train/loss_ctc": 0.403205543756485, "train/loss_error": 0.42245739698410034, "train/loss_total": 0.41860705614089966 }, { "epoch": 2.153246753246753, "step": 829, "train/loss_ctc": 0.5145460367202759, "train/loss_error": 0.3598005473613739, "train/loss_total": 0.39074966311454773 }, { "epoch": 2.155844155844156, "grad_norm": 14881.0927734375, "learning_rate": 1.7096103896103896e-05, "loss": 0.4259, "step": 830 }, { "epoch": 2.155844155844156, "step": 830, "train/loss_ctc": 0.4400797188282013, "train/loss_error": 0.4515416622161865, "train/loss_total": 0.449249267578125 }, { "epoch": 2.1584415584415586, "step": 831, "train/loss_ctc": 0.30711179971694946, "train/loss_error": 0.429286926984787, "train/loss_total": 0.40485188364982605 }, { "epoch": 2.161038961038961, "step": 832, "train/loss_ctc": 0.3907570242881775, "train/loss_error": 0.3822210729122162, "train/loss_total": 0.3839282691478729 }, { "epoch": 2.1636363636363636, "step": 833, "train/loss_ctc": 0.24183958768844604, "train/loss_error": 0.37867534160614014, "train/loss_total": 0.3513081967830658 }, { "epoch": 2.1662337662337663, "step": 834, "train/loss_ctc": 0.5783170461654663, "train/loss_error": 0.442571222782135, "train/loss_total": 0.46972039341926575 }, { "epoch": 2.168831168831169, "step": 835, "train/loss_ctc": 0.6652653217315674, "train/loss_error": 0.5225139856338501, "train/loss_total": 0.5510642528533936 }, { "epoch": 2.1714285714285713, "step": 836, "train/loss_ctc": 0.342498242855072, "train/loss_error": 0.41002875566482544, "train/loss_total": 0.3965226709842682 }, { "epoch": 2.174025974025974, "step": 837, "train/loss_ctc": 0.36259543895721436, "train/loss_error": 0.34281283617019653, "train/loss_total": 0.3467693626880646 }, { "epoch": 2.1766233766233767, "step": 838, "train/loss_ctc": 0.6713797450065613, "train/loss_error": 0.4256833791732788, "train/loss_total": 0.47482264041900635 }, { "epoch": 2.1792207792207794, "step": 839, "train/loss_ctc": 0.5897607803344727, "train/loss_error": 0.46838486194610596, "train/loss_total": 0.4926600754261017 }, { "epoch": 2.1818181818181817, "grad_norm": 14862.8828125, "learning_rate": 1.694025974025974e-05, "loss": 0.4321, "step": 840 }, { "epoch": 2.1818181818181817, "step": 840, "train/loss_ctc": 0.6177065372467041, "train/loss_error": 0.5143692493438721, "train/loss_total": 0.5350367426872253 }, { "epoch": 2.1844155844155844, "step": 841, "train/loss_ctc": 0.3139568567276001, "train/loss_error": 0.35498422384262085, "train/loss_total": 0.3467787504196167 }, { "epoch": 2.187012987012987, "step": 842, "train/loss_ctc": 0.6609050035476685, "train/loss_error": 0.459782212972641, "train/loss_total": 0.5000067949295044 }, { "epoch": 2.18961038961039, "step": 843, "train/loss_ctc": 0.4081082344055176, "train/loss_error": 0.4036921560764313, "train/loss_total": 0.404575377702713 }, { "epoch": 2.192207792207792, "step": 844, "train/loss_ctc": 0.4607420861721039, "train/loss_error": 0.3836720287799835, "train/loss_total": 0.399086058139801 }, { "epoch": 2.1948051948051948, "step": 845, "train/loss_ctc": 0.6040630340576172, "train/loss_error": 0.3919373154640198, "train/loss_total": 0.4343624711036682 }, { "epoch": 2.1974025974025975, "step": 846, "train/loss_ctc": 0.6442917585372925, "train/loss_error": 0.4252021312713623, "train/loss_total": 0.4690200686454773 }, { "epoch": 2.2, "step": 847, "train/loss_ctc": 0.4876896142959595, "train/loss_error": 0.42965224385261536, "train/loss_total": 0.4412597417831421 }, { "epoch": 2.2025974025974024, "step": 848, "train/loss_ctc": 0.383976012468338, "train/loss_error": 0.36528587341308594, "train/loss_total": 0.3690239191055298 }, { "epoch": 2.205194805194805, "step": 849, "train/loss_ctc": 0.5117840766906738, "train/loss_error": 0.4148900806903839, "train/loss_total": 0.43426889181137085 }, { "epoch": 2.207792207792208, "grad_norm": 11797.171875, "learning_rate": 1.6784415584415585e-05, "loss": 0.4333, "step": 850 }, { "epoch": 2.207792207792208, "step": 850, "train/loss_ctc": 0.5235559940338135, "train/loss_error": 0.4464889168739319, "train/loss_total": 0.46190235018730164 }, { "epoch": 2.2103896103896106, "step": 851, "train/loss_ctc": 0.43603843450546265, "train/loss_error": 0.36150071024894714, "train/loss_total": 0.37640827894210815 }, { "epoch": 2.212987012987013, "step": 852, "train/loss_ctc": 0.35369807481765747, "train/loss_error": 0.37288349866867065, "train/loss_total": 0.3690464198589325 }, { "epoch": 2.2155844155844155, "step": 853, "train/loss_ctc": 0.9099494218826294, "train/loss_error": 0.4382050633430481, "train/loss_total": 0.5325539708137512 }, { "epoch": 2.2181818181818183, "step": 854, "train/loss_ctc": 0.36349940299987793, "train/loss_error": 0.39636746048927307, "train/loss_total": 0.38979384303092957 }, { "epoch": 2.220779220779221, "step": 855, "train/loss_ctc": 0.46508005261421204, "train/loss_error": 0.4268033504486084, "train/loss_total": 0.4344586730003357 }, { "epoch": 2.2233766233766232, "step": 856, "train/loss_ctc": 0.5807335376739502, "train/loss_error": 0.4128448963165283, "train/loss_total": 0.44642263650894165 }, { "epoch": 2.225974025974026, "step": 857, "train/loss_ctc": 0.33381396532058716, "train/loss_error": 0.44872036576271057, "train/loss_total": 0.4257391095161438 }, { "epoch": 2.2285714285714286, "step": 858, "train/loss_ctc": 0.5197009444236755, "train/loss_error": 0.48383140563964844, "train/loss_total": 0.4910053312778473 }, { "epoch": 2.2311688311688314, "step": 859, "train/loss_ctc": 0.33309271931648254, "train/loss_error": 0.41821804642677307, "train/loss_total": 0.40119296312332153 }, { "epoch": 2.2337662337662336, "grad_norm": 8212.6630859375, "learning_rate": 1.662857142857143e-05, "loss": 0.4329, "step": 860 }, { "epoch": 2.2337662337662336, "step": 860, "train/loss_ctc": 0.41835877299308777, "train/loss_error": 0.3564020097255707, "train/loss_total": 0.3687933683395386 }, { "epoch": 2.2363636363636363, "step": 861, "train/loss_ctc": 0.5390868186950684, "train/loss_error": 0.38346919417381287, "train/loss_total": 0.4145927429199219 }, { "epoch": 2.238961038961039, "step": 862, "train/loss_ctc": 0.4712357521057129, "train/loss_error": 0.3405696749687195, "train/loss_total": 0.3667029142379761 }, { "epoch": 2.2415584415584417, "step": 863, "train/loss_ctc": 0.4809374213218689, "train/loss_error": 0.3645903766155243, "train/loss_total": 0.3878598213195801 }, { "epoch": 2.244155844155844, "step": 864, "train/loss_ctc": 0.5795261859893799, "train/loss_error": 0.3581917881965637, "train/loss_total": 0.40245866775512695 }, { "epoch": 2.2467532467532467, "step": 865, "train/loss_ctc": 0.4797807037830353, "train/loss_error": 0.3668404817581177, "train/loss_total": 0.3894285261631012 }, { "epoch": 2.2493506493506494, "step": 866, "train/loss_ctc": 0.44796597957611084, "train/loss_error": 0.36629074811935425, "train/loss_total": 0.3826258182525635 }, { "epoch": 2.2519480519480517, "step": 867, "train/loss_ctc": 0.4257166385650635, "train/loss_error": 0.4043627381324768, "train/loss_total": 0.4086335301399231 }, { "epoch": 2.2545454545454544, "step": 868, "train/loss_ctc": 0.4174830913543701, "train/loss_error": 0.4477106034755707, "train/loss_total": 0.4416651129722595 }, { "epoch": 2.257142857142857, "step": 869, "train/loss_ctc": 0.896233856678009, "train/loss_error": 0.4767587184906006, "train/loss_total": 0.5606537461280823 }, { "epoch": 2.25974025974026, "grad_norm": 24332.64453125, "learning_rate": 1.647272727272727e-05, "loss": 0.4123, "step": 870 }, { "epoch": 2.25974025974026, "step": 870, "train/loss_ctc": 0.3547186255455017, "train/loss_error": 0.4285808205604553, "train/loss_total": 0.4138084053993225 }, { "epoch": 2.2623376623376625, "step": 871, "train/loss_ctc": 0.5583764910697937, "train/loss_error": 0.38677778840065, "train/loss_total": 0.4210975170135498 }, { "epoch": 2.264935064935065, "step": 872, "train/loss_ctc": 0.44184964895248413, "train/loss_error": 0.33840733766555786, "train/loss_total": 0.35909581184387207 }, { "epoch": 2.2675324675324675, "step": 873, "train/loss_ctc": 0.5611216425895691, "train/loss_error": 0.43846845626831055, "train/loss_total": 0.4629991054534912 }, { "epoch": 2.27012987012987, "step": 874, "train/loss_ctc": 0.41219931840896606, "train/loss_error": 0.40702733397483826, "train/loss_total": 0.4080617427825928 }, { "epoch": 2.2727272727272725, "step": 875, "train/loss_ctc": 0.7703379392623901, "train/loss_error": 0.5288889408111572, "train/loss_total": 0.5771787762641907 }, { "epoch": 2.275324675324675, "step": 876, "train/loss_ctc": 0.40528732538223267, "train/loss_error": 0.39727142453193665, "train/loss_total": 0.3988746106624603 }, { "epoch": 2.277922077922078, "step": 877, "train/loss_ctc": 0.4908028841018677, "train/loss_error": 0.4079836905002594, "train/loss_total": 0.42454755306243896 }, { "epoch": 2.2805194805194806, "step": 878, "train/loss_ctc": 0.691383957862854, "train/loss_error": 0.4672930836677551, "train/loss_total": 0.512111246585846 }, { "epoch": 2.2831168831168833, "step": 879, "train/loss_ctc": 0.28690746426582336, "train/loss_error": 0.35537219047546387, "train/loss_total": 0.34167924523353577 }, { "epoch": 2.2857142857142856, "grad_norm": 9908.0703125, "learning_rate": 1.6316883116883116e-05, "loss": 0.4319, "step": 880 }, { "epoch": 2.2857142857142856, "step": 880, "train/loss_ctc": 0.6021571159362793, "train/loss_error": 0.4606909155845642, "train/loss_total": 0.4889841675758362 }, { "epoch": 2.2883116883116883, "step": 881, "train/loss_ctc": 0.3922659158706665, "train/loss_error": 0.3601117730140686, "train/loss_total": 0.36654260754585266 }, { "epoch": 2.290909090909091, "step": 882, "train/loss_ctc": 0.4484396278858185, "train/loss_error": 0.420206755399704, "train/loss_total": 0.42585331201553345 }, { "epoch": 2.2935064935064933, "step": 883, "train/loss_ctc": 0.3968833088874817, "train/loss_error": 0.42323756217956543, "train/loss_total": 0.41796672344207764 }, { "epoch": 2.296103896103896, "step": 884, "train/loss_ctc": 0.5683820247650146, "train/loss_error": 0.39902183413505554, "train/loss_total": 0.43289387226104736 }, { "epoch": 2.2987012987012987, "step": 885, "train/loss_ctc": 0.5323489308357239, "train/loss_error": 0.38130268454551697, "train/loss_total": 0.41151192784309387 }, { "epoch": 2.3012987012987014, "step": 886, "train/loss_ctc": 0.573399543762207, "train/loss_error": 0.4163878262042999, "train/loss_total": 0.4477901756763458 }, { "epoch": 2.303896103896104, "step": 887, "train/loss_ctc": 0.6136782169342041, "train/loss_error": 0.4958378076553345, "train/loss_total": 0.5194059014320374 }, { "epoch": 2.3064935064935064, "step": 888, "train/loss_ctc": 0.6974093914031982, "train/loss_error": 0.4070490002632141, "train/loss_total": 0.4651210904121399 }, { "epoch": 2.309090909090909, "step": 889, "train/loss_ctc": 0.3266877233982086, "train/loss_error": 0.392097532749176, "train/loss_total": 0.37901556491851807 }, { "epoch": 2.311688311688312, "grad_norm": 10028.509765625, "learning_rate": 1.616103896103896e-05, "loss": 0.4355, "step": 890 }, { "epoch": 2.311688311688312, "step": 890, "train/loss_ctc": 0.5204488039016724, "train/loss_error": 0.36929598450660706, "train/loss_total": 0.39952656626701355 }, { "epoch": 2.314285714285714, "step": 891, "train/loss_ctc": 0.48641979694366455, "train/loss_error": 0.45409759879112244, "train/loss_total": 0.4605620503425598 }, { "epoch": 2.3168831168831168, "step": 892, "train/loss_ctc": 1.0617347955703735, "train/loss_error": 0.437641978263855, "train/loss_total": 0.5624605417251587 }, { "epoch": 2.3194805194805195, "step": 893, "train/loss_ctc": 0.5074892044067383, "train/loss_error": 0.43171361088752747, "train/loss_total": 0.4468687176704407 }, { "epoch": 2.322077922077922, "step": 894, "train/loss_ctc": 0.7364961504936218, "train/loss_error": 0.5227401852607727, "train/loss_total": 0.5654913783073425 }, { "epoch": 2.324675324675325, "step": 895, "train/loss_ctc": 0.3694998025894165, "train/loss_error": 0.42485311627388, "train/loss_total": 0.4137824475765228 }, { "epoch": 2.327272727272727, "step": 896, "train/loss_ctc": 0.38385671377182007, "train/loss_error": 0.3592588007450104, "train/loss_total": 0.3641783893108368 }, { "epoch": 2.32987012987013, "step": 897, "train/loss_ctc": 0.618284285068512, "train/loss_error": 0.43958786129951477, "train/loss_total": 0.47532716393470764 }, { "epoch": 2.3324675324675326, "step": 898, "train/loss_ctc": 0.36461347341537476, "train/loss_error": 0.30372825264930725, "train/loss_total": 0.31590530276298523 }, { "epoch": 2.335064935064935, "step": 899, "train/loss_ctc": 0.4165268540382385, "train/loss_error": 0.3738512396812439, "train/loss_total": 0.38238638639450073 }, { "epoch": 2.3376623376623376, "grad_norm": 13132.94921875, "learning_rate": 1.6005194805194806e-05, "loss": 0.4386, "step": 900 }, { "epoch": 2.3376623376623376, "step": 900, "train/loss_ctc": 0.492990106344223, "train/loss_error": 0.47821927070617676, "train/loss_total": 0.48117345571517944 }, { "epoch": 2.3402597402597403, "step": 901, "train/loss_ctc": 0.5471018552780151, "train/loss_error": 0.4364902675151825, "train/loss_total": 0.4586126208305359 }, { "epoch": 2.342857142857143, "step": 902, "train/loss_ctc": 0.28577980399131775, "train/loss_error": 0.35239672660827637, "train/loss_total": 0.3390733599662781 }, { "epoch": 2.3454545454545457, "step": 903, "train/loss_ctc": 0.5022401809692383, "train/loss_error": 0.4344979524612427, "train/loss_total": 0.44804641604423523 }, { "epoch": 2.348051948051948, "step": 904, "train/loss_ctc": 0.4446657598018646, "train/loss_error": 0.41828376054763794, "train/loss_total": 0.42356014251708984 }, { "epoch": 2.3506493506493507, "step": 905, "train/loss_ctc": 0.4036804437637329, "train/loss_error": 0.3811241686344147, "train/loss_total": 0.3856354355812073 }, { "epoch": 2.3532467532467534, "step": 906, "train/loss_ctc": 0.9125467538833618, "train/loss_error": 0.4339280426502228, "train/loss_total": 0.5296517610549927 }, { "epoch": 2.3558441558441556, "step": 907, "train/loss_ctc": 0.6082503795623779, "train/loss_error": 0.4572719633579254, "train/loss_total": 0.4874676465988159 }, { "epoch": 2.3584415584415583, "step": 908, "train/loss_ctc": 0.37929612398147583, "train/loss_error": 0.3201039135456085, "train/loss_total": 0.3319423496723175 }, { "epoch": 2.361038961038961, "step": 909, "train/loss_ctc": 0.8852708339691162, "train/loss_error": 0.46233320236206055, "train/loss_total": 0.5469207763671875 }, { "epoch": 2.3636363636363638, "grad_norm": 19198.26171875, "learning_rate": 1.584935064935065e-05, "loss": 0.4432, "step": 910 }, { "epoch": 2.3636363636363638, "step": 910, "train/loss_ctc": 0.5908734798431396, "train/loss_error": 0.41531696915626526, "train/loss_total": 0.4504282772541046 }, { "epoch": 2.3662337662337665, "step": 911, "train/loss_ctc": 0.4083363711833954, "train/loss_error": 0.38240066170692444, "train/loss_total": 0.3875878155231476 }, { "epoch": 2.3688311688311687, "step": 912, "train/loss_ctc": 0.57112717628479, "train/loss_error": 0.35318848490715027, "train/loss_total": 0.3967762291431427 }, { "epoch": 2.3714285714285714, "step": 913, "train/loss_ctc": 0.45806971192359924, "train/loss_error": 0.5141322016716003, "train/loss_total": 0.5029196739196777 }, { "epoch": 2.374025974025974, "step": 914, "train/loss_ctc": 0.524329662322998, "train/loss_error": 0.4267936050891876, "train/loss_total": 0.44630083441734314 }, { "epoch": 2.3766233766233764, "step": 915, "train/loss_ctc": 0.5446964502334595, "train/loss_error": 0.4465690851211548, "train/loss_total": 0.4661945700645447 }, { "epoch": 2.379220779220779, "step": 916, "train/loss_ctc": 0.3763575255870819, "train/loss_error": 0.4421062171459198, "train/loss_total": 0.4289565086364746 }, { "epoch": 2.381818181818182, "step": 917, "train/loss_ctc": 0.4556613564491272, "train/loss_error": 0.47836756706237793, "train/loss_total": 0.4738263487815857 }, { "epoch": 2.3844155844155845, "step": 918, "train/loss_ctc": 0.40266871452331543, "train/loss_error": 0.4253237545490265, "train/loss_total": 0.42079275846481323 }, { "epoch": 2.3870129870129873, "step": 919, "train/loss_ctc": 0.5047339797019958, "train/loss_error": 0.4404617249965668, "train/loss_total": 0.45331621170043945 }, { "epoch": 2.3896103896103895, "grad_norm": 14973.8896484375, "learning_rate": 1.5693506493506492e-05, "loss": 0.4427, "step": 920 }, { "epoch": 2.3896103896103895, "step": 920, "train/loss_ctc": 0.3718258738517761, "train/loss_error": 0.37255215644836426, "train/loss_total": 0.37240689992904663 }, { "epoch": 2.3922077922077922, "step": 921, "train/loss_ctc": 0.576583981513977, "train/loss_error": 0.46216443181037903, "train/loss_total": 0.4850483536720276 }, { "epoch": 2.394805194805195, "step": 922, "train/loss_ctc": 0.3506154716014862, "train/loss_error": 0.39525043964385986, "train/loss_total": 0.3863234519958496 }, { "epoch": 2.397402597402597, "step": 923, "train/loss_ctc": 0.7546201944351196, "train/loss_error": 0.3565257787704468, "train/loss_total": 0.4361446499824524 }, { "epoch": 2.4, "step": 924, "train/loss_ctc": 0.4674195647239685, "train/loss_error": 0.35265424847602844, "train/loss_total": 0.37560734152793884 }, { "epoch": 2.4025974025974026, "step": 925, "train/loss_ctc": 0.498261034488678, "train/loss_error": 0.45255592465400696, "train/loss_total": 0.46169695258140564 }, { "epoch": 2.4051948051948053, "step": 926, "train/loss_ctc": 0.44843044877052307, "train/loss_error": 0.4227740466594696, "train/loss_total": 0.4279053509235382 }, { "epoch": 2.407792207792208, "step": 927, "train/loss_ctc": 0.3146737217903137, "train/loss_error": 0.4166557490825653, "train/loss_total": 0.3962593674659729 }, { "epoch": 2.4103896103896103, "step": 928, "train/loss_ctc": 0.6139093041419983, "train/loss_error": 0.48244068026542664, "train/loss_total": 0.508734405040741 }, { "epoch": 2.412987012987013, "step": 929, "train/loss_ctc": 0.38253405690193176, "train/loss_error": 0.35021382570266724, "train/loss_total": 0.3566778898239136 }, { "epoch": 2.4155844155844157, "grad_norm": 10677.263671875, "learning_rate": 1.5537662337662337e-05, "loss": 0.4207, "step": 930 }, { "epoch": 2.4155844155844157, "step": 930, "train/loss_ctc": 0.4841327369213104, "train/loss_error": 0.41960734128952026, "train/loss_total": 0.43251243233680725 }, { "epoch": 2.418181818181818, "step": 931, "train/loss_ctc": 0.3861841857433319, "train/loss_error": 0.421945720911026, "train/loss_total": 0.4147934019565582 }, { "epoch": 2.4207792207792207, "step": 932, "train/loss_ctc": 0.4644802510738373, "train/loss_error": 0.4056479036808014, "train/loss_total": 0.4174143671989441 }, { "epoch": 2.4233766233766234, "step": 933, "train/loss_ctc": 0.3808186948299408, "train/loss_error": 0.40607553720474243, "train/loss_total": 0.4010241627693176 }, { "epoch": 2.425974025974026, "step": 934, "train/loss_ctc": 0.32315683364868164, "train/loss_error": 0.34238365292549133, "train/loss_total": 0.3385382890701294 }, { "epoch": 2.4285714285714284, "step": 935, "train/loss_ctc": 0.6908047199249268, "train/loss_error": 0.4852575659751892, "train/loss_total": 0.5263670086860657 }, { "epoch": 2.431168831168831, "step": 936, "train/loss_ctc": 0.8386857509613037, "train/loss_error": 0.5807844996452332, "train/loss_total": 0.6323647499084473 }, { "epoch": 2.433766233766234, "step": 937, "train/loss_ctc": 0.49941080808639526, "train/loss_error": 0.41461628675460815, "train/loss_total": 0.4315751791000366 }, { "epoch": 2.4363636363636365, "step": 938, "train/loss_ctc": 0.46249252557754517, "train/loss_error": 0.3472427427768707, "train/loss_total": 0.3702927231788635 }, { "epoch": 2.4389610389610388, "step": 939, "train/loss_ctc": 0.4470609724521637, "train/loss_error": 0.3857211470603943, "train/loss_total": 0.39798909425735474 }, { "epoch": 2.4415584415584415, "grad_norm": 13424.255859375, "learning_rate": 1.5381818181818182e-05, "loss": 0.4363, "step": 940 }, { "epoch": 2.4415584415584415, "step": 940, "train/loss_ctc": 0.5267013907432556, "train/loss_error": 0.4063579738140106, "train/loss_total": 0.4304266571998596 }, { "epoch": 2.444155844155844, "step": 941, "train/loss_ctc": 0.6240761280059814, "train/loss_error": 0.4802128076553345, "train/loss_total": 0.5089855194091797 }, { "epoch": 2.446753246753247, "step": 942, "train/loss_ctc": 0.4655873775482178, "train/loss_error": 0.38219645619392395, "train/loss_total": 0.3988746404647827 }, { "epoch": 2.449350649350649, "step": 943, "train/loss_ctc": 0.4442059397697449, "train/loss_error": 0.4081078767776489, "train/loss_total": 0.4153275191783905 }, { "epoch": 2.451948051948052, "step": 944, "train/loss_ctc": 0.4524049460887909, "train/loss_error": 0.5407121181488037, "train/loss_total": 0.5230507254600525 }, { "epoch": 2.4545454545454546, "step": 945, "train/loss_ctc": 0.5099567174911499, "train/loss_error": 0.5127904415130615, "train/loss_total": 0.5122237205505371 }, { "epoch": 2.4571428571428573, "step": 946, "train/loss_ctc": 0.37007495760917664, "train/loss_error": 0.43631017208099365, "train/loss_total": 0.42306312918663025 }, { "epoch": 2.4597402597402596, "step": 947, "train/loss_ctc": 0.37737518548965454, "train/loss_error": 0.3483261168003082, "train/loss_total": 0.3541359305381775 }, { "epoch": 2.4623376623376623, "step": 948, "train/loss_ctc": 0.6110984086990356, "train/loss_error": 0.4635458290576935, "train/loss_total": 0.49305635690689087 }, { "epoch": 2.464935064935065, "step": 949, "train/loss_ctc": 0.3760339021682739, "train/loss_error": 0.34763187170028687, "train/loss_total": 0.35331228375434875 }, { "epoch": 2.4675324675324677, "grad_norm": 11634.9755859375, "learning_rate": 1.5225974025974025e-05, "loss": 0.4412, "step": 950 }, { "epoch": 2.4675324675324677, "step": 950, "train/loss_ctc": 0.4524855315685272, "train/loss_error": 0.39415672421455383, "train/loss_total": 0.4058224856853485 }, { "epoch": 2.47012987012987, "step": 951, "train/loss_ctc": 0.4797646701335907, "train/loss_error": 0.49434778094291687, "train/loss_total": 0.4914311468601227 }, { "epoch": 2.4727272727272727, "step": 952, "train/loss_ctc": 0.4318815767765045, "train/loss_error": 0.35757726430892944, "train/loss_total": 0.37243813276290894 }, { "epoch": 2.4753246753246754, "step": 953, "train/loss_ctc": 0.5445318222045898, "train/loss_error": 0.42599770426750183, "train/loss_total": 0.44970452785491943 }, { "epoch": 2.477922077922078, "step": 954, "train/loss_ctc": 0.4669070839881897, "train/loss_error": 0.46757370233535767, "train/loss_total": 0.4674403667449951 }, { "epoch": 2.4805194805194803, "step": 955, "train/loss_ctc": 0.5061408281326294, "train/loss_error": 0.3648381233215332, "train/loss_total": 0.3930986821651459 }, { "epoch": 2.483116883116883, "step": 956, "train/loss_ctc": 0.4596114754676819, "train/loss_error": 0.30807238817214966, "train/loss_total": 0.33838021755218506 }, { "epoch": 2.4857142857142858, "step": 957, "train/loss_ctc": 0.4671792984008789, "train/loss_error": 0.41620537638664246, "train/loss_total": 0.42640015482902527 }, { "epoch": 2.4883116883116885, "step": 958, "train/loss_ctc": 0.5488667488098145, "train/loss_error": 0.3809160590171814, "train/loss_total": 0.414506196975708 }, { "epoch": 2.4909090909090907, "step": 959, "train/loss_ctc": 0.45594608783721924, "train/loss_error": 0.4200165569782257, "train/loss_total": 0.4272024631500244 }, { "epoch": 2.4935064935064934, "grad_norm": 11915.2099609375, "learning_rate": 1.507012987012987e-05, "loss": 0.4186, "step": 960 }, { "epoch": 2.4935064935064934, "step": 960, "train/loss_ctc": 0.34328389167785645, "train/loss_error": 0.4276708662509918, "train/loss_total": 0.4107934832572937 }, { "epoch": 2.496103896103896, "step": 961, "train/loss_ctc": 0.36251676082611084, "train/loss_error": 0.2807380259037018, "train/loss_total": 0.29709377884864807 }, { "epoch": 2.498701298701299, "step": 962, "train/loss_ctc": 0.3776510953903198, "train/loss_error": 0.41441285610198975, "train/loss_total": 0.40706053376197815 }, { "epoch": 2.501298701298701, "step": 963, "train/loss_ctc": 0.5070983171463013, "train/loss_error": 0.42518526315689087, "train/loss_total": 0.44156786799430847 }, { "epoch": 2.503896103896104, "step": 964, "train/loss_ctc": 0.8065208196640015, "train/loss_error": 0.4169529676437378, "train/loss_total": 0.4948665499687195 }, { "epoch": 2.5064935064935066, "step": 965, "train/loss_ctc": 0.4925708472728729, "train/loss_error": 0.47087857127189636, "train/loss_total": 0.4752170443534851 }, { "epoch": 2.509090909090909, "step": 966, "train/loss_ctc": 0.33729618787765503, "train/loss_error": 0.3601418733596802, "train/loss_total": 0.35557276010513306 }, { "epoch": 2.5116883116883115, "step": 967, "train/loss_ctc": 0.37843841314315796, "train/loss_error": 0.33219054341316223, "train/loss_total": 0.3414401113986969 }, { "epoch": 2.5142857142857142, "step": 968, "train/loss_ctc": 0.5728816986083984, "train/loss_error": 0.49273768067359924, "train/loss_total": 0.5087665319442749 }, { "epoch": 2.516883116883117, "step": 969, "train/loss_ctc": 0.44609159231185913, "train/loss_error": 0.4276614487171173, "train/loss_total": 0.43134748935699463 }, { "epoch": 2.5194805194805197, "grad_norm": 13544.7978515625, "learning_rate": 1.4914285714285715e-05, "loss": 0.4164, "step": 970 }, { "epoch": 2.5194805194805197, "step": 970, "train/loss_ctc": 0.4279748201370239, "train/loss_error": 0.40015843510627747, "train/loss_total": 0.4057217240333557 }, { "epoch": 2.522077922077922, "step": 971, "train/loss_ctc": 0.6499279141426086, "train/loss_error": 0.4708511531352997, "train/loss_total": 0.5066665410995483 }, { "epoch": 2.5246753246753246, "step": 972, "train/loss_ctc": 0.7952671051025391, "train/loss_error": 0.34153228998184204, "train/loss_total": 0.4322792887687683 }, { "epoch": 2.5272727272727273, "step": 973, "train/loss_ctc": 0.4812987446784973, "train/loss_error": 0.4295656383037567, "train/loss_total": 0.43991225957870483 }, { "epoch": 2.5298701298701296, "step": 974, "train/loss_ctc": 0.5476354360580444, "train/loss_error": 0.4596945345401764, "train/loss_total": 0.47728270292282104 }, { "epoch": 2.5324675324675323, "step": 975, "train/loss_ctc": 0.3473578095436096, "train/loss_error": 0.32817837595939636, "train/loss_total": 0.332014262676239 }, { "epoch": 2.535064935064935, "step": 976, "train/loss_ctc": 0.5081796050071716, "train/loss_error": 0.3174552619457245, "train/loss_total": 0.35560014843940735 }, { "epoch": 2.5376623376623377, "step": 977, "train/loss_ctc": 0.48893746733665466, "train/loss_error": 0.340977281332016, "train/loss_total": 0.3705693185329437 }, { "epoch": 2.5402597402597404, "step": 978, "train/loss_ctc": 0.3531177043914795, "train/loss_error": 0.36193186044692993, "train/loss_total": 0.36016905307769775 }, { "epoch": 2.5428571428571427, "step": 979, "train/loss_ctc": 0.3538201153278351, "train/loss_error": 0.3532697856426239, "train/loss_total": 0.35337987542152405 }, { "epoch": 2.5454545454545454, "grad_norm": 12353.1396484375, "learning_rate": 1.475844155844156e-05, "loss": 0.4034, "step": 980 }, { "epoch": 2.5454545454545454, "step": 980, "train/loss_ctc": 0.5315302610397339, "train/loss_error": 0.4036361575126648, "train/loss_total": 0.4292149841785431 }, { "epoch": 2.548051948051948, "step": 981, "train/loss_ctc": 0.630653977394104, "train/loss_error": 0.4137430191040039, "train/loss_total": 0.4571252465248108 }, { "epoch": 2.5506493506493504, "step": 982, "train/loss_ctc": 0.6394188404083252, "train/loss_error": 0.4928719699382782, "train/loss_total": 0.5221813321113586 }, { "epoch": 2.553246753246753, "step": 983, "train/loss_ctc": 0.6692814826965332, "train/loss_error": 0.45620638132095337, "train/loss_total": 0.4988214075565338 }, { "epoch": 2.555844155844156, "step": 984, "train/loss_ctc": 1.121273159980774, "train/loss_error": 0.45556989312171936, "train/loss_total": 0.5887105464935303 }, { "epoch": 2.5584415584415585, "step": 985, "train/loss_ctc": 0.37719571590423584, "train/loss_error": 0.3684132993221283, "train/loss_total": 0.3701697885990143 }, { "epoch": 2.5610389610389612, "step": 986, "train/loss_ctc": 0.6871421933174133, "train/loss_error": 0.45496273040771484, "train/loss_total": 0.5013986229896545 }, { "epoch": 2.5636363636363635, "step": 987, "train/loss_ctc": 0.35494160652160645, "train/loss_error": 0.3817555010318756, "train/loss_total": 0.3763927221298218 }, { "epoch": 2.566233766233766, "step": 988, "train/loss_ctc": 0.5271691679954529, "train/loss_error": 0.48632434010505676, "train/loss_total": 0.494493305683136 }, { "epoch": 2.568831168831169, "step": 989, "train/loss_ctc": 0.6413418054580688, "train/loss_error": 0.4333842098712921, "train/loss_total": 0.47497573494911194 }, { "epoch": 2.571428571428571, "grad_norm": 12963.4072265625, "learning_rate": 1.4602597402597404e-05, "loss": 0.4713, "step": 990 }, { "epoch": 2.571428571428571, "step": 990, "train/loss_ctc": 0.43262797594070435, "train/loss_error": 0.3891089856624603, "train/loss_total": 0.39781278371810913 }, { "epoch": 2.574025974025974, "step": 991, "train/loss_ctc": 0.8660361170768738, "train/loss_error": 0.41862282156944275, "train/loss_total": 0.5081055164337158 }, { "epoch": 2.5766233766233766, "step": 992, "train/loss_ctc": 0.6961153745651245, "train/loss_error": 0.439606249332428, "train/loss_total": 0.49090808629989624 }, { "epoch": 2.5792207792207793, "step": 993, "train/loss_ctc": 0.6246310472488403, "train/loss_error": 0.42666202783584595, "train/loss_total": 0.4662558436393738 }, { "epoch": 2.581818181818182, "step": 994, "train/loss_ctc": 0.5790760517120361, "train/loss_error": 0.4425555467605591, "train/loss_total": 0.46985965967178345 }, { "epoch": 2.5844155844155843, "step": 995, "train/loss_ctc": 0.46272534132003784, "train/loss_error": 0.38976606726646423, "train/loss_total": 0.40435791015625 }, { "epoch": 2.587012987012987, "step": 996, "train/loss_ctc": 0.4255228042602539, "train/loss_error": 0.4421822428703308, "train/loss_total": 0.4388503432273865 }, { "epoch": 2.5896103896103897, "step": 997, "train/loss_ctc": 0.49702948331832886, "train/loss_error": 0.5097610354423523, "train/loss_total": 0.5072147250175476 }, { "epoch": 2.592207792207792, "step": 998, "train/loss_ctc": 0.5565353631973267, "train/loss_error": 0.36313632130622864, "train/loss_total": 0.40181615948677063 }, { "epoch": 2.5948051948051947, "step": 999, "train/loss_ctc": 0.4335663318634033, "train/loss_error": 0.4316859245300293, "train/loss_total": 0.432062029838562 }, { "epoch": 2.5974025974025974, "grad_norm": 13716.6767578125, "learning_rate": 1.4446753246753247e-05, "loss": 0.4517, "step": 1000 }, { "epoch": 2.5974025974025974, "step": 1000, "train/loss_ctc": 0.4987727403640747, "train/loss_error": 0.40463167428970337, "train/loss_total": 0.42345988750457764 }, { "epoch": 2.6, "step": 1001, "train/loss_ctc": 0.5270754098892212, "train/loss_error": 0.37964123487472534, "train/loss_total": 0.4091280698776245 }, { "epoch": 2.602597402597403, "step": 1002, "train/loss_ctc": 0.8018827438354492, "train/loss_error": 0.4920518100261688, "train/loss_total": 0.5540180206298828 }, { "epoch": 2.605194805194805, "step": 1003, "train/loss_ctc": 0.4729597270488739, "train/loss_error": 0.43630141019821167, "train/loss_total": 0.4436330795288086 }, { "epoch": 2.6077922077922078, "step": 1004, "train/loss_ctc": 0.5143141150474548, "train/loss_error": 0.44653376936912537, "train/loss_total": 0.46008986234664917 }, { "epoch": 2.6103896103896105, "step": 1005, "train/loss_ctc": 0.39888328313827515, "train/loss_error": 0.42317870259284973, "train/loss_total": 0.4183196425437927 }, { "epoch": 2.6129870129870127, "step": 1006, "train/loss_ctc": 0.5184609889984131, "train/loss_error": 0.4710523188114166, "train/loss_total": 0.48053407669067383 }, { "epoch": 2.6155844155844155, "step": 1007, "train/loss_ctc": 0.649005651473999, "train/loss_error": 0.4179949462413788, "train/loss_total": 0.4641970992088318 }, { "epoch": 2.618181818181818, "step": 1008, "train/loss_ctc": 0.45214372873306274, "train/loss_error": 0.4308151602745056, "train/loss_total": 0.435080885887146 }, { "epoch": 2.620779220779221, "step": 1009, "train/loss_ctc": 0.25275325775146484, "train/loss_error": 0.2914881408214569, "train/loss_total": 0.28374117612838745 }, { "epoch": 2.6233766233766236, "grad_norm": 9230.818359375, "learning_rate": 1.4290909090909092e-05, "loss": 0.4372, "step": 1010 }, { "epoch": 2.6233766233766236, "step": 1010, "train/loss_ctc": 0.3079911172389984, "train/loss_error": 0.43098753690719604, "train/loss_total": 0.4063882529735565 }, { "epoch": 2.625974025974026, "step": 1011, "train/loss_ctc": 0.42114967107772827, "train/loss_error": 0.45406582951545715, "train/loss_total": 0.4474826157093048 }, { "epoch": 2.6285714285714286, "step": 1012, "train/loss_ctc": 0.717671275138855, "train/loss_error": 0.5265818238258362, "train/loss_total": 0.5647997260093689 }, { "epoch": 2.6311688311688313, "step": 1013, "train/loss_ctc": 0.45170164108276367, "train/loss_error": 0.5215328931808472, "train/loss_total": 0.5075666308403015 }, { "epoch": 2.6337662337662335, "step": 1014, "train/loss_ctc": 0.4669455289840698, "train/loss_error": 0.42576658725738525, "train/loss_total": 0.4340023994445801 }, { "epoch": 2.6363636363636362, "step": 1015, "train/loss_ctc": 0.5613925457000732, "train/loss_error": 0.4632062315940857, "train/loss_total": 0.4828435182571411 }, { "epoch": 2.638961038961039, "step": 1016, "train/loss_ctc": 0.3198038339614868, "train/loss_error": 0.3694586753845215, "train/loss_total": 0.35952770709991455 }, { "epoch": 2.6415584415584417, "step": 1017, "train/loss_ctc": 0.6537165641784668, "train/loss_error": 0.4395303428173065, "train/loss_total": 0.4823675751686096 }, { "epoch": 2.6441558441558444, "step": 1018, "train/loss_ctc": 0.4459243714809418, "train/loss_error": 0.44902297854423523, "train/loss_total": 0.4484032690525055 }, { "epoch": 2.6467532467532466, "step": 1019, "train/loss_ctc": 0.7179641723632812, "train/loss_error": 0.4447786509990692, "train/loss_total": 0.4994157552719116 }, { "epoch": 2.6493506493506493, "grad_norm": 16578.5703125, "learning_rate": 1.4135064935064935e-05, "loss": 0.4633, "step": 1020 }, { "epoch": 2.6493506493506493, "step": 1020, "train/loss_ctc": 0.37889450788497925, "train/loss_error": 0.390621155500412, "train/loss_total": 0.3882758319377899 }, { "epoch": 2.651948051948052, "step": 1021, "train/loss_ctc": 0.4493411183357239, "train/loss_error": 0.3998917043209076, "train/loss_total": 0.4097815752029419 }, { "epoch": 2.6545454545454543, "step": 1022, "train/loss_ctc": 0.5038045048713684, "train/loss_error": 0.3954967260360718, "train/loss_total": 0.417158305644989 }, { "epoch": 2.657142857142857, "step": 1023, "train/loss_ctc": 0.4896755814552307, "train/loss_error": 0.38437139987945557, "train/loss_total": 0.40543222427368164 }, { "epoch": 2.6597402597402597, "step": 1024, "train/loss_ctc": 0.6458239555358887, "train/loss_error": 0.40622562170028687, "train/loss_total": 0.45414531230926514 }, { "epoch": 2.6623376623376624, "step": 1025, "train/loss_ctc": 0.508952260017395, "train/loss_error": 0.4792308211326599, "train/loss_total": 0.48517513275146484 }, { "epoch": 2.664935064935065, "step": 1026, "train/loss_ctc": 0.6300722360610962, "train/loss_error": 0.45385244488716125, "train/loss_total": 0.48909640312194824 }, { "epoch": 2.6675324675324674, "step": 1027, "train/loss_ctc": 0.48440277576446533, "train/loss_error": 0.43442434072494507, "train/loss_total": 0.4444200396537781 }, { "epoch": 2.67012987012987, "step": 1028, "train/loss_ctc": 0.4635459780693054, "train/loss_error": 0.4058707356452942, "train/loss_total": 0.41740578413009644 }, { "epoch": 2.672727272727273, "step": 1029, "train/loss_ctc": 0.5041993856430054, "train/loss_error": 0.2950458824634552, "train/loss_total": 0.3368765711784363 }, { "epoch": 2.675324675324675, "grad_norm": 15989.16015625, "learning_rate": 1.397922077922078e-05, "loss": 0.4248, "step": 1030 }, { "epoch": 2.675324675324675, "step": 1030, "train/loss_ctc": 0.3979530334472656, "train/loss_error": 0.37484362721443176, "train/loss_total": 0.3794655203819275 }, { "epoch": 2.677922077922078, "step": 1031, "train/loss_ctc": 0.4984753727912903, "train/loss_error": 0.37333646416664124, "train/loss_total": 0.39836427569389343 }, { "epoch": 2.6805194805194805, "step": 1032, "train/loss_ctc": 0.4586094617843628, "train/loss_error": 0.3393966555595398, "train/loss_total": 0.36323922872543335 }, { "epoch": 2.6831168831168832, "step": 1033, "train/loss_ctc": 0.43333354592323303, "train/loss_error": 0.43430858850479126, "train/loss_total": 0.43411359190940857 }, { "epoch": 2.685714285714286, "step": 1034, "train/loss_ctc": 0.6000279188156128, "train/loss_error": 0.4229491651058197, "train/loss_total": 0.45836490392684937 }, { "epoch": 2.688311688311688, "step": 1035, "train/loss_ctc": 0.3444042205810547, "train/loss_error": 0.3095646798610687, "train/loss_total": 0.31653258204460144 }, { "epoch": 2.690909090909091, "step": 1036, "train/loss_ctc": 0.6002054810523987, "train/loss_error": 0.47752419114112854, "train/loss_total": 0.5020604729652405 }, { "epoch": 2.6935064935064936, "step": 1037, "train/loss_ctc": 0.31782734394073486, "train/loss_error": 0.35844287276268005, "train/loss_total": 0.3503197729587555 }, { "epoch": 2.696103896103896, "step": 1038, "train/loss_ctc": 0.2719046473503113, "train/loss_error": 0.3130962550640106, "train/loss_total": 0.30485793948173523 }, { "epoch": 2.6987012987012986, "step": 1039, "train/loss_ctc": 0.3908427953720093, "train/loss_error": 0.387118399143219, "train/loss_total": 0.38786330819129944 }, { "epoch": 2.7012987012987013, "grad_norm": 11574.791015625, "learning_rate": 1.3823376623376625e-05, "loss": 0.3895, "step": 1040 }, { "epoch": 2.7012987012987013, "step": 1040, "train/loss_ctc": 0.5052008032798767, "train/loss_error": 0.44562774896621704, "train/loss_total": 0.457542359828949 }, { "epoch": 2.703896103896104, "step": 1041, "train/loss_ctc": 0.4696553945541382, "train/loss_error": 0.36477455496788025, "train/loss_total": 0.38575074076652527 }, { "epoch": 2.7064935064935067, "step": 1042, "train/loss_ctc": 0.4223549962043762, "train/loss_error": 0.33233892917633057, "train/loss_total": 0.35034215450286865 }, { "epoch": 2.709090909090909, "step": 1043, "train/loss_ctc": 0.4176913797855377, "train/loss_error": 0.4194674491882324, "train/loss_total": 0.41911226511001587 }, { "epoch": 2.7116883116883117, "step": 1044, "train/loss_ctc": 0.5529048442840576, "train/loss_error": 0.4541856646537781, "train/loss_total": 0.4739295244216919 }, { "epoch": 2.7142857142857144, "step": 1045, "train/loss_ctc": 0.49745607376098633, "train/loss_error": 0.3630813956260681, "train/loss_total": 0.3899563252925873 }, { "epoch": 2.7168831168831167, "step": 1046, "train/loss_ctc": 0.5026822686195374, "train/loss_error": 0.4120347201824188, "train/loss_total": 0.43016424775123596 }, { "epoch": 2.7194805194805194, "step": 1047, "train/loss_ctc": 0.6250364780426025, "train/loss_error": 0.38298454880714417, "train/loss_total": 0.43139493465423584 }, { "epoch": 2.722077922077922, "step": 1048, "train/loss_ctc": 0.6164462566375732, "train/loss_error": 0.5284151434898376, "train/loss_total": 0.5460213422775269 }, { "epoch": 2.724675324675325, "step": 1049, "train/loss_ctc": 0.5541439056396484, "train/loss_error": 0.4348126947879791, "train/loss_total": 0.4586789608001709 }, { "epoch": 2.7272727272727275, "grad_norm": 14187.2724609375, "learning_rate": 1.3667532467532468e-05, "loss": 0.4343, "step": 1050 }, { "epoch": 2.7272727272727275, "step": 1050, "train/loss_ctc": 0.780341386795044, "train/loss_error": 0.4311685264110565, "train/loss_total": 0.501003086566925 }, { "epoch": 2.72987012987013, "step": 1051, "train/loss_ctc": 1.1609972715377808, "train/loss_error": 0.4361199736595154, "train/loss_total": 0.5810954570770264 }, { "epoch": 2.7324675324675325, "step": 1052, "train/loss_ctc": 0.46462422609329224, "train/loss_error": 0.4106169641017914, "train/loss_total": 0.4214184284210205 }, { "epoch": 2.735064935064935, "step": 1053, "train/loss_ctc": 0.5143638253211975, "train/loss_error": 0.49570852518081665, "train/loss_total": 0.4994395971298218 }, { "epoch": 2.7376623376623375, "step": 1054, "train/loss_ctc": 0.3964276909828186, "train/loss_error": 0.4146256744861603, "train/loss_total": 0.410986065864563 }, { "epoch": 2.74025974025974, "step": 1055, "train/loss_ctc": 0.4950419068336487, "train/loss_error": 0.4104815125465393, "train/loss_total": 0.4273935854434967 }, { "epoch": 2.742857142857143, "step": 1056, "train/loss_ctc": 0.39985448122024536, "train/loss_error": 0.3963375687599182, "train/loss_total": 0.3970409631729126 }, { "epoch": 2.7454545454545456, "step": 1057, "train/loss_ctc": 0.572894275188446, "train/loss_error": 0.4803411364555359, "train/loss_total": 0.4988517761230469 }, { "epoch": 2.7480519480519483, "step": 1058, "train/loss_ctc": 0.5108768939971924, "train/loss_error": 0.4495455026626587, "train/loss_total": 0.46181178092956543 }, { "epoch": 2.7506493506493506, "step": 1059, "train/loss_ctc": 0.6093026995658875, "train/loss_error": 0.44449475407600403, "train/loss_total": 0.47745633125305176 }, { "epoch": 2.7532467532467533, "grad_norm": 14710.0322265625, "learning_rate": 1.3511688311688313e-05, "loss": 0.4676, "step": 1060 }, { "epoch": 2.7532467532467533, "step": 1060, "train/loss_ctc": 0.6753272414207458, "train/loss_error": 0.4035671651363373, "train/loss_total": 0.457919180393219 }, { "epoch": 2.755844155844156, "step": 1061, "train/loss_ctc": 0.4775344729423523, "train/loss_error": 0.50078946352005, "train/loss_total": 0.49613848328590393 }, { "epoch": 2.7584415584415583, "step": 1062, "train/loss_ctc": 0.4724724292755127, "train/loss_error": 0.4700593650341034, "train/loss_total": 0.47054198384284973 }, { "epoch": 2.761038961038961, "step": 1063, "train/loss_ctc": 0.4361476004123688, "train/loss_error": 0.394765704870224, "train/loss_total": 0.4030420780181885 }, { "epoch": 2.7636363636363637, "step": 1064, "train/loss_ctc": 0.39983850717544556, "train/loss_error": 0.4512430429458618, "train/loss_total": 0.44096213579177856 }, { "epoch": 2.7662337662337664, "step": 1065, "train/loss_ctc": 0.6008154153823853, "train/loss_error": 0.5169841051101685, "train/loss_total": 0.5337504148483276 }, { "epoch": 2.768831168831169, "step": 1066, "train/loss_ctc": 0.4954982399940491, "train/loss_error": 0.48771169781684875, "train/loss_total": 0.4892690181732178 }, { "epoch": 2.7714285714285714, "step": 1067, "train/loss_ctc": 0.5153289437294006, "train/loss_error": 0.5125377178192139, "train/loss_total": 0.5130959749221802 }, { "epoch": 2.774025974025974, "step": 1068, "train/loss_ctc": 0.5492093563079834, "train/loss_error": 0.4371412396430969, "train/loss_total": 0.45955488085746765 }, { "epoch": 2.7766233766233768, "step": 1069, "train/loss_ctc": 0.656535267829895, "train/loss_error": 0.4384705722332001, "train/loss_total": 0.4820834994316101 }, { "epoch": 2.779220779220779, "grad_norm": 14375.89453125, "learning_rate": 1.3355844155844156e-05, "loss": 0.4746, "step": 1070 }, { "epoch": 2.779220779220779, "step": 1070, "train/loss_ctc": 0.5079290270805359, "train/loss_error": 0.5048472285270691, "train/loss_total": 0.5054636001586914 }, { "epoch": 2.7818181818181817, "step": 1071, "train/loss_ctc": 0.602364718914032, "train/loss_error": 0.40241947770118713, "train/loss_total": 0.4424085319042206 }, { "epoch": 2.7844155844155845, "step": 1072, "train/loss_ctc": 0.4793398380279541, "train/loss_error": 0.42118212580680847, "train/loss_total": 0.4328136742115021 }, { "epoch": 2.787012987012987, "step": 1073, "train/loss_ctc": 0.4589349925518036, "train/loss_error": 0.42974212765693665, "train/loss_total": 0.4355807304382324 }, { "epoch": 2.78961038961039, "step": 1074, "train/loss_ctc": 0.4698387384414673, "train/loss_error": 0.4356905519962311, "train/loss_total": 0.4425202012062073 }, { "epoch": 2.792207792207792, "step": 1075, "train/loss_ctc": 0.6028221249580383, "train/loss_error": 0.44005629420280457, "train/loss_total": 0.4726094603538513 }, { "epoch": 2.794805194805195, "step": 1076, "train/loss_ctc": 0.7742828130722046, "train/loss_error": 0.41778162121772766, "train/loss_total": 0.48908185958862305 }, { "epoch": 2.7974025974025976, "step": 1077, "train/loss_ctc": 0.4541749954223633, "train/loss_error": 0.3864564299583435, "train/loss_total": 0.4000001549720764 }, { "epoch": 2.8, "step": 1078, "train/loss_ctc": 0.4714727997779846, "train/loss_error": 0.35577452182769775, "train/loss_total": 0.3789141774177551 }, { "epoch": 2.8025974025974025, "step": 1079, "train/loss_ctc": 0.44834038615226746, "train/loss_error": 0.4045446515083313, "train/loss_total": 0.41330379247665405 }, { "epoch": 2.8051948051948052, "grad_norm": 10078.7109375, "learning_rate": 1.32e-05, "loss": 0.4413, "step": 1080 }, { "epoch": 2.8051948051948052, "step": 1080, "train/loss_ctc": 0.3243316411972046, "train/loss_error": 0.3725133538246155, "train/loss_total": 0.3628770112991333 }, { "epoch": 2.807792207792208, "step": 1081, "train/loss_ctc": 0.40206336975097656, "train/loss_error": 0.4323243498802185, "train/loss_total": 0.4262721836566925 }, { "epoch": 2.8103896103896107, "step": 1082, "train/loss_ctc": 0.3635670840740204, "train/loss_error": 0.4551927447319031, "train/loss_total": 0.4368675947189331 }, { "epoch": 2.812987012987013, "step": 1083, "train/loss_ctc": 0.31648457050323486, "train/loss_error": 0.37347766757011414, "train/loss_total": 0.36207905411720276 }, { "epoch": 2.8155844155844156, "step": 1084, "train/loss_ctc": 0.9470053911209106, "train/loss_error": 0.4542229473590851, "train/loss_total": 0.5527794361114502 }, { "epoch": 2.8181818181818183, "step": 1085, "train/loss_ctc": 0.4470227360725403, "train/loss_error": 0.42260104417800903, "train/loss_total": 0.4274854063987732 }, { "epoch": 2.8207792207792206, "step": 1086, "train/loss_ctc": 0.36180830001831055, "train/loss_error": 0.3373108208179474, "train/loss_total": 0.3422103524208069 }, { "epoch": 2.8233766233766233, "step": 1087, "train/loss_ctc": 0.4190465807914734, "train/loss_error": 0.3839525878429413, "train/loss_total": 0.3909713923931122 }, { "epoch": 2.825974025974026, "step": 1088, "train/loss_ctc": 0.4910831153392792, "train/loss_error": 0.4252614378929138, "train/loss_total": 0.43842577934265137 }, { "epoch": 2.8285714285714287, "step": 1089, "train/loss_ctc": 0.4055614471435547, "train/loss_error": 0.4607435166835785, "train/loss_total": 0.44970712065696716 }, { "epoch": 2.8311688311688314, "grad_norm": 10268.173828125, "learning_rate": 1.3044155844155844e-05, "loss": 0.419, "step": 1090 }, { "epoch": 2.8311688311688314, "step": 1090, "train/loss_ctc": 0.5393203496932983, "train/loss_error": 0.38176965713500977, "train/loss_total": 0.41327980160713196 }, { "epoch": 2.8337662337662337, "step": 1091, "train/loss_ctc": 0.7220108509063721, "train/loss_error": 0.44992443919181824, "train/loss_total": 0.504341721534729 }, { "epoch": 2.8363636363636364, "step": 1092, "train/loss_ctc": 0.401045560836792, "train/loss_error": 0.32425710558891296, "train/loss_total": 0.3396148085594177 }, { "epoch": 2.838961038961039, "step": 1093, "train/loss_ctc": 0.5128271579742432, "train/loss_error": 0.48050743341445923, "train/loss_total": 0.486971378326416 }, { "epoch": 2.8415584415584414, "step": 1094, "train/loss_ctc": 0.5499550104141235, "train/loss_error": 0.5488672852516174, "train/loss_total": 0.5490848422050476 }, { "epoch": 2.844155844155844, "step": 1095, "train/loss_ctc": 0.8399548530578613, "train/loss_error": 0.4667408764362335, "train/loss_total": 0.541383683681488 }, { "epoch": 2.846753246753247, "step": 1096, "train/loss_ctc": 0.6348162889480591, "train/loss_error": 0.5328812003135681, "train/loss_total": 0.5532681941986084 }, { "epoch": 2.849350649350649, "step": 1097, "train/loss_ctc": 0.5650181770324707, "train/loss_error": 0.5220789313316345, "train/loss_total": 0.5306667685508728 }, { "epoch": 2.851948051948052, "step": 1098, "train/loss_ctc": 0.7270939350128174, "train/loss_error": 0.36447402834892273, "train/loss_total": 0.43699800968170166 }, { "epoch": 2.8545454545454545, "step": 1099, "train/loss_ctc": 0.44491246342658997, "train/loss_error": 0.38273072242736816, "train/loss_total": 0.3951670825481415 }, { "epoch": 2.857142857142857, "grad_norm": 11684.5927734375, "learning_rate": 1.2888311688311689e-05, "loss": 0.4751, "step": 1100 }, { "epoch": 2.857142857142857, "step": 1100, "train/loss_ctc": 0.4735669791698456, "train/loss_error": 0.41752803325653076, "train/loss_total": 0.4287358224391937 }, { "epoch": 2.85974025974026, "step": 1101, "train/loss_ctc": 0.3072929084300995, "train/loss_error": 0.4196677803993225, "train/loss_total": 0.3971928060054779 }, { "epoch": 2.862337662337662, "step": 1102, "train/loss_ctc": 0.41832083463668823, "train/loss_error": 0.3735329508781433, "train/loss_total": 0.3824905455112457 }, { "epoch": 2.864935064935065, "step": 1103, "train/loss_ctc": 0.3380206525325775, "train/loss_error": 0.3762279152870178, "train/loss_total": 0.3685864508152008 }, { "epoch": 2.8675324675324676, "step": 1104, "train/loss_ctc": 0.5267245173454285, "train/loss_error": 0.43652206659317017, "train/loss_total": 0.45456254482269287 }, { "epoch": 2.87012987012987, "step": 1105, "train/loss_ctc": 0.551058292388916, "train/loss_error": 0.43591251969337463, "train/loss_total": 0.4589416980743408 }, { "epoch": 2.8727272727272726, "step": 1106, "train/loss_ctc": 0.4001947045326233, "train/loss_error": 0.36505547165870667, "train/loss_total": 0.37208330631256104 }, { "epoch": 2.8753246753246753, "step": 1107, "train/loss_ctc": 0.833452582359314, "train/loss_error": 0.4229649305343628, "train/loss_total": 0.505062460899353 }, { "epoch": 2.877922077922078, "step": 1108, "train/loss_ctc": 0.4934249520301819, "train/loss_error": 0.3944964110851288, "train/loss_total": 0.4142821431159973 }, { "epoch": 2.8805194805194807, "step": 1109, "train/loss_ctc": 0.640194296836853, "train/loss_error": 0.41116654872894287, "train/loss_total": 0.4569721221923828 }, { "epoch": 2.883116883116883, "grad_norm": 15772.6171875, "learning_rate": 1.2732467532467533e-05, "loss": 0.4239, "step": 1110 }, { "epoch": 2.883116883116883, "step": 1110, "train/loss_ctc": 0.6515666246414185, "train/loss_error": 0.4170789420604706, "train/loss_total": 0.46397650241851807 }, { "epoch": 2.8857142857142857, "step": 1111, "train/loss_ctc": 0.45644718408584595, "train/loss_error": 0.475750595331192, "train/loss_total": 0.4718899130821228 }, { "epoch": 2.8883116883116884, "step": 1112, "train/loss_ctc": 0.4641418755054474, "train/loss_error": 0.466952383518219, "train/loss_total": 0.46639031171798706 }, { "epoch": 2.8909090909090907, "step": 1113, "train/loss_ctc": 0.40143442153930664, "train/loss_error": 0.3783647418022156, "train/loss_total": 0.3829786777496338 }, { "epoch": 2.8935064935064934, "step": 1114, "train/loss_ctc": 0.509638249874115, "train/loss_error": 0.43316298723220825, "train/loss_total": 0.44845807552337646 }, { "epoch": 2.896103896103896, "step": 1115, "train/loss_ctc": 0.32767122983932495, "train/loss_error": 0.4466158449649811, "train/loss_total": 0.42282694578170776 }, { "epoch": 2.898701298701299, "step": 1116, "train/loss_ctc": 0.31911253929138184, "train/loss_error": 0.368724524974823, "train/loss_total": 0.3588021397590637 }, { "epoch": 2.9012987012987015, "step": 1117, "train/loss_ctc": 0.5532773733139038, "train/loss_error": 0.393316388130188, "train/loss_total": 0.42530861496925354 }, { "epoch": 2.9038961038961038, "step": 1118, "train/loss_ctc": 0.6523173451423645, "train/loss_error": 0.4070286154747009, "train/loss_total": 0.45608633756637573 }, { "epoch": 2.9064935064935065, "step": 1119, "train/loss_ctc": 0.4289053678512573, "train/loss_error": 0.4377175271511078, "train/loss_total": 0.43595510721206665 }, { "epoch": 2.909090909090909, "grad_norm": 18408.3984375, "learning_rate": 1.2576623376623377e-05, "loss": 0.4333, "step": 1120 }, { "epoch": 2.909090909090909, "step": 1120, "train/loss_ctc": 0.7566837668418884, "train/loss_error": 0.4256063997745514, "train/loss_total": 0.49182188510894775 }, { "epoch": 2.9116883116883114, "step": 1121, "train/loss_ctc": 0.48620209097862244, "train/loss_error": 0.4784765839576721, "train/loss_total": 0.4800216853618622 }, { "epoch": 2.914285714285714, "step": 1122, "train/loss_ctc": 0.6565996408462524, "train/loss_error": 0.4748297929763794, "train/loss_total": 0.5111837387084961 }, { "epoch": 2.916883116883117, "step": 1123, "train/loss_ctc": 0.5651522874832153, "train/loss_error": 0.5169439911842346, "train/loss_total": 0.5265856385231018 }, { "epoch": 2.9194805194805196, "step": 1124, "train/loss_ctc": 0.6273560523986816, "train/loss_error": 0.5062985420227051, "train/loss_total": 0.5305100679397583 }, { "epoch": 2.9220779220779223, "step": 1125, "train/loss_ctc": 0.4008033573627472, "train/loss_error": 0.3787464201450348, "train/loss_total": 0.3831578195095062 }, { "epoch": 2.9246753246753245, "step": 1126, "train/loss_ctc": 0.5779982805252075, "train/loss_error": 0.47681185603141785, "train/loss_total": 0.49704915285110474 }, { "epoch": 2.9272727272727272, "step": 1127, "train/loss_ctc": 0.4759480655193329, "train/loss_error": 0.3589356243610382, "train/loss_total": 0.38233810663223267 }, { "epoch": 2.92987012987013, "step": 1128, "train/loss_ctc": 0.35516661405563354, "train/loss_error": 0.39917734265327454, "train/loss_total": 0.39037519693374634 }, { "epoch": 2.9324675324675322, "step": 1129, "train/loss_ctc": 0.3993743658065796, "train/loss_error": 0.34156861901283264, "train/loss_total": 0.3531297743320465 }, { "epoch": 2.935064935064935, "grad_norm": 10478.005859375, "learning_rate": 1.2420779220779221e-05, "loss": 0.4546, "step": 1130 }, { "epoch": 2.935064935064935, "step": 1130, "train/loss_ctc": 0.3402799367904663, "train/loss_error": 0.4956597685813904, "train/loss_total": 0.4645838141441345 }, { "epoch": 2.9376623376623376, "step": 1131, "train/loss_ctc": 0.40472668409347534, "train/loss_error": 0.3615226745605469, "train/loss_total": 0.3701635003089905 }, { "epoch": 2.9402597402597404, "step": 1132, "train/loss_ctc": 0.5316271781921387, "train/loss_error": 0.44803938269615173, "train/loss_total": 0.46475696563720703 }, { "epoch": 2.942857142857143, "step": 1133, "train/loss_ctc": 0.5957574844360352, "train/loss_error": 0.4779989719390869, "train/loss_total": 0.5015506744384766 }, { "epoch": 2.9454545454545453, "step": 1134, "train/loss_ctc": 0.3421749770641327, "train/loss_error": 0.40288326144218445, "train/loss_total": 0.39074158668518066 }, { "epoch": 2.948051948051948, "step": 1135, "train/loss_ctc": 0.45721250772476196, "train/loss_error": 0.2997426688671112, "train/loss_total": 0.3312366306781769 }, { "epoch": 2.9506493506493507, "step": 1136, "train/loss_ctc": 0.6706161499023438, "train/loss_error": 0.47212955355644226, "train/loss_total": 0.5118268728256226 }, { "epoch": 2.953246753246753, "step": 1137, "train/loss_ctc": 0.46815750002861023, "train/loss_error": 0.43193432688713074, "train/loss_total": 0.4391789734363556 }, { "epoch": 2.9558441558441557, "step": 1138, "train/loss_ctc": 0.423090398311615, "train/loss_error": 0.4449428617954254, "train/loss_total": 0.4405723810195923 }, { "epoch": 2.9584415584415584, "step": 1139, "train/loss_ctc": 0.6532402634620667, "train/loss_error": 0.43275895714759827, "train/loss_total": 0.47685521841049194 }, { "epoch": 2.961038961038961, "grad_norm": 15774.033203125, "learning_rate": 1.2264935064935064e-05, "loss": 0.4391, "step": 1140 }, { "epoch": 2.961038961038961, "step": 1140, "train/loss_ctc": 0.5413474440574646, "train/loss_error": 0.37877848744392395, "train/loss_total": 0.41129228472709656 }, { "epoch": 2.963636363636364, "step": 1141, "train/loss_ctc": 0.5238915681838989, "train/loss_error": 0.35331690311431885, "train/loss_total": 0.3874318599700928 }, { "epoch": 2.966233766233766, "step": 1142, "train/loss_ctc": 0.35574871301651, "train/loss_error": 0.34070584177970886, "train/loss_total": 0.3437144160270691 }, { "epoch": 2.968831168831169, "step": 1143, "train/loss_ctc": 0.3639603853225708, "train/loss_error": 0.3014603853225708, "train/loss_total": 0.31396037340164185 }, { "epoch": 2.9714285714285715, "step": 1144, "train/loss_ctc": 0.8587663769721985, "train/loss_error": 0.500674307346344, "train/loss_total": 0.5722927451133728 }, { "epoch": 2.974025974025974, "step": 1145, "train/loss_ctc": 0.5448176860809326, "train/loss_error": 0.3748875856399536, "train/loss_total": 0.40887361764907837 }, { "epoch": 2.9766233766233765, "step": 1146, "train/loss_ctc": 0.5990529656410217, "train/loss_error": 0.5113049149513245, "train/loss_total": 0.5288545489311218 }, { "epoch": 2.979220779220779, "step": 1147, "train/loss_ctc": 0.5857210159301758, "train/loss_error": 0.3825155198574066, "train/loss_total": 0.42315661907196045 }, { "epoch": 2.981818181818182, "step": 1148, "train/loss_ctc": 0.7220996618270874, "train/loss_error": 0.37011298537254333, "train/loss_total": 0.4405103325843811 }, { "epoch": 2.9844155844155846, "step": 1149, "train/loss_ctc": 0.5409660935401917, "train/loss_error": 0.5616151690483093, "train/loss_total": 0.5574853420257568 }, { "epoch": 2.987012987012987, "grad_norm": 11940.0712890625, "learning_rate": 1.210909090909091e-05, "loss": 0.4388, "step": 1150 }, { "epoch": 2.987012987012987, "step": 1150, "train/loss_ctc": 0.4117790460586548, "train/loss_error": 0.41203221678733826, "train/loss_total": 0.41198158264160156 }, { "epoch": 2.9896103896103896, "step": 1151, "train/loss_ctc": 0.4917967617511749, "train/loss_error": 0.5261300206184387, "train/loss_total": 0.5192633867263794 }, { "epoch": 2.9922077922077923, "step": 1152, "train/loss_ctc": 0.6187111139297485, "train/loss_error": 0.45116427540779114, "train/loss_total": 0.4846736490726471 }, { "epoch": 2.9948051948051946, "step": 1153, "train/loss_ctc": 0.40464097261428833, "train/loss_error": 0.39993995428085327, "train/loss_total": 0.40088018774986267 }, { "epoch": 2.9974025974025973, "step": 1154, "train/loss_ctc": 0.3937489688396454, "train/loss_error": 0.33331114053726196, "train/loss_total": 0.3453987240791321 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.6580639481544495, "val/loss_error": 0.15718591213226318, "val/loss_total": 0.2573615312576294 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.6621676683425903, "val/loss_error": 0.5485708713531494, "val/loss_total": 0.5712902545928955 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.13273023068904877, "val/loss_error": 0.32054683566093445, "val/loss_total": 0.28298354148864746 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.20483270287513733, "val/loss_error": 0.15497153997421265, "val/loss_total": 0.16494378447532654 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.16707952320575714, "val/loss_error": 0.22042898833751678, "val/loss_total": 0.20975908637046814 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.15507860481739044, "val/loss_error": 0.34929752349853516, "val/loss_total": 0.31045374274253845 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.18056946992874146, "val/loss_error": 0.4774301052093506, "val/loss_total": 0.41805797815322876 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.30342721939086914, "val/loss_error": 0.3142806589603424, "val/loss_total": 0.31210997700691223 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5220861434936523, "val/loss_error": 0.37400922179222107, "val/loss_total": 0.40362462401390076 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.2722392976284027, "val/loss_error": 0.4073939025402069, "val/loss_total": 0.38036298751831055 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.12215763330459595, "val/loss_error": 0.16581222414970398, "val/loss_total": 0.15708130598068237 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.2889125943183899, "val/loss_error": 0.3483034372329712, "val/loss_total": 0.3364252746105194 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.8021115064620972, "val/loss_error": 0.5139584541320801, "val/loss_total": 0.5715890526771545 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.8975286483764648, "val/loss_error": 0.5533605217933655, "val/loss_total": 0.6221941709518433 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.7002307176589966, "val/loss_error": 0.39720526337623596, "val/loss_total": 0.4578103721141815 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.622624933719635, "val/loss_error": 0.7004300355911255, "val/loss_total": 0.6848690509796143 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.1265534907579422, "val/loss_error": 0.25233903527259827, "val/loss_total": 0.22718192636966705 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.1739174723625183, "val/loss_error": 0.26358503103256226, "val/loss_total": 0.24565152823925018 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.12272503972053528, "val/loss_error": 0.3766392469406128, "val/loss_total": 0.32585641741752625 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5140892863273621, "val/loss_error": 0.4491191804409027, "val/loss_total": 0.4621132016181946 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.8124223947525024, "val/loss_error": 0.648276150226593, "val/loss_total": 0.681105375289917 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.29811644554138184, "val/loss_error": 0.5495657920837402, "val/loss_total": 0.49927595257759094 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.3216789960861206, "val/loss_error": 0.43675029277801514, "val/loss_total": 0.4137360453605652 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.16273917257785797, "val/loss_error": 0.2803690731525421, "val/loss_total": 0.25684309005737305 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.3482425808906555, "val/loss_error": 0.536351203918457, "val/loss_total": 0.4987294673919678 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.505254864692688, "val/loss_error": 0.6310588717460632, "val/loss_total": 0.6058980822563171 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.36416545510292053, "val/loss_error": 0.43008938431739807, "val/loss_total": 0.41690459847450256 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.16577476263046265, "val/loss_error": 0.29135894775390625, "val/loss_total": 0.266242116689682 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.03596179187297821, "val/loss_error": 0.10874757170677185, "val/loss_total": 0.09419041872024536 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.30921247601509094, "val/loss_error": 0.5036188364028931, "val/loss_total": 0.46473756432533264 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.12864145636558533, "val/loss_error": 0.11369780451059341, "val/loss_total": 0.11668653786182404 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.13178551197052002, "val/loss_error": 0.3102155029773712, "val/loss_total": 0.27452951669692993 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.115273617208004, "val/loss_error": 0.17199793457984924, "val/loss_total": 0.16065306961536407 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.6405043005943298, "val/loss_error": 0.8141154050827026, "val/loss_total": 0.779393196105957 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.1030980572104454, "val/loss_error": 0.3935265839099884, "val/loss_total": 0.33544087409973145 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.18573683500289917, "val/loss_error": 0.4743245542049408, "val/loss_total": 0.4166070222854614 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.4643550515174866, "val/loss_error": 0.6636830568313599, "val/loss_total": 0.6238174438476562 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5016526579856873, "val/loss_error": 0.44068288803100586, "val/loss_total": 0.45287683606147766 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.23791144788265228, "val/loss_error": 0.5324169397354126, "val/loss_total": 0.4735158681869507 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.6225698590278625, "val/loss_error": 0.6702117919921875, "val/loss_total": 0.6606834530830383 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.35752439498901367, "val/loss_error": 0.3109796345233917, "val/loss_total": 0.32028859853744507 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.3961281478404999, "val/loss_error": 0.4945213198661804, "val/loss_total": 0.47484269738197327 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.49980098009109497, "val/loss_error": 0.9358336329460144, "val/loss_total": 0.8486271500587463 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.6459805369377136, "val/loss_error": 0.6026987433433533, "val/loss_total": 0.6113550662994385 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.47025877237319946, "val/loss_error": 0.37504804134368896, "val/loss_total": 0.3940901756286621 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.42480966448783875, "val/loss_error": 0.5924115180969238, "val/loss_total": 0.5588911771774292 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.6073455810546875, "val/loss_error": 0.45560306310653687, "val/loss_total": 0.48595157265663147 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.3805772662162781, "val/loss_error": 0.5505061745643616, "val/loss_total": 0.5165203809738159 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.42129766941070557, "val/loss_error": 0.2496919333934784, "val/loss_total": 0.2840130925178528 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.7383629083633423, "val/loss_error": 0.9179896712303162, "val/loss_total": 0.8820643424987793 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.3275763988494873, "val/loss_error": 0.8340638875961304, "val/loss_total": 0.7327663898468018 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.31633949279785156, "val/loss_error": 0.469523549079895, "val/loss_total": 0.43888676166534424 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.34511521458625793, "val/loss_error": 0.4992108643054962, "val/loss_total": 0.4683917462825775 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5280628204345703, "val/loss_error": 0.8945280909538269, "val/loss_total": 0.8212350606918335 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.19480204582214355, "val/loss_error": 0.41341185569763184, "val/loss_total": 0.3696898818016052 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.1899215131998062, "val/loss_error": 0.5256856083869934, "val/loss_total": 0.45853281021118164 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.15713685750961304, "val/loss_error": 0.34616461396217346, "val/loss_total": 0.3083590865135193 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.6430149674415588, "val/loss_error": 0.36787253618240356, "val/loss_total": 0.4229010343551636 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.6692752838134766, "val/loss_error": 0.6818868517875671, "val/loss_total": 0.6793645620346069 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.018353624269366264, "val/loss_error": 0.1475881189107895, "val/loss_total": 0.12174122035503387 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.7836276888847351, "val/loss_error": 0.712173342704773, "val/loss_total": 0.7264642119407654 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.6199716925621033, "val/loss_error": 0.4252519905567169, "val/loss_total": 0.46419593691825867 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.09071795642375946, "val/loss_error": 0.11779610067605972, "val/loss_total": 0.11238047480583191 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.2051188200712204, "val/loss_error": 0.33128517866134644, "val/loss_total": 0.30605190992355347 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.14487744867801666, "val/loss_error": 0.23051516711711884, "val/loss_total": 0.2133876234292984 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.24379706382751465, "val/loss_error": 0.45275649428367615, "val/loss_total": 0.41096460819244385 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.3175213038921356, "val/loss_error": 0.3009627163410187, "val/loss_total": 0.30427443981170654 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.6426157355308533, "val/loss_error": 0.5618151426315308, "val/loss_total": 0.5779752731323242 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.47045525908470154, "val/loss_error": 0.43299204111099243, "val/loss_total": 0.4404847025871277 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5138522386550903, "val/loss_error": 0.3497912585735321, "val/loss_total": 0.3826034665107727 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.22630517184734344, "val/loss_error": 0.2883339822292328, "val/loss_total": 0.27592822909355164 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.6723560690879822, "val/loss_error": 0.5081537961959839, "val/loss_total": 0.5409942865371704 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5852820873260498, "val/loss_error": 0.7170498967170715, "val/loss_total": 0.6906963586807251 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.24082712829113007, "val/loss_error": 0.3945828080177307, "val/loss_total": 0.36383169889450073 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.258217990398407, "val/loss_error": 0.3212644159793854, "val/loss_total": 0.30865514278411865 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.059018947184085846, "val/loss_error": 0.34356769919395447, "val/loss_total": 0.2866579294204712 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.8264157772064209, "val/loss_error": 0.5909216403961182, "val/loss_total": 0.6380204558372498 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5260598063468933, "val/loss_error": 0.33366337418556213, "val/loss_total": 0.3721426725387573 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5887243747711182, "val/loss_error": 0.38573944568634033, "val/loss_total": 0.42633646726608276 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.004134493414312601, "val/loss_error": 0.41631585359573364, "val/loss_total": 0.33387959003448486 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.41651082038879395, "val/loss_error": 0.297037810087204, "val/loss_total": 0.32093241810798645 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.01626768708229065, "val/loss_error": 0.09875071048736572, "val/loss_total": 0.08225410431623459 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.24177709221839905, "val/loss_error": 0.2904958426952362, "val/loss_total": 0.2807520925998688 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.3480969965457916, "val/loss_error": 0.41882196068763733, "val/loss_total": 0.40467697381973267 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.43603867292404175, "val/loss_error": 0.6251306533813477, "val/loss_total": 0.5873122811317444 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5490698218345642, "val/loss_error": 0.6772449612617493, "val/loss_total": 0.6516099572181702 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.050307661294937134, "val/loss_error": 0.13460572063922882, "val/loss_total": 0.11774610728025436 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.49185776710510254, "val/loss_error": 0.5086323022842407, "val/loss_total": 0.5052773952484131 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.27931448817253113, "val/loss_error": 0.29421481490135193, "val/loss_total": 0.2912347614765167 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.48148366808891296, "val/loss_error": 0.41693803668022156, "val/loss_total": 0.4298471510410309 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.2782840430736542, "val/loss_error": 0.5790437459945679, "val/loss_total": 0.5188918113708496 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.7449977397918701, "val/loss_error": 0.4395989179611206, "val/loss_total": 0.5006787180900574 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.24617502093315125, "val/loss_error": 0.11400170624256134, "val/loss_total": 0.14043636620044708 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.35983026027679443, "val/loss_error": 0.31918010115623474, "val/loss_total": 0.32731014490127563 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.5144906640052795, "val/loss_error": 0.5286933779716492, "val/loss_total": 0.5258528590202332 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.26924049854278564, "val/loss_error": 0.22666649520397186, "val/loss_total": 0.2351813018321991 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.09533322602510452, "val/loss_error": 0.24223613739013672, "val/loss_total": 0.21285556256771088 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.6909613013267517, "val/loss_error": 0.5298591256141663, "val/loss_total": 0.5620795488357544 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.15339118242263794, "val/loss_error": 0.3335748314857483, "val/loss_total": 0.2975381016731262 }, { "epoch": 3.0, "step": 1155, "val/loss_ctc": 0.18729117512702942, "val/loss_error": 0.2632543742656708, "val/loss_total": 0.24806174635887146 }, { "epoch": 3.0, "eval_eval/f1_0": 0.5216346383094788, "eval_eval/f1_1": 0.9293574690818787, "eval_eval/precision_0": 0.5507614016532898, "eval_eval/precision_1": 0.9221556782722473, "eval_eval/recall_0": 0.4954337775707245, "eval_eval/recall_1": 0.9366726279258728, "eval_eval/wer": 0.15477220017647558, "eval_loss": 0.4150855541229248, "eval_runtime": 7.6035, "eval_samples_per_second": 13.152, "eval_steps_per_second": 13.152, "step": 1155 }, { "epoch": 3.0, "step": 1155, "train/loss_ctc": 0.560454785823822, "train/loss_error": 0.3775007128715515, "train/loss_total": 0.4140915274620056 }, { "epoch": 3.0025974025974027, "step": 1156, "train/loss_ctc": 0.43425455689430237, "train/loss_error": 0.3782772719860077, "train/loss_total": 0.38947272300720215 }, { "epoch": 3.005194805194805, "step": 1157, "train/loss_ctc": 0.41532307863235474, "train/loss_error": 0.413300096988678, "train/loss_total": 0.41370469331741333 }, { "epoch": 3.0077922077922077, "step": 1158, "train/loss_ctc": 0.5292662382125854, "train/loss_error": 0.4271813631057739, "train/loss_total": 0.4475983679294586 }, { "epoch": 3.0103896103896104, "step": 1159, "train/loss_ctc": 0.37234580516815186, "train/loss_error": 0.35202357172966003, "train/loss_total": 0.3560880422592163 }, { "epoch": 3.012987012987013, "grad_norm": 10511.923828125, "learning_rate": 1.1953246753246752e-05, "loss": 0.4183, "step": 1160 }, { "epoch": 3.012987012987013, "step": 1160, "train/loss_ctc": 0.2726876735687256, "train/loss_error": 0.3123156726360321, "train/loss_total": 0.3043900728225708 }, { "epoch": 3.0155844155844154, "step": 1161, "train/loss_ctc": 0.5444832444190979, "train/loss_error": 0.4198555052280426, "train/loss_total": 0.4447810649871826 }, { "epoch": 3.018181818181818, "step": 1162, "train/loss_ctc": 0.5770219564437866, "train/loss_error": 0.39468154311180115, "train/loss_total": 0.4311496317386627 }, { "epoch": 3.020779220779221, "step": 1163, "train/loss_ctc": 0.42646124958992004, "train/loss_error": 0.3965921401977539, "train/loss_total": 0.40256595611572266 }, { "epoch": 3.0233766233766235, "step": 1164, "train/loss_ctc": 0.3199550211429596, "train/loss_error": 0.3710996210575104, "train/loss_total": 0.36087071895599365 }, { "epoch": 3.0259740259740258, "step": 1165, "train/loss_ctc": 0.28304916620254517, "train/loss_error": 0.3724706470966339, "train/loss_total": 0.3545863628387451 }, { "epoch": 3.0285714285714285, "step": 1166, "train/loss_ctc": 0.49488377571105957, "train/loss_error": 0.4037034213542938, "train/loss_total": 0.421939492225647 }, { "epoch": 3.031168831168831, "step": 1167, "train/loss_ctc": 0.5108945369720459, "train/loss_error": 0.39426326751708984, "train/loss_total": 0.4175895154476166 }, { "epoch": 3.033766233766234, "step": 1168, "train/loss_ctc": 0.61514812707901, "train/loss_error": 0.3763536512851715, "train/loss_total": 0.42411255836486816 }, { "epoch": 3.036363636363636, "step": 1169, "train/loss_ctc": 0.5195884704589844, "train/loss_error": 0.34042036533355713, "train/loss_total": 0.37625399231910706 }, { "epoch": 3.038961038961039, "grad_norm": 16237.5771484375, "learning_rate": 1.1797402597402597e-05, "loss": 0.3938, "step": 1170 }, { "epoch": 3.038961038961039, "step": 1170, "train/loss_ctc": 0.5451658964157104, "train/loss_error": 0.4196418523788452, "train/loss_total": 0.4447466731071472 }, { "epoch": 3.0415584415584416, "step": 1171, "train/loss_ctc": 0.5507853031158447, "train/loss_error": 0.45364588499069214, "train/loss_total": 0.4730737805366516 }, { "epoch": 3.0441558441558443, "step": 1172, "train/loss_ctc": 0.42613035440444946, "train/loss_error": 0.5132092833518982, "train/loss_total": 0.49579352140426636 }, { "epoch": 3.0467532467532465, "step": 1173, "train/loss_ctc": 0.4548606872558594, "train/loss_error": 0.4315774738788605, "train/loss_total": 0.43623411655426025 }, { "epoch": 3.0493506493506493, "step": 1174, "train/loss_ctc": 0.4347500801086426, "train/loss_error": 0.30086764693260193, "train/loss_total": 0.32764413952827454 }, { "epoch": 3.051948051948052, "step": 1175, "train/loss_ctc": 0.4582422971725464, "train/loss_error": 0.4368739426136017, "train/loss_total": 0.4411476254463196 }, { "epoch": 3.0545454545454547, "step": 1176, "train/loss_ctc": 0.3444085419178009, "train/loss_error": 0.39689168334007263, "train/loss_total": 0.38639506697654724 }, { "epoch": 3.057142857142857, "step": 1177, "train/loss_ctc": 0.41650062799453735, "train/loss_error": 0.41012629866600037, "train/loss_total": 0.4114011526107788 }, { "epoch": 3.0597402597402596, "step": 1178, "train/loss_ctc": 0.4043588936328888, "train/loss_error": 0.3891022205352783, "train/loss_total": 0.3921535611152649 }, { "epoch": 3.0623376623376624, "step": 1179, "train/loss_ctc": 0.4342518746852875, "train/loss_error": 0.4324661195278168, "train/loss_total": 0.4328232705593109 }, { "epoch": 3.064935064935065, "grad_norm": 12345.1884765625, "learning_rate": 1.1641558441558442e-05, "loss": 0.4241, "step": 1180 }, { "epoch": 3.064935064935065, "step": 1180, "train/loss_ctc": 0.49679338932037354, "train/loss_error": 0.4205424189567566, "train/loss_total": 0.43579262495040894 }, { "epoch": 3.0675324675324673, "step": 1181, "train/loss_ctc": 0.4987674653530121, "train/loss_error": 0.461103618144989, "train/loss_total": 0.4686363935470581 }, { "epoch": 3.07012987012987, "step": 1182, "train/loss_ctc": 0.32694077491760254, "train/loss_error": 0.3923311233520508, "train/loss_total": 0.379253089427948 }, { "epoch": 3.0727272727272728, "step": 1183, "train/loss_ctc": 0.6609147191047668, "train/loss_error": 0.48893266916275024, "train/loss_total": 0.5233290791511536 }, { "epoch": 3.0753246753246755, "step": 1184, "train/loss_ctc": 0.8765750527381897, "train/loss_error": 0.4031088650226593, "train/loss_total": 0.49780207872390747 }, { "epoch": 3.0779220779220777, "step": 1185, "train/loss_ctc": 0.3634054660797119, "train/loss_error": 0.42640307545661926, "train/loss_total": 0.4138035774230957 }, { "epoch": 3.0805194805194804, "step": 1186, "train/loss_ctc": 0.30674415826797485, "train/loss_error": 0.35527849197387695, "train/loss_total": 0.3455716371536255 }, { "epoch": 3.083116883116883, "step": 1187, "train/loss_ctc": 0.6243969202041626, "train/loss_error": 0.39692750573158264, "train/loss_total": 0.44242140650749207 }, { "epoch": 3.085714285714286, "step": 1188, "train/loss_ctc": 0.4030070900917053, "train/loss_error": 0.40200433135032654, "train/loss_total": 0.4022049009799957 }, { "epoch": 3.088311688311688, "step": 1189, "train/loss_ctc": 0.43888065218925476, "train/loss_error": 0.3438030481338501, "train/loss_total": 0.36281856894493103 }, { "epoch": 3.090909090909091, "grad_norm": 14501.564453125, "learning_rate": 1.1485714285714285e-05, "loss": 0.4272, "step": 1190 }, { "epoch": 3.090909090909091, "step": 1190, "train/loss_ctc": 0.529021143913269, "train/loss_error": 0.3362174928188324, "train/loss_total": 0.37477824091911316 }, { "epoch": 3.0935064935064935, "step": 1191, "train/loss_ctc": 0.4172998070716858, "train/loss_error": 0.36731401085853577, "train/loss_total": 0.37731119990348816 }, { "epoch": 3.0961038961038962, "step": 1192, "train/loss_ctc": 0.38997727632522583, "train/loss_error": 0.30747556686401367, "train/loss_total": 0.32397592067718506 }, { "epoch": 3.0987012987012985, "step": 1193, "train/loss_ctc": 0.4798024594783783, "train/loss_error": 0.4238984286785126, "train/loss_total": 0.4350792467594147 }, { "epoch": 3.101298701298701, "step": 1194, "train/loss_ctc": 0.5723548531532288, "train/loss_error": 0.47323355078697205, "train/loss_total": 0.49305784702301025 }, { "epoch": 3.103896103896104, "step": 1195, "train/loss_ctc": 0.5121769905090332, "train/loss_error": 0.36761102080345154, "train/loss_total": 0.39652422070503235 }, { "epoch": 3.1064935064935066, "step": 1196, "train/loss_ctc": 0.486442893743515, "train/loss_error": 0.39444971084594727, "train/loss_total": 0.4128483533859253 }, { "epoch": 3.109090909090909, "step": 1197, "train/loss_ctc": 0.9496522545814514, "train/loss_error": 0.4460899829864502, "train/loss_total": 0.5468024611473083 }, { "epoch": 3.1116883116883116, "step": 1198, "train/loss_ctc": 0.3496629595756531, "train/loss_error": 0.36546438932418823, "train/loss_total": 0.36230409145355225 }, { "epoch": 3.1142857142857143, "step": 1199, "train/loss_ctc": 0.326794296503067, "train/loss_error": 0.35488560795783997, "train/loss_total": 0.3492673635482788 }, { "epoch": 3.116883116883117, "grad_norm": 11032.3857421875, "learning_rate": 1.132987012987013e-05, "loss": 0.4072, "step": 1200 }, { "epoch": 3.116883116883117, "step": 1200, "train/loss_ctc": 0.5999653935432434, "train/loss_error": 0.3345664143562317, "train/loss_total": 0.38764622807502747 }, { "epoch": 3.1194805194805193, "step": 1201, "train/loss_ctc": 0.6130951642990112, "train/loss_error": 0.42901211977005005, "train/loss_total": 0.4658287465572357 }, { "epoch": 3.122077922077922, "step": 1202, "train/loss_ctc": 0.3966794013977051, "train/loss_error": 0.3328374922275543, "train/loss_total": 0.34560590982437134 }, { "epoch": 3.1246753246753247, "step": 1203, "train/loss_ctc": 0.5416979789733887, "train/loss_error": 0.39920881390571594, "train/loss_total": 0.42770665884017944 }, { "epoch": 3.1272727272727274, "step": 1204, "train/loss_ctc": 0.3864938020706177, "train/loss_error": 0.4009104371070862, "train/loss_total": 0.39802712202072144 }, { "epoch": 3.1298701298701297, "step": 1205, "train/loss_ctc": 0.5936585068702698, "train/loss_error": 0.3365963101387024, "train/loss_total": 0.3880087733268738 }, { "epoch": 3.1324675324675324, "step": 1206, "train/loss_ctc": 0.46402329206466675, "train/loss_error": 0.3534928858280182, "train/loss_total": 0.3755989968776703 }, { "epoch": 3.135064935064935, "step": 1207, "train/loss_ctc": 0.3886217474937439, "train/loss_error": 0.4809587001800537, "train/loss_total": 0.46249133348464966 }, { "epoch": 3.137662337662338, "step": 1208, "train/loss_ctc": 0.2977336049079895, "train/loss_error": 0.5114662051200867, "train/loss_total": 0.4687196910381317 }, { "epoch": 3.14025974025974, "step": 1209, "train/loss_ctc": 0.3758595287799835, "train/loss_error": 0.3286065459251404, "train/loss_total": 0.33805716037750244 }, { "epoch": 3.142857142857143, "grad_norm": 11681.8857421875, "learning_rate": 1.1174025974025973e-05, "loss": 0.4058, "step": 1210 }, { "epoch": 3.142857142857143, "step": 1210, "train/loss_ctc": 0.6247178316116333, "train/loss_error": 0.42511194944381714, "train/loss_total": 0.4650331139564514 }, { "epoch": 3.1454545454545455, "step": 1211, "train/loss_ctc": 0.4361884593963623, "train/loss_error": 0.36450889706611633, "train/loss_total": 0.3788447976112366 }, { "epoch": 3.148051948051948, "step": 1212, "train/loss_ctc": 0.6630990505218506, "train/loss_error": 0.4281615614891052, "train/loss_total": 0.47514909505844116 }, { "epoch": 3.1506493506493505, "step": 1213, "train/loss_ctc": 0.4213089346885681, "train/loss_error": 0.3841707706451416, "train/loss_total": 0.3915984034538269 }, { "epoch": 3.153246753246753, "step": 1214, "train/loss_ctc": 0.3656207323074341, "train/loss_error": 0.360701322555542, "train/loss_total": 0.36168521642684937 }, { "epoch": 3.155844155844156, "step": 1215, "train/loss_ctc": 0.4388577342033386, "train/loss_error": 0.33401426672935486, "train/loss_total": 0.35498297214508057 }, { "epoch": 3.1584415584415586, "step": 1216, "train/loss_ctc": 0.818760097026825, "train/loss_error": 0.4197370409965515, "train/loss_total": 0.49954167008399963 }, { "epoch": 3.161038961038961, "step": 1217, "train/loss_ctc": 0.3856218755245209, "train/loss_error": 0.3314916491508484, "train/loss_total": 0.34231770038604736 }, { "epoch": 3.1636363636363636, "step": 1218, "train/loss_ctc": 0.3780733346939087, "train/loss_error": 0.3476686179637909, "train/loss_total": 0.3537495732307434 }, { "epoch": 3.1662337662337663, "step": 1219, "train/loss_ctc": 0.5582375526428223, "train/loss_error": 0.5387893915176392, "train/loss_total": 0.5426790118217468 }, { "epoch": 3.168831168831169, "grad_norm": 13356.3134765625, "learning_rate": 1.1018181818181818e-05, "loss": 0.4166, "step": 1220 }, { "epoch": 3.168831168831169, "step": 1220, "train/loss_ctc": 0.46425294876098633, "train/loss_error": 0.4198004901409149, "train/loss_total": 0.42869099974632263 }, { "epoch": 3.1714285714285713, "step": 1221, "train/loss_ctc": 0.4409901797771454, "train/loss_error": 0.41633960604667664, "train/loss_total": 0.4212697148323059 }, { "epoch": 3.174025974025974, "step": 1222, "train/loss_ctc": 0.5765416622161865, "train/loss_error": 0.46541741490364075, "train/loss_total": 0.4876422882080078 }, { "epoch": 3.1766233766233767, "step": 1223, "train/loss_ctc": 0.5923153162002563, "train/loss_error": 0.4061568081378937, "train/loss_total": 0.44338852167129517 }, { "epoch": 3.1792207792207794, "step": 1224, "train/loss_ctc": 0.8155616521835327, "train/loss_error": 0.47797515988349915, "train/loss_total": 0.5454924702644348 }, { "epoch": 3.1818181818181817, "step": 1225, "train/loss_ctc": 0.626903772354126, "train/loss_error": 0.4505944848060608, "train/loss_total": 0.4858563542366028 }, { "epoch": 3.1844155844155844, "step": 1226, "train/loss_ctc": 0.46414461731910706, "train/loss_error": 0.3575604259967804, "train/loss_total": 0.37887728214263916 }, { "epoch": 3.187012987012987, "step": 1227, "train/loss_ctc": 0.4629397392272949, "train/loss_error": 0.39003679156303406, "train/loss_total": 0.40461739897727966 }, { "epoch": 3.18961038961039, "step": 1228, "train/loss_ctc": 0.36131125688552856, "train/loss_error": 0.33203473687171936, "train/loss_total": 0.33789005875587463 }, { "epoch": 3.192207792207792, "step": 1229, "train/loss_ctc": 0.420377254486084, "train/loss_error": 0.3873375654220581, "train/loss_total": 0.39394551515579224 }, { "epoch": 3.1948051948051948, "grad_norm": 10076.8720703125, "learning_rate": 1.0862337662337663e-05, "loss": 0.4328, "step": 1230 }, { "epoch": 3.1948051948051948, "step": 1230, "train/loss_ctc": 0.3365088701248169, "train/loss_error": 0.43334028124809265, "train/loss_total": 0.41397401690483093 }, { "epoch": 3.1974025974025975, "step": 1231, "train/loss_ctc": 0.720054030418396, "train/loss_error": 0.44211381673812866, "train/loss_total": 0.49770188331604004 }, { "epoch": 3.2, "step": 1232, "train/loss_ctc": 0.3929821252822876, "train/loss_error": 0.41905149817466736, "train/loss_total": 0.41383761167526245 }, { "epoch": 3.2025974025974024, "step": 1233, "train/loss_ctc": 0.5043747425079346, "train/loss_error": 0.3679743707180023, "train/loss_total": 0.3952544629573822 }, { "epoch": 3.205194805194805, "step": 1234, "train/loss_ctc": 0.3755224347114563, "train/loss_error": 0.3394257426261902, "train/loss_total": 0.3466451168060303 }, { "epoch": 3.207792207792208, "step": 1235, "train/loss_ctc": 0.5006710886955261, "train/loss_error": 0.49816063046455383, "train/loss_total": 0.4986627399921417 }, { "epoch": 3.2103896103896106, "step": 1236, "train/loss_ctc": 0.615953803062439, "train/loss_error": 0.5012214779853821, "train/loss_total": 0.5241679549217224 }, { "epoch": 3.212987012987013, "step": 1237, "train/loss_ctc": 0.4601154327392578, "train/loss_error": 0.43039435148239136, "train/loss_total": 0.4363386034965515 }, { "epoch": 3.2155844155844155, "step": 1238, "train/loss_ctc": 0.6050335168838501, "train/loss_error": 0.40728843212127686, "train/loss_total": 0.446837455034256 }, { "epoch": 3.2181818181818183, "step": 1239, "train/loss_ctc": 0.5222371816635132, "train/loss_error": 0.4058828353881836, "train/loss_total": 0.4291537404060364 }, { "epoch": 3.220779220779221, "grad_norm": 13037.583984375, "learning_rate": 1.0706493506493506e-05, "loss": 0.4403, "step": 1240 }, { "epoch": 3.220779220779221, "step": 1240, "train/loss_ctc": 0.5334767699241638, "train/loss_error": 0.39727386832237244, "train/loss_total": 0.42451444268226624 }, { "epoch": 3.2233766233766232, "step": 1241, "train/loss_ctc": 0.3529520034790039, "train/loss_error": 0.4672631025314331, "train/loss_total": 0.4444009065628052 }, { "epoch": 3.225974025974026, "step": 1242, "train/loss_ctc": 0.28983691334724426, "train/loss_error": 0.38572806119918823, "train/loss_total": 0.36654984951019287 }, { "epoch": 3.2285714285714286, "step": 1243, "train/loss_ctc": 0.5056844353675842, "train/loss_error": 0.3730250597000122, "train/loss_total": 0.3995569348335266 }, { "epoch": 3.2311688311688314, "step": 1244, "train/loss_ctc": 0.3091326355934143, "train/loss_error": 0.37206947803497314, "train/loss_total": 0.3594821095466614 }, { "epoch": 3.2337662337662336, "step": 1245, "train/loss_ctc": 0.568239152431488, "train/loss_error": 0.4395183324813843, "train/loss_total": 0.4652625322341919 }, { "epoch": 3.2363636363636363, "step": 1246, "train/loss_ctc": 0.5412859916687012, "train/loss_error": 0.452788382768631, "train/loss_total": 0.47048792243003845 }, { "epoch": 3.238961038961039, "step": 1247, "train/loss_ctc": 0.8182598352432251, "train/loss_error": 0.4337199032306671, "train/loss_total": 0.5106278657913208 }, { "epoch": 3.2415584415584417, "step": 1248, "train/loss_ctc": 0.5448137521743774, "train/loss_error": 0.34255290031433105, "train/loss_total": 0.3830050826072693 }, { "epoch": 3.244155844155844, "step": 1249, "train/loss_ctc": 0.5024176836013794, "train/loss_error": 0.533534824848175, "train/loss_total": 0.5273114442825317 }, { "epoch": 3.2467532467532467, "grad_norm": 13136.814453125, "learning_rate": 1.055064935064935e-05, "loss": 0.4351, "step": 1250 }, { "epoch": 3.2467532467532467, "step": 1250, "train/loss_ctc": 0.46249693632125854, "train/loss_error": 0.4294285476207733, "train/loss_total": 0.43604224920272827 }, { "epoch": 3.2493506493506494, "step": 1251, "train/loss_ctc": 0.3186558187007904, "train/loss_error": 0.3508254885673523, "train/loss_total": 0.3443915545940399 }, { "epoch": 3.2519480519480517, "step": 1252, "train/loss_ctc": 0.7146618962287903, "train/loss_error": 0.4620758295059204, "train/loss_total": 0.5125930309295654 }, { "epoch": 3.2545454545454544, "step": 1253, "train/loss_ctc": 0.5081959962844849, "train/loss_error": 0.38993752002716064, "train/loss_total": 0.4135892391204834 }, { "epoch": 3.257142857142857, "step": 1254, "train/loss_ctc": 0.6742507219314575, "train/loss_error": 0.4021266996860504, "train/loss_total": 0.45655152201652527 }, { "epoch": 3.25974025974026, "step": 1255, "train/loss_ctc": 0.4914345145225525, "train/loss_error": 0.37859371304512024, "train/loss_total": 0.40116187930107117 }, { "epoch": 3.2623376623376625, "step": 1256, "train/loss_ctc": 0.4572444558143616, "train/loss_error": 0.4627815783023834, "train/loss_total": 0.46167418360710144 }, { "epoch": 3.264935064935065, "step": 1257, "train/loss_ctc": 0.35783281922340393, "train/loss_error": 0.4540521204471588, "train/loss_total": 0.43480825424194336 }, { "epoch": 3.2675324675324675, "step": 1258, "train/loss_ctc": 0.6624414920806885, "train/loss_error": 0.4156232476234436, "train/loss_total": 0.4649869203567505 }, { "epoch": 3.27012987012987, "step": 1259, "train/loss_ctc": 0.5141432285308838, "train/loss_error": 0.4313735067844391, "train/loss_total": 0.44792747497558594 }, { "epoch": 3.2727272727272725, "grad_norm": 12299.7529296875, "learning_rate": 1.0394805194805195e-05, "loss": 0.4374, "step": 1260 }, { "epoch": 3.2727272727272725, "step": 1260, "train/loss_ctc": 0.46616455912590027, "train/loss_error": 0.40131938457489014, "train/loss_total": 0.41428840160369873 }, { "epoch": 3.275324675324675, "step": 1261, "train/loss_ctc": 0.5997427701950073, "train/loss_error": 0.36976075172424316, "train/loss_total": 0.4157571792602539 }, { "epoch": 3.277922077922078, "step": 1262, "train/loss_ctc": 0.4911749064922333, "train/loss_error": 0.39728835225105286, "train/loss_total": 0.41606566309928894 }, { "epoch": 3.2805194805194806, "step": 1263, "train/loss_ctc": 0.2909751534461975, "train/loss_error": 0.28966426849365234, "train/loss_total": 0.2899264395236969 }, { "epoch": 3.2831168831168833, "step": 1264, "train/loss_ctc": 0.3588910698890686, "train/loss_error": 0.3279787302017212, "train/loss_total": 0.3341611921787262 }, { "epoch": 3.2857142857142856, "step": 1265, "train/loss_ctc": 0.574114978313446, "train/loss_error": 0.39110296964645386, "train/loss_total": 0.42770540714263916 }, { "epoch": 3.2883116883116883, "step": 1266, "train/loss_ctc": 0.6516112089157104, "train/loss_error": 0.38768792152404785, "train/loss_total": 0.4404726028442383 }, { "epoch": 3.290909090909091, "step": 1267, "train/loss_ctc": 0.3163907527923584, "train/loss_error": 0.40727323293685913, "train/loss_total": 0.389096736907959 }, { "epoch": 3.2935064935064933, "step": 1268, "train/loss_ctc": 0.7065231800079346, "train/loss_error": 0.45859625935554504, "train/loss_total": 0.5081816911697388 }, { "epoch": 3.296103896103896, "step": 1269, "train/loss_ctc": 0.44237497448921204, "train/loss_error": 0.4235139787197113, "train/loss_total": 0.42728617787361145 }, { "epoch": 3.2987012987012987, "grad_norm": 9740.2607421875, "learning_rate": 1.023896103896104e-05, "loss": 0.4063, "step": 1270 }, { "epoch": 3.2987012987012987, "step": 1270, "train/loss_ctc": 0.5570170879364014, "train/loss_error": 0.3965860605239868, "train/loss_total": 0.42867225408554077 }, { "epoch": 3.3012987012987014, "step": 1271, "train/loss_ctc": 0.5257846713066101, "train/loss_error": 0.4143438935279846, "train/loss_total": 0.43663203716278076 }, { "epoch": 3.303896103896104, "step": 1272, "train/loss_ctc": 0.3378147482872009, "train/loss_error": 0.3820844888687134, "train/loss_total": 0.37323057651519775 }, { "epoch": 3.3064935064935064, "step": 1273, "train/loss_ctc": 0.3689268231391907, "train/loss_error": 0.39379894733428955, "train/loss_total": 0.3888245224952698 }, { "epoch": 3.309090909090909, "step": 1274, "train/loss_ctc": 0.6763808131217957, "train/loss_error": 0.39020654559135437, "train/loss_total": 0.4474413990974426 }, { "epoch": 3.311688311688312, "step": 1275, "train/loss_ctc": 0.6483334302902222, "train/loss_error": 0.47395285964012146, "train/loss_total": 0.5088289976119995 }, { "epoch": 3.314285714285714, "step": 1276, "train/loss_ctc": 0.4048723578453064, "train/loss_error": 0.3965241312980652, "train/loss_total": 0.3981937766075134 }, { "epoch": 3.3168831168831168, "step": 1277, "train/loss_ctc": 0.4436143636703491, "train/loss_error": 0.36391201615333557, "train/loss_total": 0.3798525035381317 }, { "epoch": 3.3194805194805195, "step": 1278, "train/loss_ctc": 0.5177027583122253, "train/loss_error": 0.38325178623199463, "train/loss_total": 0.4101420044898987 }, { "epoch": 3.322077922077922, "step": 1279, "train/loss_ctc": 0.5004217624664307, "train/loss_error": 0.40793928503990173, "train/loss_total": 0.42643579840660095 }, { "epoch": 3.324675324675325, "grad_norm": 12232.8115234375, "learning_rate": 1.0083116883116883e-05, "loss": 0.4198, "step": 1280 }, { "epoch": 3.324675324675325, "step": 1280, "train/loss_ctc": 0.37940269708633423, "train/loss_error": 0.37407946586608887, "train/loss_total": 0.3751441240310669 }, { "epoch": 3.327272727272727, "step": 1281, "train/loss_ctc": 0.41623684763908386, "train/loss_error": 0.5096496343612671, "train/loss_total": 0.4909670650959015 }, { "epoch": 3.32987012987013, "step": 1282, "train/loss_ctc": 0.7898601293563843, "train/loss_error": 0.42296648025512695, "train/loss_total": 0.4963452219963074 }, { "epoch": 3.3324675324675326, "step": 1283, "train/loss_ctc": 0.4783031940460205, "train/loss_error": 0.4562467038631439, "train/loss_total": 0.4606580138206482 }, { "epoch": 3.335064935064935, "step": 1284, "train/loss_ctc": 0.4100441038608551, "train/loss_error": 0.44881489872932434, "train/loss_total": 0.44106072187423706 }, { "epoch": 3.3376623376623376, "step": 1285, "train/loss_ctc": 0.4775768518447876, "train/loss_error": 0.39635786414146423, "train/loss_total": 0.41260167956352234 }, { "epoch": 3.3402597402597403, "step": 1286, "train/loss_ctc": 0.37349218130111694, "train/loss_error": 0.413324773311615, "train/loss_total": 0.40535828471183777 }, { "epoch": 3.342857142857143, "step": 1287, "train/loss_ctc": 0.41993486881256104, "train/loss_error": 0.3417554795742035, "train/loss_total": 0.357391357421875 }, { "epoch": 3.3454545454545457, "step": 1288, "train/loss_ctc": 0.6609988212585449, "train/loss_error": 0.37006327509880066, "train/loss_total": 0.42825040221214294 }, { "epoch": 3.348051948051948, "step": 1289, "train/loss_ctc": 0.6980478763580322, "train/loss_error": 0.42777106165885925, "train/loss_total": 0.48182642459869385 }, { "epoch": 3.3506493506493507, "grad_norm": 14929.916015625, "learning_rate": 9.927272727272728e-06, "loss": 0.435, "step": 1290 }, { "epoch": 3.3506493506493507, "step": 1290, "train/loss_ctc": 0.4908629357814789, "train/loss_error": 0.3994499146938324, "train/loss_total": 0.4177325367927551 }, { "epoch": 3.3532467532467534, "step": 1291, "train/loss_ctc": 0.4317016899585724, "train/loss_error": 0.481991171836853, "train/loss_total": 0.4719332754611969 }, { "epoch": 3.3558441558441556, "step": 1292, "train/loss_ctc": 0.6678430438041687, "train/loss_error": 0.4736858010292053, "train/loss_total": 0.5125172734260559 }, { "epoch": 3.3584415584415583, "step": 1293, "train/loss_ctc": 0.49064499139785767, "train/loss_error": 0.35784977674484253, "train/loss_total": 0.3844088315963745 }, { "epoch": 3.361038961038961, "step": 1294, "train/loss_ctc": 0.5663731694221497, "train/loss_error": 0.37192246317863464, "train/loss_total": 0.4108126163482666 }, { "epoch": 3.3636363636363638, "step": 1295, "train/loss_ctc": 0.5950202941894531, "train/loss_error": 0.44790658354759216, "train/loss_total": 0.4773293435573578 }, { "epoch": 3.3662337662337665, "step": 1296, "train/loss_ctc": 0.522273063659668, "train/loss_error": 0.3731810748577118, "train/loss_total": 0.4029994606971741 }, { "epoch": 3.3688311688311687, "step": 1297, "train/loss_ctc": 0.394092321395874, "train/loss_error": 0.4422683119773865, "train/loss_total": 0.4326331317424774 }, { "epoch": 3.3714285714285714, "step": 1298, "train/loss_ctc": 0.49276983737945557, "train/loss_error": 0.47283902764320374, "train/loss_total": 0.47682517766952515 }, { "epoch": 3.374025974025974, "step": 1299, "train/loss_ctc": 0.3718269467353821, "train/loss_error": 0.4154662489891052, "train/loss_total": 0.40673840045928955 }, { "epoch": 3.3766233766233764, "grad_norm": 10983.310546875, "learning_rate": 9.771428571428573e-06, "loss": 0.4394, "step": 1300 }, { "epoch": 3.3766233766233764, "step": 1300, "train/loss_ctc": 0.44570451974868774, "train/loss_error": 0.40567293763160706, "train/loss_total": 0.41367924213409424 }, { "epoch": 3.379220779220779, "step": 1301, "train/loss_ctc": 1.1163398027420044, "train/loss_error": 0.530860424041748, "train/loss_total": 0.6479563117027283 }, { "epoch": 3.381818181818182, "step": 1302, "train/loss_ctc": 0.35902124643325806, "train/loss_error": 0.38113662600517273, "train/loss_total": 0.3767135739326477 }, { "epoch": 3.3844155844155845, "step": 1303, "train/loss_ctc": 0.5398958921432495, "train/loss_error": 0.4046563506126404, "train/loss_total": 0.43170425295829773 }, { "epoch": 3.3870129870129873, "step": 1304, "train/loss_ctc": 0.672162652015686, "train/loss_error": 0.4779109060764313, "train/loss_total": 0.5167612433433533 }, { "epoch": 3.3896103896103895, "step": 1305, "train/loss_ctc": 0.45505499839782715, "train/loss_error": 0.4058052599430084, "train/loss_total": 0.4156551957130432 }, { "epoch": 3.3922077922077922, "step": 1306, "train/loss_ctc": 0.3671361804008484, "train/loss_error": 0.5467464923858643, "train/loss_total": 0.51082444190979 }, { "epoch": 3.394805194805195, "step": 1307, "train/loss_ctc": 0.6639873385429382, "train/loss_error": 0.45730531215667725, "train/loss_total": 0.4986417293548584 }, { "epoch": 3.397402597402597, "step": 1308, "train/loss_ctc": 0.46235421299934387, "train/loss_error": 0.38930630683898926, "train/loss_total": 0.4039159119129181 }, { "epoch": 3.4, "step": 1309, "train/loss_ctc": 0.5102282762527466, "train/loss_error": 0.42744767665863037, "train/loss_total": 0.44400379061698914 }, { "epoch": 3.4025974025974026, "grad_norm": 15437.8193359375, "learning_rate": 9.615584415584416e-06, "loss": 0.466, "step": 1310 }, { "epoch": 3.4025974025974026, "step": 1310, "train/loss_ctc": 0.4986162781715393, "train/loss_error": 0.31437283754348755, "train/loss_total": 0.3512215316295624 }, { "epoch": 3.4051948051948053, "step": 1311, "train/loss_ctc": 0.6417385339736938, "train/loss_error": 0.38858091831207275, "train/loss_total": 0.439212441444397 }, { "epoch": 3.407792207792208, "step": 1312, "train/loss_ctc": 0.5926350355148315, "train/loss_error": 0.3834102749824524, "train/loss_total": 0.4252552390098572 }, { "epoch": 3.4103896103896103, "step": 1313, "train/loss_ctc": 0.462144136428833, "train/loss_error": 0.42203712463378906, "train/loss_total": 0.4300585389137268 }, { "epoch": 3.412987012987013, "step": 1314, "train/loss_ctc": 0.6734921932220459, "train/loss_error": 0.5617452263832092, "train/loss_total": 0.5840946435928345 }, { "epoch": 3.4155844155844157, "step": 1315, "train/loss_ctc": 0.3855300843715668, "train/loss_error": 0.34425055980682373, "train/loss_total": 0.35250648856163025 }, { "epoch": 3.418181818181818, "step": 1316, "train/loss_ctc": 0.3795663118362427, "train/loss_error": 0.3978242576122284, "train/loss_total": 0.39417266845703125 }, { "epoch": 3.4207792207792207, "step": 1317, "train/loss_ctc": 0.5374703407287598, "train/loss_error": 0.4686153531074524, "train/loss_total": 0.48238635063171387 }, { "epoch": 3.4233766233766234, "step": 1318, "train/loss_ctc": 0.756875216960907, "train/loss_error": 0.4496529996395111, "train/loss_total": 0.5110974311828613 }, { "epoch": 3.425974025974026, "step": 1319, "train/loss_ctc": 0.48215383291244507, "train/loss_error": 0.43357813358306885, "train/loss_total": 0.4432933032512665 }, { "epoch": 3.4285714285714284, "grad_norm": 13290.212890625, "learning_rate": 9.459740259740261e-06, "loss": 0.4413, "step": 1320 }, { "epoch": 3.4285714285714284, "step": 1320, "train/loss_ctc": 0.3514283299446106, "train/loss_error": 0.3546869456768036, "train/loss_total": 0.35403522849082947 }, { "epoch": 3.431168831168831, "step": 1321, "train/loss_ctc": 0.2643968164920807, "train/loss_error": 0.34223490953445435, "train/loss_total": 0.32666730880737305 }, { "epoch": 3.433766233766234, "step": 1322, "train/loss_ctc": 0.5083020925521851, "train/loss_error": 0.47165971994400024, "train/loss_total": 0.4789882004261017 }, { "epoch": 3.4363636363636365, "step": 1323, "train/loss_ctc": 0.545059323310852, "train/loss_error": 0.39247584342956543, "train/loss_total": 0.4229925274848938 }, { "epoch": 3.4389610389610388, "step": 1324, "train/loss_ctc": 0.4754941463470459, "train/loss_error": 0.3996099829673767, "train/loss_total": 0.41478681564331055 }, { "epoch": 3.4415584415584415, "step": 1325, "train/loss_ctc": 0.5911481380462646, "train/loss_error": 0.44727757573127747, "train/loss_total": 0.4760516881942749 }, { "epoch": 3.444155844155844, "step": 1326, "train/loss_ctc": 0.6006382703781128, "train/loss_error": 0.3866845667362213, "train/loss_total": 0.4294753074645996 }, { "epoch": 3.446753246753247, "step": 1327, "train/loss_ctc": 0.4470299482345581, "train/loss_error": 0.35541054606437683, "train/loss_total": 0.37373441457748413 }, { "epoch": 3.449350649350649, "step": 1328, "train/loss_ctc": 0.5041096210479736, "train/loss_error": 0.40694907307624817, "train/loss_total": 0.4263811707496643 }, { "epoch": 3.451948051948052, "step": 1329, "train/loss_ctc": 0.6270551681518555, "train/loss_error": 0.43547409772872925, "train/loss_total": 0.47379031777381897 }, { "epoch": 3.4545454545454546, "grad_norm": 17538.173828125, "learning_rate": 9.303896103896104e-06, "loss": 0.4177, "step": 1330 }, { "epoch": 3.4545454545454546, "step": 1330, "train/loss_ctc": 0.40062427520751953, "train/loss_error": 0.41459962725639343, "train/loss_total": 0.41180455684661865 }, { "epoch": 3.4571428571428573, "step": 1331, "train/loss_ctc": 0.5082250833511353, "train/loss_error": 0.44025489687919617, "train/loss_total": 0.4538489580154419 }, { "epoch": 3.4597402597402596, "step": 1332, "train/loss_ctc": 0.44792577624320984, "train/loss_error": 0.39377331733703613, "train/loss_total": 0.4046038091182709 }, { "epoch": 3.4623376623376623, "step": 1333, "train/loss_ctc": 0.580712080001831, "train/loss_error": 0.4066387116909027, "train/loss_total": 0.44145339727401733 }, { "epoch": 3.464935064935065, "step": 1334, "train/loss_ctc": 0.4221975803375244, "train/loss_error": 0.3806561529636383, "train/loss_total": 0.38896444439888 }, { "epoch": 3.4675324675324677, "step": 1335, "train/loss_ctc": 0.9225654602050781, "train/loss_error": 0.544942319393158, "train/loss_total": 0.620466947555542 }, { "epoch": 3.47012987012987, "step": 1336, "train/loss_ctc": 0.44546401500701904, "train/loss_error": 0.3228594660758972, "train/loss_total": 0.3473803997039795 }, { "epoch": 3.4727272727272727, "step": 1337, "train/loss_ctc": 0.3946203589439392, "train/loss_error": 0.3790142834186554, "train/loss_total": 0.3821355104446411 }, { "epoch": 3.4753246753246754, "step": 1338, "train/loss_ctc": 0.6760531663894653, "train/loss_error": 0.5030357837677002, "train/loss_total": 0.5376392602920532 }, { "epoch": 3.477922077922078, "step": 1339, "train/loss_ctc": 0.4389795660972595, "train/loss_error": 0.41250374913215637, "train/loss_total": 0.4177989065647125 }, { "epoch": 3.4805194805194803, "grad_norm": 10591.1220703125, "learning_rate": 9.148051948051949e-06, "loss": 0.4406, "step": 1340 }, { "epoch": 3.4805194805194803, "step": 1340, "train/loss_ctc": 0.4503701329231262, "train/loss_error": 0.32020512223243713, "train/loss_total": 0.3462381362915039 }, { "epoch": 3.483116883116883, "step": 1341, "train/loss_ctc": 0.48344287276268005, "train/loss_error": 0.46861082315444946, "train/loss_total": 0.4715772271156311 }, { "epoch": 3.4857142857142858, "step": 1342, "train/loss_ctc": 0.27351266145706177, "train/loss_error": 0.40275803208351135, "train/loss_total": 0.37690895795822144 }, { "epoch": 3.4883116883116885, "step": 1343, "train/loss_ctc": 0.8590928912162781, "train/loss_error": 0.45168089866638184, "train/loss_total": 0.53316330909729 }, { "epoch": 3.4909090909090907, "step": 1344, "train/loss_ctc": 0.8572540879249573, "train/loss_error": 0.48417994379997253, "train/loss_total": 0.5587947964668274 }, { "epoch": 3.4935064935064934, "step": 1345, "train/loss_ctc": 0.5597317814826965, "train/loss_error": 0.39446502923965454, "train/loss_total": 0.427518367767334 }, { "epoch": 3.496103896103896, "step": 1346, "train/loss_ctc": 0.33393463492393494, "train/loss_error": 0.4543595314025879, "train/loss_total": 0.4302745461463928 }, { "epoch": 3.498701298701299, "step": 1347, "train/loss_ctc": 0.6497001647949219, "train/loss_error": 0.5072917938232422, "train/loss_total": 0.535773515701294 }, { "epoch": 3.501298701298701, "step": 1348, "train/loss_ctc": 0.5615580081939697, "train/loss_error": 0.3630445599555969, "train/loss_total": 0.402747243642807 }, { "epoch": 3.503896103896104, "step": 1349, "train/loss_ctc": 0.46369943022727966, "train/loss_error": 0.3468446433544159, "train/loss_total": 0.37021559476852417 }, { "epoch": 3.5064935064935066, "grad_norm": 10766.6416015625, "learning_rate": 8.992207792207792e-06, "loss": 0.4453, "step": 1350 }, { "epoch": 3.5064935064935066, "step": 1350, "train/loss_ctc": 0.43675553798675537, "train/loss_error": 0.34126827120780945, "train/loss_total": 0.36036574840545654 }, { "epoch": 3.509090909090909, "step": 1351, "train/loss_ctc": 0.5456709861755371, "train/loss_error": 0.3517710864543915, "train/loss_total": 0.39055106043815613 }, { "epoch": 3.5116883116883115, "step": 1352, "train/loss_ctc": 0.6604889631271362, "train/loss_error": 0.36581891775131226, "train/loss_total": 0.42475295066833496 }, { "epoch": 3.5142857142857142, "step": 1353, "train/loss_ctc": 0.3232189416885376, "train/loss_error": 0.4600088596343994, "train/loss_total": 0.4326508939266205 }, { "epoch": 3.516883116883117, "step": 1354, "train/loss_ctc": 0.3550971746444702, "train/loss_error": 0.3585284352302551, "train/loss_total": 0.35784220695495605 }, { "epoch": 3.5194805194805197, "step": 1355, "train/loss_ctc": 0.3749171495437622, "train/loss_error": 0.38022908568382263, "train/loss_total": 0.37916672229766846 }, { "epoch": 3.522077922077922, "step": 1356, "train/loss_ctc": 0.39525923132896423, "train/loss_error": 0.3647264242172241, "train/loss_total": 0.37083300948143005 }, { "epoch": 3.5246753246753246, "step": 1357, "train/loss_ctc": 0.4746891260147095, "train/loss_error": 0.36589351296424866, "train/loss_total": 0.3876526355743408 }, { "epoch": 3.5272727272727273, "step": 1358, "train/loss_ctc": 0.5385679006576538, "train/loss_error": 0.40191203355789185, "train/loss_total": 0.42924320697784424 }, { "epoch": 3.5298701298701296, "step": 1359, "train/loss_ctc": 0.256634384393692, "train/loss_error": 0.3773106038570404, "train/loss_total": 0.3531753718852997 }, { "epoch": 3.5324675324675323, "grad_norm": 12734.982421875, "learning_rate": 8.836363636363637e-06, "loss": 0.3886, "step": 1360 }, { "epoch": 3.5324675324675323, "step": 1360, "train/loss_ctc": 0.38704490661621094, "train/loss_error": 0.35041192173957825, "train/loss_total": 0.35773855447769165 }, { "epoch": 3.535064935064935, "step": 1361, "train/loss_ctc": 0.5839318037033081, "train/loss_error": 0.42442864179611206, "train/loss_total": 0.4563292860984802 }, { "epoch": 3.5376623376623377, "step": 1362, "train/loss_ctc": 0.446794331073761, "train/loss_error": 0.34034475684165955, "train/loss_total": 0.36163467168807983 }, { "epoch": 3.5402597402597404, "step": 1363, "train/loss_ctc": 0.45931872725486755, "train/loss_error": 0.3165898025035858, "train/loss_total": 0.3451355993747711 }, { "epoch": 3.5428571428571427, "step": 1364, "train/loss_ctc": 0.43039393424987793, "train/loss_error": 0.3589068353176117, "train/loss_total": 0.3732042610645294 }, { "epoch": 3.5454545454545454, "step": 1365, "train/loss_ctc": 0.5546805262565613, "train/loss_error": 0.4339548349380493, "train/loss_total": 0.45809999108314514 }, { "epoch": 3.548051948051948, "step": 1366, "train/loss_ctc": 0.3784133791923523, "train/loss_error": 0.33629080653190613, "train/loss_total": 0.34471532702445984 }, { "epoch": 3.5506493506493504, "step": 1367, "train/loss_ctc": 0.3197287321090698, "train/loss_error": 0.2383388727903366, "train/loss_total": 0.2546168565750122 }, { "epoch": 3.553246753246753, "step": 1368, "train/loss_ctc": 0.5376799702644348, "train/loss_error": 0.4587138593196869, "train/loss_total": 0.47450709342956543 }, { "epoch": 3.555844155844156, "step": 1369, "train/loss_ctc": 0.5051330327987671, "train/loss_error": 0.4690857231616974, "train/loss_total": 0.4762951731681824 }, { "epoch": 3.5584415584415585, "grad_norm": 13162.830078125, "learning_rate": 8.680519480519482e-06, "loss": 0.3902, "step": 1370 }, { "epoch": 3.5584415584415585, "step": 1370, "train/loss_ctc": 0.37672877311706543, "train/loss_error": 0.3609672784805298, "train/loss_total": 0.36411958932876587 }, { "epoch": 3.5610389610389612, "step": 1371, "train/loss_ctc": 0.515227735042572, "train/loss_error": 0.4646449089050293, "train/loss_total": 0.4747614860534668 }, { "epoch": 3.5636363636363635, "step": 1372, "train/loss_ctc": 0.7646007537841797, "train/loss_error": 0.43994227051734924, "train/loss_total": 0.5048739910125732 }, { "epoch": 3.566233766233766, "step": 1373, "train/loss_ctc": 0.5104670524597168, "train/loss_error": 0.5218390226364136, "train/loss_total": 0.5195646286010742 }, { "epoch": 3.568831168831169, "step": 1374, "train/loss_ctc": 0.4515654444694519, "train/loss_error": 0.43272829055786133, "train/loss_total": 0.43649572134017944 }, { "epoch": 3.571428571428571, "step": 1375, "train/loss_ctc": 0.3106382489204407, "train/loss_error": 0.4127144515514374, "train/loss_total": 0.39229920506477356 }, { "epoch": 3.574025974025974, "step": 1376, "train/loss_ctc": 0.7020418643951416, "train/loss_error": 0.4687291979789734, "train/loss_total": 0.5153917670249939 }, { "epoch": 3.5766233766233766, "step": 1377, "train/loss_ctc": 0.3654468357563019, "train/loss_error": 0.4061749279499054, "train/loss_total": 0.39802929759025574 }, { "epoch": 3.5792207792207793, "step": 1378, "train/loss_ctc": 0.497866690158844, "train/loss_error": 0.42423057556152344, "train/loss_total": 0.4389578104019165 }, { "epoch": 3.581818181818182, "step": 1379, "train/loss_ctc": 0.5566498637199402, "train/loss_error": 0.40213513374328613, "train/loss_total": 0.4330380856990814 }, { "epoch": 3.5844155844155843, "grad_norm": 11336.7578125, "learning_rate": 8.524675324675325e-06, "loss": 0.4478, "step": 1380 }, { "epoch": 3.5844155844155843, "step": 1380, "train/loss_ctc": 0.3255332410335541, "train/loss_error": 0.3790300190448761, "train/loss_total": 0.3683306872844696 }, { "epoch": 3.587012987012987, "step": 1381, "train/loss_ctc": 0.5046752691268921, "train/loss_error": 0.4328015446662903, "train/loss_total": 0.4471762776374817 }, { "epoch": 3.5896103896103897, "step": 1382, "train/loss_ctc": 0.49085733294487, "train/loss_error": 0.3971039354801178, "train/loss_total": 0.41585463285446167 }, { "epoch": 3.592207792207792, "step": 1383, "train/loss_ctc": 0.6112737059593201, "train/loss_error": 0.36587288975715637, "train/loss_total": 0.4149530529975891 }, { "epoch": 3.5948051948051947, "step": 1384, "train/loss_ctc": 0.5688763856887817, "train/loss_error": 0.42813825607299805, "train/loss_total": 0.45628589391708374 }, { "epoch": 3.5974025974025974, "step": 1385, "train/loss_ctc": 0.3224748969078064, "train/loss_error": 0.3272251784801483, "train/loss_total": 0.326275110244751 }, { "epoch": 3.6, "step": 1386, "train/loss_ctc": 0.5913050174713135, "train/loss_error": 0.4292895793914795, "train/loss_total": 0.4616926908493042 }, { "epoch": 3.602597402597403, "step": 1387, "train/loss_ctc": 0.7407423257827759, "train/loss_error": 0.434255450963974, "train/loss_total": 0.49555283784866333 }, { "epoch": 3.605194805194805, "step": 1388, "train/loss_ctc": 0.5003228783607483, "train/loss_error": 0.47073861956596375, "train/loss_total": 0.4766554832458496 }, { "epoch": 3.6077922077922078, "step": 1389, "train/loss_ctc": 0.6905198097229004, "train/loss_error": 0.4245204031467438, "train/loss_total": 0.4777202904224396 }, { "epoch": 3.6103896103896105, "grad_norm": 15288.2861328125, "learning_rate": 8.36883116883117e-06, "loss": 0.434, "step": 1390 }, { "epoch": 3.6103896103896105, "step": 1390, "train/loss_ctc": 0.4472474455833435, "train/loss_error": 0.4082481265068054, "train/loss_total": 0.41604799032211304 }, { "epoch": 3.6129870129870127, "step": 1391, "train/loss_ctc": 0.557290256023407, "train/loss_error": 0.4391278028488159, "train/loss_total": 0.4627602994441986 }, { "epoch": 3.6155844155844155, "step": 1392, "train/loss_ctc": 0.6675815582275391, "train/loss_error": 0.4403178095817566, "train/loss_total": 0.4857705533504486 }, { "epoch": 3.618181818181818, "step": 1393, "train/loss_ctc": 0.6746823787689209, "train/loss_error": 0.4554252326488495, "train/loss_total": 0.49927666783332825 }, { "epoch": 3.620779220779221, "step": 1394, "train/loss_ctc": 0.4473112225532532, "train/loss_error": 0.43102678656578064, "train/loss_total": 0.43428367376327515 }, { "epoch": 3.6233766233766236, "step": 1395, "train/loss_ctc": 0.42694568634033203, "train/loss_error": 0.3278711438179016, "train/loss_total": 0.3476860523223877 }, { "epoch": 3.625974025974026, "step": 1396, "train/loss_ctc": 0.5764203071594238, "train/loss_error": 0.4277254045009613, "train/loss_total": 0.45746439695358276 }, { "epoch": 3.6285714285714286, "step": 1397, "train/loss_ctc": 0.3895387351512909, "train/loss_error": 0.44900065660476685, "train/loss_total": 0.43710827827453613 }, { "epoch": 3.6311688311688313, "step": 1398, "train/loss_ctc": 0.5785950422286987, "train/loss_error": 0.4300076961517334, "train/loss_total": 0.45972517132759094 }, { "epoch": 3.6337662337662335, "step": 1399, "train/loss_ctc": 0.8157651424407959, "train/loss_error": 0.4445559084415436, "train/loss_total": 0.518797755241394 }, { "epoch": 3.6363636363636362, "grad_norm": 15820.4658203125, "learning_rate": 8.212987012987013e-06, "loss": 0.4519, "step": 1400 }, { "epoch": 3.6363636363636362, "step": 1400, "train/loss_ctc": 0.5646628141403198, "train/loss_error": 0.46041586995124817, "train/loss_total": 0.48126527667045593 }, { "epoch": 3.638961038961039, "step": 1401, "train/loss_ctc": 0.30494779348373413, "train/loss_error": 0.3020995855331421, "train/loss_total": 0.3026692271232605 }, { "epoch": 3.6415584415584417, "step": 1402, "train/loss_ctc": 0.6238329410552979, "train/loss_error": 0.44345027208328247, "train/loss_total": 0.4795268177986145 }, { "epoch": 3.6441558441558444, "step": 1403, "train/loss_ctc": 0.3201274871826172, "train/loss_error": 0.4151129424571991, "train/loss_total": 0.39611583948135376 }, { "epoch": 3.6467532467532466, "step": 1404, "train/loss_ctc": 0.35374191403388977, "train/loss_error": 0.41359952092170715, "train/loss_total": 0.4016280174255371 }, { "epoch": 3.6493506493506493, "step": 1405, "train/loss_ctc": 0.45421504974365234, "train/loss_error": 0.39983153343200684, "train/loss_total": 0.4107082486152649 }, { "epoch": 3.651948051948052, "step": 1406, "train/loss_ctc": 0.4598768949508667, "train/loss_error": 0.4419255256652832, "train/loss_total": 0.44551581144332886 }, { "epoch": 3.6545454545454543, "step": 1407, "train/loss_ctc": 0.335799902677536, "train/loss_error": 0.3549775779247284, "train/loss_total": 0.3511420488357544 }, { "epoch": 3.657142857142857, "step": 1408, "train/loss_ctc": 0.5666134357452393, "train/loss_error": 0.36912938952445984, "train/loss_total": 0.4086261987686157 }, { "epoch": 3.6597402597402597, "step": 1409, "train/loss_ctc": 0.3912401795387268, "train/loss_error": 0.4052352011203766, "train/loss_total": 0.40243619680404663 }, { "epoch": 3.6623376623376624, "grad_norm": 10796.84765625, "learning_rate": 8.057142857142857e-06, "loss": 0.408, "step": 1410 }, { "epoch": 3.6623376623376624, "step": 1410, "train/loss_ctc": 1.0309879779815674, "train/loss_error": 0.47392842173576355, "train/loss_total": 0.5853403210639954 }, { "epoch": 3.664935064935065, "step": 1411, "train/loss_ctc": 0.2898538112640381, "train/loss_error": 0.30645471811294556, "train/loss_total": 0.303134560585022 }, { "epoch": 3.6675324675324674, "step": 1412, "train/loss_ctc": 0.35842272639274597, "train/loss_error": 0.38387030363082886, "train/loss_total": 0.3787807822227478 }, { "epoch": 3.67012987012987, "step": 1413, "train/loss_ctc": 0.5414354801177979, "train/loss_error": 0.42592766880989075, "train/loss_total": 0.44902923703193665 }, { "epoch": 3.672727272727273, "step": 1414, "train/loss_ctc": 0.40043193101882935, "train/loss_error": 0.40091991424560547, "train/loss_total": 0.40082231163978577 }, { "epoch": 3.675324675324675, "step": 1415, "train/loss_ctc": 0.5032666325569153, "train/loss_error": 0.4040027856826782, "train/loss_total": 0.4238555431365967 }, { "epoch": 3.677922077922078, "step": 1416, "train/loss_ctc": 0.5427212119102478, "train/loss_error": 0.3651111125946045, "train/loss_total": 0.40063315629959106 }, { "epoch": 3.6805194805194805, "step": 1417, "train/loss_ctc": 0.47717124223709106, "train/loss_error": 0.3902820646762848, "train/loss_total": 0.4076599180698395 }, { "epoch": 3.6831168831168832, "step": 1418, "train/loss_ctc": 0.5362117290496826, "train/loss_error": 0.4230179786682129, "train/loss_total": 0.44565674662590027 }, { "epoch": 3.685714285714286, "step": 1419, "train/loss_ctc": 0.45786672830581665, "train/loss_error": 0.44217920303344727, "train/loss_total": 0.44531673192977905 }, { "epoch": 3.688311688311688, "grad_norm": 16781.263671875, "learning_rate": 7.901298701298702e-06, "loss": 0.424, "step": 1420 }, { "epoch": 3.688311688311688, "step": 1420, "train/loss_ctc": 0.5698102712631226, "train/loss_error": 0.4144764244556427, "train/loss_total": 0.44554319977760315 }, { "epoch": 3.690909090909091, "step": 1421, "train/loss_ctc": 0.6487504243850708, "train/loss_error": 0.47411787509918213, "train/loss_total": 0.5090444087982178 }, { "epoch": 3.6935064935064936, "step": 1422, "train/loss_ctc": 0.5871405601501465, "train/loss_error": 0.4368927776813507, "train/loss_total": 0.46694234013557434 }, { "epoch": 3.696103896103896, "step": 1423, "train/loss_ctc": 0.4532585144042969, "train/loss_error": 0.37517809867858887, "train/loss_total": 0.39079421758651733 }, { "epoch": 3.6987012987012986, "step": 1424, "train/loss_ctc": 0.49880021810531616, "train/loss_error": 0.40076935291290283, "train/loss_total": 0.4203755557537079 }, { "epoch": 3.7012987012987013, "step": 1425, "train/loss_ctc": 0.3956424593925476, "train/loss_error": 0.3611103594303131, "train/loss_total": 0.3680168092250824 }, { "epoch": 3.703896103896104, "step": 1426, "train/loss_ctc": 0.3885621130466461, "train/loss_error": 0.4131113290786743, "train/loss_total": 0.4082014858722687 }, { "epoch": 3.7064935064935067, "step": 1427, "train/loss_ctc": 0.4581233263015747, "train/loss_error": 0.5167943239212036, "train/loss_total": 0.5050601363182068 }, { "epoch": 3.709090909090909, "step": 1428, "train/loss_ctc": 0.4600987434387207, "train/loss_error": 0.4267243444919586, "train/loss_total": 0.4333992600440979 }, { "epoch": 3.7116883116883117, "step": 1429, "train/loss_ctc": 0.38975995779037476, "train/loss_error": 0.29827314615249634, "train/loss_total": 0.316570520401001 }, { "epoch": 3.7142857142857144, "grad_norm": 12175.5380859375, "learning_rate": 7.745454545454545e-06, "loss": 0.4264, "step": 1430 }, { "epoch": 3.7142857142857144, "step": 1430, "train/loss_ctc": 0.6894023418426514, "train/loss_error": 0.4512873888015747, "train/loss_total": 0.49891039729118347 }, { "epoch": 3.7168831168831167, "step": 1431, "train/loss_ctc": 0.8012465238571167, "train/loss_error": 0.4062129557132721, "train/loss_total": 0.48521965742111206 }, { "epoch": 3.7194805194805194, "step": 1432, "train/loss_ctc": 0.45521044731140137, "train/loss_error": 0.49656787514686584, "train/loss_total": 0.48829641938209534 }, { "epoch": 3.722077922077922, "step": 1433, "train/loss_ctc": 0.5171545743942261, "train/loss_error": 0.4993803799152374, "train/loss_total": 0.5029352307319641 }, { "epoch": 3.724675324675325, "step": 1434, "train/loss_ctc": 0.4472656846046448, "train/loss_error": 0.5465297102928162, "train/loss_total": 0.5266768932342529 }, { "epoch": 3.7272727272727275, "step": 1435, "train/loss_ctc": 0.3133692443370819, "train/loss_error": 0.34040024876594543, "train/loss_total": 0.3349940776824951 }, { "epoch": 3.72987012987013, "step": 1436, "train/loss_ctc": 0.25406908988952637, "train/loss_error": 0.3748210668563843, "train/loss_total": 0.3506706953048706 }, { "epoch": 3.7324675324675325, "step": 1437, "train/loss_ctc": 0.6867108941078186, "train/loss_error": 0.4483608603477478, "train/loss_total": 0.49603086709976196 }, { "epoch": 3.735064935064935, "step": 1438, "train/loss_ctc": 0.5657088756561279, "train/loss_error": 0.3839317858219147, "train/loss_total": 0.42028722167015076 }, { "epoch": 3.7376623376623375, "step": 1439, "train/loss_ctc": 0.449011892080307, "train/loss_error": 0.40440833568573, "train/loss_total": 0.4133290648460388 }, { "epoch": 3.74025974025974, "grad_norm": 11321.240234375, "learning_rate": 7.589610389610389e-06, "loss": 0.4517, "step": 1440 }, { "epoch": 3.74025974025974, "step": 1440, "train/loss_ctc": 0.4721786379814148, "train/loss_error": 0.48838260769844055, "train/loss_total": 0.4851418137550354 }, { "epoch": 3.742857142857143, "step": 1441, "train/loss_ctc": 0.3114035427570343, "train/loss_error": 0.37567320466041565, "train/loss_total": 0.36281928420066833 }, { "epoch": 3.7454545454545456, "step": 1442, "train/loss_ctc": 0.7000479698181152, "train/loss_error": 0.40031948685646057, "train/loss_total": 0.46026521921157837 }, { "epoch": 3.7480519480519483, "step": 1443, "train/loss_ctc": 0.29460468888282776, "train/loss_error": 0.31263771653175354, "train/loss_total": 0.3090311288833618 }, { "epoch": 3.7506493506493506, "step": 1444, "train/loss_ctc": 0.36168384552001953, "train/loss_error": 0.4160008728504181, "train/loss_total": 0.40513747930526733 }, { "epoch": 3.7532467532467533, "step": 1445, "train/loss_ctc": 0.42898809909820557, "train/loss_error": 0.5227046608924866, "train/loss_total": 0.5039613842964172 }, { "epoch": 3.755844155844156, "step": 1446, "train/loss_ctc": 0.35844823718070984, "train/loss_error": 0.41169288754463196, "train/loss_total": 0.40104395151138306 }, { "epoch": 3.7584415584415583, "step": 1447, "train/loss_ctc": 0.4387092590332031, "train/loss_error": 0.4516884982585907, "train/loss_total": 0.44909265637397766 }, { "epoch": 3.761038961038961, "step": 1448, "train/loss_ctc": 0.48273342847824097, "train/loss_error": 0.4323335587978363, "train/loss_total": 0.4424135386943817 }, { "epoch": 3.7636363636363637, "step": 1449, "train/loss_ctc": 0.5046195387840271, "train/loss_error": 0.4315757751464844, "train/loss_total": 0.44618451595306396 }, { "epoch": 3.7662337662337664, "grad_norm": 16787.5703125, "learning_rate": 7.433766233766234e-06, "loss": 0.4265, "step": 1450 }, { "epoch": 3.7662337662337664, "step": 1450, "train/loss_ctc": 0.5612258315086365, "train/loss_error": 0.4192807972431183, "train/loss_total": 0.4476698040962219 }, { "epoch": 3.768831168831169, "step": 1451, "train/loss_ctc": 0.5382845997810364, "train/loss_error": 0.3314781188964844, "train/loss_total": 0.37283942103385925 }, { "epoch": 3.7714285714285714, "step": 1452, "train/loss_ctc": 0.5074520111083984, "train/loss_error": 0.47251978516578674, "train/loss_total": 0.479506254196167 }, { "epoch": 3.774025974025974, "step": 1453, "train/loss_ctc": 0.5169554948806763, "train/loss_error": 0.3748437166213989, "train/loss_total": 0.4032661020755768 }, { "epoch": 3.7766233766233768, "step": 1454, "train/loss_ctc": 0.515274167060852, "train/loss_error": 0.4041530191898346, "train/loss_total": 0.42637723684310913 }, { "epoch": 3.779220779220779, "step": 1455, "train/loss_ctc": 0.5233120322227478, "train/loss_error": 0.4097946584224701, "train/loss_total": 0.43249815702438354 }, { "epoch": 3.7818181818181817, "step": 1456, "train/loss_ctc": 0.5449364185333252, "train/loss_error": 0.4075853228569031, "train/loss_total": 0.43505555391311646 }, { "epoch": 3.7844155844155845, "step": 1457, "train/loss_ctc": 0.3191568851470947, "train/loss_error": 0.40666520595550537, "train/loss_total": 0.3891635537147522 }, { "epoch": 3.787012987012987, "step": 1458, "train/loss_ctc": 0.4815218448638916, "train/loss_error": 0.3968192934989929, "train/loss_total": 0.41375982761383057 }, { "epoch": 3.78961038961039, "step": 1459, "train/loss_ctc": 0.523328959941864, "train/loss_error": 0.45490288734436035, "train/loss_total": 0.46858811378479004 }, { "epoch": 3.792207792207792, "grad_norm": 13933.0810546875, "learning_rate": 7.277922077922078e-06, "loss": 0.4269, "step": 1460 }, { "epoch": 3.792207792207792, "step": 1460, "train/loss_ctc": 0.37731295824050903, "train/loss_error": 0.37495696544647217, "train/loss_total": 0.3754281997680664 }, { "epoch": 3.794805194805195, "step": 1461, "train/loss_ctc": 0.5567022562026978, "train/loss_error": 0.40683430433273315, "train/loss_total": 0.43680790066719055 }, { "epoch": 3.7974025974025976, "step": 1462, "train/loss_ctc": 0.486639142036438, "train/loss_error": 0.5111755132675171, "train/loss_total": 0.5062682628631592 }, { "epoch": 3.8, "step": 1463, "train/loss_ctc": 0.38781651854515076, "train/loss_error": 0.4818308651447296, "train/loss_total": 0.4630280137062073 }, { "epoch": 3.8025974025974025, "step": 1464, "train/loss_ctc": 0.43217724561691284, "train/loss_error": 0.42803990840911865, "train/loss_total": 0.4288673996925354 }, { "epoch": 3.8051948051948052, "step": 1465, "train/loss_ctc": 0.4360421895980835, "train/loss_error": 0.4322177767753601, "train/loss_total": 0.4329826831817627 }, { "epoch": 3.807792207792208, "step": 1466, "train/loss_ctc": 0.47328248620033264, "train/loss_error": 0.3976363241672516, "train/loss_total": 0.4127655625343323 }, { "epoch": 3.8103896103896107, "step": 1467, "train/loss_ctc": 0.4223981499671936, "train/loss_error": 0.34246116876602173, "train/loss_total": 0.3584485650062561 }, { "epoch": 3.812987012987013, "step": 1468, "train/loss_ctc": 0.39979201555252075, "train/loss_error": 0.4264717400074005, "train/loss_total": 0.421135812997818 }, { "epoch": 3.8155844155844156, "step": 1469, "train/loss_ctc": 0.4895024597644806, "train/loss_error": 0.3674772083759308, "train/loss_total": 0.3918822407722473 }, { "epoch": 3.8181818181818183, "grad_norm": 14612.7509765625, "learning_rate": 7.122077922077922e-06, "loss": 0.4228, "step": 1470 }, { "epoch": 3.8181818181818183, "step": 1470, "train/loss_ctc": 0.46757209300994873, "train/loss_error": 0.4810556471347809, "train/loss_total": 0.4783589243888855 }, { "epoch": 3.8207792207792206, "step": 1471, "train/loss_ctc": 0.3550802767276764, "train/loss_error": 0.3572498559951782, "train/loss_total": 0.3568159341812134 }, { "epoch": 3.8233766233766233, "step": 1472, "train/loss_ctc": 0.39336639642715454, "train/loss_error": 0.35499048233032227, "train/loss_total": 0.36266565322875977 }, { "epoch": 3.825974025974026, "step": 1473, "train/loss_ctc": 0.5391318798065186, "train/loss_error": 0.46770206093788147, "train/loss_total": 0.4819880425930023 }, { "epoch": 3.8285714285714287, "step": 1474, "train/loss_ctc": 0.2936549782752991, "train/loss_error": 0.4027458131313324, "train/loss_total": 0.3809276521205902 }, { "epoch": 3.8311688311688314, "step": 1475, "train/loss_ctc": 0.6962833404541016, "train/loss_error": 0.4368145763874054, "train/loss_total": 0.4887083172798157 }, { "epoch": 3.8337662337662337, "step": 1476, "train/loss_ctc": 0.2895939350128174, "train/loss_error": 0.3557896316051483, "train/loss_total": 0.34255048632621765 }, { "epoch": 3.8363636363636364, "step": 1477, "train/loss_ctc": 0.3695833683013916, "train/loss_error": 0.4004504978656769, "train/loss_total": 0.39427706599235535 }, { "epoch": 3.838961038961039, "step": 1478, "train/loss_ctc": 0.5398925542831421, "train/loss_error": 0.47912779450416565, "train/loss_total": 0.49128076434135437 }, { "epoch": 3.8415584415584414, "step": 1479, "train/loss_ctc": 0.4987926483154297, "train/loss_error": 0.4171009063720703, "train/loss_total": 0.4334392547607422 }, { "epoch": 3.844155844155844, "grad_norm": 18708.177734375, "learning_rate": 6.966233766233766e-06, "loss": 0.4211, "step": 1480 }, { "epoch": 3.844155844155844, "step": 1480, "train/loss_ctc": 0.5047426819801331, "train/loss_error": 0.37874582409858704, "train/loss_total": 0.4039452075958252 }, { "epoch": 3.846753246753247, "step": 1481, "train/loss_ctc": 0.5098236799240112, "train/loss_error": 0.39315474033355713, "train/loss_total": 0.41648852825164795 }, { "epoch": 3.849350649350649, "step": 1482, "train/loss_ctc": 0.49802619218826294, "train/loss_error": 0.3883044123649597, "train/loss_total": 0.4102487564086914 }, { "epoch": 3.851948051948052, "step": 1483, "train/loss_ctc": 0.40937894582748413, "train/loss_error": 0.42439088225364685, "train/loss_total": 0.42138850688934326 }, { "epoch": 3.8545454545454545, "step": 1484, "train/loss_ctc": 0.5877373218536377, "train/loss_error": 0.4332084655761719, "train/loss_total": 0.464114248752594 }, { "epoch": 3.857142857142857, "step": 1485, "train/loss_ctc": 0.47797253727912903, "train/loss_error": 0.37816038727760315, "train/loss_total": 0.3981228470802307 }, { "epoch": 3.85974025974026, "step": 1486, "train/loss_ctc": 0.5813621878623962, "train/loss_error": 0.3506884276866913, "train/loss_total": 0.3968231976032257 }, { "epoch": 3.862337662337662, "step": 1487, "train/loss_ctc": 0.5496196746826172, "train/loss_error": 0.44364410638809204, "train/loss_total": 0.46483922004699707 }, { "epoch": 3.864935064935065, "step": 1488, "train/loss_ctc": 0.5808078050613403, "train/loss_error": 0.3599795699119568, "train/loss_total": 0.404145210981369 }, { "epoch": 3.8675324675324676, "step": 1489, "train/loss_ctc": 0.6398704051971436, "train/loss_error": 0.4422135651111603, "train/loss_total": 0.4817449450492859 }, { "epoch": 3.87012987012987, "grad_norm": 11231.513671875, "learning_rate": 6.81038961038961e-06, "loss": 0.4262, "step": 1490 }, { "epoch": 3.87012987012987, "step": 1490, "train/loss_ctc": 0.2935243546962738, "train/loss_error": 0.38904476165771484, "train/loss_total": 0.36994069814682007 }, { "epoch": 3.8727272727272726, "step": 1491, "train/loss_ctc": 0.7100521326065063, "train/loss_error": 0.497321218252182, "train/loss_total": 0.5398674011230469 }, { "epoch": 3.8753246753246753, "step": 1492, "train/loss_ctc": 0.3543411195278168, "train/loss_error": 0.29659634828567505, "train/loss_total": 0.30814528465270996 }, { "epoch": 3.877922077922078, "step": 1493, "train/loss_ctc": 0.35123926401138306, "train/loss_error": 0.43275320529937744, "train/loss_total": 0.4164504408836365 }, { "epoch": 3.8805194805194807, "step": 1494, "train/loss_ctc": 0.6395795941352844, "train/loss_error": 0.5007463693618774, "train/loss_total": 0.5285130143165588 }, { "epoch": 3.883116883116883, "step": 1495, "train/loss_ctc": 0.5788543224334717, "train/loss_error": 0.4813002943992615, "train/loss_total": 0.5008111000061035 }, { "epoch": 3.8857142857142857, "step": 1496, "train/loss_ctc": 0.4450996518135071, "train/loss_error": 0.3721869885921478, "train/loss_total": 0.38676953315734863 }, { "epoch": 3.8883116883116884, "step": 1497, "train/loss_ctc": 0.49926120042800903, "train/loss_error": 0.37830880284309387, "train/loss_total": 0.4024992883205414 }, { "epoch": 3.8909090909090907, "step": 1498, "train/loss_ctc": 0.8566545248031616, "train/loss_error": 0.500335693359375, "train/loss_total": 0.5715994834899902 }, { "epoch": 3.8935064935064934, "step": 1499, "train/loss_ctc": 0.3867664337158203, "train/loss_error": 0.4216660261154175, "train/loss_total": 0.4146861135959625 }, { "epoch": 3.896103896103896, "grad_norm": 9199.6572265625, "learning_rate": 6.654545454545455e-06, "loss": 0.4439, "step": 1500 }, { "epoch": 3.896103896103896, "step": 1500, "train/loss_ctc": 0.32998013496398926, "train/loss_error": 0.39471861720085144, "train/loss_total": 0.38177090883255005 }, { "epoch": 3.898701298701299, "step": 1501, "train/loss_ctc": 0.42260193824768066, "train/loss_error": 0.382793128490448, "train/loss_total": 0.39075490832328796 }, { "epoch": 3.9012987012987015, "step": 1502, "train/loss_ctc": 0.5001608729362488, "train/loss_error": 0.44733765721321106, "train/loss_total": 0.45790231227874756 }, { "epoch": 3.9038961038961038, "step": 1503, "train/loss_ctc": 0.4188244640827179, "train/loss_error": 0.4478384852409363, "train/loss_total": 0.4420356750488281 }, { "epoch": 3.9064935064935065, "step": 1504, "train/loss_ctc": 0.6335625648498535, "train/loss_error": 0.40799739956855774, "train/loss_total": 0.4531104564666748 }, { "epoch": 3.909090909090909, "step": 1505, "train/loss_ctc": 0.46383342146873474, "train/loss_error": 0.26981157064437866, "train/loss_total": 0.30861595273017883 }, { "epoch": 3.9116883116883114, "step": 1506, "train/loss_ctc": 0.49284762144088745, "train/loss_error": 0.23819245398044586, "train/loss_total": 0.2891234755516052 }, { "epoch": 3.914285714285714, "step": 1507, "train/loss_ctc": 0.425432950258255, "train/loss_error": 0.3591070771217346, "train/loss_total": 0.37237223982810974 }, { "epoch": 3.916883116883117, "step": 1508, "train/loss_ctc": 0.39870548248291016, "train/loss_error": 0.4102773070335388, "train/loss_total": 0.40796294808387756 }, { "epoch": 3.9194805194805196, "step": 1509, "train/loss_ctc": 0.6582106351852417, "train/loss_error": 0.4826628863811493, "train/loss_total": 0.5177724361419678 }, { "epoch": 3.9220779220779223, "grad_norm": 14761.3408203125, "learning_rate": 6.498701298701299e-06, "loss": 0.4021, "step": 1510 }, { "epoch": 3.9220779220779223, "step": 1510, "train/loss_ctc": 0.4112743139266968, "train/loss_error": 0.32479268312454224, "train/loss_total": 0.3420889973640442 }, { "epoch": 3.9246753246753245, "step": 1511, "train/loss_ctc": 0.3973589539527893, "train/loss_error": 0.4141470789909363, "train/loss_total": 0.41078945994377136 }, { "epoch": 3.9272727272727272, "step": 1512, "train/loss_ctc": 0.7635411620140076, "train/loss_error": 0.46641677618026733, "train/loss_total": 0.5258416533470154 }, { "epoch": 3.92987012987013, "step": 1513, "train/loss_ctc": 0.5567386150360107, "train/loss_error": 0.3974088132381439, "train/loss_total": 0.4292747974395752 }, { "epoch": 3.9324675324675322, "step": 1514, "train/loss_ctc": 0.2930745482444763, "train/loss_error": 0.30011415481567383, "train/loss_total": 0.2987062335014343 }, { "epoch": 3.935064935064935, "step": 1515, "train/loss_ctc": 0.4414609670639038, "train/loss_error": 0.42454543709754944, "train/loss_total": 0.4279285669326782 }, { "epoch": 3.9376623376623376, "step": 1516, "train/loss_ctc": 0.3780791163444519, "train/loss_error": 0.33023571968078613, "train/loss_total": 0.33980441093444824 }, { "epoch": 3.9402597402597404, "step": 1517, "train/loss_ctc": 0.4977274239063263, "train/loss_error": 0.4035957157611847, "train/loss_total": 0.42242205142974854 }, { "epoch": 3.942857142857143, "step": 1518, "train/loss_ctc": 0.6129897236824036, "train/loss_error": 0.45368143916130066, "train/loss_total": 0.4855431318283081 }, { "epoch": 3.9454545454545453, "step": 1519, "train/loss_ctc": 0.4518822729587555, "train/loss_error": 0.45044559240341187, "train/loss_total": 0.450732946395874 }, { "epoch": 3.948051948051948, "grad_norm": 10289.5517578125, "learning_rate": 6.342857142857144e-06, "loss": 0.4133, "step": 1520 }, { "epoch": 3.948051948051948, "step": 1520, "train/loss_ctc": 0.572801947593689, "train/loss_error": 0.44346287846565247, "train/loss_total": 0.46933069825172424 }, { "epoch": 3.9506493506493507, "step": 1521, "train/loss_ctc": 0.3563143014907837, "train/loss_error": 0.43155738711357117, "train/loss_total": 0.4165087938308716 }, { "epoch": 3.953246753246753, "step": 1522, "train/loss_ctc": 0.5450352430343628, "train/loss_error": 0.4770386517047882, "train/loss_total": 0.49063798785209656 }, { "epoch": 3.9558441558441557, "step": 1523, "train/loss_ctc": 0.5573639869689941, "train/loss_error": 0.451957643032074, "train/loss_total": 0.473038911819458 }, { "epoch": 3.9584415584415584, "step": 1524, "train/loss_ctc": 0.4832117557525635, "train/loss_error": 0.4072534441947937, "train/loss_total": 0.4224451184272766 }, { "epoch": 3.961038961038961, "step": 1525, "train/loss_ctc": 0.4205090403556824, "train/loss_error": 0.42762401700019836, "train/loss_total": 0.4262010455131531 }, { "epoch": 3.963636363636364, "step": 1526, "train/loss_ctc": 0.31964755058288574, "train/loss_error": 0.3959290385246277, "train/loss_total": 0.38067275285720825 }, { "epoch": 3.966233766233766, "step": 1527, "train/loss_ctc": 0.5690191388130188, "train/loss_error": 0.455673485994339, "train/loss_total": 0.47834262251853943 }, { "epoch": 3.968831168831169, "step": 1528, "train/loss_ctc": 0.48689353466033936, "train/loss_error": 0.37712734937667847, "train/loss_total": 0.3990805745124817 }, { "epoch": 3.9714285714285715, "step": 1529, "train/loss_ctc": 0.3908514976501465, "train/loss_error": 0.45667821168899536, "train/loss_total": 0.443512886762619 }, { "epoch": 3.974025974025974, "grad_norm": 9723.720703125, "learning_rate": 6.1870129870129876e-06, "loss": 0.44, "step": 1530 }, { "epoch": 3.974025974025974, "step": 1530, "train/loss_ctc": 0.6744332313537598, "train/loss_error": 0.430682510137558, "train/loss_total": 0.4794326722621918 }, { "epoch": 3.9766233766233765, "step": 1531, "train/loss_ctc": 0.5600382685661316, "train/loss_error": 0.4492310881614685, "train/loss_total": 0.47139251232147217 }, { "epoch": 3.979220779220779, "step": 1532, "train/loss_ctc": 0.30666473507881165, "train/loss_error": 0.364956796169281, "train/loss_total": 0.3532983958721161 }, { "epoch": 3.981818181818182, "step": 1533, "train/loss_ctc": 0.5514463186264038, "train/loss_error": 0.5221889019012451, "train/loss_total": 0.5280404090881348 }, { "epoch": 3.9844155844155846, "step": 1534, "train/loss_ctc": 0.6900286674499512, "train/loss_error": 0.4526161551475525, "train/loss_total": 0.500098705291748 }, { "epoch": 3.987012987012987, "step": 1535, "train/loss_ctc": 0.42351359128952026, "train/loss_error": 0.35769641399383545, "train/loss_total": 0.37085986137390137 }, { "epoch": 3.9896103896103896, "step": 1536, "train/loss_ctc": 0.46932822465896606, "train/loss_error": 0.4409582316875458, "train/loss_total": 0.4466322660446167 }, { "epoch": 3.9922077922077923, "step": 1537, "train/loss_ctc": 0.5213923454284668, "train/loss_error": 0.4886096119880676, "train/loss_total": 0.49516618251800537 }, { "epoch": 3.9948051948051946, "step": 1538, "train/loss_ctc": 0.4128141403198242, "train/loss_error": 0.3992675244808197, "train/loss_total": 0.4019768536090851 }, { "epoch": 3.9974025974025973, "step": 1539, "train/loss_ctc": 0.5601736307144165, "train/loss_error": 0.34997522830963135, "train/loss_total": 0.39201492071151733 }, { "epoch": 4.0, "grad_norm": 26715.30078125, "learning_rate": 6.0311688311688315e-06, "loss": 0.4439, "step": 1540 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.6617535352706909, "val/loss_error": 0.16989634931087494, "val/loss_total": 0.26826781034469604 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.728701114654541, "val/loss_error": 0.5203905701637268, "val/loss_total": 0.5620526671409607 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.11360479891300201, "val/loss_error": 0.31346747279167175, "val/loss_total": 0.2734949588775635 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.2514151930809021, "val/loss_error": 0.17457735538482666, "val/loss_total": 0.18994493782520294 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.14210864901542664, "val/loss_error": 0.20571152865886688, "val/loss_total": 0.1929909586906433 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.15178018808364868, "val/loss_error": 0.3483298122882843, "val/loss_total": 0.30901989340782166 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.25142228603363037, "val/loss_error": 0.4727407395839691, "val/loss_total": 0.42847704887390137 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.3079298436641693, "val/loss_error": 0.31410062313079834, "val/loss_total": 0.3128664791584015 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.4399119019508362, "val/loss_error": 0.337258517742157, "val/loss_total": 0.35778921842575073 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.22461305558681488, "val/loss_error": 0.4047263562679291, "val/loss_total": 0.3687037229537964 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.10807893425226212, "val/loss_error": 0.21439780294895172, "val/loss_total": 0.19313403964042664 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.2788211405277252, "val/loss_error": 0.3565574884414673, "val/loss_total": 0.3410102128982544 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.7789663076400757, "val/loss_error": 0.5235947966575623, "val/loss_total": 0.5746691226959229 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.861347496509552, "val/loss_error": 0.5771816372871399, "val/loss_total": 0.6340148448944092 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.6487600803375244, "val/loss_error": 0.3974345028400421, "val/loss_total": 0.4476996064186096 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5977194309234619, "val/loss_error": 0.7737212777137756, "val/loss_total": 0.7385209202766418 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.10736650973558426, "val/loss_error": 0.2259344607591629, "val/loss_total": 0.2022208720445633 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.13795901834964752, "val/loss_error": 0.25657233595848083, "val/loss_total": 0.23284968733787537 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.13741354644298553, "val/loss_error": 0.3523638844490051, "val/loss_total": 0.3093738257884979 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5045650601387024, "val/loss_error": 0.42334964871406555, "val/loss_total": 0.43959274888038635 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.8175076842308044, "val/loss_error": 0.6448903679847717, "val/loss_total": 0.6794138550758362 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.300775408744812, "val/loss_error": 0.6094077825546265, "val/loss_total": 0.5476813316345215 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.32990729808807373, "val/loss_error": 0.4366808533668518, "val/loss_total": 0.41532617807388306 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.15740440785884857, "val/loss_error": 0.20769312977790833, "val/loss_total": 0.19763538241386414 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.37323302030563354, "val/loss_error": 0.5226756930351257, "val/loss_total": 0.4927871823310852 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5465680956840515, "val/loss_error": 0.5488367080688477, "val/loss_total": 0.5483829975128174 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.31341129541397095, "val/loss_error": 0.42610886693000793, "val/loss_total": 0.40356937050819397 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.150761678814888, "val/loss_error": 0.27393725514411926, "val/loss_total": 0.24930214881896973 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.03695406764745712, "val/loss_error": 0.1086018905043602, "val/loss_total": 0.09427232295274734 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.26823049783706665, "val/loss_error": 0.5265489220619202, "val/loss_total": 0.4748852252960205 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.13018304109573364, "val/loss_error": 0.15401695668697357, "val/loss_total": 0.14925017952919006 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.10223007947206497, "val/loss_error": 0.29738447070121765, "val/loss_total": 0.258353590965271 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.08936381340026855, "val/loss_error": 0.21162228286266327, "val/loss_total": 0.1871705949306488 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5371357202529907, "val/loss_error": 0.7444442510604858, "val/loss_total": 0.7029825448989868 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.09369838982820511, "val/loss_error": 0.37970998883247375, "val/loss_total": 0.32250767946243286 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.17889684438705444, "val/loss_error": 0.5269924998283386, "val/loss_total": 0.45737335085868835 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5139767527580261, "val/loss_error": 0.5937736630439758, "val/loss_total": 0.5778142809867859 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5091070532798767, "val/loss_error": 0.46460092067718506, "val/loss_total": 0.47350215911865234 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.3064963221549988, "val/loss_error": 0.5008012652397156, "val/loss_total": 0.46194028854370117 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.6507424712181091, "val/loss_error": 0.6463766694068909, "val/loss_total": 0.6472498178482056 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.2994008958339691, "val/loss_error": 0.2836393117904663, "val/loss_total": 0.2867916226387024 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.35666516423225403, "val/loss_error": 0.41541433334350586, "val/loss_total": 0.4036645293235779 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.4809524714946747, "val/loss_error": 0.923035204410553, "val/loss_total": 0.8346186876296997 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5202158689498901, "val/loss_error": 0.5686102509498596, "val/loss_total": 0.5589313507080078 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.4443039894104004, "val/loss_error": 0.2885812520980835, "val/loss_total": 0.31972581148147583 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.3730580508708954, "val/loss_error": 0.6476048231124878, "val/loss_total": 0.5926954746246338 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5763376355171204, "val/loss_error": 0.4377492368221283, "val/loss_total": 0.4654669165611267 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.3788464069366455, "val/loss_error": 0.5326692461967468, "val/loss_total": 0.5019046664237976 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.467325896024704, "val/loss_error": 0.24621173739433289, "val/loss_total": 0.2904345691204071 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.6445432901382446, "val/loss_error": 0.8567513823509216, "val/loss_total": 0.81430983543396 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.39789605140686035, "val/loss_error": 0.8898370862007141, "val/loss_total": 0.7914488911628723 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.3593446910381317, "val/loss_error": 0.6718379259109497, "val/loss_total": 0.6093392968177795 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.40874215960502625, "val/loss_error": 0.461180716753006, "val/loss_total": 0.4506930112838745 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5181517004966736, "val/loss_error": 0.7669851183891296, "val/loss_total": 0.7172184586524963 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.21528922021389008, "val/loss_error": 0.3790895938873291, "val/loss_total": 0.3463295102119446 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.18945302069187164, "val/loss_error": 0.49214398860931396, "val/loss_total": 0.43160581588745117 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.10099285840988159, "val/loss_error": 0.28537607192993164, "val/loss_total": 0.24849942326545715 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5843119025230408, "val/loss_error": 0.2994202673435211, "val/loss_total": 0.3563985824584961 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.6918607354164124, "val/loss_error": 0.749823272228241, "val/loss_total": 0.73823082447052 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.028993625193834305, "val/loss_error": 0.14553886651992798, "val/loss_total": 0.12222982197999954 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.7680637836456299, "val/loss_error": 0.6475275158882141, "val/loss_total": 0.6716347932815552 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.6182758808135986, "val/loss_error": 0.4241107106208801, "val/loss_total": 0.4629437327384949 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.07980868220329285, "val/loss_error": 0.09977161139249802, "val/loss_total": 0.09577902406454086 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.1811370849609375, "val/loss_error": 0.3383423984050751, "val/loss_total": 0.30690133571624756 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.15370319783687592, "val/loss_error": 0.215968057513237, "val/loss_total": 0.20351508259773254 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.2021399885416031, "val/loss_error": 0.4220103919506073, "val/loss_total": 0.3780363202095032 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.3132476508617401, "val/loss_error": 0.2874320149421692, "val/loss_total": 0.29259514808654785 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.6389130353927612, "val/loss_error": 0.5433400273323059, "val/loss_total": 0.5624546408653259 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.48333314061164856, "val/loss_error": 0.3835921287536621, "val/loss_total": 0.40354034304618835 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.45895469188690186, "val/loss_error": 0.4433237612247467, "val/loss_total": 0.44644996523857117 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.2539585828781128, "val/loss_error": 0.26099011301994324, "val/loss_total": 0.25958380103111267 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5580675005912781, "val/loss_error": 0.5163927674293518, "val/loss_total": 0.5247277021408081 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.6225395798683167, "val/loss_error": 0.715599775314331, "val/loss_total": 0.6969877481460571 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.26466992497444153, "val/loss_error": 0.38213858008384705, "val/loss_total": 0.35864487290382385 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.22129249572753906, "val/loss_error": 0.30307045578956604, "val/loss_total": 0.2867148518562317 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.05767582729458809, "val/loss_error": 0.34325796365737915, "val/loss_total": 0.2861415445804596 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.7981420755386353, "val/loss_error": 0.5394331812858582, "val/loss_total": 0.5911749601364136 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5135515928268433, "val/loss_error": 0.4057385325431824, "val/loss_total": 0.42730116844177246 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.6202462315559387, "val/loss_error": 0.355093777179718, "val/loss_total": 0.40812426805496216 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.002891520969569683, "val/loss_error": 0.3968760371208191, "val/loss_total": 0.3180791437625885 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.4503142833709717, "val/loss_error": 0.2887035012245178, "val/loss_total": 0.32102566957473755 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.023583410307765007, "val/loss_error": 0.1322137713432312, "val/loss_total": 0.11048769950866699 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.24838019907474518, "val/loss_error": 0.2989429235458374, "val/loss_total": 0.28883036971092224 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.3718937635421753, "val/loss_error": 0.43577879667282104, "val/loss_total": 0.42300179600715637 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.4976947605609894, "val/loss_error": 0.6130363941192627, "val/loss_total": 0.5899680852890015 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5241696834564209, "val/loss_error": 0.7146698236465454, "val/loss_total": 0.6765698194503784 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.04895176365971565, "val/loss_error": 0.14816626906394958, "val/loss_total": 0.12832336127758026 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5202515721321106, "val/loss_error": 0.566058874130249, "val/loss_total": 0.5568974018096924 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.23264066874980927, "val/loss_error": 0.29135820269584656, "val/loss_total": 0.27961471676826477 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.40376585721969604, "val/loss_error": 0.3744681775569916, "val/loss_total": 0.3803277313709259 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.2882551848888397, "val/loss_error": 0.5340330004692078, "val/loss_total": 0.48487743735313416 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.8112992644309998, "val/loss_error": 0.4655640721321106, "val/loss_total": 0.5347111225128174 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.23670744895935059, "val/loss_error": 0.11240798234939575, "val/loss_total": 0.13726787269115448 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.3064383268356323, "val/loss_error": 0.3146789073944092, "val/loss_total": 0.31303080916404724 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5443935394287109, "val/loss_error": 0.5496771335601807, "val/loss_total": 0.5486204028129578 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.2250211536884308, "val/loss_error": 0.1999150812625885, "val/loss_total": 0.20493629574775696 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.10269328951835632, "val/loss_error": 0.21716509759426117, "val/loss_total": 0.19427074491977692 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.5945166349411011, "val/loss_error": 0.48969778418540955, "val/loss_total": 0.5106615424156189 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.12880685925483704, "val/loss_error": 0.32451310753822327, "val/loss_total": 0.285371869802475 }, { "epoch": 4.0, "step": 1540, "val/loss_ctc": 0.19679027795791626, "val/loss_error": 0.2668781876564026, "val/loss_total": 0.2528606057167053 }, { "epoch": 4.0, "eval_eval/f1_0": 0.5210466384887695, "eval_eval/f1_1": 0.9246464967727661, "eval_eval/precision_0": 0.5192743539810181, "eval_eval/precision_1": 0.9251432418823242, "eval_eval/recall_0": 0.5228310227394104, "eval_eval/recall_1": 0.924150288105011, "eval_eval/wer": 0.15065286699802058, "eval_loss": 0.4080162048339844, "eval_runtime": 7.5365, "eval_samples_per_second": 13.269, "eval_steps_per_second": 13.269, "step": 1540 }, { "epoch": 4.0, "step": 1540, "train/loss_ctc": 0.44728848338127136, "train/loss_error": 0.2837469279766083, "train/loss_total": 0.31645524501800537 }, { "epoch": 4.002597402597402, "step": 1541, "train/loss_ctc": 0.33993011713027954, "train/loss_error": 0.3521200120449066, "train/loss_total": 0.3496820330619812 }, { "epoch": 4.005194805194805, "step": 1542, "train/loss_ctc": 0.5578049421310425, "train/loss_error": 0.42556700110435486, "train/loss_total": 0.45201459527015686 }, { "epoch": 4.007792207792208, "step": 1543, "train/loss_ctc": 0.3568418025970459, "train/loss_error": 0.3568305969238281, "train/loss_total": 0.3568328619003296 }, { "epoch": 4.01038961038961, "step": 1544, "train/loss_ctc": 0.4682139754295349, "train/loss_error": 0.4272368252277374, "train/loss_total": 0.4354322552680969 }, { "epoch": 4.012987012987013, "step": 1545, "train/loss_ctc": 0.43908804655075073, "train/loss_error": 0.3775036633014679, "train/loss_total": 0.38982054591178894 }, { "epoch": 4.015584415584415, "step": 1546, "train/loss_ctc": 0.43925487995147705, "train/loss_error": 0.3995489180088043, "train/loss_total": 0.4074901342391968 }, { "epoch": 4.0181818181818185, "step": 1547, "train/loss_ctc": 0.5202529430389404, "train/loss_error": 0.3462075889110565, "train/loss_total": 0.38101667165756226 }, { "epoch": 4.020779220779221, "step": 1548, "train/loss_ctc": 0.7903116941452026, "train/loss_error": 0.45761147141456604, "train/loss_total": 0.5241515636444092 }, { "epoch": 4.023376623376623, "step": 1549, "train/loss_ctc": 0.4383309483528137, "train/loss_error": 0.45943158864974976, "train/loss_total": 0.45521146059036255 }, { "epoch": 4.025974025974026, "grad_norm": 11191.2705078125, "learning_rate": 5.8753246753246755e-06, "loss": 0.4068, "step": 1550 }, { "epoch": 4.025974025974026, "step": 1550, "train/loss_ctc": 0.48474645614624023, "train/loss_error": 0.3601175844669342, "train/loss_total": 0.3850433826446533 }, { "epoch": 4.0285714285714285, "step": 1551, "train/loss_ctc": 0.3908858895301819, "train/loss_error": 0.3725205659866333, "train/loss_total": 0.376193642616272 }, { "epoch": 4.031168831168831, "step": 1552, "train/loss_ctc": 0.4636493921279907, "train/loss_error": 0.3609965741634369, "train/loss_total": 0.3815271258354187 }, { "epoch": 4.033766233766234, "step": 1553, "train/loss_ctc": 0.3941332697868347, "train/loss_error": 0.3800765872001648, "train/loss_total": 0.38288792967796326 }, { "epoch": 4.036363636363636, "step": 1554, "train/loss_ctc": 0.3735249638557434, "train/loss_error": 0.36883053183555603, "train/loss_total": 0.369769424200058 }, { "epoch": 4.038961038961039, "step": 1555, "train/loss_ctc": 0.4460533857345581, "train/loss_error": 0.34056663513183594, "train/loss_total": 0.3616639971733093 }, { "epoch": 4.041558441558442, "step": 1556, "train/loss_ctc": 0.361951619386673, "train/loss_error": 0.4181339144706726, "train/loss_total": 0.4068974554538727 }, { "epoch": 4.044155844155844, "step": 1557, "train/loss_ctc": 0.5610405206680298, "train/loss_error": 0.44677823781967163, "train/loss_total": 0.4696306884288788 }, { "epoch": 4.046753246753247, "step": 1558, "train/loss_ctc": 0.3876192569732666, "train/loss_error": 0.3447537124156952, "train/loss_total": 0.35332682728767395 }, { "epoch": 4.049350649350649, "step": 1559, "train/loss_ctc": 0.5319012403488159, "train/loss_error": 0.34266600012779236, "train/loss_total": 0.380513072013855 }, { "epoch": 4.0519480519480515, "grad_norm": 13894.2021484375, "learning_rate": 5.7194805194805195e-06, "loss": 0.3867, "step": 1560 }, { "epoch": 4.0519480519480515, "step": 1560, "train/loss_ctc": 0.6479262709617615, "train/loss_error": 0.38441526889801025, "train/loss_total": 0.43711745738983154 }, { "epoch": 4.054545454545455, "step": 1561, "train/loss_ctc": 0.5157873034477234, "train/loss_error": 0.4672612249851227, "train/loss_total": 0.4769664406776428 }, { "epoch": 4.057142857142857, "step": 1562, "train/loss_ctc": 0.5684306025505066, "train/loss_error": 0.48560914397239685, "train/loss_total": 0.5021734237670898 }, { "epoch": 4.05974025974026, "step": 1563, "train/loss_ctc": 0.493461012840271, "train/loss_error": 0.4287630319595337, "train/loss_total": 0.44170263409614563 }, { "epoch": 4.062337662337662, "step": 1564, "train/loss_ctc": 0.4504014253616333, "train/loss_error": 0.33643248677253723, "train/loss_total": 0.3592262864112854 }, { "epoch": 4.064935064935065, "step": 1565, "train/loss_ctc": 0.4581126272678375, "train/loss_error": 0.4554077386856079, "train/loss_total": 0.45594874024391174 }, { "epoch": 4.067532467532468, "step": 1566, "train/loss_ctc": 0.4601333737373352, "train/loss_error": 0.3614419400691986, "train/loss_total": 0.3811802268028259 }, { "epoch": 4.07012987012987, "step": 1567, "train/loss_ctc": 0.5407822728157043, "train/loss_error": 0.43608999252319336, "train/loss_total": 0.45702844858169556 }, { "epoch": 4.072727272727272, "step": 1568, "train/loss_ctc": 0.5015274286270142, "train/loss_error": 0.3592326045036316, "train/loss_total": 0.38769158720970154 }, { "epoch": 4.0753246753246755, "step": 1569, "train/loss_ctc": 0.4069700539112091, "train/loss_error": 0.3925088047981262, "train/loss_total": 0.3954010605812073 }, { "epoch": 4.077922077922078, "grad_norm": 10221.6064453125, "learning_rate": 5.563636363636363e-06, "loss": 0.4294, "step": 1570 }, { "epoch": 4.077922077922078, "step": 1570, "train/loss_ctc": 0.48584502935409546, "train/loss_error": 0.38152995705604553, "train/loss_total": 0.4023929834365845 }, { "epoch": 4.080519480519481, "step": 1571, "train/loss_ctc": 0.3485802412033081, "train/loss_error": 0.3694388270378113, "train/loss_total": 0.3652670979499817 }, { "epoch": 4.083116883116883, "step": 1572, "train/loss_ctc": 0.5651668310165405, "train/loss_error": 0.47883740067481995, "train/loss_total": 0.49610328674316406 }, { "epoch": 4.085714285714285, "step": 1573, "train/loss_ctc": 0.46150147914886475, "train/loss_error": 0.48256394267082214, "train/loss_total": 0.4783514440059662 }, { "epoch": 4.088311688311689, "step": 1574, "train/loss_ctc": 0.5440318584442139, "train/loss_error": 0.4915340840816498, "train/loss_total": 0.5020336508750916 }, { "epoch": 4.090909090909091, "step": 1575, "train/loss_ctc": 0.4876243770122528, "train/loss_error": 0.42722564935684204, "train/loss_total": 0.4393053948879242 }, { "epoch": 4.093506493506493, "step": 1576, "train/loss_ctc": 0.5023910403251648, "train/loss_error": 0.4454450011253357, "train/loss_total": 0.45683419704437256 }, { "epoch": 4.096103896103896, "step": 1577, "train/loss_ctc": 0.4737335443496704, "train/loss_error": 0.4416500926017761, "train/loss_total": 0.4480668008327484 }, { "epoch": 4.0987012987012985, "step": 1578, "train/loss_ctc": 0.3735518753528595, "train/loss_error": 0.4272896945476532, "train/loss_total": 0.4165421426296234 }, { "epoch": 4.101298701298702, "step": 1579, "train/loss_ctc": 0.40125492215156555, "train/loss_error": 0.40831443667411804, "train/loss_total": 0.4069025218486786 }, { "epoch": 4.103896103896104, "grad_norm": 9379.501953125, "learning_rate": 5.407792207792208e-06, "loss": 0.4412, "step": 1580 }, { "epoch": 4.103896103896104, "step": 1580, "train/loss_ctc": 0.43084612488746643, "train/loss_error": 0.36669936776161194, "train/loss_total": 0.3795287013053894 }, { "epoch": 4.106493506493506, "step": 1581, "train/loss_ctc": 0.3975585103034973, "train/loss_error": 0.36577895283699036, "train/loss_total": 0.37213486433029175 }, { "epoch": 4.109090909090909, "step": 1582, "train/loss_ctc": 0.39290276169776917, "train/loss_error": 0.3306809663772583, "train/loss_total": 0.3431253433227539 }, { "epoch": 4.111688311688312, "step": 1583, "train/loss_ctc": 0.5077006220817566, "train/loss_error": 0.46390801668167114, "train/loss_total": 0.47266653180122375 }, { "epoch": 4.114285714285714, "step": 1584, "train/loss_ctc": 0.7144367694854736, "train/loss_error": 0.41789722442626953, "train/loss_total": 0.4772051274776459 }, { "epoch": 4.116883116883117, "step": 1585, "train/loss_ctc": 0.49341708421707153, "train/loss_error": 0.43007761240005493, "train/loss_total": 0.44274550676345825 }, { "epoch": 4.119480519480519, "step": 1586, "train/loss_ctc": 0.2514399588108063, "train/loss_error": 0.2877013087272644, "train/loss_total": 0.2804490327835083 }, { "epoch": 4.1220779220779225, "step": 1587, "train/loss_ctc": 0.5884899497032166, "train/loss_error": 0.4623771905899048, "train/loss_total": 0.4875997304916382 }, { "epoch": 4.124675324675325, "step": 1588, "train/loss_ctc": 0.6157239079475403, "train/loss_error": 0.39911890029907227, "train/loss_total": 0.4424399137496948 }, { "epoch": 4.127272727272727, "step": 1589, "train/loss_ctc": 0.3998509347438812, "train/loss_error": 0.37147119641304016, "train/loss_total": 0.3771471381187439 }, { "epoch": 4.12987012987013, "grad_norm": 12735.2802734375, "learning_rate": 5.251948051948052e-06, "loss": 0.4075, "step": 1590 }, { "epoch": 4.12987012987013, "step": 1590, "train/loss_ctc": 0.46051138639450073, "train/loss_error": 0.39918115735054016, "train/loss_total": 0.4114472270011902 }, { "epoch": 4.132467532467532, "step": 1591, "train/loss_ctc": 0.42278534173965454, "train/loss_error": 0.436718225479126, "train/loss_total": 0.4339316487312317 }, { "epoch": 4.135064935064935, "step": 1592, "train/loss_ctc": 0.4530753195285797, "train/loss_error": 0.46379774808883667, "train/loss_total": 0.4616532623767853 }, { "epoch": 4.137662337662338, "step": 1593, "train/loss_ctc": 0.40551578998565674, "train/loss_error": 0.3236544132232666, "train/loss_total": 0.3400266766548157 }, { "epoch": 4.14025974025974, "step": 1594, "train/loss_ctc": 0.35734593868255615, "train/loss_error": 0.439620703458786, "train/loss_total": 0.42316576838493347 }, { "epoch": 4.142857142857143, "step": 1595, "train/loss_ctc": 0.5600236058235168, "train/loss_error": 0.37465324997901917, "train/loss_total": 0.41172733902931213 }, { "epoch": 4.1454545454545455, "step": 1596, "train/loss_ctc": 0.4336416721343994, "train/loss_error": 0.42611825466156006, "train/loss_total": 0.4276229441165924 }, { "epoch": 4.148051948051948, "step": 1597, "train/loss_ctc": 0.5018081665039062, "train/loss_error": 0.4369862377643585, "train/loss_total": 0.449950635433197 }, { "epoch": 4.150649350649351, "step": 1598, "train/loss_ctc": 0.523415207862854, "train/loss_error": 0.41863375902175903, "train/loss_total": 0.43959006667137146 }, { "epoch": 4.153246753246753, "step": 1599, "train/loss_ctc": 0.4025014638900757, "train/loss_error": 0.4370766878128052, "train/loss_total": 0.43016165494918823 }, { "epoch": 4.1558441558441555, "grad_norm": 11996.931640625, "learning_rate": 5.096103896103896e-06, "loss": 0.4229, "step": 1600 }, { "epoch": 4.1558441558441555, "step": 1600, "train/loss_ctc": 0.4182925224304199, "train/loss_error": 0.4002408981323242, "train/loss_total": 0.4038512408733368 }, { "epoch": 4.158441558441559, "step": 1601, "train/loss_ctc": 0.2615950107574463, "train/loss_error": 0.3249712586402893, "train/loss_total": 0.3122960031032562 }, { "epoch": 4.161038961038961, "step": 1602, "train/loss_ctc": 0.3944246172904968, "train/loss_error": 0.3401116728782654, "train/loss_total": 0.35097426176071167 }, { "epoch": 4.163636363636364, "step": 1603, "train/loss_ctc": 0.27969756722450256, "train/loss_error": 0.3691646456718445, "train/loss_total": 0.35127124190330505 }, { "epoch": 4.166233766233766, "step": 1604, "train/loss_ctc": 0.670106053352356, "train/loss_error": 0.5200679302215576, "train/loss_total": 0.5500755310058594 }, { "epoch": 4.1688311688311686, "step": 1605, "train/loss_ctc": 0.7936950325965881, "train/loss_error": 0.4646320641040802, "train/loss_total": 0.5304446816444397 }, { "epoch": 4.171428571428572, "step": 1606, "train/loss_ctc": 0.5702555179595947, "train/loss_error": 0.437995582818985, "train/loss_total": 0.46444758772850037 }, { "epoch": 4.174025974025974, "step": 1607, "train/loss_ctc": 0.4498525857925415, "train/loss_error": 0.38979214429855347, "train/loss_total": 0.40180423855781555 }, { "epoch": 4.176623376623376, "step": 1608, "train/loss_ctc": 0.6648669838905334, "train/loss_error": 0.40260517597198486, "train/loss_total": 0.4550575613975525 }, { "epoch": 4.179220779220779, "step": 1609, "train/loss_ctc": 0.47706130146980286, "train/loss_error": 0.4777873754501343, "train/loss_total": 0.47764214873313904 }, { "epoch": 4.181818181818182, "grad_norm": 10859.3115234375, "learning_rate": 4.94025974025974e-06, "loss": 0.4298, "step": 1610 }, { "epoch": 4.181818181818182, "step": 1610, "train/loss_ctc": 0.46089085936546326, "train/loss_error": 0.3787146806716919, "train/loss_total": 0.39514991641044617 }, { "epoch": 4.184415584415585, "step": 1611, "train/loss_ctc": 0.43337777256965637, "train/loss_error": 0.3337256908416748, "train/loss_total": 0.3536561131477356 }, { "epoch": 4.187012987012987, "step": 1612, "train/loss_ctc": 0.8412022590637207, "train/loss_error": 0.4106912314891815, "train/loss_total": 0.4967934489250183 }, { "epoch": 4.189610389610389, "step": 1613, "train/loss_ctc": 0.308860719203949, "train/loss_error": 0.44591182470321655, "train/loss_total": 0.418501615524292 }, { "epoch": 4.1922077922077925, "step": 1614, "train/loss_ctc": 0.4158588647842407, "train/loss_error": 0.42332571744918823, "train/loss_total": 0.4218323528766632 }, { "epoch": 4.194805194805195, "step": 1615, "train/loss_ctc": 0.5158162117004395, "train/loss_error": 0.4596414864063263, "train/loss_total": 0.47087642550468445 }, { "epoch": 4.197402597402597, "step": 1616, "train/loss_ctc": 0.6086522340774536, "train/loss_error": 0.48550671339035034, "train/loss_total": 0.5101358294487 }, { "epoch": 4.2, "step": 1617, "train/loss_ctc": 0.5287953019142151, "train/loss_error": 0.37548917531967163, "train/loss_total": 0.4061504006385803 }, { "epoch": 4.202597402597402, "step": 1618, "train/loss_ctc": 0.5239232182502747, "train/loss_error": 0.3801216185092926, "train/loss_total": 0.40888193249702454 }, { "epoch": 4.205194805194806, "step": 1619, "train/loss_ctc": 0.5638959407806396, "train/loss_error": 0.4515092968940735, "train/loss_total": 0.4739866554737091 }, { "epoch": 4.207792207792208, "grad_norm": 12409.9716796875, "learning_rate": 4.784415584415584e-06, "loss": 0.4356, "step": 1620 }, { "epoch": 4.207792207792208, "step": 1620, "train/loss_ctc": 0.37979698181152344, "train/loss_error": 0.37259534001350403, "train/loss_total": 0.37403565645217896 }, { "epoch": 4.21038961038961, "step": 1621, "train/loss_ctc": 0.42325448989868164, "train/loss_error": 0.4010305106639862, "train/loss_total": 0.40547531843185425 }, { "epoch": 4.212987012987013, "step": 1622, "train/loss_ctc": 0.3292677402496338, "train/loss_error": 0.4404807984828949, "train/loss_total": 0.41823822259902954 }, { "epoch": 4.2155844155844155, "step": 1623, "train/loss_ctc": 0.7017526030540466, "train/loss_error": 0.47742581367492676, "train/loss_total": 0.5222911834716797 }, { "epoch": 4.218181818181818, "step": 1624, "train/loss_ctc": 1.149817705154419, "train/loss_error": 0.40308552980422974, "train/loss_total": 0.5524319410324097 }, { "epoch": 4.220779220779221, "step": 1625, "train/loss_ctc": 0.6307495832443237, "train/loss_error": 0.41504135727882385, "train/loss_total": 0.45818302035331726 }, { "epoch": 4.223376623376623, "step": 1626, "train/loss_ctc": 0.32428622245788574, "train/loss_error": 0.3071080148220062, "train/loss_total": 0.31054365634918213 }, { "epoch": 4.225974025974026, "step": 1627, "train/loss_ctc": 0.3088032007217407, "train/loss_error": 0.32718053460121155, "train/loss_total": 0.32350507378578186 }, { "epoch": 4.228571428571429, "step": 1628, "train/loss_ctc": 0.5026681423187256, "train/loss_error": 0.37638935446739197, "train/loss_total": 0.40164512395858765 }, { "epoch": 4.231168831168831, "step": 1629, "train/loss_ctc": 0.5197674036026001, "train/loss_error": 0.3120137155056, "train/loss_total": 0.35356444120407104 }, { "epoch": 4.233766233766234, "grad_norm": 16095.087890625, "learning_rate": 4.628571428571428e-06, "loss": 0.412, "step": 1630 }, { "epoch": 4.233766233766234, "step": 1630, "train/loss_ctc": 0.36405235528945923, "train/loss_error": 0.3438868224620819, "train/loss_total": 0.34791994094848633 }, { "epoch": 4.236363636363636, "step": 1631, "train/loss_ctc": 0.3384685516357422, "train/loss_error": 0.3303252160549164, "train/loss_total": 0.33195388317108154 }, { "epoch": 4.238961038961039, "step": 1632, "train/loss_ctc": 0.6216474771499634, "train/loss_error": 0.368144154548645, "train/loss_total": 0.4188448488712311 }, { "epoch": 4.241558441558442, "step": 1633, "train/loss_ctc": 0.3957550525665283, "train/loss_error": 0.4315589666366577, "train/loss_total": 0.42439818382263184 }, { "epoch": 4.244155844155844, "step": 1634, "train/loss_ctc": 0.3257419168949127, "train/loss_error": 0.31330662965774536, "train/loss_total": 0.3157936930656433 }, { "epoch": 4.246753246753247, "step": 1635, "train/loss_ctc": 0.5598452687263489, "train/loss_error": 0.41056543588638306, "train/loss_total": 0.4404214024543762 }, { "epoch": 4.249350649350649, "step": 1636, "train/loss_ctc": 0.4490984082221985, "train/loss_error": 0.42322590947151184, "train/loss_total": 0.4284003973007202 }, { "epoch": 4.251948051948052, "step": 1637, "train/loss_ctc": 0.4497813582420349, "train/loss_error": 0.3666553497314453, "train/loss_total": 0.38328057527542114 }, { "epoch": 4.254545454545455, "step": 1638, "train/loss_ctc": 0.6072348356246948, "train/loss_error": 0.4662221372127533, "train/loss_total": 0.4944246709346771 }, { "epoch": 4.257142857142857, "step": 1639, "train/loss_ctc": 0.32474473118782043, "train/loss_error": 0.3173506259918213, "train/loss_total": 0.3188294470310211 }, { "epoch": 4.259740259740259, "grad_norm": 10105.94921875, "learning_rate": 4.472727272727273e-06, "loss": 0.3904, "step": 1640 }, { "epoch": 4.259740259740259, "step": 1640, "train/loss_ctc": 0.4680470824241638, "train/loss_error": 0.4192682206630707, "train/loss_total": 0.42902401089668274 }, { "epoch": 4.2623376623376625, "step": 1641, "train/loss_ctc": 0.39283865690231323, "train/loss_error": 0.5256423950195312, "train/loss_total": 0.49908167123794556 }, { "epoch": 4.264935064935065, "step": 1642, "train/loss_ctc": 0.4835870563983917, "train/loss_error": 0.3875761926174164, "train/loss_total": 0.40677836537361145 }, { "epoch": 4.267532467532467, "step": 1643, "train/loss_ctc": 0.2942894697189331, "train/loss_error": 0.4287721514701843, "train/loss_total": 0.4018756151199341 }, { "epoch": 4.27012987012987, "step": 1644, "train/loss_ctc": 0.49964627623558044, "train/loss_error": 0.4224177598953247, "train/loss_total": 0.43786346912384033 }, { "epoch": 4.2727272727272725, "step": 1645, "train/loss_ctc": 0.6860910058021545, "train/loss_error": 0.46901416778564453, "train/loss_total": 0.5124295353889465 }, { "epoch": 4.275324675324676, "step": 1646, "train/loss_ctc": 0.5364892482757568, "train/loss_error": 0.3683001697063446, "train/loss_total": 0.4019380211830139 }, { "epoch": 4.277922077922078, "step": 1647, "train/loss_ctc": 0.5363167524337769, "train/loss_error": 0.4605674147605896, "train/loss_total": 0.47571730613708496 }, { "epoch": 4.28051948051948, "step": 1648, "train/loss_ctc": 0.6160451173782349, "train/loss_error": 0.3972871005535126, "train/loss_total": 0.44103872776031494 }, { "epoch": 4.283116883116883, "step": 1649, "train/loss_ctc": 0.39867615699768066, "train/loss_error": 0.30253803730010986, "train/loss_total": 0.321765661239624 }, { "epoch": 4.285714285714286, "grad_norm": 11140.4521484375, "learning_rate": 4.316883116883117e-06, "loss": 0.4328, "step": 1650 }, { "epoch": 4.285714285714286, "step": 1650, "train/loss_ctc": 0.7433724403381348, "train/loss_error": 0.3987838625907898, "train/loss_total": 0.46770158410072327 }, { "epoch": 4.288311688311689, "step": 1651, "train/loss_ctc": 0.36789196729660034, "train/loss_error": 0.35997140407562256, "train/loss_total": 0.3615555167198181 }, { "epoch": 4.290909090909091, "step": 1652, "train/loss_ctc": 0.2641940712928772, "train/loss_error": 0.3438020646572113, "train/loss_total": 0.32788047194480896 }, { "epoch": 4.293506493506493, "step": 1653, "train/loss_ctc": 0.4652871787548065, "train/loss_error": 0.431561678647995, "train/loss_total": 0.4383068084716797 }, { "epoch": 4.296103896103896, "step": 1654, "train/loss_ctc": 0.5921127796173096, "train/loss_error": 0.4597608149051666, "train/loss_total": 0.4862312376499176 }, { "epoch": 4.298701298701299, "step": 1655, "train/loss_ctc": 0.424400269985199, "train/loss_error": 0.38208717107772827, "train/loss_total": 0.39054980874061584 }, { "epoch": 4.301298701298701, "step": 1656, "train/loss_ctc": 0.479650616645813, "train/loss_error": 0.3677009046077728, "train/loss_total": 0.39009085297584534 }, { "epoch": 4.303896103896104, "step": 1657, "train/loss_ctc": 0.46297314763069153, "train/loss_error": 0.39550021290779114, "train/loss_total": 0.40899479389190674 }, { "epoch": 4.306493506493506, "step": 1658, "train/loss_ctc": 0.4782988429069519, "train/loss_error": 0.4013978838920593, "train/loss_total": 0.4167780876159668 }, { "epoch": 4.309090909090909, "step": 1659, "train/loss_ctc": 0.5576987862586975, "train/loss_error": 0.45059677958488464, "train/loss_total": 0.47201716899871826 }, { "epoch": 4.311688311688312, "grad_norm": 18920.322265625, "learning_rate": 4.161038961038962e-06, "loss": 0.416, "step": 1660 }, { "epoch": 4.311688311688312, "step": 1660, "train/loss_ctc": 0.7321990728378296, "train/loss_error": 0.41831091046333313, "train/loss_total": 0.4810885488986969 }, { "epoch": 4.314285714285714, "step": 1661, "train/loss_ctc": 0.4431873559951782, "train/loss_error": 0.4506557583808899, "train/loss_total": 0.449162095785141 }, { "epoch": 4.316883116883117, "step": 1662, "train/loss_ctc": 0.44919925928115845, "train/loss_error": 0.34464094042778015, "train/loss_total": 0.3655526041984558 }, { "epoch": 4.3194805194805195, "step": 1663, "train/loss_ctc": 0.5462825298309326, "train/loss_error": 0.38956624269485474, "train/loss_total": 0.42090949416160583 }, { "epoch": 4.322077922077922, "step": 1664, "train/loss_ctc": 0.4589868187904358, "train/loss_error": 0.4111640453338623, "train/loss_total": 0.4207286238670349 }, { "epoch": 4.324675324675325, "step": 1665, "train/loss_ctc": 0.680091381072998, "train/loss_error": 0.4720194935798645, "train/loss_total": 0.5136338472366333 }, { "epoch": 4.327272727272727, "step": 1666, "train/loss_ctc": 0.49509912729263306, "train/loss_error": 0.38280367851257324, "train/loss_total": 0.4052627682685852 }, { "epoch": 4.32987012987013, "step": 1667, "train/loss_ctc": 0.4927162826061249, "train/loss_error": 0.41866883635520935, "train/loss_total": 0.43347832560539246 }, { "epoch": 4.332467532467533, "step": 1668, "train/loss_ctc": 0.8025676012039185, "train/loss_error": 0.4362216591835022, "train/loss_total": 0.5094908475875854 }, { "epoch": 4.335064935064935, "step": 1669, "train/loss_ctc": 0.38901716470718384, "train/loss_error": 0.38118377327919006, "train/loss_total": 0.3827504515647888 }, { "epoch": 4.337662337662338, "grad_norm": 9597.1279296875, "learning_rate": 4.005194805194806e-06, "loss": 0.4382, "step": 1670 }, { "epoch": 4.337662337662338, "step": 1670, "train/loss_ctc": 0.6099722385406494, "train/loss_error": 0.4531960189342499, "train/loss_total": 0.48455125093460083 }, { "epoch": 4.34025974025974, "step": 1671, "train/loss_ctc": 0.4789498746395111, "train/loss_error": 0.40995314717292786, "train/loss_total": 0.42375248670578003 }, { "epoch": 4.3428571428571425, "step": 1672, "train/loss_ctc": 0.5187481641769409, "train/loss_error": 0.4532058537006378, "train/loss_total": 0.46631431579589844 }, { "epoch": 4.345454545454546, "step": 1673, "train/loss_ctc": 0.3731723725795746, "train/loss_error": 0.4530022442340851, "train/loss_total": 0.43703627586364746 }, { "epoch": 4.348051948051948, "step": 1674, "train/loss_ctc": 0.5049423575401306, "train/loss_error": 0.3909023404121399, "train/loss_total": 0.413710355758667 }, { "epoch": 4.35064935064935, "step": 1675, "train/loss_ctc": 0.37596723437309265, "train/loss_error": 0.4668675661087036, "train/loss_total": 0.44868749380111694 }, { "epoch": 4.353246753246753, "step": 1676, "train/loss_ctc": 0.6322377920150757, "train/loss_error": 0.4216672480106354, "train/loss_total": 0.46378135681152344 }, { "epoch": 4.355844155844156, "step": 1677, "train/loss_ctc": 0.4999816417694092, "train/loss_error": 0.45436885952949524, "train/loss_total": 0.46349141001701355 }, { "epoch": 4.358441558441559, "step": 1678, "train/loss_ctc": 0.39354032278060913, "train/loss_error": 0.3116876482963562, "train/loss_total": 0.3280581831932068 }, { "epoch": 4.361038961038961, "step": 1679, "train/loss_ctc": 0.7283447980880737, "train/loss_error": 0.40437814593315125, "train/loss_total": 0.4691714644432068 }, { "epoch": 4.363636363636363, "grad_norm": 14079.9609375, "learning_rate": 3.84935064935065e-06, "loss": 0.4399, "step": 1680 }, { "epoch": 4.363636363636363, "step": 1680, "train/loss_ctc": 0.4025823771953583, "train/loss_error": 0.3394223153591156, "train/loss_total": 0.3520543575286865 }, { "epoch": 4.3662337662337665, "step": 1681, "train/loss_ctc": 0.35161861777305603, "train/loss_error": 0.3780398964881897, "train/loss_total": 0.37275564670562744 }, { "epoch": 4.368831168831169, "step": 1682, "train/loss_ctc": 0.41160714626312256, "train/loss_error": 0.3507396876811981, "train/loss_total": 0.36291319131851196 }, { "epoch": 4.371428571428572, "step": 1683, "train/loss_ctc": 0.6920864582061768, "train/loss_error": 0.4114500880241394, "train/loss_total": 0.46757733821868896 }, { "epoch": 4.374025974025974, "step": 1684, "train/loss_ctc": 0.3691822290420532, "train/loss_error": 0.32810500264167786, "train/loss_total": 0.3363204598426819 }, { "epoch": 4.376623376623376, "step": 1685, "train/loss_ctc": 0.6888706684112549, "train/loss_error": 0.3917597234249115, "train/loss_total": 0.45118191838264465 }, { "epoch": 4.37922077922078, "step": 1686, "train/loss_ctc": 0.5622658729553223, "train/loss_error": 0.4183824956417084, "train/loss_total": 0.44715917110443115 }, { "epoch": 4.381818181818182, "step": 1687, "train/loss_ctc": 0.5454022884368896, "train/loss_error": 0.3454376459121704, "train/loss_total": 0.38543057441711426 }, { "epoch": 4.384415584415584, "step": 1688, "train/loss_ctc": 0.4156094193458557, "train/loss_error": 0.3945930302143097, "train/loss_total": 0.39879631996154785 }, { "epoch": 4.387012987012987, "step": 1689, "train/loss_ctc": 0.4039286971092224, "train/loss_error": 0.2685530185699463, "train/loss_total": 0.295628160238266 }, { "epoch": 4.3896103896103895, "grad_norm": 12527.7353515625, "learning_rate": 3.6935064935064935e-06, "loss": 0.387, "step": 1690 }, { "epoch": 4.3896103896103895, "step": 1690, "train/loss_ctc": 0.6871685981750488, "train/loss_error": 0.3324653208255768, "train/loss_total": 0.40340596437454224 }, { "epoch": 4.392207792207792, "step": 1691, "train/loss_ctc": 0.4176119267940521, "train/loss_error": 0.38088834285736084, "train/loss_total": 0.3882330656051636 }, { "epoch": 4.394805194805195, "step": 1692, "train/loss_ctc": 0.4891428053379059, "train/loss_error": 0.3208713233470917, "train/loss_total": 0.354525625705719 }, { "epoch": 4.397402597402597, "step": 1693, "train/loss_ctc": 0.6305591464042664, "train/loss_error": 0.4995543658733368, "train/loss_total": 0.5257553458213806 }, { "epoch": 4.4, "step": 1694, "train/loss_ctc": 0.35013407468795776, "train/loss_error": 0.38673391938209534, "train/loss_total": 0.3794139623641968 }, { "epoch": 4.402597402597403, "step": 1695, "train/loss_ctc": 0.5550258159637451, "train/loss_error": 0.4466528594493866, "train/loss_total": 0.46832746267318726 }, { "epoch": 4.405194805194805, "step": 1696, "train/loss_ctc": 0.6697174310684204, "train/loss_error": 0.49217113852500916, "train/loss_total": 0.5276803970336914 }, { "epoch": 4.407792207792208, "step": 1697, "train/loss_ctc": 0.6221614480018616, "train/loss_error": 0.38120949268341064, "train/loss_total": 0.42939990758895874 }, { "epoch": 4.41038961038961, "step": 1698, "train/loss_ctc": 0.43096983432769775, "train/loss_error": 0.40122532844543457, "train/loss_total": 0.4071742594242096 }, { "epoch": 4.4129870129870135, "step": 1699, "train/loss_ctc": 0.2886575162410736, "train/loss_error": 0.34019163250923157, "train/loss_total": 0.3298848271369934 }, { "epoch": 4.415584415584416, "grad_norm": 8562.3798828125, "learning_rate": 3.537662337662338e-06, "loss": 0.4214, "step": 1700 }, { "epoch": 4.415584415584416, "step": 1700, "train/loss_ctc": 0.3704932928085327, "train/loss_error": 0.365607887506485, "train/loss_total": 0.3665849566459656 }, { "epoch": 4.418181818181818, "step": 1701, "train/loss_ctc": 0.416189581155777, "train/loss_error": 0.39605867862701416, "train/loss_total": 0.40008485317230225 }, { "epoch": 4.420779220779221, "step": 1702, "train/loss_ctc": 0.43610960245132446, "train/loss_error": 0.4793318808078766, "train/loss_total": 0.4706874191761017 }, { "epoch": 4.423376623376623, "step": 1703, "train/loss_ctc": 0.31687837839126587, "train/loss_error": 0.35905832052230835, "train/loss_total": 0.35062235593795776 }, { "epoch": 4.425974025974026, "step": 1704, "train/loss_ctc": 0.6250123977661133, "train/loss_error": 0.3830348551273346, "train/loss_total": 0.4314303994178772 }, { "epoch": 4.428571428571429, "step": 1705, "train/loss_ctc": 0.487000048160553, "train/loss_error": 0.33476415276527405, "train/loss_total": 0.3652113378047943 }, { "epoch": 4.431168831168831, "step": 1706, "train/loss_ctc": 0.9864441156387329, "train/loss_error": 0.5735663771629333, "train/loss_total": 0.6561419367790222 }, { "epoch": 4.433766233766233, "step": 1707, "train/loss_ctc": 0.5658236145973206, "train/loss_error": 0.4811282157897949, "train/loss_total": 0.49806728959083557 }, { "epoch": 4.4363636363636365, "step": 1708, "train/loss_ctc": 0.6318638324737549, "train/loss_error": 0.3563104271888733, "train/loss_total": 0.41142112016677856 }, { "epoch": 4.438961038961039, "step": 1709, "train/loss_ctc": 0.49430540204048157, "train/loss_error": 0.3753451406955719, "train/loss_total": 0.3991371989250183 }, { "epoch": 4.441558441558442, "grad_norm": 13849.1279296875, "learning_rate": 3.381818181818182e-06, "loss": 0.4349, "step": 1710 }, { "epoch": 4.441558441558442, "step": 1710, "train/loss_ctc": 0.5612739324569702, "train/loss_error": 0.5349518656730652, "train/loss_total": 0.5402162671089172 }, { "epoch": 4.444155844155844, "step": 1711, "train/loss_ctc": 0.6361967325210571, "train/loss_error": 0.501406729221344, "train/loss_total": 0.5283647775650024 }, { "epoch": 4.4467532467532465, "step": 1712, "train/loss_ctc": 0.3498649001121521, "train/loss_error": 0.2844845950603485, "train/loss_total": 0.2975606620311737 }, { "epoch": 4.44935064935065, "step": 1713, "train/loss_ctc": 0.40254178643226624, "train/loss_error": 0.35230594873428345, "train/loss_total": 0.36235311627388 }, { "epoch": 4.451948051948052, "step": 1714, "train/loss_ctc": 0.6035752296447754, "train/loss_error": 0.46148520708084106, "train/loss_total": 0.48990321159362793 }, { "epoch": 4.454545454545454, "step": 1715, "train/loss_ctc": 0.3573164939880371, "train/loss_error": 0.3684881329536438, "train/loss_total": 0.3662537932395935 }, { "epoch": 4.457142857142857, "step": 1716, "train/loss_ctc": 0.6574375629425049, "train/loss_error": 0.31553274393081665, "train/loss_total": 0.38391372561454773 }, { "epoch": 4.45974025974026, "step": 1717, "train/loss_ctc": 0.5028491020202637, "train/loss_error": 0.40921610593795776, "train/loss_total": 0.42794269323349 }, { "epoch": 4.462337662337663, "step": 1718, "train/loss_ctc": 0.546451985836029, "train/loss_error": 0.4058191478252411, "train/loss_total": 0.4339457154273987 }, { "epoch": 4.464935064935065, "step": 1719, "train/loss_ctc": 0.4210110306739807, "train/loss_error": 0.23059865832328796, "train/loss_total": 0.268681138753891 }, { "epoch": 4.467532467532467, "grad_norm": 13452.884765625, "learning_rate": 3.2259740259740263e-06, "loss": 0.4099, "step": 1720 }, { "epoch": 4.467532467532467, "step": 1720, "train/loss_ctc": 0.3525708317756653, "train/loss_error": 0.3275449573993683, "train/loss_total": 0.33255013823509216 }, { "epoch": 4.47012987012987, "step": 1721, "train/loss_ctc": 0.608899712562561, "train/loss_error": 0.40682706236839294, "train/loss_total": 0.4472416043281555 }, { "epoch": 4.472727272727273, "step": 1722, "train/loss_ctc": 0.498076856136322, "train/loss_error": 0.3439390957355499, "train/loss_total": 0.37476664781570435 }, { "epoch": 4.475324675324675, "step": 1723, "train/loss_ctc": 0.46164166927337646, "train/loss_error": 0.5311197638511658, "train/loss_total": 0.517224133014679 }, { "epoch": 4.477922077922078, "step": 1724, "train/loss_ctc": 0.4563804864883423, "train/loss_error": 0.480836421251297, "train/loss_total": 0.47594526410102844 }, { "epoch": 4.48051948051948, "step": 1725, "train/loss_ctc": 0.5559326410293579, "train/loss_error": 0.42457884550094604, "train/loss_total": 0.45084962248802185 }, { "epoch": 4.4831168831168835, "step": 1726, "train/loss_ctc": 0.42514926195144653, "train/loss_error": 0.4044025242328644, "train/loss_total": 0.4085518717765808 }, { "epoch": 4.485714285714286, "step": 1727, "train/loss_ctc": 0.41433119773864746, "train/loss_error": 0.3757173418998718, "train/loss_total": 0.38344013690948486 }, { "epoch": 4.488311688311688, "step": 1728, "train/loss_ctc": 0.3874393701553345, "train/loss_error": 0.33628594875335693, "train/loss_total": 0.3465166389942169 }, { "epoch": 4.490909090909091, "step": 1729, "train/loss_ctc": 0.3528135418891907, "train/loss_error": 0.33384019136428833, "train/loss_total": 0.3376348912715912 }, { "epoch": 4.4935064935064934, "grad_norm": 8829.7880859375, "learning_rate": 3.0701298701298702e-06, "loss": 0.4075, "step": 1730 }, { "epoch": 4.4935064935064934, "step": 1730, "train/loss_ctc": 0.47491544485092163, "train/loss_error": 0.36466360092163086, "train/loss_total": 0.38671398162841797 }, { "epoch": 4.496103896103896, "step": 1731, "train/loss_ctc": 0.4142327308654785, "train/loss_error": 0.40858083963394165, "train/loss_total": 0.40971124172210693 }, { "epoch": 4.498701298701299, "step": 1732, "train/loss_ctc": 0.42105358839035034, "train/loss_error": 0.36249229311943054, "train/loss_total": 0.3742045760154724 }, { "epoch": 4.501298701298701, "step": 1733, "train/loss_ctc": 0.4030689597129822, "train/loss_error": 0.41209545731544495, "train/loss_total": 0.4102901518344879 }, { "epoch": 4.503896103896103, "step": 1734, "train/loss_ctc": 0.45742279291152954, "train/loss_error": 0.4029315412044525, "train/loss_total": 0.4138298034667969 }, { "epoch": 4.5064935064935066, "step": 1735, "train/loss_ctc": 0.554865300655365, "train/loss_error": 0.4018315374851227, "train/loss_total": 0.43243828415870667 }, { "epoch": 4.509090909090909, "step": 1736, "train/loss_ctc": 0.5024821758270264, "train/loss_error": 0.4249308109283447, "train/loss_total": 0.4404411017894745 }, { "epoch": 4.511688311688312, "step": 1737, "train/loss_ctc": 0.3954819440841675, "train/loss_error": 0.41346365213394165, "train/loss_total": 0.4098673462867737 }, { "epoch": 4.514285714285714, "step": 1738, "train/loss_ctc": 0.4807076156139374, "train/loss_error": 0.45058926939964294, "train/loss_total": 0.4566129446029663 }, { "epoch": 4.5168831168831165, "step": 1739, "train/loss_ctc": 0.6101760268211365, "train/loss_error": 0.4575943052768707, "train/loss_total": 0.48811066150665283 }, { "epoch": 4.51948051948052, "grad_norm": 12719.87890625, "learning_rate": 2.9142857142857142e-06, "loss": 0.4222, "step": 1740 }, { "epoch": 4.51948051948052, "step": 1740, "train/loss_ctc": 0.4305347204208374, "train/loss_error": 0.467781662940979, "train/loss_total": 0.46033230423927307 }, { "epoch": 4.522077922077922, "step": 1741, "train/loss_ctc": 0.6758469343185425, "train/loss_error": 0.42959633469581604, "train/loss_total": 0.4788464605808258 }, { "epoch": 4.524675324675325, "step": 1742, "train/loss_ctc": 0.5602647662162781, "train/loss_error": 0.4433674216270447, "train/loss_total": 0.46674689650535583 }, { "epoch": 4.527272727272727, "step": 1743, "train/loss_ctc": 0.4906679689884186, "train/loss_error": 0.45124351978302, "train/loss_total": 0.45912840962409973 }, { "epoch": 4.52987012987013, "step": 1744, "train/loss_ctc": 0.3910750448703766, "train/loss_error": 0.3891662359237671, "train/loss_total": 0.38954800367355347 }, { "epoch": 4.532467532467533, "step": 1745, "train/loss_ctc": 0.4285149574279785, "train/loss_error": 0.3757133483886719, "train/loss_total": 0.38627368211746216 }, { "epoch": 4.535064935064935, "step": 1746, "train/loss_ctc": 0.4912792444229126, "train/loss_error": 0.4229351878166199, "train/loss_total": 0.43660399317741394 }, { "epoch": 4.537662337662337, "step": 1747, "train/loss_ctc": 0.42213737964630127, "train/loss_error": 0.3543930947780609, "train/loss_total": 0.3679419457912445 }, { "epoch": 4.54025974025974, "step": 1748, "train/loss_ctc": 0.3922780752182007, "train/loss_error": 0.41687464714050293, "train/loss_total": 0.4119553565979004 }, { "epoch": 4.542857142857143, "step": 1749, "train/loss_ctc": 0.537169337272644, "train/loss_error": 0.3526870608329773, "train/loss_total": 0.3895835280418396 }, { "epoch": 4.545454545454545, "grad_norm": 11913.6611328125, "learning_rate": 2.7584415584415586e-06, "loss": 0.4247, "step": 1750 }, { "epoch": 4.545454545454545, "step": 1750, "train/loss_ctc": 0.529628336429596, "train/loss_error": 0.3992043435573578, "train/loss_total": 0.4252891540527344 }, { "epoch": 4.548051948051948, "step": 1751, "train/loss_ctc": 0.622391402721405, "train/loss_error": 0.43632254004478455, "train/loss_total": 0.47353631258010864 }, { "epoch": 4.55064935064935, "step": 1752, "train/loss_ctc": 0.39429783821105957, "train/loss_error": 0.34144994616508484, "train/loss_total": 0.3520195186138153 }, { "epoch": 4.5532467532467535, "step": 1753, "train/loss_ctc": 0.4136980473995209, "train/loss_error": 0.4455699622631073, "train/loss_total": 0.4391956031322479 }, { "epoch": 4.555844155844156, "step": 1754, "train/loss_ctc": 0.6087924242019653, "train/loss_error": 0.3671843409538269, "train/loss_total": 0.415505975484848 }, { "epoch": 4.558441558441558, "step": 1755, "train/loss_ctc": 0.48968926072120667, "train/loss_error": 0.3776126801967621, "train/loss_total": 0.4000279903411865 }, { "epoch": 4.561038961038961, "step": 1756, "train/loss_ctc": 0.3174304962158203, "train/loss_error": 0.3986155092716217, "train/loss_total": 0.3823785185813904 }, { "epoch": 4.5636363636363635, "step": 1757, "train/loss_ctc": 0.3660723567008972, "train/loss_error": 0.4541739821434021, "train/loss_total": 0.4365536570549011 }, { "epoch": 4.566233766233767, "step": 1758, "train/loss_ctc": 0.4780247211456299, "train/loss_error": 0.3283267021179199, "train/loss_total": 0.35826632380485535 }, { "epoch": 4.568831168831169, "step": 1759, "train/loss_ctc": 0.41531550884246826, "train/loss_error": 0.4451141357421875, "train/loss_total": 0.43915441632270813 }, { "epoch": 4.571428571428571, "grad_norm": 12934.853515625, "learning_rate": 2.6025974025974026e-06, "loss": 0.4122, "step": 1760 }, { "epoch": 4.571428571428571, "step": 1760, "train/loss_ctc": 0.3596014380455017, "train/loss_error": 0.38482457399368286, "train/loss_total": 0.3797799348831177 }, { "epoch": 4.574025974025974, "step": 1761, "train/loss_ctc": 0.6066306829452515, "train/loss_error": 0.3828592002391815, "train/loss_total": 0.4276135265827179 }, { "epoch": 4.576623376623377, "step": 1762, "train/loss_ctc": 0.48325496912002563, "train/loss_error": 0.43881839513778687, "train/loss_total": 0.4477057158946991 }, { "epoch": 4.579220779220779, "step": 1763, "train/loss_ctc": 0.46529173851013184, "train/loss_error": 0.3866559565067291, "train/loss_total": 0.40238311886787415 }, { "epoch": 4.581818181818182, "step": 1764, "train/loss_ctc": 0.5763528347015381, "train/loss_error": 0.36261361837387085, "train/loss_total": 0.40536147356033325 }, { "epoch": 4.584415584415584, "step": 1765, "train/loss_ctc": 0.28888753056526184, "train/loss_error": 0.2866673767566681, "train/loss_total": 0.28711140155792236 }, { "epoch": 4.5870129870129865, "step": 1766, "train/loss_ctc": 0.5300920605659485, "train/loss_error": 0.30801689624786377, "train/loss_total": 0.35243192315101624 }, { "epoch": 4.58961038961039, "step": 1767, "train/loss_ctc": 0.5020685791969299, "train/loss_error": 0.34364694356918335, "train/loss_total": 0.3753312826156616 }, { "epoch": 4.592207792207792, "step": 1768, "train/loss_ctc": 0.4285567104816437, "train/loss_error": 0.4013385772705078, "train/loss_total": 0.40678220987319946 }, { "epoch": 4.594805194805195, "step": 1769, "train/loss_ctc": 0.5500727295875549, "train/loss_error": 0.4233303964138031, "train/loss_total": 0.4486788809299469 }, { "epoch": 4.597402597402597, "grad_norm": 16538.291015625, "learning_rate": 2.4467532467532465e-06, "loss": 0.3933, "step": 1770 }, { "epoch": 4.597402597402597, "step": 1770, "train/loss_ctc": 0.6107359528541565, "train/loss_error": 0.4841046631336212, "train/loss_total": 0.5094309449195862 }, { "epoch": 4.6, "step": 1771, "train/loss_ctc": 0.44464391469955444, "train/loss_error": 0.3955104947090149, "train/loss_total": 0.4053371846675873 }, { "epoch": 4.602597402597403, "step": 1772, "train/loss_ctc": 0.2665215730667114, "train/loss_error": 0.39486560225486755, "train/loss_total": 0.3691968023777008 }, { "epoch": 4.605194805194805, "step": 1773, "train/loss_ctc": 0.5397481322288513, "train/loss_error": 0.41320595145225525, "train/loss_total": 0.4385144114494324 }, { "epoch": 4.607792207792208, "step": 1774, "train/loss_ctc": 0.498008131980896, "train/loss_error": 0.43639108538627625, "train/loss_total": 0.4487144947052002 }, { "epoch": 4.6103896103896105, "step": 1775, "train/loss_ctc": 0.6303338408470154, "train/loss_error": 0.4264262616634369, "train/loss_total": 0.46720778942108154 }, { "epoch": 4.612987012987013, "step": 1776, "train/loss_ctc": 0.5422765016555786, "train/loss_error": 0.32599231600761414, "train/loss_total": 0.369249165058136 }, { "epoch": 4.615584415584416, "step": 1777, "train/loss_ctc": 0.38659849762916565, "train/loss_error": 0.40521731972694397, "train/loss_total": 0.4014935791492462 }, { "epoch": 4.618181818181818, "step": 1778, "train/loss_ctc": 0.5778908729553223, "train/loss_error": 0.3328360915184021, "train/loss_total": 0.3818470537662506 }, { "epoch": 4.62077922077922, "step": 1779, "train/loss_ctc": 0.4343201518058777, "train/loss_error": 0.3299342095851898, "train/loss_total": 0.3508113920688629 }, { "epoch": 4.623376623376624, "grad_norm": 14157.12890625, "learning_rate": 2.2909090909090913e-06, "loss": 0.4142, "step": 1780 }, { "epoch": 4.623376623376624, "step": 1780, "train/loss_ctc": 0.45170024037361145, "train/loss_error": 0.30942586064338684, "train/loss_total": 0.3378807306289673 }, { "epoch": 4.625974025974026, "step": 1781, "train/loss_ctc": 0.45851850509643555, "train/loss_error": 0.40199539065361023, "train/loss_total": 0.4133000373840332 }, { "epoch": 4.628571428571428, "step": 1782, "train/loss_ctc": 0.48552459478378296, "train/loss_error": 0.45342355966567993, "train/loss_total": 0.4598437547683716 }, { "epoch": 4.631168831168831, "step": 1783, "train/loss_ctc": 0.42787641286849976, "train/loss_error": 0.3734860420227051, "train/loss_total": 0.38436412811279297 }, { "epoch": 4.6337662337662335, "step": 1784, "train/loss_ctc": 0.45452770590782166, "train/loss_error": 0.3887462913990021, "train/loss_total": 0.40190258622169495 }, { "epoch": 4.636363636363637, "step": 1785, "train/loss_ctc": 0.364932119846344, "train/loss_error": 0.3593684136867523, "train/loss_total": 0.3604811728000641 }, { "epoch": 4.638961038961039, "step": 1786, "train/loss_ctc": 0.4075302481651306, "train/loss_error": 0.3669377565383911, "train/loss_total": 0.37505626678466797 }, { "epoch": 4.641558441558441, "step": 1787, "train/loss_ctc": 0.44159018993377686, "train/loss_error": 0.41686153411865234, "train/loss_total": 0.42180728912353516 }, { "epoch": 4.644155844155844, "step": 1788, "train/loss_ctc": 0.38681837916374207, "train/loss_error": 0.2963608205318451, "train/loss_total": 0.3144523501396179 }, { "epoch": 4.646753246753247, "step": 1789, "train/loss_ctc": 0.6940243244171143, "train/loss_error": 0.3601498007774353, "train/loss_total": 0.4269247055053711 }, { "epoch": 4.64935064935065, "grad_norm": 14768.4033203125, "learning_rate": 2.1350649350649353e-06, "loss": 0.3896, "step": 1790 }, { "epoch": 4.64935064935065, "step": 1790, "train/loss_ctc": 0.3987659215927124, "train/loss_error": 0.48365139961242676, "train/loss_total": 0.4666743278503418 }, { "epoch": 4.651948051948052, "step": 1791, "train/loss_ctc": 0.650017261505127, "train/loss_error": 0.466369092464447, "train/loss_total": 0.503098726272583 }, { "epoch": 4.654545454545454, "step": 1792, "train/loss_ctc": 0.6692866086959839, "train/loss_error": 0.43099290132522583, "train/loss_total": 0.47865164279937744 }, { "epoch": 4.6571428571428575, "step": 1793, "train/loss_ctc": 0.5923020839691162, "train/loss_error": 0.41762059926986694, "train/loss_total": 0.45255690813064575 }, { "epoch": 4.65974025974026, "step": 1794, "train/loss_ctc": 0.41764506697654724, "train/loss_error": 0.4280316233634949, "train/loss_total": 0.42595434188842773 }, { "epoch": 4.662337662337662, "step": 1795, "train/loss_ctc": 0.4216887056827545, "train/loss_error": 0.405958890914917, "train/loss_total": 0.4091048538684845 }, { "epoch": 4.664935064935065, "step": 1796, "train/loss_ctc": 0.40112197399139404, "train/loss_error": 0.33450230956077576, "train/loss_total": 0.3478262424468994 }, { "epoch": 4.667532467532467, "step": 1797, "train/loss_ctc": 0.23543350398540497, "train/loss_error": 0.36350369453430176, "train/loss_total": 0.3378896713256836 }, { "epoch": 4.67012987012987, "step": 1798, "train/loss_ctc": 0.358320951461792, "train/loss_error": 0.42216747999191284, "train/loss_total": 0.4093981683254242 }, { "epoch": 4.672727272727273, "step": 1799, "train/loss_ctc": 0.4593353569507599, "train/loss_error": 0.41028428077697754, "train/loss_total": 0.42009449005126953 }, { "epoch": 4.675324675324675, "grad_norm": 11767.9873046875, "learning_rate": 1.9792207792207793e-06, "loss": 0.4251, "step": 1800 }, { "epoch": 4.675324675324675, "step": 1800, "train/loss_ctc": 0.2721851170063019, "train/loss_error": 0.25583362579345703, "train/loss_total": 0.259103924036026 }, { "epoch": 4.677922077922078, "step": 1801, "train/loss_ctc": 0.6109359860420227, "train/loss_error": 0.4427731931209564, "train/loss_total": 0.4764057695865631 }, { "epoch": 4.6805194805194805, "step": 1802, "train/loss_ctc": 0.5636656880378723, "train/loss_error": 0.34381234645843506, "train/loss_total": 0.3877830505371094 }, { "epoch": 4.683116883116883, "step": 1803, "train/loss_ctc": 0.45244261622428894, "train/loss_error": 0.41688936948776245, "train/loss_total": 0.4240000247955322 }, { "epoch": 4.685714285714286, "step": 1804, "train/loss_ctc": 0.18645305931568146, "train/loss_error": 0.31460708379745483, "train/loss_total": 0.2889762818813324 }, { "epoch": 4.688311688311688, "step": 1805, "train/loss_ctc": 0.537310004234314, "train/loss_error": 0.38115590810775757, "train/loss_total": 0.4123867154121399 }, { "epoch": 4.690909090909091, "step": 1806, "train/loss_ctc": 0.4874435067176819, "train/loss_error": 0.37078288197517395, "train/loss_total": 0.39411500096321106 }, { "epoch": 4.693506493506494, "step": 1807, "train/loss_ctc": 0.3692941665649414, "train/loss_error": 0.32215356826782227, "train/loss_total": 0.3315816819667816 }, { "epoch": 4.696103896103896, "step": 1808, "train/loss_ctc": 0.42017021775245667, "train/loss_error": 0.29740217328071594, "train/loss_total": 0.3219558000564575 }, { "epoch": 4.698701298701299, "step": 1809, "train/loss_ctc": 0.44699788093566895, "train/loss_error": 0.3187485337257385, "train/loss_total": 0.3443984091281891 }, { "epoch": 4.701298701298701, "grad_norm": 10598.0908203125, "learning_rate": 1.8233766233766235e-06, "loss": 0.3641, "step": 1810 }, { "epoch": 4.701298701298701, "step": 1810, "train/loss_ctc": 0.40421250462532043, "train/loss_error": 0.3522477447986603, "train/loss_total": 0.36264070868492126 }, { "epoch": 4.703896103896104, "step": 1811, "train/loss_ctc": 0.5873057842254639, "train/loss_error": 0.39997804164886475, "train/loss_total": 0.4374436140060425 }, { "epoch": 4.706493506493507, "step": 1812, "train/loss_ctc": 0.4854055643081665, "train/loss_error": 0.4108489453792572, "train/loss_total": 0.42576029896736145 }, { "epoch": 4.709090909090909, "step": 1813, "train/loss_ctc": 0.3329662084579468, "train/loss_error": 0.36934006214141846, "train/loss_total": 0.36206531524658203 }, { "epoch": 4.711688311688311, "step": 1814, "train/loss_ctc": 0.666250467300415, "train/loss_error": 0.38215726613998413, "train/loss_total": 0.4389759302139282 }, { "epoch": 4.714285714285714, "step": 1815, "train/loss_ctc": 0.4717218279838562, "train/loss_error": 0.4293256998062134, "train/loss_total": 0.4378049373626709 }, { "epoch": 4.716883116883117, "step": 1816, "train/loss_ctc": 0.30764102935791016, "train/loss_error": 0.4356011152267456, "train/loss_total": 0.41000911593437195 }, { "epoch": 4.71948051948052, "step": 1817, "train/loss_ctc": 0.9043107032775879, "train/loss_error": 0.47621363401412964, "train/loss_total": 0.5618330836296082 }, { "epoch": 4.722077922077922, "step": 1818, "train/loss_ctc": 0.795289158821106, "train/loss_error": 0.44868195056915283, "train/loss_total": 0.5180034041404724 }, { "epoch": 4.724675324675324, "step": 1819, "train/loss_ctc": 0.7346314191818237, "train/loss_error": 0.5433991551399231, "train/loss_total": 0.5816456079483032 }, { "epoch": 4.7272727272727275, "grad_norm": 13950.4033203125, "learning_rate": 1.6675324675324676e-06, "loss": 0.4536, "step": 1820 }, { "epoch": 4.7272727272727275, "step": 1820, "train/loss_ctc": 0.45166438817977905, "train/loss_error": 0.41668760776519775, "train/loss_total": 0.4236829876899719 }, { "epoch": 4.72987012987013, "step": 1821, "train/loss_ctc": 0.3628138303756714, "train/loss_error": 0.3585377037525177, "train/loss_total": 0.3593929409980774 }, { "epoch": 4.732467532467533, "step": 1822, "train/loss_ctc": 0.35199734568595886, "train/loss_error": 0.3580729067325592, "train/loss_total": 0.3568578064441681 }, { "epoch": 4.735064935064935, "step": 1823, "train/loss_ctc": 0.5250440239906311, "train/loss_error": 0.3526321053504944, "train/loss_total": 0.3871144950389862 }, { "epoch": 4.7376623376623375, "step": 1824, "train/loss_ctc": 0.44682836532592773, "train/loss_error": 0.49489620327949524, "train/loss_total": 0.48528265953063965 }, { "epoch": 4.740259740259741, "step": 1825, "train/loss_ctc": 0.4703420400619507, "train/loss_error": 0.4679144322872162, "train/loss_total": 0.4683999717235565 }, { "epoch": 4.742857142857143, "step": 1826, "train/loss_ctc": 0.29085803031921387, "train/loss_error": 0.37240836024284363, "train/loss_total": 0.3560982942581177 }, { "epoch": 4.745454545454545, "step": 1827, "train/loss_ctc": 0.5689529180526733, "train/loss_error": 0.35649827122688293, "train/loss_total": 0.398989200592041 }, { "epoch": 4.748051948051948, "step": 1828, "train/loss_ctc": 0.5215883255004883, "train/loss_error": 0.3889677822589874, "train/loss_total": 0.41549190878868103 }, { "epoch": 4.750649350649351, "step": 1829, "train/loss_ctc": 0.6459214687347412, "train/loss_error": 0.5158162713050842, "train/loss_total": 0.5418373346328735 }, { "epoch": 4.753246753246753, "grad_norm": 13600.9736328125, "learning_rate": 1.5116883116883116e-06, "loss": 0.4193, "step": 1830 }, { "epoch": 4.753246753246753, "step": 1830, "train/loss_ctc": 0.3940260708332062, "train/loss_error": 0.39378416538238525, "train/loss_total": 0.3938325345516205 }, { "epoch": 4.755844155844156, "step": 1831, "train/loss_ctc": 0.2937958240509033, "train/loss_error": 0.3042256236076355, "train/loss_total": 0.30213966965675354 }, { "epoch": 4.758441558441558, "step": 1832, "train/loss_ctc": 0.44967037439346313, "train/loss_error": 0.3066394627094269, "train/loss_total": 0.33524563908576965 }, { "epoch": 4.761038961038961, "step": 1833, "train/loss_ctc": 0.43896055221557617, "train/loss_error": 0.5146698951721191, "train/loss_total": 0.49952805042266846 }, { "epoch": 4.763636363636364, "step": 1834, "train/loss_ctc": 0.45440661907196045, "train/loss_error": 0.3453032970428467, "train/loss_total": 0.36712396144866943 }, { "epoch": 4.766233766233766, "step": 1835, "train/loss_ctc": 0.5594608783721924, "train/loss_error": 0.42681166529655457, "train/loss_total": 0.453341543674469 }, { "epoch": 4.768831168831169, "step": 1836, "train/loss_ctc": 0.3738638758659363, "train/loss_error": 0.37625327706336975, "train/loss_total": 0.37577539682388306 }, { "epoch": 4.771428571428571, "step": 1837, "train/loss_ctc": 0.7779686450958252, "train/loss_error": 0.43587419390678406, "train/loss_total": 0.5042930841445923 }, { "epoch": 4.7740259740259745, "step": 1838, "train/loss_ctc": 0.47671836614608765, "train/loss_error": 0.4314018487930298, "train/loss_total": 0.44046515226364136 }, { "epoch": 4.776623376623377, "step": 1839, "train/loss_ctc": 0.5031704306602478, "train/loss_error": 0.4254395067691803, "train/loss_total": 0.44098570942878723 }, { "epoch": 4.779220779220779, "grad_norm": 14358.6005859375, "learning_rate": 1.355844155844156e-06, "loss": 0.4113, "step": 1840 }, { "epoch": 4.779220779220779, "step": 1840, "train/loss_ctc": 0.7423962950706482, "train/loss_error": 0.48851969838142395, "train/loss_total": 0.5392950177192688 }, { "epoch": 4.781818181818182, "step": 1841, "train/loss_ctc": 0.3954870104789734, "train/loss_error": 0.3796749413013458, "train/loss_total": 0.38283735513687134 }, { "epoch": 4.7844155844155845, "step": 1842, "train/loss_ctc": 0.4111485481262207, "train/loss_error": 0.3698723614215851, "train/loss_total": 0.3781276047229767 }, { "epoch": 4.787012987012987, "step": 1843, "train/loss_ctc": 0.47818723320961, "train/loss_error": 0.378029465675354, "train/loss_total": 0.39806100726127625 }, { "epoch": 4.78961038961039, "step": 1844, "train/loss_ctc": 0.7258540987968445, "train/loss_error": 0.4567834138870239, "train/loss_total": 0.5105975866317749 }, { "epoch": 4.792207792207792, "step": 1845, "train/loss_ctc": 0.500343918800354, "train/loss_error": 0.4799771010875702, "train/loss_total": 0.484050452709198 }, { "epoch": 4.794805194805194, "step": 1846, "train/loss_ctc": 0.38486918807029724, "train/loss_error": 0.3260926306247711, "train/loss_total": 0.3378479480743408 }, { "epoch": 4.797402597402598, "step": 1847, "train/loss_ctc": 0.3234502971172333, "train/loss_error": 0.34794551134109497, "train/loss_total": 0.3430464565753937 }, { "epoch": 4.8, "step": 1848, "train/loss_ctc": 0.4288945198059082, "train/loss_error": 0.3093739449977875, "train/loss_total": 0.3332780599594116 }, { "epoch": 4.802597402597403, "step": 1849, "train/loss_ctc": 0.4350263476371765, "train/loss_error": 0.36069318652153015, "train/loss_total": 0.37555980682373047 }, { "epoch": 4.805194805194805, "grad_norm": 16117.974609375, "learning_rate": 1.2000000000000002e-06, "loss": 0.4083, "step": 1850 }, { "epoch": 4.805194805194805, "step": 1850, "train/loss_ctc": 0.5243316888809204, "train/loss_error": 0.3830089867115021, "train/loss_total": 0.4112735390663147 }, { "epoch": 4.8077922077922075, "step": 1851, "train/loss_ctc": 0.2723351716995239, "train/loss_error": 0.2727559804916382, "train/loss_total": 0.27267181873321533 }, { "epoch": 4.810389610389611, "step": 1852, "train/loss_ctc": 0.47798627614974976, "train/loss_error": 0.4455215036869049, "train/loss_total": 0.4520144760608673 }, { "epoch": 4.812987012987013, "step": 1853, "train/loss_ctc": 0.4455527067184448, "train/loss_error": 0.42685651779174805, "train/loss_total": 0.4305957853794098 }, { "epoch": 4.815584415584416, "step": 1854, "train/loss_ctc": 0.676306962966919, "train/loss_error": 0.4849454462528229, "train/loss_total": 0.5232177376747131 }, { "epoch": 4.818181818181818, "step": 1855, "train/loss_ctc": 0.41996970772743225, "train/loss_error": 0.3644717335700989, "train/loss_total": 0.3755713403224945 }, { "epoch": 4.820779220779221, "step": 1856, "train/loss_ctc": 0.28689825534820557, "train/loss_error": 0.3436351716518402, "train/loss_total": 0.33228781819343567 }, { "epoch": 4.823376623376624, "step": 1857, "train/loss_ctc": 0.5259252786636353, "train/loss_error": 0.4450709819793701, "train/loss_total": 0.46124184131622314 }, { "epoch": 4.825974025974026, "step": 1858, "train/loss_ctc": 0.5113155245780945, "train/loss_error": 0.35005709528923035, "train/loss_total": 0.3823087811470032 }, { "epoch": 4.828571428571428, "step": 1859, "train/loss_ctc": 0.3418533205986023, "train/loss_error": 0.2906495928764343, "train/loss_total": 0.30089032649993896 }, { "epoch": 4.8311688311688314, "grad_norm": 9893.8330078125, "learning_rate": 1.0441558441558441e-06, "loss": 0.3942, "step": 1860 }, { "epoch": 4.8311688311688314, "step": 1860, "train/loss_ctc": 0.43650534749031067, "train/loss_error": 0.39821240305900574, "train/loss_total": 0.4058710038661957 }, { "epoch": 4.833766233766234, "step": 1861, "train/loss_ctc": 0.5439802408218384, "train/loss_error": 0.406249463558197, "train/loss_total": 0.43379563093185425 }, { "epoch": 4.836363636363636, "step": 1862, "train/loss_ctc": 0.4093051850795746, "train/loss_error": 0.32606041431427, "train/loss_total": 0.34270939230918884 }, { "epoch": 4.838961038961039, "step": 1863, "train/loss_ctc": 0.44684267044067383, "train/loss_error": 0.3681526184082031, "train/loss_total": 0.38389062881469727 }, { "epoch": 4.841558441558441, "step": 1864, "train/loss_ctc": 0.6503671407699585, "train/loss_error": 0.5347073078155518, "train/loss_total": 0.5578392744064331 }, { "epoch": 4.8441558441558445, "step": 1865, "train/loss_ctc": 0.33074527978897095, "train/loss_error": 0.45157864689826965, "train/loss_total": 0.4274119734764099 }, { "epoch": 4.846753246753247, "step": 1866, "train/loss_ctc": 0.5600571632385254, "train/loss_error": 0.42231297492980957, "train/loss_total": 0.4498618245124817 }, { "epoch": 4.849350649350649, "step": 1867, "train/loss_ctc": 0.49751558899879456, "train/loss_error": 0.37560558319091797, "train/loss_total": 0.3999876081943512 }, { "epoch": 4.851948051948052, "step": 1868, "train/loss_ctc": 0.5959056615829468, "train/loss_error": 0.44987836480140686, "train/loss_total": 0.4790838360786438 }, { "epoch": 4.8545454545454545, "step": 1869, "train/loss_ctc": 0.4731236696243286, "train/loss_error": 0.47707507014274597, "train/loss_total": 0.47628480195999146 }, { "epoch": 4.857142857142857, "grad_norm": 12025.32421875, "learning_rate": 8.883116883116884e-07, "loss": 0.4357, "step": 1870 }, { "epoch": 4.857142857142857, "step": 1870, "train/loss_ctc": 0.4817366600036621, "train/loss_error": 0.32546359300613403, "train/loss_total": 0.3567182123661041 }, { "epoch": 4.85974025974026, "step": 1871, "train/loss_ctc": 0.4113258719444275, "train/loss_error": 0.40184876322746277, "train/loss_total": 0.4037441909313202 }, { "epoch": 4.862337662337662, "step": 1872, "train/loss_ctc": 0.519696831703186, "train/loss_error": 0.37214145064353943, "train/loss_total": 0.4016525149345398 }, { "epoch": 4.8649350649350644, "step": 1873, "train/loss_ctc": 0.44471684098243713, "train/loss_error": 0.3455350697040558, "train/loss_total": 0.365371435880661 }, { "epoch": 4.867532467532468, "step": 1874, "train/loss_ctc": 0.8185798525810242, "train/loss_error": 0.4783753752708435, "train/loss_total": 0.5464162826538086 }, { "epoch": 4.87012987012987, "step": 1875, "train/loss_ctc": 0.5225358009338379, "train/loss_error": 0.3638836741447449, "train/loss_total": 0.3956140875816345 }, { "epoch": 4.872727272727273, "step": 1876, "train/loss_ctc": 0.48338356614112854, "train/loss_error": 0.4586475193500519, "train/loss_total": 0.4635947346687317 }, { "epoch": 4.875324675324675, "step": 1877, "train/loss_ctc": 0.7772195339202881, "train/loss_error": 0.45796075463294983, "train/loss_total": 0.5218125581741333 }, { "epoch": 4.8779220779220775, "step": 1878, "train/loss_ctc": 0.2940283417701721, "train/loss_error": 0.38359159231185913, "train/loss_total": 0.36567896604537964 }, { "epoch": 4.880519480519481, "step": 1879, "train/loss_ctc": 0.7187150716781616, "train/loss_error": 0.4605528712272644, "train/loss_total": 0.5121853351593018 }, { "epoch": 4.883116883116883, "grad_norm": 15109.8779296875, "learning_rate": 7.324675324675325e-07, "loss": 0.4333, "step": 1880 }, { "epoch": 4.883116883116883, "step": 1880, "train/loss_ctc": 0.37175801396369934, "train/loss_error": 0.32657864689826965, "train/loss_total": 0.33561453223228455 }, { "epoch": 4.885714285714286, "step": 1881, "train/loss_ctc": 0.5146573781967163, "train/loss_error": 0.40422186255455017, "train/loss_total": 0.4263089597225189 }, { "epoch": 4.888311688311688, "step": 1882, "train/loss_ctc": 0.5688270330429077, "train/loss_error": 0.40185514092445374, "train/loss_total": 0.43524953722953796 }, { "epoch": 4.890909090909091, "step": 1883, "train/loss_ctc": 0.5565567016601562, "train/loss_error": 0.42284533381462097, "train/loss_total": 0.4495876133441925 }, { "epoch": 4.893506493506494, "step": 1884, "train/loss_ctc": 0.5107940435409546, "train/loss_error": 0.3993544280529022, "train/loss_total": 0.42164236307144165 }, { "epoch": 4.896103896103896, "step": 1885, "train/loss_ctc": 0.4956749677658081, "train/loss_error": 0.4507199823856354, "train/loss_total": 0.4597110152244568 }, { "epoch": 4.898701298701298, "step": 1886, "train/loss_ctc": 0.4455299377441406, "train/loss_error": 0.4165908992290497, "train/loss_total": 0.4223787188529968 }, { "epoch": 4.9012987012987015, "step": 1887, "train/loss_ctc": 0.386932373046875, "train/loss_error": 0.38424068689346313, "train/loss_total": 0.38477903604507446 }, { "epoch": 4.903896103896104, "step": 1888, "train/loss_ctc": 0.34606727957725525, "train/loss_error": 0.36450520157814026, "train/loss_total": 0.3608176112174988 }, { "epoch": 4.906493506493506, "step": 1889, "train/loss_ctc": 0.6680192351341248, "train/loss_error": 0.5147483348846436, "train/loss_total": 0.5454025268554688 }, { "epoch": 4.909090909090909, "grad_norm": 15392.7685546875, "learning_rate": 5.766233766233767e-07, "loss": 0.4241, "step": 1890 }, { "epoch": 4.909090909090909, "step": 1890, "train/loss_ctc": 0.40479379892349243, "train/loss_error": 0.37960389256477356, "train/loss_total": 0.3846418857574463 }, { "epoch": 4.911688311688311, "step": 1891, "train/loss_ctc": 0.43567898869514465, "train/loss_error": 0.30012571811676025, "train/loss_total": 0.3272363841533661 }, { "epoch": 4.914285714285715, "step": 1892, "train/loss_ctc": 0.3917548358440399, "train/loss_error": 0.41418343782424927, "train/loss_total": 0.4096977114677429 }, { "epoch": 4.916883116883117, "step": 1893, "train/loss_ctc": 0.2613515555858612, "train/loss_error": 0.30400964617729187, "train/loss_total": 0.29547804594039917 }, { "epoch": 4.919480519480519, "step": 1894, "train/loss_ctc": 0.6088802814483643, "train/loss_error": 0.4068549573421478, "train/loss_total": 0.4472600221633911 }, { "epoch": 4.922077922077922, "step": 1895, "train/loss_ctc": 0.7018362283706665, "train/loss_error": 0.42145031690597534, "train/loss_total": 0.4775274991989136 }, { "epoch": 4.9246753246753245, "step": 1896, "train/loss_ctc": 0.3957171142101288, "train/loss_error": 0.34479427337646484, "train/loss_total": 0.35497885942459106 }, { "epoch": 4.927272727272728, "step": 1897, "train/loss_ctc": 0.4906078577041626, "train/loss_error": 0.4616478681564331, "train/loss_total": 0.4674398899078369 }, { "epoch": 4.92987012987013, "step": 1898, "train/loss_ctc": 0.5299292802810669, "train/loss_error": 0.30312639474868774, "train/loss_total": 0.348486989736557 }, { "epoch": 4.932467532467532, "step": 1899, "train/loss_ctc": 0.43524155020713806, "train/loss_error": 0.39527642726898193, "train/loss_total": 0.4032694697380066 }, { "epoch": 4.935064935064935, "grad_norm": 11673.2197265625, "learning_rate": 4.2077922077922083e-07, "loss": 0.3916, "step": 1900 }, { "epoch": 4.935064935064935, "step": 1900, "train/loss_ctc": 0.453904390335083, "train/loss_error": 0.41712236404418945, "train/loss_total": 0.42447876930236816 }, { "epoch": 4.937662337662338, "step": 1901, "train/loss_ctc": 0.44959139823913574, "train/loss_error": 0.33315643668174744, "train/loss_total": 0.3564434349536896 }, { "epoch": 4.94025974025974, "step": 1902, "train/loss_ctc": 0.42419469356536865, "train/loss_error": 0.2955370843410492, "train/loss_total": 0.32126861810684204 }, { "epoch": 4.942857142857143, "step": 1903, "train/loss_ctc": 0.7139376997947693, "train/loss_error": 0.45672932267189026, "train/loss_total": 0.508171021938324 }, { "epoch": 4.945454545454545, "step": 1904, "train/loss_ctc": 0.421017050743103, "train/loss_error": 0.4752630591392517, "train/loss_total": 0.4644138813018799 }, { "epoch": 4.948051948051948, "step": 1905, "train/loss_ctc": 0.4474524259567261, "train/loss_error": 0.4232032001018524, "train/loss_total": 0.428053081035614 }, { "epoch": 4.950649350649351, "step": 1906, "train/loss_ctc": 0.41201627254486084, "train/loss_error": 0.3682951331138611, "train/loss_total": 0.37703937292099 }, { "epoch": 4.953246753246753, "step": 1907, "train/loss_ctc": 0.4575275778770447, "train/loss_error": 0.4655357599258423, "train/loss_total": 0.46393415331840515 }, { "epoch": 4.955844155844156, "step": 1908, "train/loss_ctc": 0.3569350838661194, "train/loss_error": 0.42653876543045044, "train/loss_total": 0.4126180410385132 }, { "epoch": 4.958441558441558, "step": 1909, "train/loss_ctc": 0.6006650328636169, "train/loss_error": 0.3361223042011261, "train/loss_total": 0.3890308737754822 }, { "epoch": 4.961038961038961, "grad_norm": 14275.7197265625, "learning_rate": 2.6493506493506495e-07, "loss": 0.4145, "step": 1910 }, { "epoch": 4.961038961038961, "step": 1910, "train/loss_ctc": 0.6299038529396057, "train/loss_error": 0.41153228282928467, "train/loss_total": 0.45520663261413574 }, { "epoch": 4.963636363636364, "step": 1911, "train/loss_ctc": 0.3179910480976105, "train/loss_error": 0.479941725730896, "train/loss_total": 0.4475516080856323 }, { "epoch": 4.966233766233766, "step": 1912, "train/loss_ctc": 0.46699059009552, "train/loss_error": 0.38453376293182373, "train/loss_total": 0.4010251462459564 }, { "epoch": 4.968831168831169, "step": 1913, "train/loss_ctc": 0.40125471353530884, "train/loss_error": 0.35703063011169434, "train/loss_total": 0.3658754527568817 }, { "epoch": 4.9714285714285715, "step": 1914, "train/loss_ctc": 0.38433295488357544, "train/loss_error": 0.36180973052978516, "train/loss_total": 0.3663143813610077 }, { "epoch": 4.974025974025974, "step": 1915, "train/loss_ctc": 0.4995887875556946, "train/loss_error": 0.3907327950000763, "train/loss_total": 0.41250401735305786 }, { "epoch": 4.976623376623377, "step": 1916, "train/loss_ctc": 0.4206276834011078, "train/loss_error": 0.3582029342651367, "train/loss_total": 0.37068790197372437 }, { "epoch": 4.979220779220779, "step": 1917, "train/loss_ctc": 0.44448038935661316, "train/loss_error": 0.28626909852027893, "train/loss_total": 0.3179113566875458 }, { "epoch": 4.9818181818181815, "step": 1918, "train/loss_ctc": 0.6283921599388123, "train/loss_error": 0.3826175928115845, "train/loss_total": 0.43177253007888794 }, { "epoch": 4.984415584415585, "step": 1919, "train/loss_ctc": 0.610848069190979, "train/loss_error": 0.34024232625961304, "train/loss_total": 0.39436349272727966 }, { "epoch": 4.987012987012987, "grad_norm": 16801.103515625, "learning_rate": 1.0909090909090909e-07, "loss": 0.3963, "step": 1920 }, { "epoch": 4.987012987012987, "step": 1920, "train/loss_ctc": 0.4774647355079651, "train/loss_error": 0.37500667572021484, "train/loss_total": 0.3954983055591583 }, { "epoch": 4.989610389610389, "step": 1921, "train/loss_ctc": 0.5116007924079895, "train/loss_error": 0.45227694511413574, "train/loss_total": 0.46414172649383545 }, { "epoch": 4.992207792207792, "step": 1922, "train/loss_ctc": 0.5191208720207214, "train/loss_error": 0.4154641926288605, "train/loss_total": 0.4361955225467682 }, { "epoch": 4.994805194805195, "step": 1923, "train/loss_ctc": 0.525776743888855, "train/loss_error": 0.3379178047180176, "train/loss_total": 0.37548959255218506 }, { "epoch": 4.997402597402598, "step": 1924, "train/loss_ctc": 0.6903512477874756, "train/loss_error": 0.4174653887748718, "train/loss_total": 0.4720425605773926 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.6502947211265564, "val/loss_error": 0.15290777385234833, "val/loss_total": 0.2523851692676544 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.6631668210029602, "val/loss_error": 0.5799373388290405, "val/loss_total": 0.5965832471847534 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.11478433758020401, "val/loss_error": 0.31586939096450806, "val/loss_total": 0.2756524085998535 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.229105606675148, "val/loss_error": 0.15639075636863708, "val/loss_total": 0.17093372344970703 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.13749413192272186, "val/loss_error": 0.21137957274913788, "val/loss_total": 0.1966024935245514 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.12278622388839722, "val/loss_error": 0.36896276473999023, "val/loss_total": 0.31972745060920715 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.23199063539505005, "val/loss_error": 0.4704458713531494, "val/loss_total": 0.42275482416152954 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.3239595293998718, "val/loss_error": 0.30951204895973206, "val/loss_total": 0.31240156292915344 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.4183884263038635, "val/loss_error": 0.360627681016922, "val/loss_total": 0.3721798360347748 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.23248715698719025, "val/loss_error": 0.4173451066017151, "val/loss_total": 0.3803735375404358 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.10391907393932343, "val/loss_error": 0.1880558878183365, "val/loss_total": 0.1712285280227661 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.26372236013412476, "val/loss_error": 0.32752957940101624, "val/loss_total": 0.31476813554763794 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.8064568042755127, "val/loss_error": 0.5086299777030945, "val/loss_total": 0.5681953430175781 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.8164549469947815, "val/loss_error": 0.5065262317657471, "val/loss_total": 0.568511962890625 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.6289647221565247, "val/loss_error": 0.34262317419052124, "val/loss_total": 0.3998914957046509 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.6032286882400513, "val/loss_error": 0.6972810626029968, "val/loss_total": 0.6784706115722656 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.09187769144773483, "val/loss_error": 0.2184145152568817, "val/loss_total": 0.19310715794563293 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.1299634575843811, "val/loss_error": 0.2550605535507202, "val/loss_total": 0.23004113137722015 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.1296682208776474, "val/loss_error": 0.3483753800392151, "val/loss_total": 0.3046339750289917 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.49893778562545776, "val/loss_error": 0.3982142508029938, "val/loss_total": 0.4183589816093445 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.8159948587417603, "val/loss_error": 0.6457027792930603, "val/loss_total": 0.6797611713409424 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.25967609882354736, "val/loss_error": 0.6173902750015259, "val/loss_total": 0.5458474159240723 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.288772314786911, "val/loss_error": 0.4518970251083374, "val/loss_total": 0.4192720949649811 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.14191141724586487, "val/loss_error": 0.2576793134212494, "val/loss_total": 0.23452574014663696 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.2846895754337311, "val/loss_error": 0.5006047487258911, "val/loss_total": 0.4574217200279236 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5001627206802368, "val/loss_error": 0.5600447654724121, "val/loss_total": 0.5480683445930481 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.30296531319618225, "val/loss_error": 0.42431819438934326, "val/loss_total": 0.40004763007164 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.13814088702201843, "val/loss_error": 0.2638839781284332, "val/loss_total": 0.2387353628873825 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.03582386672496796, "val/loss_error": 0.0767209604382515, "val/loss_total": 0.06854154169559479 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.27288225293159485, "val/loss_error": 0.546875536441803, "val/loss_total": 0.49207690358161926 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.12284929305315018, "val/loss_error": 0.13607552647590637, "val/loss_total": 0.13343028724193573 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.11668069660663605, "val/loss_error": 0.2883216142654419, "val/loss_total": 0.253993421792984 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.09969229996204376, "val/loss_error": 0.21388676762580872, "val/loss_total": 0.19104787707328796 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5423118472099304, "val/loss_error": 0.783321738243103, "val/loss_total": 0.7351198196411133 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.08951950818300247, "val/loss_error": 0.40348732471466064, "val/loss_total": 0.34069377183914185 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.17088420689105988, "val/loss_error": 0.5543215870857239, "val/loss_total": 0.47763413190841675 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.6904881000518799, "val/loss_error": 0.6362645626068115, "val/loss_total": 0.64710932970047 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.4612067639827728, "val/loss_error": 0.4902573525905609, "val/loss_total": 0.4844472408294678 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.25377747416496277, "val/loss_error": 0.49611663818359375, "val/loss_total": 0.447648823261261 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5957915186882019, "val/loss_error": 0.6822612881660461, "val/loss_total": 0.6649673581123352 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.2793625593185425, "val/loss_error": 0.25771364569664, "val/loss_total": 0.26204341650009155 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.3350191116333008, "val/loss_error": 0.4277181625366211, "val/loss_total": 0.40917834639549255 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.4772638976573944, "val/loss_error": 1.0240321159362793, "val/loss_total": 0.9146785140037537 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5851114988327026, "val/loss_error": 0.6034877896308899, "val/loss_total": 0.5998125076293945 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.4475010633468628, "val/loss_error": 0.3314063549041748, "val/loss_total": 0.35462531447410583 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.3755498230457306, "val/loss_error": 0.6695669293403625, "val/loss_total": 0.6107634902000427 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5778792500495911, "val/loss_error": 0.43751898407936096, "val/loss_total": 0.46559104323387146 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.34823736548423767, "val/loss_error": 0.5152190923690796, "val/loss_total": 0.4818227291107178 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.4455154240131378, "val/loss_error": 0.19940124452114105, "val/loss_total": 0.24862408638000488 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.6562138199806213, "val/loss_error": 0.9544947743415833, "val/loss_total": 0.8948385715484619 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.3831235468387604, "val/loss_error": 0.8966730237007141, "val/loss_total": 0.7939631342887878 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.3680100440979004, "val/loss_error": 0.6100391149520874, "val/loss_total": 0.561633288860321 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.3583754301071167, "val/loss_error": 0.49149662256240845, "val/loss_total": 0.4648723900318146 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.40126779675483704, "val/loss_error": 0.8697590827941895, "val/loss_total": 0.7760608196258545 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.22149023413658142, "val/loss_error": 0.39145952463150024, "val/loss_total": 0.3574656844139099 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.17794276773929596, "val/loss_error": 0.5295842289924622, "val/loss_total": 0.45925596356391907 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.1085384413599968, "val/loss_error": 0.25839513540267944, "val/loss_total": 0.22842378914356232 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5732911229133606, "val/loss_error": 0.265715628862381, "val/loss_total": 0.3272307217121124 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.6587561964988708, "val/loss_error": 0.6754524111747742, "val/loss_total": 0.6721131801605225 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.03536437451839447, "val/loss_error": 0.13251085579395294, "val/loss_total": 0.11308155953884125 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.7567411065101624, "val/loss_error": 0.6709885597229004, "val/loss_total": 0.6881390810012817 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.57301926612854, "val/loss_error": 0.39529183506965637, "val/loss_total": 0.43083733320236206 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.07780729234218597, "val/loss_error": 0.0865296870470047, "val/loss_total": 0.08478521555662155 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.19986696541309357, "val/loss_error": 0.33145901560783386, "val/loss_total": 0.3051406145095825 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.14966769516468048, "val/loss_error": 0.2055303305387497, "val/loss_total": 0.19435781240463257 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.22932296991348267, "val/loss_error": 0.4233166575431824, "val/loss_total": 0.3845179080963135 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.3186458349227905, "val/loss_error": 0.28708696365356445, "val/loss_total": 0.29339873790740967 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5898680686950684, "val/loss_error": 0.5499787926673889, "val/loss_total": 0.5579566359519958 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5298444032669067, "val/loss_error": 0.389118492603302, "val/loss_total": 0.4172636866569519 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.513893187046051, "val/loss_error": 0.46043524146080017, "val/loss_total": 0.47112685441970825 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.25581544637680054, "val/loss_error": 0.26743218302726746, "val/loss_total": 0.2651088237762451 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5866863131523132, "val/loss_error": 0.5855690836906433, "val/loss_total": 0.5857925415039062 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5859644412994385, "val/loss_error": 0.7163224816322327, "val/loss_total": 0.6902508735656738 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.24186214804649353, "val/loss_error": 0.4175693392753601, "val/loss_total": 0.3824279010295868 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.19488407671451569, "val/loss_error": 0.2964746356010437, "val/loss_total": 0.2761565148830414 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.06159788742661476, "val/loss_error": 0.30907735228538513, "val/loss_total": 0.25958144664764404 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.7841247320175171, "val/loss_error": 0.5653871297836304, "val/loss_total": 0.6091346740722656 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5376716256141663, "val/loss_error": 0.38945773243904114, "val/loss_total": 0.4191005229949951 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5759063363075256, "val/loss_error": 0.36601004004478455, "val/loss_total": 0.4079892933368683 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.0035563695710152388, "val/loss_error": 0.36763373017311096, "val/loss_total": 0.294818252325058 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.43786129355430603, "val/loss_error": 0.33720481395721436, "val/loss_total": 0.3573361039161682 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.02497098222374916, "val/loss_error": 0.11212052404880524, "val/loss_total": 0.09469062089920044 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.2607932388782501, "val/loss_error": 0.32497018575668335, "val/loss_total": 0.3121348023414612 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.38751450181007385, "val/loss_error": 0.447648286819458, "val/loss_total": 0.4356215298175812 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.43410155177116394, "val/loss_error": 0.6312932968139648, "val/loss_total": 0.5918549299240112 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5017808079719543, "val/loss_error": 0.778891384601593, "val/loss_total": 0.7234692573547363 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.04612242057919502, "val/loss_error": 0.11171171069145203, "val/loss_total": 0.09859385341405869 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.4868955612182617, "val/loss_error": 0.5803483128547668, "val/loss_total": 0.5616577863693237 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.25387170910835266, "val/loss_error": 0.27565282583236694, "val/loss_total": 0.2712966203689575 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.45272281765937805, "val/loss_error": 0.38942041993141174, "val/loss_total": 0.4020808935165405 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.2676068842411041, "val/loss_error": 0.5528354644775391, "val/loss_total": 0.4957897365093231 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.8200857639312744, "val/loss_error": 0.4931739270687103, "val/loss_total": 0.558556318283081 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.26992934942245483, "val/loss_error": 0.0941166952252388, "val/loss_total": 0.129279226064682 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.3150992691516876, "val/loss_error": 0.3400745987892151, "val/loss_total": 0.335079550743103 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.5154405236244202, "val/loss_error": 0.523764431476593, "val/loss_total": 0.5220996737480164 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.20549026131629944, "val/loss_error": 0.18917234241962433, "val/loss_total": 0.19243593513965607 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.10733981430530548, "val/loss_error": 0.2395693063735962, "val/loss_total": 0.2131234109401703 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.6431707739830017, "val/loss_error": 0.4878285527229309, "val/loss_total": 0.5188969969749451 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.13839367032051086, "val/loss_error": 0.3303104341030121, "val/loss_total": 0.2919270694255829 }, { "epoch": 5.0, "step": 1925, "val/loss_ctc": 0.19230690598487854, "val/loss_error": 0.27221062779426575, "val/loss_total": 0.25622987747192383 }, { "epoch": 5.0, "eval_eval/f1_0": 0.5210466384887695, "eval_eval/f1_1": 0.9246464967727661, "eval_eval/precision_0": 0.5192743539810181, "eval_eval/precision_1": 0.9251432418823242, "eval_eval/recall_0": 0.5228310227394104, "eval_eval/recall_1": 0.924150288105011, "eval_eval/wer": 0.15024975007655647, "eval_loss": 0.4098331034183502, "eval_runtime": 7.7337, "eval_samples_per_second": 12.93, "eval_steps_per_second": 12.93, "step": 1925 } ], "logging_steps": 10, "max_steps": 1925, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }