@@ -93,18 +93,20 @@ def py():
93
93
ctc_model = ctc_model , prior = prior , prior_scale = prior_scale , language_model = lm , lm_scale = lm_scale
94
94
)
95
95
for name , opts in [
96
- (
97
- "beam1024-beamToken128-cache1024" ,
98
- {
99
- "n_best" : 32 ,
100
- "beam_size" : 1024 ,
101
- "beam_size_token" : 128 ,
102
- "beam_threshold" : 14 ,
103
- "batch_size" : 5_000 * ctc_model .definition .batch_size_factor ,
104
- "torch_amp" : {"dtype" : "bfloat16" },
105
- "lm_state_lru_initial_cache_size" : 1024 ,
106
- },
107
- ),
96
+ # This takes forever (more than 2h for only the first (longest) seq of the corpora),
97
+ # and then at some point runs out of CPU memory (OOM killer kills it).
98
+ # (
99
+ # "beam1024-beamToken128-cache1024",
100
+ # {
101
+ # "n_best": 32,
102
+ # "beam_size": 1024,
103
+ # "beam_size_token": 128,
104
+ # "beam_threshold": 14,
105
+ # "batch_size": 5_000 * ctc_model.definition.batch_size_factor,
106
+ # "torch_amp": {"dtype": "bfloat16"},
107
+ # "lm_state_lru_initial_cache_size": 1024,
108
+ # },
109
+ # ),
108
110
( # {"dev-clean": 3.51, "dev-other": 5.79, "test-clean": 3.66, "test-other": 6.27}
109
111
"beam16-beamToken16-cache1024" ,
110
112
{
0 commit comments