Skip to content

Commit 244aab3

Browse files
committed
more
1 parent 927fe75 commit 244aab3

File tree

1 file changed

+14
-12
lines changed

1 file changed

+14
-12
lines changed

users/zeyer/experiments/exp2024_04_23_baselines/ctc_recog_ext.py

Lines changed: 14 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -93,18 +93,20 @@ def py():
9393
ctc_model=ctc_model, prior=prior, prior_scale=prior_scale, language_model=lm, lm_scale=lm_scale
9494
)
9595
for name, opts in [
96-
(
97-
"beam1024-beamToken128-cache1024",
98-
{
99-
"n_best": 32,
100-
"beam_size": 1024,
101-
"beam_size_token": 128,
102-
"beam_threshold": 14,
103-
"batch_size": 5_000 * ctc_model.definition.batch_size_factor,
104-
"torch_amp": {"dtype": "bfloat16"},
105-
"lm_state_lru_initial_cache_size": 1024,
106-
},
107-
),
96+
# This takes forever (more than 2h for only the first (longest) seq of the corpora),
97+
# and then at some point runs out of CPU memory (OOM killer kills it).
98+
# (
99+
# "beam1024-beamToken128-cache1024",
100+
# {
101+
# "n_best": 32,
102+
# "beam_size": 1024,
103+
# "beam_size_token": 128,
104+
# "beam_threshold": 14,
105+
# "batch_size": 5_000 * ctc_model.definition.batch_size_factor,
106+
# "torch_amp": {"dtype": "bfloat16"},
107+
# "lm_state_lru_initial_cache_size": 1024,
108+
# },
109+
# ),
108110
( # {"dev-clean": 3.51, "dev-other": 5.79, "test-clean": 3.66, "test-other": 6.27}
109111
"beam16-beamToken16-cache1024",
110112
{

0 commit comments

Comments
 (0)