Skip to content

Commit df153b6

Browse files
trax-robotcopybara-github
authored andcommitted
remove deprecated assertDictContainsSubset
PiperOrigin-RevId: 715816929
1 parent d3b3435 commit df153b6

File tree

1 file changed

+6
-4
lines changed

1 file changed

+6
-4
lines changed

trax/data/tokenizer_test.py

Lines changed: 6 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -78,15 +78,16 @@ def test_corpus_token_counts_split_on_newlines(self):
7878
u"my": 2,
7979
u"pajamas": 2,
8080
}
81-
self.assertDictContainsSubset(expected, token_counts)
81+
self.assertEqual(token_counts, {**token_counts, **expected})
8282
self.assertNotIn(u".\n\n", token_counts)
8383
self.assertNotIn(u"\n", token_counts)
8484

8585
def test_corpus_token_counts_no_split_on_newlines(self):
8686
token_counts = tokenizer.corpus_token_counts(
8787
self.corpus_path, corpus_max_lines=0, split_on_newlines=False)
8888

89-
self.assertDictContainsSubset({u".\n\n": 2, u"\n": 3}, token_counts)
89+
expected_subset = {u".\n\n": 2, u"\n": 3}
90+
self.assertEqual(token_counts, {**token_counts, **expected_subset})
9091

9192
def test_corpus_token_counts_split_with_max_lines(self):
9293
token_counts = tokenizer.corpus_token_counts(
@@ -101,11 +102,12 @@ def test_corpus_token_counts_no_split_with_max_lines(self):
101102

102103
self.assertIn(u"slept", token_counts)
103104
self.assertNotIn(u"Mitch", token_counts)
104-
self.assertDictContainsSubset({
105+
expected_subset = {
105106
u".\n\n": 1,
106107
u"\n": 2,
107108
u".\n": 1
108-
}, token_counts)
109+
}
110+
self.assertEqual(token_counts, {**token_counts, **expected_subset})
109111

110112
def test_vocab_token_counts(self):
111113
token_counts = tokenizer.vocab_token_counts(self.vocab_path, 0)

0 commit comments

Comments
 (0)