A small experiment using both Mecab and Tinysegmenter to create a tokenized list of Japanese sentences in JSON, taken from the Tatoeba corpus.
-
Updated
Mar 25, 2021 - Python
A small experiment using both Mecab and Tinysegmenter to create a tokenized list of Japanese sentences in JSON, taken from the Tatoeba corpus.
An app for automating the creation of cloze (fill-in-the-blank) vocabulary and grammar activities. Powered by the Tatoeba corpus.
Add a description, image, and links to the tatoeba-corpus topic page so that developers can more easily learn about it.
To associate your repository with the tatoeba-corpus topic, visit your repo's landing page and select "manage topics."