Skip to content

A sentence level multi-lingual embedding pre-trained model based on BERT was download from https://huggingface.co/sentence-transformers/paraphrase-multilingual-mpnet-base-v2 for a further fine-turning on million level chinese train-dataset. now we posted the fine-turning model shared with you here. now posted the fine-turning model.

License

Notifications You must be signed in to change notification settings

zaqwsxcderfvbgtyhnmjuik/sentence-bert-fine-turning

Repository files navigation

sentence-bert-fine-turning

A sentence level multi-lingual embedding pre-trained model based on BERT was download from https://huggingface.co/sentence-transformers/paraphrase-multilingual-mpnet-base-v2 for a further fine-turning on million level chinese train-dataset. now posted the fine-turning model.

how to use this fine-turing pre-trained model

step one

Due to the limitation of one file size which must less than 25 MB in github repository. we choose to cut the big model file into many small files for avoiding this problem. So when you git clone the repository from remote end into the local end, the first thing you need to do is to merge the small files into one big file by executing Python script named initialize.py. Note that you should promise that the all files in local reposity are not opened by file explorer, or the process of executing Python script named initialize.py will generate a error deleting small files with a failure.

step two

To execute the Python scripts which are load_model_after_fine_turning.py and load_model_before_fine_turning.py is the usage of this pre-trained model.

About

A sentence level multi-lingual embedding pre-trained model based on BERT was download from https://huggingface.co/sentence-transformers/paraphrase-multilingual-mpnet-base-v2 for a further fine-turning on million level chinese train-dataset. now we posted the fine-turning model shared with you here. now posted the fine-turning model.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages