About PT #518
Henry-Joen
started this conversation in
General
About PT
#518
Replies: 1 comment 1 reply
-
|
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Dear author, thanks for your hard-working on code updations recently! But I was puzzled about the problem this days, that if I want a model to perform better in a certain field via Chinese, such as the medical field, is it all right that I should do further pre-training on the Chinese-Llama-2 to get the LoRA, instead of on the Chinese-Alpaca-2?
In my understanding, the Chinese-Alpaca-2 is a instructed fine-tuned model via GAtt Method like Llama2-chat, which requires labeled dataset to format inputs between [INST] and [/INST], is thus less sensitive to those unlabeled plain text data.
Beta Was this translation helpful? Give feedback.
All reactions