Weights version | v1.1 | v0 |
---|---|---|
Link | 7B, 13B | 7B, 13B |
Separator | </s> |
### |
FastChat PyPI package compatibility | >= v0.2.1 | <= v0.1.10 |
FastChat source code compatibility | after tag v0.2.1 | tag v0.1.10 |
Major updates of weights v1.1
- Refactor the tokenization and separator. In Vicuna v1.1, the separator has been changed from
###
to the EOS token</s>
. This change makes it easier to determine the generation stop criteria and enables better compatibility with other libraries. - Fix the supervised fine-tuning loss computation for better model quality.
A chat between a user and an assistant.
USER: Hello!
ASSISTANT: Hello!</s>
USER: How are you?
ASSISTANT: I am good.</s>
See a full prompt template here.
A chat between a human and an assistant.
### Human: Hello!
### Assistant: Hello!
### Human: How are you?
### Assistant: I am good.
See the full prompt template here.
There are some frequently asked tokenizer issues (lm-sys#408). Some of them are not only related to FastChat or Vicuna weights but are also related to how you convert the base llama model.
We suggest that you use transformers>=4.28.0
and redo the weight conversion for the base llama model.
After applying the delta, you should have a file named special_tokens_map.json
in your converted weight folder for either v0 or v1.1.
The contents of this file should be the same as this file: https://huggingface.co/lmsys/vicuna-13b-delta-v0/blob/main/special_tokens_map.json.
If the file is not present, please copy the special_tokens_map.json
and tokenizer_config.json
files from https://huggingface.co/lmsys/vicuna-13b-delta-v0/tree/main to your converted weight folder. This works for both v0 and v1.1.