Skip to content

Latest commit

 

History

History
59 lines (46 loc) · 7.21 KB

internlm_20b.md

File metadata and controls

59 lines (46 loc) · 7.21 KB

InternLM-20B

Introduction

InternLM-20B was pre-trained on over 2.3T Tokens containing high-quality English, Chinese, and code data. Additionally, the Chat version has undergone SFT and RLHF training, enabling it to better and more securely meet users' needs.

In terms of model structure, InternLM-20B opted for a deeper architecture, with a depth set at 60 layers. This surpasses the conventional 7B and 13B models that utilize 32 or 40 layers. When parameters are limited, increasing the number of layers can enhance the model's overall capability. Furthermore, compared to InternLM-7B, the pre-training data used for InternLM-20B underwent higher quality cleansing and was supplemented with data rich in knowledge and designed for reinforcing understanding and reasoning capabilities. As a result, it exhibits significant improvements in understanding, reasoning, mathematical, and programming abilities—all of which test the technical proficiency of language models. Overall, InternLM-20B features the following characteristics:

  • Outstanding overall performance
  • Strong utility invocation capability
  • Supports a 16k context length (Through inference extrapolation)
  • Better value alignment.

Model Zoo

Model Transformers(HF) ModelScope(HF) OpenXLab(HF) OpenXLab(Original) Release Date
InternLM Chat 20B 🤗internlm/internlm-chat-20b Shanghai_AI_Laboratory/internlm-chat-20b Open in OpenXLab Open in OpenXLab 2023-12-12
InternLM 20B 🤗internlm/internlm-20b Shanghai_AI_Laboratory/internlm-20b Open in OpenXLab Open in OpenXLab 2023-09-20

Performance Evaluation

On the 5 capability dimensions proposed by OpenCompass, InternLM-20B has achieved excellent results (the bolded scores represent the best performances within the 13B-33B parameter range).

Capability Llama-13B Llama2-13B Baichuan2-13B InternLM-20B Llama-33B Llama-65B Llama2-70B
Language 42.5 47 47.5 55 44.6 47.1 51.6
Knowledge 58.2 58.3 48.9 60.1 64 66 67.7
Understanding 45.5 50.9 58.1 67.3 50.6 54.2 60.8
Reasoning 42.7 43.6 44.2 54.9 46.4 49.8 55
Examination 37.3 45.2 51.8 62.5 47.4 49.7 57.3
Overall 43.8 47.3 49.4 59.2 48.9 51.9 57.4

The table below compares the performance of mainstream open-source models on some influential and typical datasets.

Benchmarks Llama-13B Llama2-13B Baichuan2-13B InternLM-20B Llama-33B Llama-65B Llama2-70B
Examination MMLU 47.73 54.99 59.55 62.05 58.73 63.71 69.75
C-Eval (val) 31.83 41.4 59.01 58.8 37.47 40.36 50.13
AGI-Eval 22.03 30.93 37.37 44.58 33.53 33.92 40.02
Knowledge BoolQ 78.75 82.42 67 87.46 84.43 86.61 87.74
TriviaQA 52.47 59.36 46.61 57.26 66.24 69.79 70.71
NaturalQuestions 20.17 24.85 16.32 25.15 30.89 33.41 34.16
Understanding CMRC 9.26 31.59 29.85 68.78 14.17 34.73 43.74
CSL 55 58.75 63.12 65.62 57.5 59.38 60
RACE (middle) 53.41 63.02 68.94 86.35 64.55 72.35 81.55
RACE (high) 47.63 58.86 67.18 83.28 62.61 68.01 79.93
XSum 20.37 23.37 25.23 35.54 20.55 19.91 25.38
Reasoning WinoGrande 64.64 64.01 67.32 69.38 66.85 69.38 69.77
BBH 37.93 45.62 48.98 52.51 49.98 58.38 64.91
GSM8K 20.32 29.57 52.62 52.62 42.3 54.44 63.31
PIQA 79.71 79.76 78.07 80.25 81.34 82.15 82.54
Programming HumanEval 14.02 18.9 17.07 25.61 17.68 18.9 26.22
MBPP 20.6 26.8 30.8 35.6 28.4 33.6 39.6

Overall, InternLM-20B comprehensively outperforms open-source models in the 13B parameter range in terms of overall capabilities, and on inference evaluation sets, it approaches or even surpasses the performance of Llama-65B.

  • The evaluation results were obtained from OpenCompass 20230920.
  • The evaluation data may have numerical differences due to the version iteration of OpenCompass, so please refer to the latest evaluation results of OpenCompass.