Chinchilla: Training Compute-Optimal Large Language Models |
https://arxiv.org/abs/2203.15556 |
https://cartinoe5930.tistory.com/entry/%EC%A7%80%EA%B8%88-%EA%B9%8C%EC%A7%80%EC%9D%98-LM-Scaling-Law%EC%97%90%EB%8A%94-%EB%AC%B8%EC%A0%9C%EC%A0%90%EC%9D%B4-%EC%9E%88%EB%8B%A4-%F0%9F%98%B6%E2%80%8D%F0%9F%8C%AB%EF%B8%8F-Chinchilla-Training-Compute-Optimal-Large-Language-Models-%EB%85%BC%EB%AC%B8-%EB%A6%AC%EB%B7%B0 |
Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling |
https://arxiv.org/abs/2304.01373 |
No plan! |
LIMA: Less Is More for Alignment |
https://arxiv.org/abs/2305.11206 |
https://cartinoe5930.tistory.com/entry/LIMA-Less-Is-More-for-Alignment-%EB%85%BC%EB%AC%B8-%EB%A6%AC%EB%B7%B0 |
LLaMA: Open and Efficient Foundation Language Models |
https://arxiv.org/abs/2302.13971 |
https://cartinoe5930.tistory.com/entry/LLaMA-Open-and-Efficient-Foundation-Language-Models-%EB%85%BC%EB%AC%B8-%EB%A6%AC%EB%B7%B0 |
WizardLM: Empowering Large Language Models to Follow Complex Instructions |
https://arxiv.org/abs/2304.12244 |
https://cartinoe5930.tistory.com/entry/Open-domain-instruction%EC%9D%98-%ED%9A%A8%EA%B3%BC-%F0%9F%AA%84-WizardLM-Empowering-Large-Language-Models-to-Follow-Complex-Instructions-%EB%85%BC%EB%AC%B8-%EB%A6%AC%EB%B7%B0 |
WizardCoder: Empowering Code Large Language Models with Evol-Instruct |
https://arxiv.org/abs/2306.08568 |
https://huggingface.co/WizardLM/WizardCoder-15B-V1.0 |
WizardMath: Empowering Mathematical Reasoning for Large Language Models via Reinforced Evol-Instruct |
https://arxiv.org/abs/2308.09583 |
https://huggingface.co/WizardLM/WizardMath-70B-V1.0 |
Alpaca: A Strong, Replicable Instruction-Following Model |
https://crfm.stanford.edu/2023/03/13/alpaca.html |
https://cartinoe5930.tistory.com/entry/Alpaca-A-Strong-Replicable-Instruction-Following-Model-%EB%A6%AC%EB%B7%B0 |
Vicuna: An Open-Source Chatbot Impressing GPT-4 |
https://lmsys.org/blog/2023-03-30-vicuna/ |
https://cartinoe5930.tistory.com/entry/Vicuna-An-Open-Source-Chatbot-Impressing-GPT-4-%EB%A6%AC%EB%B7%B0 |
Koala: A Dialogue Model for Academic Research |
https://bair.berkeley.edu/blog/2023/04/03/koala/ |
https://cartinoe5930.tistory.com/entry/%EC%A4%91%EC%9A%94%ED%95%9C-%EA%B1%B4-%EA%BA%BE%EC%9D%B4%EC%A7%80-%EC%95%8A%EB%8A%94-high-quality-data-Koala%F0%9F%90%A8-A-Dialogue-Model-for-Academic-Researc |
Baize: An Open-Source Chat Model with Parameter-Efficient Tuning on Self-Chat Data |
https://arxiv.org/abs/2304.01196 |
https://cartinoe5930.tistory.com/entry/%F0%9F%90%B2Baize-An-Open-Source-Chat-Model-with-Parameter-Efficient-Tuning-on-Self-Chat-Data-%EB%85%BC%EB%AC%B8-%EB%A6%AC%EB%B7%B0 |
Scaling Data-Constrained Language Models |
https://arxiv.org/abs/2305.16264 |
https://www.youtube.com/watch?v=TK0-sitkCMw&pp=ygUgaHR0cHM6Ly9hcnhpdi5vcmcvYWJzLzIzMDUuMTYyNjQ%3D |
Falcon & RefinedWeb |
https://arxiv.org/abs/2306.01116 |
https://cartinoe5930.tistory.com/entry/Open-LLM-Leaderboard%EB%A5%BC-%ED%9C%A9%EC%93%B4-Falcon%F0%9F%A6%85-LLM-Falcon-RefinedWeb |
Orca: Progressive Learning from Complex Explanation Traces of GPT-4 |
https://arxiv.org/pdf/2306.02707 |
https://cartinoe5930.tistory.com/entry/%F0%9F%90%ACOrca-Progressive-Learning-from-Complex-Explanation-Traces-of-GPT-4-%EB%85%BC%EB%AC%B8-%EB%A6%AC%EB%B7%B0 |
phi-1: Text Books Are All You Need |
https://arxiv.org/abs/2306.11644 |
https://cartinoe5930.tistory.com/entry/%ED%95%84%EC%9A%94%ED%95%9C-%EA%B1%B4-%EC%98%A4%EC%A7%81-%EA%B5%90%EA%B3%BC%EC%84%9C-%EC%88%98%EC%A4%80%EC%9D%98-%EB%8D%B0%EC%9D%B4%ED%84%B0%EB%BF%90-%F0%9F%93%96-phi-1-Textbooks-Are-All-You-Need-%EB%85%BC%EB%AC%B8-%EB%A6%AC%EB%B7%B0 |
AlpaGasus: Training a Better Alpaca with Fewer Data |
https://arxiv.org/abs/2307.08701 |
Will be uploaded later! |
Llama 2: Open Foundation and Fine-Tuned Chat Models |
https://arxiv.org/abs/2307.09288 |
https://cartinoe5930.tistory.com/entry/The-hopes-of-researchers-Open-source-%F0%9F%A4%97-%EC%97%B0%EA%B5%AC%EC%9E%90%EB%93%A4%EC%9D%98-%ED%9D%AC%EB%A7%9D-Open-source-%F0%9F%A4%97 |
Platypus: Quick, Cheap, and Powerful Refinement of LLMs |
https://arxiv.org/abs/2308.07317 |
Will be uploaded later! |
Code Llama: Open Foundation Models for Code |
https://arxiv.org/abs/2308.12950 |
No plan |
FLM-101B: An Open LLM and How to Train It with $100K Budget |
https://arxiv.org/pdf/2309.03852 |
No plan! |
Textbooks are All You Need II: phi-1.5 technical report |
https://arxiv.org/abs/2309.05463 |
https://huggingface.co/microsoft/phi-1_5 |
OpenChat: Advancing Open-Source Language Models with Mixed-Quality Data |
https://arxiv.org/abs/2309.11235 |
https://github.com/imoneoi/openchat |
Mistral 7B |
https://arxiv.org/abs/2310.06825 |
https://mistral.ai/news/announcing-mistral-7b/ |
Prometheus: Inducing Fine-grained Evaluation Capability in Language Models |
https://arxiv.org/abs/2310.08491 |
https://huggingface.co/papers/2310.08491#652a8e7f30355beba68c1be6 |
Zephyr: Direct Distillation of LM Alignment |
https://arxiv.org/abs/2310.16944 |
https://www.youtube.com/watch?v=TkZBg3mKsIo |
Orca2: Teaching Small Language Models How to Reason |
https://arxiv.org/abs/2311.11045 |
https://www.microsoft.com/en-us/research/blog/orca-2-teaching-small-language-models-how-to-reason/ |
The Falcon Series of Open Language Models |
https://arxiv.org/abs/2311.16867 |
No plan! |
SOLAR 10.7B: Scaling Large Language Models with Simple yet Effective Depth Up-Scaling |
https://arxiv.org/abs/2312.15166 |
No plan! |