├── .gitignore
├── LICENSE
└── README.md
/.gitignore:
--------------------------------------------------------------------------------
1 | # Byte-compiled / optimized / DLL files
2 | __pycache__/
3 | *.py[cod]
4 | *$py.class
5 |
6 | # C extensions
7 | *.so
8 |
9 | # Distribution / packaging
10 | .Python
11 | build/
12 | develop-eggs/
13 | dist/
14 | downloads/
15 | eggs/
16 | .eggs/
17 | lib/
18 | lib64/
19 | parts/
20 | sdist/
21 | var/
22 | wheels/
23 | share/python-wheels/
24 | *.egg-info/
25 | .installed.cfg
26 | *.egg
27 | MANIFEST
28 |
29 | # PyInstaller
30 | # Usually these files are written by a python script from a template
31 | # before PyInstaller builds the exe, so as to inject date/other infos into it.
32 | *.manifest
33 | *.spec
34 |
35 | # Installer logs
36 | pip-log.txt
37 | pip-delete-this-directory.txt
38 |
39 | # Unit test / coverage reports
40 | htmlcov/
41 | .tox/
42 | .nox/
43 | .coverage
44 | .coverage.*
45 | .cache
46 | nosetests.xml
47 | coverage.xml
48 | *.cover
49 | *.py,cover
50 | .hypothesis/
51 | .pytest_cache/
52 | cover/
53 |
54 | # Translations
55 | *.mo
56 | *.pot
57 |
58 | # Django stuff:
59 | *.log
60 | local_settings.py
61 | db.sqlite3
62 | db.sqlite3-journal
63 |
64 | # Flask stuff:
65 | instance/
66 | .webassets-cache
67 |
68 | # Scrapy stuff:
69 | .scrapy
70 |
71 | # Sphinx documentation
72 | docs/_build/
73 |
74 | # PyBuilder
75 | .pybuilder/
76 | target/
77 |
78 | # Jupyter Notebook
79 | .ipynb_checkpoints
80 |
81 | # IPython
82 | profile_default/
83 | ipython_config.py
84 |
85 | # pyenv
86 | # For a library or package, you might want to ignore these files since the code is
87 | # intended to run in multiple environments; otherwise, check them in:
88 | # .python-version
89 |
90 | # pipenv
91 | # According to pypa/pipenv#598, it is recommended to include Pipfile.lock in version control.
92 | # However, in case of collaboration, if having platform-specific dependencies or dependencies
93 | # having no cross-platform support, pipenv may install dependencies that don't work, or not
94 | # install all needed dependencies.
95 | #Pipfile.lock
96 |
97 | # poetry
98 | # Similar to Pipfile.lock, it is generally recommended to include poetry.lock in version control.
99 | # This is especially recommended for binary packages to ensure reproducibility, and is more
100 | # commonly ignored for libraries.
101 | # https://python-poetry.org/docs/basic-usage/#commit-your-poetrylock-file-to-version-control
102 | #poetry.lock
103 |
104 | # pdm
105 | # Similar to Pipfile.lock, it is generally recommended to include pdm.lock in version control.
106 | #pdm.lock
107 | # pdm stores project-wide configurations in .pdm.toml, but it is recommended to not include it
108 | # in version control.
109 | # https://pdm.fming.dev/#use-with-ide
110 | .pdm.toml
111 |
112 | # PEP 582; used by e.g. github.com/David-OConnor/pyflow and github.com/pdm-project/pdm
113 | __pypackages__/
114 |
115 | # Celery stuff
116 | celerybeat-schedule
117 | celerybeat.pid
118 |
119 | # SageMath parsed files
120 | *.sage.py
121 |
122 | # Environments
123 | .env
124 | .venv
125 | env/
126 | venv/
127 | ENV/
128 | env.bak/
129 | venv.bak/
130 |
131 | # Spyder project settings
132 | .spyderproject
133 | .spyproject
134 |
135 | # Rope project settings
136 | .ropeproject
137 |
138 | # mkdocs documentation
139 | /site
140 |
141 | # mypy
142 | .mypy_cache/
143 | .dmypy.json
144 | dmypy.json
145 |
146 | # Pyre type checker
147 | .pyre/
148 |
149 | # pytype static type analyzer
150 | .pytype/
151 |
152 | # Cython debug symbols
153 | cython_debug/
154 |
155 | # PyCharm
156 | # JetBrains specific template is maintained in a separate JetBrains.gitignore that can
157 | # be found at https://github.com/github/gitignore/blob/main/Global/JetBrains.gitignore
158 | # and can be added to the global gitignore or merged into this file. For a more nuclear
159 | # option (not recommended) you can uncomment the following to ignore the entire idea folder.
160 | #.idea/
161 |
--------------------------------------------------------------------------------
/LICENSE:
--------------------------------------------------------------------------------
1 | MIT License
2 |
3 | Copyright (c) 2023 DAMO-NLP-SG
4 |
5 | Permission is hereby granted, free of charge, to any person obtaining a copy
6 | of this software and associated documentation files (the "Software"), to deal
7 | in the Software without restriction, including without limitation the rights
8 | to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
9 | copies of the Software, and to permit persons to whom the Software is
10 | furnished to do so, subject to the following conditions:
11 |
12 | The above copyright notice and this permission notice shall be included in all
13 | copies or substantial portions of the Software.
14 |
15 | THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 | IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 | FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
18 | AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19 | LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
20 | OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
21 | SOFTWARE.
22 |
--------------------------------------------------------------------------------
/README.md:
--------------------------------------------------------------------------------
1 | # 🏰 LLM Zoo
2 |
3 | As new animal species are being discovered in the world of natural language processing (NLP) 🌍 every day, it becomes necessary to establish a zoo 🦁 to accommodate them.
4 |
5 | This project collects below information of various open- and closed-source LLMs (after the release of ChatGPT):
6 |
7 | - Release time
8 | - Model size
9 | - Languages supported
10 | - Domain
11 | - Training data
12 | - Links to resources: GitHub, HuggingFace, Demo, Paper, Official blog
13 |
14 | ## 📰 News
15 |
16 | - [2023.05.03] First release! We will regularly update 🔄 the repository to keep track of the latest LLMs. We welcome 👐 any contributions to this project. Please feel free to open an issue or submit a pull request to include new LLMs or update the information of existing LLMs 🙏.
17 |
18 | ## 📖 Open-Sourced LLMs
19 |
20 | | Release Time | Model | Version | Size | Backbone | Langs | Domain | Training Data | GitHub | HF | Paper | Demo | Official Blog |
21 | | :------------: | :------------: | :------------------------------: | :--------------: | :------------------: | :--------------: | :--------: | :-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------: | :------------------------------------------------------------------: | :-------------------------------------------------------------------------------: | :------------------------------------------: | :----------------------------------------------------------------: | :---------------------------------------------------------------------------------------------------------------: |
22 | | 2023.02.27 | LLaMA | llama-7b/13b/33b/65b | 7B/13B/33B/65B | \- | en | General | detail
1T tokens (English CommonCrawl, C4, Github, Wikipedia, Gutenberg and Books3, ArXiv, Stack Exchange) | [[link](https://github.com/facebookresearch/llama)] | [[link](https://huggingface.co/decapoda-research/llama-7b-hf)] | [[link](https://arxiv.org/abs/2302.13971)] | \- | [[link](https://ai.facebook.com/blog/large-language-model-llama-meta-ai/)] |
23 | | 2023.03.13 | Alpaca | alpaca-7b/13b | 7B/13B | LLaMA | en | General | detail
52k instruction-following data generated by InstructGPT [[link](https://github.com/tatsu-lab/stanford_alpaca/blob/main/alpaca_data.json)] | [[link](https://github.com/tatsu-lab/stanford_alpaca)] | [[link](https://huggingface.co/tatsu-lab/alpaca-7b-wdiff)] | \- | [[link](https://alpaca-ai.ngrok.io/)] | [[link](https://crfm.stanford.edu/2023/03/13/alpaca.html)] |
24 | | 2023.03.13 | Vicuna | vicuna-7b/13b-delta-v1.1 | 7B/13B | LLaMA | en | General | detail
70K samples from sharedGPT | [[link](https://github.com/lm-sys/FastChat)] | [[link](https://huggingface.co/lmsys/vicuna-7b-delta-v1.1)] | \- | [[link](https://chat.lmsys.org/)] | [[link](https://vicuna.lmsys.org/)] |
25 | | 2023.03.14 | ChatGLM | chatglm-6b | 6B | GLM | zh, en | General | detail
supervised fine-tuning, feedback bootstrap, and reinforcement learning with human feedback | [[link](https://github.com/THUDM/ChatGLM-6B)] | [[link](https://huggingface.co/THUDM/chatglm-6b)] | \- | \- | [[link](https://chatglm.cn/blog)] |
26 | | 2023.03.14 | ChatGLM | chatglm-130b | 130B | GLM | zh, en | General | detail
supervised fine-tuning, feedback bootstrap, and reinforcement learning with human feedback | [[link](https://github.com/THUDM/GLM-130B)] | \- | [[link](https://arxiv.org/abs/2210.02414)] | [[link](https://chatglm.cn/login?redirect=%2F)] | [[link](https://chatglm.cn/blog)] |
27 | | 2023.03.16 | Guanaco | \- | 7B | LLaMA | ja, zh, en, de | General | detail
multilingual datasets [[link](https://huggingface.co/datasets/JosephusCheung/GuanacoDataset)] | [[link](https://github.com/Guanaco-Model/Guanaco-Model.github.io)] | [[link](https://huggingface.co/JosephusCheung/Guanaco)] | \- | \- | \- |
28 | | 2023.03.24 | Dolly | dolly-v1-6b | 6B | GPT-J-6B | en | General | detail
52k stanford alpaca instruction-following data [[link](https://github.com/tatsu-lab/stanford_alpaca/blob/main/alpaca_data.json)] | \- | [[link](https://huggingface.co/databricks/dolly-v1-6b)] | \- | \- | [[link](https://www.databricks.com/blog/2023/03/24/hello-dolly-democratizing-magic-chatgpt-open-models.html)] |
29 | | 2023.03.24 | ChatDoctor | \- | 7B | LLaMA | en | Medicine | detail
52K stanford alpaca [[link](https://github.com/tatsu-lab/stanford_alpaca/blob/main/alpaca_data.json)], 100K HealthCareMagic [[link](https://drive.google.com/file/d/1lyfqIwlLSClhgrCutWuEe_IACNq6XNUt/view)], 10K icliniq [[link](https://drive.google.com/file/d/1ZKbqgYqWc7DJHs3N9TQYQVPdDQmZaClA/view)], 5K GenMedGPT-5k [[link](https://drive.google.com/file/d/1nDTKZ3wZbZWTkFMBkxlamrzbNz0frugg/view)] | [[link](https://github.com/Kent0n-Li/ChatDoctor)] | \- | [[link](https://arxiv.org/abs/2303.14070)] | [[link](https://huggingface.co/spaces/kenton-li/chatdoctor_csv)] | \- |
30 | | 2023.03.25 | LuoTuo | Chinese-alpaca-lora | 7B | LLaMA | zh, en | General | detail
Translated 52k stanford alpaca instruction-following data [[link](https://github.com/tatsu-lab/stanford_alpaca/blob/main/alpaca_data.json)], guanaco [[link](https://huggingface.co/datasets/JosephusCheung/GuanacoDataset)] | [[link](https://github.com/LC1332/Chinese-alpaca-lora)] | [[link](https://huggingface.co/silk-road/luotuo-lora-7b-1.0)] | \- | \- | \- |
31 | | 2023.03.26 | BELLE | BELLE-7B-0.2M/0.6M/1M/2M | 7B | BLOOMZ-7B1-mt | zh, en | General | detail
0.2M/0.6M/1M/2M Chinese data [[link](https://github.com/LianjiaTech/BELLE/tree/main/data)], 52k stanford alpaca instruction-following data [[link](https://github.com/tatsu-lab/stanford_alpaca/blob/main/alpaca_data.json)] | [[link](https://github.com/LianjiaTech/BELLE)] | [[link](https://huggingface.co/BelleGroup/BELLE-7B-0.2M)] | [[link](https://arxiv.org/abs/2303.14742)] | \- | \- |
32 | | 2023.03.28 | Linly (伶荔) | Linly-Chinese-LLaMA 7b/13b/33b | 7B/13B/33B | LLaMA | zh | General | detail
Chinese-English parallel corpora [[link](https://statmt.org/wmt18/translation-task.html#download)], Chinese Wikipedia, community interaction, news data [[link](https://github.com/CLUEbenchmark/CLUECorpus2020)], scientific literature [[link](https://github.com/ydli-ai/CSL)] | [[link](https://github.com/CVI-SZU/Linly)] | [[link](https://huggingface.co/P01son/Linly-Chinese-LLaMA-7b-hf)] | \- | \- | \- |
33 | | 2023.03.28 | Linly (伶荔) | Linly-ChatFlow 7b/13b | 7B/13B | LLaMA | zh | General | detail
BELLE [[link](https://github.com/LianjiaTech/BELLE)], pCLUE [[link](https://github.com/CLUEbenchmark/pCLUE)], CSL [[link](https://github.com/ydli-ai/CSL)], GuanacoDataset [[link](https://huggingface.co/datasets/JosephusCheung/GuanacoDataset)], Chain-of-Thought [[link](https://github.com/CVI-SZU/Linly/tree/main/instructions)], news_commentary [[link](https://github.com/CVI-SZU/Linly/tree/main/instructions)], firefly [[link](https://huggingface.co/datasets/YeungNLP/firefly-train-1.1M)] | [[link](https://github.com/CVI-SZU/Linly)] | [[link](https://huggingface.co/P01son/Linly-ChatFlow-13b-hf)] | \- | \- | [[link](https://zhuanlan.zhihu.com/p/616748134)] |
34 | | 2023.04.01 | BAIZE | baize-7B/13B/30B | 7B/13B/30B | LLaMA | en | General | detail
52K Stanford Alpaca [[link](https://github.com/tatsu-lab/stanford_alpaca/blob/main/alpaca_data.json)], 54K Quora [[link](https://github.com/project-baize/baize-chatbot/blob/main/data/quora_chat_data.json)], 57K StackOverFlow [[link](https://github.com/project-baize/baize-chatbot/blob/main/data/stackoverflow_chat_data.json)] | [[link](https://github.com/project-baize/baize-chatbot)] | [[link](https://huggingface.co/project-baize/baize-lora-7B)] | [[link](https://arxiv.org/abs/2304.01196)] | [[link](https://huggingface.co/spaces/project-baize/Baize-7B)] | \- |
35 | | 2023.04.03 | Koala | \- | 13B | LLaMA | en | General | detail
ShareGPT, HC3 [[link](https://arxiv.org/abs/2301.07597)], OIG [[link](https://laion.ai/blog/oig-dataset/)], Stanford alpaca [[link](https://github.com/tatsu-lab/stanford_alpaca/blob/main/alpaca_data.json)], Anthropic HH [[link](https://huggingface.co/datasets/Anthropic/hh-rlhf)], OpenAI WebGPT [[link](https://huggingface.co/datasets/openai/webgpt_comparisons)], OpenAI Summarization [[link](https://huggingface.co/datasets/openai/summarize_from_feedback)] | \- | [[link](https://huggingface.co/young-geng/koala)] | \- | [[link](https://chat.lmsys.org/)] | [[link](https://bair.berkeley.edu/blog/2023/04/03/koala/)] |
36 | | 2023.04.03 | BAIZE | baize-healthcare-7b | 7B | LLaMA | en | Medicine | detail
54K Quora [[link](https://github.com/project-baize/baize-chatbot/blob/main/data/quora_chat_data.json)], 47K medical dialogs [[link](https://github.com/project-baize/baize-chatbot/blob/main/data/medical_chat_data.json)] | [[link](https://github.com/project-baize/baize-chatbot)] | [[link](https://huggingface.co/project-baize/baize-healthcare-lora-7B)] | \- | \- | \- |
37 | | 2023.04.06 | Firefly (流萤) | firefly-1b4/2b6 | 1.4B/2.6B | BLOOM-ZH | zh | General | detail
Chinese question-answering pairs [[link](https://huggingface.co/datasets/YeungNLP/firefly-train-1.1M)], [[link](https://huggingface.co/datasets/BelleGroup/train_0.5M_CN)] | [[link](https://github.com/yangjianxin1/Firefly)] | [[link](https://huggingface.co/YeungNLP/firefly-2b6)] | \- | \- | \- |
38 | | 2023.04.08 | Phoenix | Phoenix-chat-7b | 7B | BLOOMZ | multi | General | detail
conversation data [[link](https://huggingface.co/datasets/FreedomIntelligence/phoenix-sft-data-v1)] | [[link](https://github.com/FreedomIntelligence/LLMZoo)] | [[link](https://huggingface.co/FreedomIntelligence/phoenix-chat-7b)] | \- | \- | \- |
39 | | 2023.04.09 | Phoenix | Phoenix-inst-chat-7b | 7B | BLOOMZ | multi | General | detail
conversation data [[link](https://huggingface.co/datasets/FreedomIntelligence/phoenix-sft-data-v1)], instruction data | [[link](https://github.com/FreedomIntelligence/LLMZoo)] | [[link](https://huggingface.co/FreedomIntelligence/phoenix-inst-chat-7b)] | \- | \- | \- |
40 | | 2023.04.10 | Chimera | chimera-chat-7b/13b | 7B/13B | LLaMA | latin | General | detail
conversation data [[link](https://huggingface.co/datasets/FreedomIntelligence/phoenix-sft-data-v1)] | [[link](https://github.com/FreedomIntelligence/LLMZoo)] | [[link](https://huggingface.co/FreedomIntelligence/chimera-chat-7b-delta)] | \- | \- | \- |
41 | | 2023.04.11 | Chimera | chimera-inst-chat-7b/13b | 7B/13B | LLaMA | latin | General | detail
conversation data [[link](https://huggingface.co/datasets/FreedomIntelligence/phoenix-sft-data-v1)], instruction data | [[link](https://github.com/FreedomIntelligence/LLMZoo)] | [[link](https://huggingface.co/FreedomIntelligence/chimera-inst-chat-7b-delta)] | \- | \- | \- |
42 | | 2023.04.12 | Dolly | dolly-v2-12b | 12B | pythia-12b | en | General | detail
15k human-generated prompt/response pairs [[link](https://huggingface.co/datasets/databricks/databricks-dolly-15k)] | [[link](https://github.com/databrickslabs/dolly)] | [[link](https://huggingface.co/databricks/dolly-v2-12b)] | \- | \- | [[link](https://www.databricks.com/blog/2023/04/12/dolly-first-open-commercially-viable-instruction-tuned-llm)] |
43 | | 2023.04.14 | MedAlpaca | medalpaca 7b/13b | 7B/13B | LLaMA | en | Medicine | detail
question-answering pairs from flash card, wikidoc, stackexchange and ChatDoctor | [[link](https://github.com/kbressem/medAlpaca)] | [[link](https://huggingface.co/medalpaca/medalpaca-7b)] | [[link](https://arxiv.org/abs/2304.08247)] | \- | \- |
44 | | 2023.04.19 | BELLE | BELLE-LLaMA-7B/13B-2M | 7B/13B | LLaMA | zh, en | General | detail
2M Chinese data [[link](https://github.com/LianjiaTech/BELLE/tree/main/data)], 52k stanford alpaca instruction-following data [[link](https://github.com/tatsu-lab/stanford_alpaca/blob/main/alpaca_data.json)] | [[link](https://github.com/LianjiaTech/BELLE)] | [[link](https://huggingface.co/BelleGroup/BELLE-LLaMA-7B-2M-enc)] | [[link](https://arxiv.org/abs/2303.14742)] | \- | \- |
45 | | 2023.04.21 | MOSS | moss-moon-003-base | 16B | CodeGen | zh, en | General | detail
100B Chinese tokens and 20B English tokens | [[link](https://github.com/OpenLMLab/MOSS)] | [[link](https://huggingface.co/fnlp/moss-moon-003-base)] | \- | [[link](https://moss.fastnlp.top/)] | [[link](https://txsun1997.github.io/blogs/moss.html)] |
46 | | 2023.04.21 | MOSS | moss-moon-003-sft | 16B | moss-moon-003-base | zh, en | General | detail
1.1M multi-turn conversational data (generated from ChatGPT) [[link](https://github.com/OpenLMLab/MOSS/tree/main/SFT_data/conversations/conversation_without_plugins)] | [[link](https://github.com/OpenLMLab/MOSS)] | [[link](https://huggingface.co/fnlp/moss-moon-003-sft)] | \- | [[link](https://moss.fastnlp.top/)] | [[link](https://txsun1997.github.io/blogs/moss.html)] |
47 | | 2023.04.21 | MOSS | moss-moon-003-sft-plugin | 16B | moss-moon-003-base | zh, en | General | detail
1.1M multi-turn conversational data [[link](https://github.com/OpenLMLab/MOSS/tree/main/SFT_data/conversations/conversation_without_plugins)], 300K plugin-augmented data (generated by InstructGPT) [[link](https://github.com/OpenLMLab/MOSS/tree/main/SFT_data/conversations/conversation_with_plugins)] | [[link](https://github.com/OpenLMLab/MOSS)] | [[link](https://huggingface.co/fnlp/moss-moon-003-sft-plugin)] | \- | [[link](https://moss.fastnlp.top/)] | [[link](https://txsun1997.github.io/blogs/moss.html)] |
48 | | 2023.04.22 | HuggingChat | oasst-sft-6-llama-30b | 30B | LLaMA | multi | General | detail
human-generated, human-annotated assistant-style conversation corpus consisting of 161k messages in 35 languages [[link](https://huggingface.co/datasets/OpenAssistant/oasst1)] | [[link](https://github.com/LAION-AI/Open-Assistant)] | [[link](https://huggingface.co/OpenAssistant/oasst-sft-6-llama-30b-xor)] | \- | [[link](https://huggingface.co/chat/)] | \- |
49 | | 2023.06.19 | KnowLM | zhixi-13b | 13B | LLaMA | zh, en | General | detail
human-generated, machine-generated and Knowledge Graph-generated in Chinese and English [[link](https://huggingface.co/datasets/OpenAssistant/oasst1)] | [[link](https://github.com/zjunlp/KnowLM)] | [[link](https://huggingface.co/zjunlp/zhixi-13b-diff)] | \- | \- | \- |
50 | | 2023.06.21 | BayLing(百聆) | BayLing-7b/13b | 7B/13B | LLaMA | zh, en | General | detail
160K human-generated, machine-generated multi-turn interactive translation corpus, alpaca instructions and sharegpt conversations [[link](https://github.com/ictnlp/BayLing)] | [[link](https://github.com/ictnlp/BayLing)] | [[link](https://huggingface.co/ICTNLP/bayling-13b-v1.1)] | [[link](https://arxiv.org/abs/2306.10968)] | [[link](http://nlp.ict.ac.cn/bayling/demo)] | [[link](http://nlp.ict.ac.cn/bayling)] |
51 | | 2023.07.18 | LLaMA 2 | llama-2-7b/13b/70b-(chat) | 7B/13B/70B | \- | en | General | detail
2T tokens (Most in English, a new mix of data from publicly available sources) | [[link](https://github.com/facebookresearch/llama)] | [[link](https://huggingface.co/meta-llama)] | [[link](https://scontent-hkg4-1.xx.fbcdn.net/v/t39.2365-6/10000000_662098952474184_2584067087619170692_n.pdf?_nc_cat=105&ccb=1-7&_nc_sid=3c67a6&_nc_ohc=SvLw337Y0sQAX87v4v0&_nc_ht=scontent-hkg4-1.xx&oh=00_AfDsUaFec_5aYbIxirPYP2RaDQTDGrpukyDQrbI0wB1trA&oe=64CE38FF)] | \- | [[link](https://ai.meta.com/llama/)] |
52 |
53 |
54 | ## 📕 Closed-Sourced LLMs
55 |
56 | | Release Time | Model | Version | Size | Langs | Domain | Demo | Official Blog | Paper |
57 | | :----------: | :---: | :-----: | :---:| :---: | :----: | :----: | :---: | :---: |
58 | | 2022.11.30 | ChatGPT | gpt-3.5-turbo | \- | multi | general | [[link](https://chat.openai.com/)] | [[link](https://openai.com/blog/chatgpt)] | \- |
59 | | 2023.03.14 | Claude | Claude InstantClaude-v1 | \- | multi | general | [[link](http://www.poe.com/)] | [[link](https://www.anthropic.com/index/introducing-claude)] | \- |
60 | | 2023.03.14 | GPT | gpt-4 | \- | multi | general | [[link](https://chat.openai.com/)] | [[link](https://openai.com/research/gpt-4)] | [[link](https://arxiv.org/abs/2303.08774)] |
61 | | 2023.03.16 | Ernie Bot (文心一言) | \- | \- | zh, en | general | [[link](https://yiyan.baidu.com/)] | [[link](https://yiyan.baidu.com/welcome)] | \- |
62 | | 2023.03.21 | Bard | \- | \- | multi | general | [[link](https://bard.google.com/)] | [[link](https://blog.google/technology/ai/bard-google-ai-search-updates/)] | \- |
63 | | 2023.03.30 | BloombergGPT | \- | 50B | en | finance | \- | [[link](https://www.bloomberg.com/company/press/bloomberggpt-50-billion-parameter-llm-tuned-finance/)] | [[link](https://arxiv.org/abs/2303.17564)] |
64 | | 2023.04.11 | Tongyi Qianwen (通义千问) | \- | \- | multi | general | [[link](https://tongyi.aliyun.com/)] | [[link](https://tongyi.aliyun.com/)] | \- |
65 | | 2023.07.07 | OmModel(欧姆大模型) | \- | \- | multi | general | [[link](https://om.linker.cc)] | [[link](https://om.linker.cc)] | \- |
66 | | 2023.07.11 | Claude 2 | Claude-v2 | \- | multi | general | \- | [[link](https://www.anthropic.com/index/claude-2)] | [[link](https://www-files.anthropic.com/production/images/Model-Card-Claude-2.pdf)] |
67 |
68 | ## 🏗 TODO List
69 |
70 | - [x] Include open-sourced LLMs
71 | - [x] Include closed-sourced LLMs
72 | - [ ] Include a systematic review of common training data
73 | - [ ] Include interesting use cases of various LLMs
74 | - [ ] Performance of LLMs on various evaluation tasks
75 |
76 | ## 📝 Citation
77 |
78 | If you find this repository useful, please consider citing.
79 |
80 | ```
81 | @software{li2023llmzoo,
82 | title = {LLM Zoo}
83 | author = {Li, Xingxuan and Zhang, Wenxuan and Bing, Lidong},
84 | url = {https://github.com/DAMO-NLP-SG/LLM-Zoo},
85 | year = {2023}
86 | }
87 | ```
88 |
--------------------------------------------------------------------------------