├── .gitignore ├── LICENSE └── README.md /.gitignore: -------------------------------------------------------------------------------- 1 | # Byte-compiled / optimized / DLL files 2 | __pycache__/ 3 | *.py[cod] 4 | *$py.class 5 | 6 | # C extensions 7 | *.so 8 | 9 | # Distribution / packaging 10 | .Python 11 | build/ 12 | develop-eggs/ 13 | dist/ 14 | downloads/ 15 | eggs/ 16 | .eggs/ 17 | lib/ 18 | lib64/ 19 | parts/ 20 | sdist/ 21 | var/ 22 | wheels/ 23 | share/python-wheels/ 24 | *.egg-info/ 25 | .installed.cfg 26 | *.egg 27 | MANIFEST 28 | 29 | # PyInstaller 30 | # Usually these files are written by a python script from a template 31 | # before PyInstaller builds the exe, so as to inject date/other infos into it. 32 | *.manifest 33 | *.spec 34 | 35 | # Installer logs 36 | pip-log.txt 37 | pip-delete-this-directory.txt 38 | 39 | # Unit test / coverage reports 40 | htmlcov/ 41 | .tox/ 42 | .nox/ 43 | .coverage 44 | .coverage.* 45 | .cache 46 | nosetests.xml 47 | coverage.xml 48 | *.cover 49 | *.py,cover 50 | .hypothesis/ 51 | .pytest_cache/ 52 | cover/ 53 | 54 | # Translations 55 | *.mo 56 | *.pot 57 | 58 | # Django stuff: 59 | *.log 60 | local_settings.py 61 | db.sqlite3 62 | db.sqlite3-journal 63 | 64 | # Flask stuff: 65 | instance/ 66 | .webassets-cache 67 | 68 | # Scrapy stuff: 69 | .scrapy 70 | 71 | # Sphinx documentation 72 | docs/_build/ 73 | 74 | # PyBuilder 75 | .pybuilder/ 76 | target/ 77 | 78 | # Jupyter Notebook 79 | .ipynb_checkpoints 80 | 81 | # IPython 82 | profile_default/ 83 | ipython_config.py 84 | 85 | # pyenv 86 | # For a library or package, you might want to ignore these files since the code is 87 | # intended to run in multiple environments; otherwise, check them in: 88 | # .python-version 89 | 90 | # pipenv 91 | # According to pypa/pipenv#598, it is recommended to include Pipfile.lock in version control. 92 | # However, in case of collaboration, if having platform-specific dependencies or dependencies 93 | # having no cross-platform support, pipenv may install dependencies that don't work, or not 94 | # install all needed dependencies. 95 | #Pipfile.lock 96 | 97 | # poetry 98 | # Similar to Pipfile.lock, it is generally recommended to include poetry.lock in version control. 99 | # This is especially recommended for binary packages to ensure reproducibility, and is more 100 | # commonly ignored for libraries. 101 | # https://python-poetry.org/docs/basic-usage/#commit-your-poetrylock-file-to-version-control 102 | #poetry.lock 103 | 104 | # pdm 105 | # Similar to Pipfile.lock, it is generally recommended to include pdm.lock in version control. 106 | #pdm.lock 107 | # pdm stores project-wide configurations in .pdm.toml, but it is recommended to not include it 108 | # in version control. 109 | # https://pdm.fming.dev/#use-with-ide 110 | .pdm.toml 111 | 112 | # PEP 582; used by e.g. github.com/David-OConnor/pyflow and github.com/pdm-project/pdm 113 | __pypackages__/ 114 | 115 | # Celery stuff 116 | celerybeat-schedule 117 | celerybeat.pid 118 | 119 | # SageMath parsed files 120 | *.sage.py 121 | 122 | # Environments 123 | .env 124 | .venv 125 | env/ 126 | venv/ 127 | ENV/ 128 | env.bak/ 129 | venv.bak/ 130 | 131 | # Spyder project settings 132 | .spyderproject 133 | .spyproject 134 | 135 | # Rope project settings 136 | .ropeproject 137 | 138 | # mkdocs documentation 139 | /site 140 | 141 | # mypy 142 | .mypy_cache/ 143 | .dmypy.json 144 | dmypy.json 145 | 146 | # Pyre type checker 147 | .pyre/ 148 | 149 | # pytype static type analyzer 150 | .pytype/ 151 | 152 | # Cython debug symbols 153 | cython_debug/ 154 | 155 | # PyCharm 156 | # JetBrains specific template is maintained in a separate JetBrains.gitignore that can 157 | # be found at https://github.com/github/gitignore/blob/main/Global/JetBrains.gitignore 158 | # and can be added to the global gitignore or merged into this file. For a more nuclear 159 | # option (not recommended) you can uncomment the following to ignore the entire idea folder. 160 | #.idea/ 161 | -------------------------------------------------------------------------------- /LICENSE: -------------------------------------------------------------------------------- 1 | MIT License 2 | 3 | Copyright (c) 2023 DAMO-NLP-SG 4 | 5 | Permission is hereby granted, free of charge, to any person obtaining a copy 6 | of this software and associated documentation files (the "Software"), to deal 7 | in the Software without restriction, including without limitation the rights 8 | to use, copy, modify, merge, publish, distribute, sublicense, and/or sell 9 | copies of the Software, and to permit persons to whom the Software is 10 | furnished to do so, subject to the following conditions: 11 | 12 | The above copyright notice and this permission notice shall be included in all 13 | copies or substantial portions of the Software. 14 | 15 | THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 16 | IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 17 | FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE 18 | AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 19 | LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, 20 | OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE 21 | SOFTWARE. 22 | -------------------------------------------------------------------------------- /README.md: -------------------------------------------------------------------------------- 1 | # 🏰 LLM Zoo 2 | 3 | As new animal species are being discovered in the world of natural language processing (NLP) 🌍 every day, it becomes necessary to establish a zoo 🦁 to accommodate them. 4 | 5 | This project collects below information of various open- and closed-source LLMs (after the release of ChatGPT): 6 | 7 | - Release time 8 | - Model size 9 | - Languages supported 10 | - Domain 11 | - Training data 12 | - Links to resources: GitHub, HuggingFace, Demo, Paper, Official blog 13 | 14 | ## 📰 News 15 | 16 | - [2023.05.03] First release! We will regularly update 🔄 the repository to keep track of the latest LLMs. We welcome 👐 any contributions to this project. Please feel free to open an issue or submit a pull request to include new LLMs or update the information of existing LLMs 🙏. 17 | 18 | ## 📖 Open-Sourced LLMs 19 | 20 | | Release Time | Model | Version | Size | Backbone | Langs | Domain | Training Data | GitHub | HF | Paper | Demo | Official Blog | 21 | | :------------: | :------------: | :------------------------------: | :--------------: | :------------------: | :--------------: | :--------: | :-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------: | :------------------------------------------------------------------: | :-------------------------------------------------------------------------------: | :------------------------------------------: | :----------------------------------------------------------------: | :---------------------------------------------------------------------------------------------------------------: | 22 | | 2023.02.27 | LLaMA | llama-7b/13b/33b/65b | 7B/13B/33B/65B | \- | en | General |
detail1T tokens (English CommonCrawl, C4, Github, Wikipedia, Gutenberg and Books3, ArXiv, Stack Exchange)
| [[link](https://github.com/facebookresearch/llama)] | [[link](https://huggingface.co/decapoda-research/llama-7b-hf)] | [[link](https://arxiv.org/abs/2302.13971)] | \- | [[link](https://ai.facebook.com/blog/large-language-model-llama-meta-ai/)] | 23 | | 2023.03.13 | Alpaca | alpaca-7b/13b | 7B/13B | LLaMA | en | General |
detail52k instruction-following data generated by InstructGPT [[link](https://github.com/tatsu-lab/stanford_alpaca/blob/main/alpaca_data.json)]
| [[link](https://github.com/tatsu-lab/stanford_alpaca)] | [[link](https://huggingface.co/tatsu-lab/alpaca-7b-wdiff)] | \- | [[link](https://alpaca-ai.ngrok.io/)] | [[link](https://crfm.stanford.edu/2023/03/13/alpaca.html)] | 24 | | 2023.03.13 | Vicuna | vicuna-7b/13b-delta-v1.1 | 7B/13B | LLaMA | en | General |
detail70K samples from sharedGPT
| [[link](https://github.com/lm-sys/FastChat)] | [[link](https://huggingface.co/lmsys/vicuna-7b-delta-v1.1)] | \- | [[link](https://chat.lmsys.org/)] | [[link](https://vicuna.lmsys.org/)] | 25 | | 2023.03.14 | ChatGLM | chatglm-6b | 6B | GLM | zh, en | General |
detailsupervised fine-tuning, feedback bootstrap, and reinforcement learning with human feedback
| [[link](https://github.com/THUDM/ChatGLM-6B)] | [[link](https://huggingface.co/THUDM/chatglm-6b)] | \- | \- | [[link](https://chatglm.cn/blog)] | 26 | | 2023.03.14 | ChatGLM | chatglm-130b | 130B | GLM | zh, en | General |
detailsupervised fine-tuning, feedback bootstrap, and reinforcement learning with human feedback
| [[link](https://github.com/THUDM/GLM-130B)] | \- | [[link](https://arxiv.org/abs/2210.02414)] | [[link](https://chatglm.cn/login?redirect=%2F)] | [[link](https://chatglm.cn/blog)] | 27 | | 2023.03.16 | Guanaco | \- | 7B | LLaMA | ja, zh, en, de | General |
detailmultilingual datasets [[link](https://huggingface.co/datasets/JosephusCheung/GuanacoDataset)]
| [[link](https://github.com/Guanaco-Model/Guanaco-Model.github.io)] | [[link](https://huggingface.co/JosephusCheung/Guanaco)] | \- | \- | \- | 28 | | 2023.03.24 | Dolly | dolly-v1-6b | 6B | GPT-J-6B | en | General |
detail52k stanford alpaca instruction-following data [[link](https://github.com/tatsu-lab/stanford_alpaca/blob/main/alpaca_data.json)]
| \- | [[link](https://huggingface.co/databricks/dolly-v1-6b)] | \- | \- | [[link](https://www.databricks.com/blog/2023/03/24/hello-dolly-democratizing-magic-chatgpt-open-models.html)] | 29 | | 2023.03.24 | ChatDoctor | \- | 7B | LLaMA | en | Medicine |
detail52K stanford alpaca [[link](https://github.com/tatsu-lab/stanford_alpaca/blob/main/alpaca_data.json)], 100K HealthCareMagic [[link](https://drive.google.com/file/d/1lyfqIwlLSClhgrCutWuEe_IACNq6XNUt/view)], 10K icliniq [[link](https://drive.google.com/file/d/1ZKbqgYqWc7DJHs3N9TQYQVPdDQmZaClA/view)], 5K GenMedGPT-5k [[link](https://drive.google.com/file/d/1nDTKZ3wZbZWTkFMBkxlamrzbNz0frugg/view)]
| [[link](https://github.com/Kent0n-Li/ChatDoctor)] | \- | [[link](https://arxiv.org/abs/2303.14070)] | [[link](https://huggingface.co/spaces/kenton-li/chatdoctor_csv)] | \- | 30 | | 2023.03.25 | LuoTuo | Chinese-alpaca-lora | 7B | LLaMA | zh, en | General |
detailTranslated 52k stanford alpaca instruction-following data [[link](https://github.com/tatsu-lab/stanford_alpaca/blob/main/alpaca_data.json)], guanaco [[link](https://huggingface.co/datasets/JosephusCheung/GuanacoDataset)]
| [[link](https://github.com/LC1332/Chinese-alpaca-lora)] | [[link](https://huggingface.co/silk-road/luotuo-lora-7b-1.0)] | \- | \- | \- | 31 | | 2023.03.26 | BELLE | BELLE-7B-0.2M/0.6M/1M/2M | 7B | BLOOMZ-7B1-mt | zh, en | General |
detail0.2M/0.6M/1M/2M Chinese data [[link](https://github.com/LianjiaTech/BELLE/tree/main/data)], 52k stanford alpaca instruction-following data [[link](https://github.com/tatsu-lab/stanford_alpaca/blob/main/alpaca_data.json)]
| [[link](https://github.com/LianjiaTech/BELLE)] | [[link](https://huggingface.co/BelleGroup/BELLE-7B-0.2M)] | [[link](https://arxiv.org/abs/2303.14742)] | \- | \- | 32 | | 2023.03.28 | Linly (伶荔) | Linly-Chinese-LLaMA 7b/13b/33b | 7B/13B/33B | LLaMA | zh | General |
detailChinese-English parallel corpora [[link](https://statmt.org/wmt18/translation-task.html#download)], Chinese Wikipedia, community interaction, news data [[link](https://github.com/CLUEbenchmark/CLUECorpus2020)], scientific literature [[link](https://github.com/ydli-ai/CSL)]
| [[link](https://github.com/CVI-SZU/Linly)] | [[link](https://huggingface.co/P01son/Linly-Chinese-LLaMA-7b-hf)] | \- | \- | \- | 33 | | 2023.03.28 | Linly (伶荔) | Linly-ChatFlow 7b/13b | 7B/13B | LLaMA | zh | General |
detailBELLE [[link](https://github.com/LianjiaTech/BELLE)], pCLUE [[link](https://github.com/CLUEbenchmark/pCLUE)], CSL [[link](https://github.com/ydli-ai/CSL)], GuanacoDataset [[link](https://huggingface.co/datasets/JosephusCheung/GuanacoDataset)], Chain-of-Thought [[link](https://github.com/CVI-SZU/Linly/tree/main/instructions)], news_commentary [[link](https://github.com/CVI-SZU/Linly/tree/main/instructions)], firefly [[link](https://huggingface.co/datasets/YeungNLP/firefly-train-1.1M)]
| [[link](https://github.com/CVI-SZU/Linly)] | [[link](https://huggingface.co/P01son/Linly-ChatFlow-13b-hf)] | \- | \- | [[link](https://zhuanlan.zhihu.com/p/616748134)] | 34 | | 2023.04.01 | BAIZE | baize-7B/13B/30B | 7B/13B/30B | LLaMA | en | General |
detail52K Stanford Alpaca [[link](https://github.com/tatsu-lab/stanford_alpaca/blob/main/alpaca_data.json)], 54K Quora [[link](https://github.com/project-baize/baize-chatbot/blob/main/data/quora_chat_data.json)], 57K StackOverFlow [[link](https://github.com/project-baize/baize-chatbot/blob/main/data/stackoverflow_chat_data.json)]
| [[link](https://github.com/project-baize/baize-chatbot)] | [[link](https://huggingface.co/project-baize/baize-lora-7B)] | [[link](https://arxiv.org/abs/2304.01196)] | [[link](https://huggingface.co/spaces/project-baize/Baize-7B)] | \- | 35 | | 2023.04.03 | Koala | \- | 13B | LLaMA | en | General |
detailShareGPT, HC3 [[link](https://arxiv.org/abs/2301.07597)], OIG [[link](https://laion.ai/blog/oig-dataset/)], Stanford alpaca [[link](https://github.com/tatsu-lab/stanford_alpaca/blob/main/alpaca_data.json)], Anthropic HH [[link](https://huggingface.co/datasets/Anthropic/hh-rlhf)], OpenAI WebGPT [[link](https://huggingface.co/datasets/openai/webgpt_comparisons)], OpenAI Summarization [[link](https://huggingface.co/datasets/openai/summarize_from_feedback)]
| \- | [[link](https://huggingface.co/young-geng/koala)] | \- | [[link](https://chat.lmsys.org/)] | [[link](https://bair.berkeley.edu/blog/2023/04/03/koala/)] | 36 | | 2023.04.03 | BAIZE | baize-healthcare-7b | 7B | LLaMA | en | Medicine |
detail54K Quora [[link](https://github.com/project-baize/baize-chatbot/blob/main/data/quora_chat_data.json)], 47K medical dialogs [[link](https://github.com/project-baize/baize-chatbot/blob/main/data/medical_chat_data.json)]
| [[link](https://github.com/project-baize/baize-chatbot)] | [[link](https://huggingface.co/project-baize/baize-healthcare-lora-7B)] | \- | \- | \- | 37 | | 2023.04.06 | Firefly (流萤) | firefly-1b4/2b6 | 1.4B/2.6B | BLOOM-ZH | zh | General |
detailChinese question-answering pairs [[link](https://huggingface.co/datasets/YeungNLP/firefly-train-1.1M)], [[link](https://huggingface.co/datasets/BelleGroup/train_0.5M_CN)]
| [[link](https://github.com/yangjianxin1/Firefly)] | [[link](https://huggingface.co/YeungNLP/firefly-2b6)] | \- | \- | \- | 38 | | 2023.04.08 | Phoenix | Phoenix-chat-7b | 7B | BLOOMZ | multi | General |
detailconversation data [[link](https://huggingface.co/datasets/FreedomIntelligence/phoenix-sft-data-v1)]
| [[link](https://github.com/FreedomIntelligence/LLMZoo)] | [[link](https://huggingface.co/FreedomIntelligence/phoenix-chat-7b)] | \- | \- | \- | 39 | | 2023.04.09 | Phoenix | Phoenix-inst-chat-7b | 7B | BLOOMZ | multi | General |
detailconversation data [[link](https://huggingface.co/datasets/FreedomIntelligence/phoenix-sft-data-v1)], instruction data
| [[link](https://github.com/FreedomIntelligence/LLMZoo)] | [[link](https://huggingface.co/FreedomIntelligence/phoenix-inst-chat-7b)] | \- | \- | \- | 40 | | 2023.04.10 | Chimera | chimera-chat-7b/13b | 7B/13B | LLaMA | latin | General |
detailconversation data [[link](https://huggingface.co/datasets/FreedomIntelligence/phoenix-sft-data-v1)]
| [[link](https://github.com/FreedomIntelligence/LLMZoo)] | [[link](https://huggingface.co/FreedomIntelligence/chimera-chat-7b-delta)] | \- | \- | \- | 41 | | 2023.04.11 | Chimera | chimera-inst-chat-7b/13b | 7B/13B | LLaMA | latin | General |
detailconversation data [[link](https://huggingface.co/datasets/FreedomIntelligence/phoenix-sft-data-v1)], instruction data
| [[link](https://github.com/FreedomIntelligence/LLMZoo)] | [[link](https://huggingface.co/FreedomIntelligence/chimera-inst-chat-7b-delta)] | \- | \- | \- | 42 | | 2023.04.12 | Dolly | dolly-v2-12b | 12B | pythia-12b | en | General |
detail15k human-generated prompt/response pairs [[link](https://huggingface.co/datasets/databricks/databricks-dolly-15k)]
| [[link](https://github.com/databrickslabs/dolly)] | [[link](https://huggingface.co/databricks/dolly-v2-12b)] | \- | \- | [[link](https://www.databricks.com/blog/2023/04/12/dolly-first-open-commercially-viable-instruction-tuned-llm)] | 43 | | 2023.04.14 | MedAlpaca | medalpaca 7b/13b | 7B/13B | LLaMA | en | Medicine |
detailquestion-answering pairs from flash card, wikidoc, stackexchange and ChatDoctor
| [[link](https://github.com/kbressem/medAlpaca)] | [[link](https://huggingface.co/medalpaca/medalpaca-7b)] | [[link](https://arxiv.org/abs/2304.08247)] | \- | \- | 44 | | 2023.04.19 | BELLE | BELLE-LLaMA-7B/13B-2M | 7B/13B | LLaMA | zh, en | General |
detail2M Chinese data [[link](https://github.com/LianjiaTech/BELLE/tree/main/data)], 52k stanford alpaca instruction-following data [[link](https://github.com/tatsu-lab/stanford_alpaca/blob/main/alpaca_data.json)]
| [[link](https://github.com/LianjiaTech/BELLE)] | [[link](https://huggingface.co/BelleGroup/BELLE-LLaMA-7B-2M-enc)] | [[link](https://arxiv.org/abs/2303.14742)] | \- | \- | 45 | | 2023.04.21 | MOSS | moss-moon-003-base | 16B | CodeGen | zh, en | General |
detail100B Chinese tokens and 20B English tokens
| [[link](https://github.com/OpenLMLab/MOSS)] | [[link](https://huggingface.co/fnlp/moss-moon-003-base)] | \- | [[link](https://moss.fastnlp.top/)] | [[link](https://txsun1997.github.io/blogs/moss.html)] | 46 | | 2023.04.21 | MOSS | moss-moon-003-sft | 16B | moss-moon-003-base | zh, en | General |
detail1.1M multi-turn conversational data (generated from ChatGPT) [[link](https://github.com/OpenLMLab/MOSS/tree/main/SFT_data/conversations/conversation_without_plugins)]
| [[link](https://github.com/OpenLMLab/MOSS)] | [[link](https://huggingface.co/fnlp/moss-moon-003-sft)] | \- | [[link](https://moss.fastnlp.top/)] | [[link](https://txsun1997.github.io/blogs/moss.html)] | 47 | | 2023.04.21 | MOSS | moss-moon-003-sft-plugin | 16B | moss-moon-003-base | zh, en | General |
detail1.1M multi-turn conversational data [[link](https://github.com/OpenLMLab/MOSS/tree/main/SFT_data/conversations/conversation_without_plugins)], 300K plugin-augmented data (generated by InstructGPT) [[link](https://github.com/OpenLMLab/MOSS/tree/main/SFT_data/conversations/conversation_with_plugins)]
| [[link](https://github.com/OpenLMLab/MOSS)] | [[link](https://huggingface.co/fnlp/moss-moon-003-sft-plugin)] | \- | [[link](https://moss.fastnlp.top/)] | [[link](https://txsun1997.github.io/blogs/moss.html)] | 48 | | 2023.04.22 | HuggingChat | oasst-sft-6-llama-30b | 30B | LLaMA | multi | General |
detailhuman-generated, human-annotated assistant-style conversation corpus consisting of 161k messages in 35 languages [[link](https://huggingface.co/datasets/OpenAssistant/oasst1)]
| [[link](https://github.com/LAION-AI/Open-Assistant)] | [[link](https://huggingface.co/OpenAssistant/oasst-sft-6-llama-30b-xor)] | \- | [[link](https://huggingface.co/chat/)] | \- | 49 | | 2023.06.19 | KnowLM | zhixi-13b | 13B | LLaMA | zh, en | General |
detailhuman-generated, machine-generated and Knowledge Graph-generated in Chinese and English [[link](https://huggingface.co/datasets/OpenAssistant/oasst1)]
| [[link](https://github.com/zjunlp/KnowLM)] | [[link](https://huggingface.co/zjunlp/zhixi-13b-diff)] | \- | \- | \- | 50 | | 2023.06.21 | BayLing(百聆) | BayLing-7b/13b | 7B/13B | LLaMA | zh, en | General |
detail160K human-generated, machine-generated multi-turn interactive translation corpus, alpaca instructions and sharegpt conversations [[link](https://github.com/ictnlp/BayLing)]
| [[link](https://github.com/ictnlp/BayLing)] | [[link](https://huggingface.co/ICTNLP/bayling-13b-v1.1)] | [[link](https://arxiv.org/abs/2306.10968)] | [[link](http://nlp.ict.ac.cn/bayling/demo)] | [[link](http://nlp.ict.ac.cn/bayling)] | 51 | | 2023.07.18 | LLaMA 2 | llama-2-7b/13b/70b-(chat) | 7B/13B/70B | \- | en | General |
detail2T tokens (Most in English, a new mix of data from publicly available sources)
| [[link](https://github.com/facebookresearch/llama)] | [[link](https://huggingface.co/meta-llama)] | [[link](https://scontent-hkg4-1.xx.fbcdn.net/v/t39.2365-6/10000000_662098952474184_2584067087619170692_n.pdf?_nc_cat=105&ccb=1-7&_nc_sid=3c67a6&_nc_ohc=SvLw337Y0sQAX87v4v0&_nc_ht=scontent-hkg4-1.xx&oh=00_AfDsUaFec_5aYbIxirPYP2RaDQTDGrpukyDQrbI0wB1trA&oe=64CE38FF)] | \- | [[link](https://ai.meta.com/llama/)] | 52 | 53 | 54 | ## 📕 Closed-Sourced LLMs 55 | 56 | | Release Time | Model | Version | Size | Langs | Domain | Demo | Official Blog | Paper | 57 | | :----------: | :---: | :-----: | :---:| :---: | :----: | :----: | :---: | :---: | 58 | | 2022.11.30 | ChatGPT | gpt-3.5-turbo | \- | multi | general | [[link](https://chat.openai.com/)] | [[link](https://openai.com/blog/chatgpt)] | \- | 59 | | 2023.03.14 | Claude | Claude Instant
Claude-v1 | \- | multi | general | [[link](http://www.poe.com/)] | [[link](https://www.anthropic.com/index/introducing-claude)] | \- | 60 | | 2023.03.14 | GPT | gpt-4 | \- | multi | general | [[link](https://chat.openai.com/)] | [[link](https://openai.com/research/gpt-4)] | [[link](https://arxiv.org/abs/2303.08774)] | 61 | | 2023.03.16 | Ernie Bot (文心一言) | \- | \- | zh, en | general | [[link](https://yiyan.baidu.com/)] | [[link](https://yiyan.baidu.com/welcome)] | \- | 62 | | 2023.03.21 | Bard | \- | \- | multi | general | [[link](https://bard.google.com/)] | [[link](https://blog.google/technology/ai/bard-google-ai-search-updates/)] | \- | 63 | | 2023.03.30 | BloombergGPT | \- | 50B | en | finance | \- | [[link](https://www.bloomberg.com/company/press/bloomberggpt-50-billion-parameter-llm-tuned-finance/)] | [[link](https://arxiv.org/abs/2303.17564)] | 64 | | 2023.04.11 | Tongyi Qianwen (通义千问) | \- | \- | multi | general | [[link](https://tongyi.aliyun.com/)] | [[link](https://tongyi.aliyun.com/)] | \- | 65 | | 2023.07.07 | OmModel(欧姆大模型) | \- | \- | multi | general | [[link](https://om.linker.cc)] | [[link](https://om.linker.cc)] | \- | 66 | | 2023.07.11 | Claude 2 | Claude-v2 | \- | multi | general | \- | [[link](https://www.anthropic.com/index/claude-2)] | [[link](https://www-files.anthropic.com/production/images/Model-Card-Claude-2.pdf)] | 67 | 68 | ## 🏗 TODO List 69 | 70 | - [x] Include open-sourced LLMs 71 | - [x] Include closed-sourced LLMs 72 | - [ ] Include a systematic review of common training data 73 | - [ ] Include interesting use cases of various LLMs 74 | - [ ] Performance of LLMs on various evaluation tasks 75 | 76 | ## 📝 Citation 77 | 78 | If you find this repository useful, please consider citing. 79 | 80 | ``` 81 | @software{li2023llmzoo, 82 | title = {LLM Zoo} 83 | author = {Li, Xingxuan and Zhang, Wenxuan and Bing, Lidong}, 84 | url = {https://github.com/DAMO-NLP-SG/LLM-Zoo}, 85 | year = {2023} 86 | } 87 | ``` 88 | --------------------------------------------------------------------------------