├── LICENSE ├── README.md ├── assets └── overview.png └── scripts └── trans.py /LICENSE: -------------------------------------------------------------------------------- 1 | Apache License 2 | Version 2.0, January 2004 3 | http://www.apache.org/licenses/ 4 | 5 | TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION 6 | 7 | 1. Definitions. 8 | 9 | "License" shall mean the terms and conditions for use, reproduction, 10 | and distribution as defined by Sections 1 through 9 of this document. 11 | 12 | "Licensor" shall mean the copyright owner or entity authorized by 13 | the copyright owner that is granting the License. 14 | 15 | "Legal Entity" shall mean the union of the acting entity and all 16 | other entities that control, are controlled by, or are under common 17 | control with that entity. For the purposes of this definition, 18 | "control" means (i) the power, direct or indirect, to cause the 19 | direction or management of such entity, whether by contract or 20 | otherwise, or (ii) ownership of fifty percent (50%) or more of the 21 | outstanding shares, or (iii) beneficial ownership of such entity. 22 | 23 | "You" (or "Your") shall mean an individual or Legal Entity 24 | exercising permissions granted by this License. 25 | 26 | "Source" form shall mean the preferred form for making modifications, 27 | including but not limited to software source code, documentation 28 | source, and configuration files. 29 | 30 | "Object" form shall mean any form resulting from mechanical 31 | transformation or translation of a Source form, including but 32 | not limited to compiled object code, generated documentation, 33 | and conversions to other media types. 34 | 35 | "Work" shall mean the work of authorship, whether in Source or 36 | Object form, made available under the License, as indicated by a 37 | copyright notice that is included in or attached to the work 38 | (an example is provided in the Appendix below). 39 | 40 | "Derivative Works" shall mean any work, whether in Source or Object 41 | form, that is based on (or derived from) the Work and for which the 42 | editorial revisions, annotations, elaborations, or other modifications 43 | represent, as a whole, an original work of authorship. For the purposes 44 | of this License, Derivative Works shall not include works that remain 45 | separable from, or merely link (or bind by name) to the interfaces of, 46 | the Work and Derivative Works thereof. 47 | 48 | "Contribution" shall mean any work of authorship, including 49 | the original version of the Work and any modifications or additions 50 | to that Work or Derivative Works thereof, that is intentionally 51 | submitted to Licensor for inclusion in the Work by the copyright owner 52 | or by an individual or Legal Entity authorized to submit on behalf of 53 | the copyright owner. For the purposes of this definition, "submitted" 54 | means any form of electronic, verbal, or written communication sent 55 | to the Licensor or its representatives, including but not limited to 56 | communication on electronic mailing lists, source code control systems, 57 | and issue tracking systems that are managed by, or on behalf of, the 58 | Licensor for the purpose of discussing and improving the Work, but 59 | excluding communication that is conspicuously marked or otherwise 60 | designated in writing by the copyright owner as "Not a Contribution." 61 | 62 | "Contributor" shall mean Licensor and any individual or Legal Entity 63 | on behalf of whom a Contribution has been received by Licensor and 64 | subsequently incorporated within the Work. 65 | 66 | 2. Grant of Copyright License. Subject to the terms and conditions of 67 | this License, each Contributor hereby grants to You a perpetual, 68 | worldwide, non-exclusive, no-charge, royalty-free, irrevocable 69 | copyright license to reproduce, prepare Derivative Works of, 70 | publicly display, publicly perform, sublicense, and distribute the 71 | Work and such Derivative Works in Source or Object form. 72 | 73 | 3. Grant of Patent License. Subject to the terms and conditions of 74 | this License, each Contributor hereby grants to You a perpetual, 75 | worldwide, non-exclusive, no-charge, royalty-free, irrevocable 76 | (except as stated in this section) patent license to make, have made, 77 | use, offer to sell, sell, import, and otherwise transfer the Work, 78 | where such license applies only to those patent claims licensable 79 | by such Contributor that are necessarily infringed by their 80 | Contribution(s) alone or by combination of their Contribution(s) 81 | with the Work to which such Contribution(s) was submitted. If You 82 | institute patent litigation against any entity (including a 83 | cross-claim or counterclaim in a lawsuit) alleging that the Work 84 | or a Contribution incorporated within the Work constitutes direct 85 | or contributory patent infringement, then any patent licenses 86 | granted to You under this License for that Work shall terminate 87 | as of the date such litigation is filed. 88 | 89 | 4. Redistribution. You may reproduce and distribute copies of the 90 | Work or Derivative Works thereof in any medium, with or without 91 | modifications, and in Source or Object form, provided that You 92 | meet the following conditions: 93 | 94 | (a) You must give any other recipients of the Work or 95 | Derivative Works a copy of this License; and 96 | 97 | (b) You must cause any modified files to carry prominent notices 98 | stating that You changed the files; and 99 | 100 | (c) You must retain, in the Source form of any Derivative Works 101 | that You distribute, all copyright, patent, trademark, and 102 | attribution notices from the Source form of the Work, 103 | excluding those notices that do not pertain to any part of 104 | the Derivative Works; and 105 | 106 | (d) If the Work includes a "NOTICE" text file as part of its 107 | distribution, then any Derivative Works that You distribute must 108 | include a readable copy of the attribution notices contained 109 | within such NOTICE file, excluding those notices that do not 110 | pertain to any part of the Derivative Works, in at least one 111 | of the following places: within a NOTICE text file distributed 112 | as part of the Derivative Works; within the Source form or 113 | documentation, if provided along with the Derivative Works; or, 114 | within a display generated by the Derivative Works, if and 115 | wherever such third-party notices normally appear. The contents 116 | of the NOTICE file are for informational purposes only and 117 | do not modify the License. You may add Your own attribution 118 | notices within Derivative Works that You distribute, alongside 119 | or as an addendum to the NOTICE text from the Work, provided 120 | that such additional attribution notices cannot be construed 121 | as modifying the License. 122 | 123 | You may add Your own copyright statement to Your modifications and 124 | may provide additional or different license terms and conditions 125 | for use, reproduction, or distribution of Your modifications, or 126 | for any such Derivative Works as a whole, provided Your use, 127 | reproduction, and distribution of the Work otherwise complies with 128 | the conditions stated in this License. 129 | 130 | 5. Submission of Contributions. Unless You explicitly state otherwise, 131 | any Contribution intentionally submitted for inclusion in the Work 132 | by You to the Licensor shall be under the terms and conditions of 133 | this License, without any additional terms or conditions. 134 | Notwithstanding the above, nothing herein shall supersede or modify 135 | the terms of any separate license agreement you may have executed 136 | with Licensor regarding such Contributions. 137 | 138 | 6. Trademarks. This License does not grant permission to use the trade 139 | names, trademarks, service marks, or product names of the Licensor, 140 | except as required for reasonable and customary use in describing the 141 | origin of the Work and reproducing the content of the NOTICE file. 142 | 143 | 7. Disclaimer of Warranty. Unless required by applicable law or 144 | agreed to in writing, Licensor provides the Work (and each 145 | Contributor provides its Contributions) on an "AS IS" BASIS, 146 | WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or 147 | implied, including, without limitation, any warranties or conditions 148 | of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A 149 | PARTICULAR PURPOSE. You are solely responsible for determining the 150 | appropriateness of using or redistributing the Work and assume any 151 | risks associated with Your exercise of permissions under this License. 152 | 153 | 8. Limitation of Liability. In no event and under no legal theory, 154 | whether in tort (including negligence), contract, or otherwise, 155 | unless required by applicable law (such as deliberate and grossly 156 | negligent acts) or agreed to in writing, shall any Contributor be 157 | liable to You for damages, including any direct, indirect, special, 158 | incidental, or consequential damages of any character arising as a 159 | result of this License or out of the use or inability to use the 160 | Work (including but not limited to damages for loss of goodwill, 161 | work stoppage, computer failure or malfunction, or any and all 162 | other commercial damages or losses), even if such Contributor 163 | has been advised of the possibility of such damages. 164 | 165 | 9. Accepting Warranty or Additional Liability. While redistributing 166 | the Work or Derivative Works thereof, You may choose to offer, 167 | and charge a fee for, acceptance of support, warranty, indemnity, 168 | or other liability obligations and/or rights consistent with this 169 | License. However, in accepting such obligations, You may act only 170 | on Your own behalf and on Your sole responsibility, not on behalf 171 | of any other Contributor, and only if You agree to indemnify, 172 | defend, and hold each Contributor harmless for any liability 173 | incurred by, or claims asserted against, such Contributor by reason 174 | of your accepting any such warranty or additional liability. 175 | 176 | END OF TERMS AND CONDITIONS 177 | 178 | APPENDIX: How to apply the Apache License to your work. 179 | 180 | To apply the Apache License to your work, attach the following 181 | boilerplate notice, with the fields enclosed by brackets "[]" 182 | replaced with your own identifying information. (Don't include 183 | the brackets!) The text should be enclosed in the appropriate 184 | comment syntax for the file format. We also recommend that a 185 | file or class name and description of purpose be included on the 186 | same "printed page" as the copyright notice for easier 187 | identification within third-party archives. 188 | 189 | Copyright [yyyy] [name of copyright owner] 190 | 191 | Licensed under the Apache License, Version 2.0 (the "License"); 192 | you may not use this file except in compliance with the License. 193 | You may obtain a copy of the License at 194 | 195 | http://www.apache.org/licenses/LICENSE-2.0 196 | 197 | Unless required by applicable law or agreed to in writing, software 198 | distributed under the License is distributed on an "AS IS" BASIS, 199 | WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 200 | See the License for the specific language governing permissions and 201 | limitations under the License. 202 | -------------------------------------------------------------------------------- /README.md: -------------------------------------------------------------------------------- 1 | # Interactive NLP Papers🤖+👨‍💼📚🤗⚒️🌏 2 | 3 | Must-read papers on [Interactive NLP](https://arxiv.org/abs/2305.13246): a new paradigm of NLP. 4 | 5 | 6 | 7 | ## Content 8 | 9 | - [What is Interactive NLP?](#what-is-interactive-nlp) 10 | - [Keywords Convention](#keywords-convention) 11 | - [Paper](#paper) 12 | - [Surveys and Position Papers](#surveys-and-position-papers) 13 | - [Human-LM Interaction](#human-lm-interaction) 14 | - [KB-LM Interaction](#kb-lm-interaction) 15 | - [Model/Tool-LM Interaction](#model-tool-lm-interaction) 16 | - [Environment-LM Interaction](#environment-lm-interaction) 17 | - [Evaluation](#evaluation) 18 | - [Application](#application) 19 | - [Related Projects](#related-projects) 20 | - [Contribution](#contribution) 21 | - [Contributors](#contributors) 22 | - [Contributing to this paper list](#contributing-to-this-paper-list) 23 | 24 | 25 | 26 | ## What is Interactive NLP? 27 | 28 | **Interactive Natural Language Processing (iNLP)** considers language models as agents capable of observing, acting, and receiving feedback in a loop with external objects such as humans, knowledge bases, tools, models, and environments, where: 29 | 30 | - **Observation** involves all kinds of inputs to language models. 31 | - **Action** involves all kinds of outputs of language models such as text generation, requesting for external objects, text editing, etc. 32 | - **Feedback** involves feedback messages passed from external objects to language models such as scoring from humans. 33 | 34 | 35 | 36 | In iNLP, language models can interact with four kinds of objects (i.e., entities): 37 | 38 | - interact with **humans** for better understanding and addressing user needs, personalizing responses, aligning with human values, and improving the overall user experience; 39 | - interact with **knowledge bases** for enriching language representations with factual knowledge, enhancing the contextual relevance of responses, and dynamically leveraging external information to generate more accurate and informed responses; 40 | - interact with **models/tools** for effectively decomposing and addressing complex tasks, leveraging specialized expertise for specific subtasks, and fostering the simulation of social behaviors; 41 | - interact with **environments** for learning grounded representations of language, and effectively tackling embodied tasks such as reasoning, planning, and decision-making in response to environmental observations. 42 | 43 | 44 | 45 | ## Keywords Convention 46 | 47 | ![img](https://img.shields.io/badge/-InstructGPT-blue) The abbreviation of the work. 48 | 49 | ![img](https://img.shields.io/badge/-Prompting%20Chaining-orange) The interaction method used by the work. 50 | 51 | ![img](https://img.shields.io/badge/-formal%20language-lightgrey) The interaction interface used by the work. 52 | 53 | ![img](https://img.shields.io/badge/Other-green) Other important information of the work. 54 | 55 | 56 | 57 | ## Paper 58 | 59 | ### 🔭Surveys and Position Papers 60 | 61 | - **[Interactive Natural Language Processing](https://arxiv.org/abs/2305.13246)**, 2023.05 ![img](https://img.shields.io/badge/Interactive_Learning-green) 62 | 63 | *Zekun Wang, Ge Zhang, Kexin Yang, Ning Shi, Wangchunshu Zhou, Shaochun Hao, Guangzheng Xiong, Yizhi Li, Mong Yuan Sim, Xiuying Chen, Qingqing Zhu, Zhenzhu Yang, Adam Nik, Qi Liu, Chenghua Lin, Shi Wang, Ruibo Liu, Wenhu Chen, Ke Xu, Dayiheng Liu, Yike Guo, Jie Fu*. 64 | 65 | - **[Tool Learning with Foundation Models](https://arxiv.org/abs/2304.08354)**, 2023.04 ![img](https://img.shields.io/badge/-Tool--use-green) 66 | 67 | *Yujia Qin, Shengding Hu, Yankai Lin, Weize Chen, Ning Ding, Ganqu Cui, Zheni Zeng, Yufei Huang, Chaojun Xiao, Chi Han, Yi Ren Fung, Yusheng Su, Huadong Wang, Cheng Qian, Runchu Tian, Kunlun Zhu, Shihao Liang, Xingyu Shen, Bokai Xu, Zhen Zhang, Yining Ye, Bowen Li, Ziwei Tang, Jing Yi, Yuzhang Zhu, Zhenning Dai, Lan Yan, Xin Cong, Yaxi Lu, Weilin Zhao, Yuxiang Huang, Junxi Yan, Xu Han, Xian Sun, Dahai Li, Jason Phang, Cheng Yang, Tongshuang Wu, Heng Ji, Zhiyuan Liu, Maosong Sun*. 68 | 69 | - **[Augmented Language Models: a Survey](https://arxiv.org/abs/2302.07842)**, 2023.02 ![img](https://img.shields.io/badge/-Reasoning-green) ![img](https://img.shields.io/badge/-Tool--use-green) 70 | 71 | *Grégoire Mialon, Roberto Dessì, Maria Lomeli, Christoforos Nalmpantis, Ram Pasunuru, Roberta Raileanu, Baptiste Rozière, Timo Schick, Jane Dwivedi-Yu, Asli Celikyilmaz, Edouard Grave, Yann LeCun, Thomas Scialom*. 72 | 73 | - **[Foundation Models for Decision Making: Problems, Methods, and Opportunities](https://arxiv.org/abs/2303.04129)**, 2023.03 ![img](https://img.shields.io/badge/-Tool--use-green) ![img](https://img.shields.io/badge/-Decision%20Making-green) 74 | 75 | *Sherry Yang, Ofir Nachum, Yilun Du, Jason Wei, Pieter Abbeel, Dale Schuurmans*. 76 | 77 | ### 👨‍💼Human-LM Interaction 78 | 79 | - **[Training language models to follow instructions with human feedback](https://arxiv.org/abs/2203.02155)**, 2022.03 ![img](https://img.shields.io/badge/-InstructGPT-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/Feedback-green) 80 | 81 | *Long Ouyang, Jeff Wu, Xu Jiang, Diogo Almeida, Carroll L. Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, John Schulman, Jacob Hilton, Fraser Kelton, Luke Miller, Maddie Simens, Amanda Askell, Peter Welinder, Paul Christiano, Jan Leike, Ryan Lowe*. 82 | 83 | - **[Deep reinforcement learning from human preferences](https://arxiv.org/abs/1706.03741)**, 2017.06 ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/Feedback-green) 84 | 85 | *Paul Christiano, Jan Leike, Tom B. Brown, Miljan Martic, Shane Legg, Dario Amodei*. 86 | 87 | - **[Improving alignment of dialogue agents via targeted human judgements](https://arxiv.org/abs/2209.14375)**, 2022.09 ![img](https://img.shields.io/badge/-Sparrow-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/Feedback-green) 88 | 89 | *Amelia Glaese, Nat McAleese, Maja Trębacz, John Aslanides, Vlad Firoiu, Timo Ewalds, Maribeth Rauh, Laura Weidinger, Martin Chadwick, Phoebe Thacker, Lucy Campbell-Gillingham, Jonathan Uesato, Po-Sen Huang, Ramona Comanescu, Fan Yang, Abigail See, Sumanth Dathathri, Rory Greig, Charlie Chen, Doug Fritz, Jaume Sanchez Elias, Richard Green, Soňa Mokrá, Nicholas Fernando, Boxi Wu, Rachel Foley, Susannah Young, Iason Gabriel, William Isaac, John Mellor, Demis Hassabis, Koray Kavukcuoglu, Lisa Anne Hendricks, Geoffrey Irving*. 90 | 91 | - **[AI Chains: Transparent and Controllable Human-AI Interaction by Chaining Large Language Model Prompts](https://arxiv.org/abs/2110.01691)**, 2021.10 ![img](https://img.shields.io/badge/-AI%20Chains-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Configuration-green) 92 | 93 | *Tongshuang Wu, Michael Terry, Carrie J. Cai*. 94 | 95 | - **[Interactive Text Generation](https://arxiv.org/abs/2303.00908)**, 2023.03 ![img](https://img.shields.io/badge/-ITG-blue) ![img](https://img.shields.io/badge/-Imitation%20Learning-orange) ![img](https://img.shields.io/badge/-Edits-lightgrey) ![img](https://img.shields.io/badge/Simulation-green) 96 | 97 | *Felix Faltings, Michel Galley, Baolin Peng, Kianté Brantley, Weixin Cai, Yizhe Zhang, Jianfeng Gao, Bill Dolan*. 98 | 99 | - **[PromptChainer: Chaining Large Language Model Prompts through Visual Programming](https://arxiv.org/abs/2203.06566)**, 2022.03 ![img](https://img.shields.io/badge/-PromptChainer-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Configuration-green) 100 | 101 | *Tongshuang Wu, Ellen Jiang, Aaron Donsbach, Jeff Gray, Alejandra Molina, Michael Terry, Carrie J Cai*. 102 | 103 | - **[RRHF: Rank Responses to Align Language Models with Human Feedback without tears](https://arxiv.org/abs/2304.05302)**, 2023.04 ![img](https://img.shields.io/badge/-RRHF-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/Feedback-green) 104 | 105 | *Zheng Yuan, Hongyi Yuan, Chuanqi Tan, Wei Wang, Songfang Huang, Fei Huang*. 106 | 107 | - **[RAFT: Reward rAnked FineTuning for Generative Foundation Model Alignment](https://arxiv.org/abs/2304.06767)**, 2023.04 ![img](https://img.shields.io/badge/-RAFT-blue) ![img](https://img.shields.io/badge/Feedback-green) 108 | 109 | *Hanze Dong, Wei Xiong, Deepanshu Goyal, Rui Pan, Shizhe Diao, Jipeng Zhang, Kashun Shum, Tong Zhang*. 110 | 111 | - **[Interactive Language: Talking to Robots in Real Time](https://arxiv.org/abs/2210.06407)**, 2022.10 ![img](https://img.shields.io/badge/-Interactive%20Language-blue) ![img](https://img.shields.io/badge/-Imitation%20Learning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Chat-green) 112 | 113 | *Corey Lynch, Ayzaan Wahid, Jonathan Tompson, Tianli Ding, James Betker, Robert Baruch, Travis Armstrong, Pete Florence*. 114 | 115 | - **[Improving Grounded Language Understanding in a Collaborative Environment by Interacting with Agents Through Help Feedback](https://arxiv.org/abs/2304.10750)**, 2023.04 ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Chat,%20Feedback-green) 116 | 117 | *Nikhil Mehta, Milagro Teruel, Patricio Figueroa Sanz, Xin Deng, Ahmed Hassan Awadallah, Julia Kiseleva*. 118 | 119 | - **[Is Reinforcement Learning (Not) for Natural Language Processing: Benchmarks, Baselines, and Building Blocks for Natural Language Policy Optimization](https://arxiv.org/abs/2210.01241)**, 2022.10 ![img](https://img.shields.io/badge/-RL4LMs-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/Feedback-green) 120 | 121 | *Rajkumar Ramamurthy, Prithviraj Ammanabrolu, Kianté Brantley, Jack Hessel, Rafet Sifa, Christian Bauckhage, Hannaneh Hajishirzi, Yejin Choi*. 122 | 123 | - **[Improving Multimodal Interactive Agents with Reinforcement Learning from Human Feedback](https://arxiv.org/abs/2211.11602)**, 2022.11 ![img](https://img.shields.io/badge/-IBT-blue) ![img](https://img.shields.io/badge/-Imitation%20Learning,%20Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/Feedback-green) 124 | 125 | *Josh Abramson, Arun Ahuja, Federico Carnevale, Petko Georgiev, Alex Goldin, Alden Hung, Jessica Landon, Jirka Lhotka, Timothy Lillicrap, Alistair Muldal, George Powell, Adam Santoro, Guy Scully, Sanjana Srivastava, Tamara von Glehn, Greg Wayne, Nathaniel Wong, Chen Yan, Rui Zhu*. 126 | 127 | - **[Towards Teachable Reasoning Systems: Using a Dynamic Memory of User Feedback for Continual System Improvement](https://arxiv.org/abs/2204.13074)**, 2022.04 ![img](https://img.shields.io/badge/-TeachMe-blue) ![img](https://img.shields.io/badge/-Continual%20Learning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Feedback-green) 128 | 129 | *Bhavana Dalvi Mishra, Oyvind Tafjord, Peter Clark*. 130 | 131 | - **[MemPrompt: Memory-assisted Prompt Editing with User Feedback](https://aclanthology.org/2022.emnlp-main.183.pdf)**, 2022.12 ![img](https://img.shields.io/badge/-MemPrompt-blue) ![img](https://img.shields.io/badge/-Active%20Learning-orange) ![img](https://img.shields.io/badge/-Edits-lightgrey) ![img](https://img.shields.io/badge/Personalization,%20Feedback-green) 132 | 133 | *Aman Madaan, Niket Tandon, Peter Clark, Yiming Yang*. 134 | 135 | - **[Constitutional Al:learning from ai feedback](https://arxiv.org/abs/2212.08073)**, 2022.12 ![img](https://img.shields.io/badge/-Constitutional%20AI-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/Feedback-green) 136 | 137 | *Yuntao Bai, Saurav Kadavath, Sandipan Kundu, Amanda Askell, Jackson Kernion, Andy Jones, Anna Chen, Anna Goldie, Azalia Mirhoseini, Cameron McKinnon, Carol Chen, Catherine Olsson, Christopher Olah, Danny Hernandez, Dawn Drain, Deep Ganguli, Dustin Li, Eli Tran-Johnson, Ethan Perez, Jamie Kerr, Jared Mueller, Jeffrey Ladish, Joshua Landau, Kamal Ndousse, Kamile Lukosuite, Liane Lovitt, Michael Sellitto, Nelson Elhage, Nicholas Schiefer, Noemi Mercado, Nova DasSarma, Robert Lasenby, Robin Larson, Sam Ringer, Scott Johnston, Shauna Kravec, Sheer El Showk, Stanislav Fort, Tamera Lanham, Timothy Telleen-Lawton, Tom Conerly, Tom Henighan, Tristan Hume, Samuel R. Bowman, Zac Hatfield-Dodds, Ben Mann, Dario Amodei, Nicholas Joseph, Sam McCandlish, Tom Brown, Jared Kaplan*. 138 | 139 | - **[Craft an Iron Sword: Dynamically Generating Interactive Game Characters by Prompting Large Language Models Tuned on Code](https://aclanthology.org/2022.wordplay-1.3/)**, 2022.01 ![img](https://img.shields.io/badge/Game,%20Chat-green) 140 | 141 | *Volum, Ryan and Rao, Sudha and Xu, Michael and DesGarennes, Gabriel A and Brockett, Chris and Van Durme, Benjamin and Deng, Olivia and Malhotra, Akanksha and Dolan, Bill*. 142 | 143 | - **[LaMP: When Large Language Models Meet Personalization](https://arxiv.org/abs/2304.11406)**, 2023.04 ![img](https://img.shields.io/badge/-LaMP-blue) ![img](https://img.shields.io/badge/Personalization-green) 144 | 145 | *Alireza Salemi, Sheshera Mysore, Michael Bendersky, Hamed Zamani*. 146 | 147 | - **[Languages are Rewards: Hindsight Finetuning using Human Feedback](https://arxiv.org/abs/2302.02676)**, 2023.02 ![img](https://img.shields.io/badge/-Chain%20of%20Hindsight-blue) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Feedback-green) 148 | 149 | *Hao Liu, Carmelo Sferrazza, Pieter Abbeel*. 150 | 151 | - **[InternChat: Solving Vision-Centric Tasks by Interacting with Chatbots Beyond Language](https://arxiv.org/abs/2305.05662)**, 2023.05 ![img](https://img.shields.io/badge/-InternChat-blue) ![img](https://img.shields.io/badge/-Instruction%20Tuning-orange) ![img](https://img.shields.io/badge/Chat-green) 152 | 153 | *Zhaoyang Liu, Yinan He, Wenhai Wang, Weiyun Wang, Yi Wang, Shoufa Chen, Qinglong Zhang, Yang Yang, Qingyun Li, Jiashuo Yu, Kunchang Li, Zhe Chen, Xue Yang, Xizhou Zhu, Yali Wang, Limin Wang, Ping Luo, Jifeng Dai, Yu Qiao*. 154 | 155 | - **[Improving Code Generation by Training with Natural Language Feedback](https://arxiv.org/abs/2303.16749)**, 2023.03 ![img](https://img.shields.io/badge/-ILF-blue) ![img](https://img.shields.io/badge/-Imitation%20Learning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Feedback-green) 156 | 157 | *Angelica Chen, Jérémy Scheurer, Tomasz Korbak, Jon Ander Campos, Jun Shern Chan, Samuel R. Bowman, Kyunghyun Cho, Ethan Perez*. 158 | 159 | 160 | ### 📚KB-LM Interaction 161 | - **[Atlas: Few-shot learning with retrieval augmented language models](https://arxiv.org/pdf/2208.03299.pdf)**, 2022.08 ![img](https://img.shields.io/badge/-Atlas-blue) ![img](https://img.shields.io/badge/-Corpus%20Knowledge-green) 162 | 163 | *Gautier Izacard, Patrick Lewis, Maria Lomeli, Lucas Hosseini, Fabio Petroni, Timo Schick, Jane Dwivedi-Yu, Armand Joulin, Sebastian Riedel, Edouard Grave* 164 | 165 | - **[MineDojo: Building Open-Ended Embodied Agents with Internet-Scale Knowledge](https://openreview.net/pdf?id=rc8o_j8I8PX)**, 2022.09 ![img](https://img.shields.io/badge/-MineDojo-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Internet%20Knowledge-green) 166 | 167 | *Linxi Fan, Guanzhi Wang, Yunfan Jiang, Ajay Mandlekar, Yuncong Yang, Haoyi Zhu, Andrew Tang, De-An Huang, Yuke Zhu, Anima Anandkumar* 168 | 169 | - **[WebGPT: Browser-assisted question-answering with human feedback](https://arxiv.org/pdf/2112.09332.pdf)**, 2021.12 ![img](https://img.shields.io/badge/-WebGPT-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Internet%20Knowledge-green) 170 | 171 | *Reiichiro Nakano, Jacob Hilton, Suchir Balaji, Jeff Wu, Long Ouyang, Christina Kim, Christopher Hesse, Shantanu Jain, Vineet Kosaraju, William Saunders, Xu Jiang, Karl Cobbe, Tyna Eloundou, Gretchen Krueger, Kevin Button, Matthew Knight, Benjamin Chess, John Schulman* 172 | 173 | - **[Improving language models by retrieving from trillions of tokens](https://arxiv.org/pdf/2112.04426.pdf)**, 2022.07 ![img](https://img.shields.io/badge/-Retro-blue) ![img](https://img.shields.io/badge/-Message%20Fusion-orange) ![img](https://img.shields.io/badge/-Corpus%20Knowledge-green) 174 | 175 | *Sebastian Borgeaud, Arthur Mensch, Jordan Hoffmann, Trevor Cai, Eliza Rutherford, Katie Millican, George van den Driessche, Jean-Baptiste Lespiau, Bogdan Damoc, Aidan Clark, Diego de Las Casas, Aurelia Guy, Jacob Menick, Roman Ring, Tom Hennigan, Saffron Huang, Loren Maggiore, Chris Jones, Albin Cassirer, Andy Brock, Michela Paganini, Geoffrey Irving, Oriol Vinyals, Simon Osindero, Karen Simonyan, Jack W. Rae, Erich Elsen, Laurent Sifre* 176 | 177 | - **[REALM: retrieval-augmented language model pre-training](https://dl.acm.org/doi/pdf/10.5555/3524938.3525306)**, 2020.07 ![img](https://img.shields.io/badge/-REALM-blue) ![img](https://img.shields.io/badge/-Message%20Fusion-orange) ![img](https://img.shields.io/badge/-Corpus%20Knowledge-green) 178 | 179 | *Kelvin Guu, Kenton Lee, Zora Tung, Panupong Pasupat, Ming-Wei Chang* 180 | 181 | - **[KELM: Knowledge Enhanced Pre-Trained Language Representations with Message Passing on Hierarchical Relational Graphs](https://openreview.net/pdf?id=FkG-sX5CE_)**, 2022.04 ![img](https://img.shields.io/badge/-KELM-blue) ![img](https://img.shields.io/badge/-Message%20Fusion-orange) ![img](https://img.shields.io/badge/-Corpus%20Knowledge-green) 182 | 183 | *Yinquan Lu, Haonan Lu, Guirong Fu, Qun Liu* 184 | 185 | - **[K-Adapter: Infusing Knowledge into Pre-Trained Models with Adapters](https://aclanthology.org/2021.findings-acl.121.pdf)**, 2021.06 ![img](https://img.shields.io/badge/-K--Adapter-blue) ![img](https://img.shields.io/badge/-Parameter%20Efficient%20Fine--Tuning-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey) ![img](https://img.shields.io/badge/-Corpus%20Knowledge-green) 186 | 187 | *Ruize Wang, Duyu Tang, Nan Duan, Zhongyu Wei, Xuanjing Huang, Jianshu Ji, Guihong Cao, Daxin Jiang, Ming Zhou* 188 | 189 | - **[Binding Language Models in Symbolic Languages](https://arxiv.org/pdf/2210.02875.pdf)**, 2023.05 ![img](https://img.shields.io/badge/-Binder-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey) ![img](https://img.shields.io/badge/-Corpus%20Knowledge-green) 190 | 191 | *Zhoujun Cheng, Tianbao Xie, Peng Shi, Chengzu Li, Rahul Nadkarni, Yushi Hu, Caiming Xiong, Dragomir Radev, Mari Ostendorf, Luke Zettlemoyer, Noah A. Smith, Tao Yu* 192 | 193 | - **[Teaching language models to support answers with verified quotes](https://arxiv.org/pdf/2203.11147.pdf)**, 2022.03 ![img](https://img.shields.io/badge/-GopherCite-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Internet%20Knowledge-green) 194 | 195 | *Jacob Menick, Maja Trebacz, Vladimir Mikulik, John Aslanides, Francis Song, Martin Chadwick, Mia Glaese, Susannah Young, Lucy Campbell-Gillingham, Geoffrey Irving, Nat McAleese* 196 | 197 | - **[ERNIE: Enhanced Representation through Knowledge Integration](https://arxiv.org/pdf/1904.09223)**, 2019.04 ![img](https://img.shields.io/badge/-ERNIE-blue) ![img](https://img.shields.io/badge/-Corpus%20Knowledge-green) 198 | 199 | *Yu Sun, Shuohuan Wang, Yukun Li, Shikun Feng, Xuyi Chen, Han Zhang, Xin Tian, Danxiang Zhu, Hao Tian, Hua Wu* 200 | 201 | - **[K-BERT: Enabling Language Representation with Knowledge Graph](https://ojs.aaai.org/index.php/AAAI/article/view/5681/5537)**, 2020.02 ![img](https://img.shields.io/badge/-K--BERT-blue) ![img](https://img.shields.io/badge/-Message%20Fusion-orange) ![img](https://img.shields.io/badge/-Corpus%20Knowledge-green) 202 | 203 | *Weijie Liu, Peng Zhou, Zhe Zhao, Zhiruo Wang, Qi Ju, Haotang Deng, Ping Wang* 204 | 205 | - **[Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks](https://arxiv.org/pdf/2005.11401)**, 2020.12 ![img](https://img.shields.io/badge/-RAG-blue) ![img](https://img.shields.io/badge/-Corpus%20Knowledge-green) 206 | 207 | *Patrick Lewis, Ethan Perez, Aleksandra Piktus, Fabio Petroni, Vladimir Karpukhin, Naman Goyal, Heinrich Küttler, Mike Lewis, Wen-tau Yih, Tim Rocktäschel, Sebastian Riedel, Douwe Kiela* 208 | 209 | - **[DocPrompting: Generating Code by Retrieving the Docs](https://openreview.net/pdf?id=ZTCxT2t2Ru)**, 2023.05 ![img](https://img.shields.io/badge/-DocPrompting-blue) ![img](https://img.shields.io/badge/-Standard%20Prompting-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey) ![img](https://img.shields.io/badge/-Internet%20Knowledge-green) 210 | 211 | *Shuyan Zhou, Uri Alon, Frank F. Xu, Zhiruo Wang, Zhengbao Jiang, Graham Neubig* 212 | 213 | - **[Recitation-Augmented Language Models](https://openreview.net/pdf?id=-cqvvvb-NkI)**, 2023.05 ![img](https://img.shields.io/badge/-RECITE-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting-orange)![img](https://img.shields.io/badge/-Corpus%20Knowledge-green) 214 | 215 | *Zhiqing Sun, Xuezhi Wang, Yi Tay, Yiming Yang, Denny Zhou* 216 | 217 | - **[Mind's Eye: Grounded Language Model Reasoning through Simulation](https://openreview.net/pdf?id=4rXMRuoJlai)**, 2023.05 ![img](https://img.shields.io/badge/-Mind%E2%80%99s%20Eye-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey) 218 | 219 | *Ruibo Liu, Jason Wei, Shixiang Shane Gu, Te-Yen Wu, Soroush Vosoughi, Claire Cui, Denny Zhou, Andrew M Dai* 220 | 221 | - **[Don't Generate, Discriminate: A Proposal for Grounding Language Models to Real-World Environments](https://arxiv.org/pdf/2212.09736)**, 2023.10 ![img](https://img.shields.io/badge/-Pangu-blue) ![img](https://img.shields.io/badge/-Message%20Fusion-orange) ![img](https://img.shields.io/badge/-Internet%20Knowledge-green) 222 | 223 | *Yu Gu, Xiang Deng, Yu Su* 224 | 225 | - **[Check Your Facts and Try Again: Improving Large Language Models with External Knowledge and Automated Feedback](https://arxiv.org/pdf/2302.12813)**, 2023.02 ![img](https://img.shields.io/badge/-LLM--Augmenter-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Internet%20Knowledge-green) 226 | 227 | *Baolin Peng, Michel Galley, Pengcheng He, Hao Cheng, Yujia Xie, Yu Hu, Qiuyuan Huang, Lars Liden, Zhou Yu, Weizhu Chen, Jianfeng Gao* 228 | 229 | ### 🤖Model/🛠Tool-LM Interaction 230 | 231 | - **[Generative Agents: Interactive Simulacra of Human Behavior](https://arxiv.org/abs/2304.03442)**, 2023.04 ![img](https://img.shields.io/badge/-Generative%20Agents-blue) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Society-green) 232 | 233 | *Joon Sung Park, Joseph C. O'Brien, Carrie J. Cai, Meredith Ringel Morris, Percy Liang, Michael S. Bernstein*. 234 | 235 | - **[ReAct: Synergizing reasoning and acting in language models](https://arxiv.org/abs/2210.03629)**, 2022.10 ![img](https://img.shields.io/badge/-ReAct-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Chain%20of%20Thought,%20Tool--use-green) 236 | 237 | *Shunyu Yao, Jeffrey Zhao, Dian Yu, Nan Du, Izhak Shafran, Karthik Narasimhan, Yuan Cao*. 238 | 239 | - **[Visual ChatGPT: Talking, Drawing and Editing with Visual Foundation Models](https://arxiv.org/abs/2303.04671)**, 2023.05 ![img](https://img.shields.io/badge/-Visual%20ChatGPT-blue) ![img](https://img.shields.io/badge/Chain%20of%20Thought,%20Tool--use-green) 240 | 241 | *Chenfei Wu, Shengming Yin, Weizhen Qi, Xiaodong Wang, Zecheng Tang, Nan Duan*. 242 | 243 | - **[HuggingGPT: Solving AI Tasks with ChatGPT and its Friends in Hugging Face](https://arxiv.org/abs/2303.17580)**, 2023.05 ![img](https://img.shields.io/badge/-HuggingGPT-blue) ![img](https://img.shields.io/badge/Tool--use-green) 244 | 245 | *Yongliang Shen, Kaitao Song, Xu Tan, Dongsheng Li, Weiming Lu, Yueting Zhuang*. 246 | 247 | - **[CAMEL: Communicative Agents for "Mind" Exploration of Large Scale Language Model Society](https://arxiv.org/abs/2303.17760)**, 2023.03 ![img](https://img.shields.io/badge/-CAMEL-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Society-green) 248 | 249 | *Guohao Li, Hasan Abed Al Kader Hammoud, Hani Itani, Dmitrii Khizbullin, Bernard Ghanem*. 250 | 251 | - **[Socratic Models: Composing Zero-Shot Multimodal Reasoning with Language](https://arxiv.org/abs/2204.00598)**, 2022.05 ![img](https://img.shields.io/badge/-Socratic%20Model-blue) ![img](https://img.shields.io/badge/Communication-green) 252 | 253 | *Andy Zeng, Maria Attarian, Brian Ichter, Krzysztof Choromanski, Adrian Wong, Stefan Welker, Federico Tombari, Aveek Purohit, Michael Ryoo, Vikas Sindhwani, Johnny Lee, Vincent Vanhoucke, Pete Florence*. 254 | 255 | - **[MindCraft: Theory of Mind Modeling for Situated Dialogue in Collaborative Tasks](https://arxiv.org/abs/2109.06275)**, 2021.09 ![img](https://img.shields.io/badge/-MindCraft-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Theory%20of%20Mind,%20Communication-green) 256 | 257 | *Cristian-Paul Bara, Sky CH-Wang, Joyce Chai*. 258 | 259 | - **[Program of Thoughts Prompting: Disentangling Computation from Reasoning for Numerical Reasoning Tasks](https://arxiv.org/abs/2211.12588)**, 2022.11 ![img](https://img.shields.io/badge/-PoT%20Prompting-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use,%20Chain%20of%20Thought-green) 260 | 261 | *Wenhu Chen, Xueguang Ma, Xinyi Wang, William W. Cohen*. 262 | 263 | - **[Toolformer: Language Models Can Teach Themselves to Use Tools](https://arxiv.org/abs/2302.04761)**, 2023.02 ![img](https://img.shields.io/badge/-Toolformer-blue) ![img](https://img.shields.io/badge/Tool--use-green) 264 | 265 | *Timo Schick, Jane Dwivedi-Yu, Roberto Dessì, Roberta Raileanu, Maria Lomeli, Luke Zettlemoyer, Nicola Cancedda, Thomas Scialom*. 266 | 267 | - **[ART: Automatic multi-step reasoning and tool-use for large language models](https://arxiv.org/abs/2303.09014)**, 2023.03 ![img](https://img.shields.io/badge/-ART-blue) ![img](https://img.shields.io/badge/-Formal%20Language,%20Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use,%20Chain%20of%20Thought-green) 268 | 269 | *Bhargavi Paranjape, Scott Lundberg, Sameer Singh, Hannaneh Hajishirzi, Luke Zettlemoyer, Marco Tulio Ribeiro*. 270 | 271 | - **[Small Models are valuable Plug-ins for large language models](https://arxiv.org/pdf/2305.08848.pdf)**, 2023.05 ![img](https://img.shields.io/badge/-SuperICL-blue) ![img](https://img.shields.io/badge/-Standard%20Prompting-orange) 272 | 273 | *Canwen Xu, Yichong Xu, Shuohang Wang, Yang Liu, Chenguang Zhu, Julian McAuley*. 274 | 275 | - **[LEAST-TO-MOST PROMPTINGENABLESCOMPLEXREASONING IN LARGE LANGUAGE MODELS](https://arxiv.org/abs/2205.10625)**, 2022.05 ![img](https://img.shields.io/badge/-least--to--most%20prompting-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green) 276 | 277 | *Denny Zhou, Nathanael Schärli, Le Hou, Jason Wei, Nathan Scales, Xuezhi Wang, Dale Schuurmans, Claire Cui, Olivier Bousquet, Quoc Le, Ed Chi*. 278 | 279 | - **[Decomposed Prompting: A Modular Approach for Solving Complex Tasks](https://arxiv.org/abs/2210.02406)**, 2022.10 ![img](https://img.shields.io/badge/-Decomposed%20Prompting-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green) 280 | 281 | *Tushar Khot, Harsh Trivedi, Matthew Finlayson, Yao Fu, Kyle Richardson, Peter Clark, Ashish Sabharwal*. 282 | 283 | - **[ViperGPT: Visual Inference via Python Execution for Reasoning](https://arxiv.org/pdf/2303.08128.pdf)**, 2023.03 ![img](https://img.shields.io/badge/-ViperGPT-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green) 284 | 285 | *Dídac Surís,  Sachit Menon, Carl Vondrick*. 286 | 287 | - **[See,Think,Confirm:Interactive Prompting Between Vision and Language Models for Knowledge-based Visual Reasoning](https://arxiv.org/pdf/2301.05226.pdf)**, 2023.01 ![img](https://img.shields.io/badge/-IPVR-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Tool--use-green) 288 | 289 | *Zhenfang Chen, Qinhong Zhou, Yikang Shen, Yining Hong, Hao Zhang, Chuang Gan*. 290 | 291 | - **[Large Language Models Are Reasoning Teachers](https://arxiv.org/abs/2212.10071)**, 2022.12 ![img](https://img.shields.io/badge/-Fine--tune--CoT-blue) ![img](https://img.shields.io/badge/-Semi--Supervised%20Fine--Tuning,%20Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green) 292 | 293 | *Namgyu Ho, Laura Schmid, Se-Young Yun*. 294 | 295 | - **[STaR:Self-Taught ReasonerBootstrapping Reasoning With Reasoning](https://arxiv.org/pdf/2203.14465.pdf)**, 2022.03 ![img](https://img.shields.io/badge/-STaR-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting,%20Semi--Supervised%20Fine--Tuning-orange) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green) 296 | 297 | *Eric Zelikman, Yuhuai Wu, Jesse Mu, Noah D. Goodman*. 298 | 299 | - **[Tree of Thoughts: Deliberate Problem Solving with Large Language Models](https://arxiv.org/pdf/2305.10601.pdf)**, 2023.05 ![img](https://img.shields.io/badge/-ToT-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green) 300 | 301 | *Shunyu Yao, Dian Yu, Jeffrey Zhao, Izhak Shafran, Thomas L. Griffiths, Yuan Cao, Karthik Narasimhan*. 302 | 303 | - **[Search-in-the-Chain: Towards Accurate, Credible and Traceable Large Language Models for Knowledge-intensive Tasks](https://arxiv.org/pdf/2304.14732v4.pdf)**, 2023.04 ![img](https://img.shields.io/badge/-SearChain-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Tool--use,%20Chain%20of%20Thought-green) 304 | 305 | *Shicheng Xu, Liang Pang, HuaWei Shen, Xueqi Cheng, Tat-Seng Chua*. 306 | 307 | - **[RECURRENTGPT: Interactive Generation of (Arbitrarily) Long Text](https://arxiv.org/pdf/2305.13304v1.pdf)**, 2023.05 ![img](https://img.shields.io/badge/-RecurrentGPT-blue) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) 308 | 309 | *Wangchunshu Zhou, Yuchen Eleanor Jiang, Peng Cui, Tiannan Wang, Zhenxin Xiao, Yifan Hou, Ryan Cotterell, Mrinmaya Sachan*. 310 | 311 | - **[PAL: Program-aided Language Models](https://arxiv.org/pdf/2211.10435v2.pdf)**, 2022.11 ![img](https://img.shields.io/badge/-PAL-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey) ![img](https://img.shields.io/badge/Chain%20of%20Thought,%20Tool--use-green) 312 | 313 | *Luyu Gao, Aman Madaan, Shuyan Zhou, Uri Alon, PengFei Liu, Yiming Yang, Jamie Callan, Graham Neubig*. 314 | 315 | - **[Internet-augmented language models through few-shot prompting for open-domain question answering](https://arxiv.org/pdf/2203.05115v2.pdf)**, 2023.05 ![img](https://img.shields.io/badge/Tool--use-green) 316 | 317 | *Angeliki Lazaridou, Elena Gribovskaya, Wojciech Stokowiec, Nikolai Grigorev*. 318 | 319 | - **[Recitation-Augmented Language Models](https://arxiv.org/pdf/2210.01296v2.pdf)**, 2022.10 ![img](https://img.shields.io/badge/-RECITE-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green) 320 | 321 | *Zhiqing Sun, Xuezhi Wang, Yi Tay, Yiming Yang, Denny Zhou*. 322 | 323 | - **[Iteratively Prompt Pre-trained Language Models for Chain of Thought](https://arxiv.org/pdf/2203.08383v3.pdf)**, 2022.03 ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green) 324 | 325 | *Boshi Wang, Xiang Deng, Huan Sun*. 326 | 327 | - **[MEASURING AND NARROWING THE COMPOSITIONALITY GAP IN LANGUAGE MODELS](https://arxiv.org/pdf/2210.03350v2.pdf)**, 2022.10 ![img](https://img.shields.io/badge/-self--ask-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Tool--use,%20Chain%20of%20Thought-green) 328 | 329 | *Ofir Press, Muru Zhang, Sewon Min, Ludwig Schmidt, Noah A. Smith, Mike Lewis*. 330 | 331 | - **[Self-Refine: Iterative Refinement with Self-Feedback](https://arxiv.org/pdf/2303.17651v1.pdf)**, 2023.03 ![img](https://img.shields.io/badge/-self--refine-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Self--Interaction-green) 332 | 333 | *Aman Madaan, Niket Tandon, Prakhar Gupta, Skyler Hallinan, Luyu Gao, Sarah Wiegreffe, Uri Alon, Nouha Dziri, Shrimai Prabhumoye, Yiming Yang, Sean Welleck, Bodhisattwa Prasad Majumder, Shashank Gupta, Amir Yazdanbakhsh, Peter Clark*. 334 | 335 | - **[LEVER: Learning to Verify Language-to-Code Generation with Execution](https://arxiv.org/pdf/2302.08468v1.pdf)**, 2023.02 ![img](https://img.shields.io/badge/-LEVER-blue) ![img](https://img.shields.io/badge/-Standard%20Prompting-orange) ![img](https://img.shields.io/badge/-Formal%20Language,%20Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green) 336 | 337 | *Ansong Ni, Srini Iyer, Dragomir Radev, Ves Stoyanov, Wen-tau Yih, Sida I. Wang, Xi Victoria Lin*. 338 | 339 | - **[Computational Language Acquisition with Theory of Mind](https://arxiv.org/pdf/2303.01502v1.pdf)**, 2023.03 ![img](https://img.shields.io/badge/Theory%20of%20Mind-green) 340 | 341 | *Andy Liu, Hao Zhu, Emmy Liu, Yonatan Bisk, Graham Neubig*. 342 | 343 | - **[Few-shot Language Coordination by Modeling Theory of Mind](https://arxiv.org/pdf/2107.05697v1.pdf)**, 2021.07 ![img](https://img.shields.io/badge/Theory%20of%20Mind-green) 344 | 345 | *Hao Zhu, Graham Neubig, Yonatan Bisk*. 346 | 347 | - **[OpenAGI: When LLM Meets Domain Experts](https://arxiv.org/pdf/2304.04370v2.pdf)**, 2023.04 ![img](https://img.shields.io/badge/-OpenAGI-blue) ![img](https://img.shields.io/badge/-Standard%20Prompting,%20Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green) 348 | 349 | *Yingqiang Ge, Wenyue Hua, Jianchao Ji, Juntao Tan, Shuyuan Xu, Yongfeng Zhang*. 350 | 351 | - **[MM-REACT: Prompting ChatGPT for Multimodal Reasoning and Action](https://arxiv.org/pdf/2303.11381v1.pdf)**, 2023.03 ![img](https://img.shields.io/badge/-MM--ReAct-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Tool--use-green) 352 | 353 | *Zhengyuan Yang, Linjie Li, JianFeng Wang, Kevin Lin, Ehsan Azarnasab, Faisal Ahmed, Zicheng Liu, Ce Liu, Michael Zeng, Lijuan Wang*. 354 | 355 | - **[Prompt, Generate, then Cache: Cascade of Foundation Models makes Strong Few-shot Learners](https://arxiv.org/pdf/2303.02151v1.pdf)**, 2023.03 ![img](https://img.shields.io/badge/-CaFo-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language,%20Machine%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green) 356 | 357 | *Renrui Zhang, Xiangfei Hu, Bohao Li, Siyuan Huang, Hanqiu Deng, Hongsheng Li, Yu Qiao, Peng Gao*. 358 | 359 | - **[Principle-Driven Self-Alignment of Language Models from Scratch with Minimal Human Supervision](https://arxiv.org/pdf/2305.03047v1.pdf)**, 2023.05 ![img](https://img.shields.io/badge/-SELF--ALIGN-blue) ![img](https://img.shields.io/badge/-Instruction%20Tuning,Semi--Supervised%20Fine--Tuning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Alignment-green) 360 | 361 | *Zhiqing Sun, Yikang Shen, Qinhong Zhou, Hongxin Zhang, Zhenfang Chen, David Cox, Yiming Yang, Chuang Gan*. 362 | 363 | - **[Think Before You Act: Unified Policy for Interleaving Language Reasoning with Actions](https://arxiv.org/pdf/2304.11063v1.pdf)**, 2023.04 ![img](https://img.shields.io/badge/-Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Natural%20Language,%20Machine%20Language-lightgrey) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green) 364 | 365 | *Lina Mezghani, Piotr Bojanowski, Karteek Alahari, Sainbayar Sukhbaatar*. 366 | 367 | - **[TALM: Tool Augmented Language Models](https://arxiv.org/pdf/2205.12255v1.pdf)**, 2022.05 ![img](https://img.shields.io/badge/-TALM-blue) ![img](https://img.shields.io/badge/-Instruction%20Tuning,Semi--Supervised%20Fine--Tuning-orange) ![img](https://img.shields.io/badge/-Formal%20Language,%20Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green) 368 | 369 | *Aaron Parisi, Yao Zhao, Noah Fiedel*. 370 | 371 | - **[Successive Prompting for Decomposing Complex Questions](https://arxiv.org/pdf/2212.04092v1.pdf)**, 2022.12 ![img](https://img.shields.io/badge/-Successive%20Prompting-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green) 372 | 373 | *Dheeru Dua, Shivanshu Gupta, Sameer Singh, Matt Gardner*. 374 | 375 | - **[REFINER: Reasoning Feedback on Intermediate Representations](https://arxiv.org/pdf/2304.01904v1.pdf)**, 2023.04 ![img](https://img.shields.io/badge/-REFINER-blue) ![img](https://img.shields.io/badge/-Semi--Supervised%20Fine--Tuning,%20Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Chain%20of%20Thought,%20Reasoning-green) 376 | 377 | *Debjit Paul, Mete Ismayilzada, Maxime Peyrard, Beatriz Borges, Antoine Bosselut, Robert West, Boi Faltings*. 378 | 379 | - **[LeTI: Learning to Generate from Textual Interactions](https://arxiv.org/pdf/2305.10314v1.pdf)**, 2023.05 ![img](https://img.shields.io/badge/-LeTI-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Formal%20Language,%20Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green) 380 | 381 | *Xingyao Wang, Hao Peng, Reyhaneh Jabbarvand, Heng Ji*. 382 | 383 | - **[InternGPT: Solving Vision-Centric Tasks by Interacting with Chatbots Beyond Language](https://arxiv.org/pdf/2305.05662v3.pdf)**, 2023.05 ![img](https://img.shields.io/badge/-InternGPT-blue) ![img](https://img.shields.io/badge/-Formal%20Language,%20Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green) 384 | 385 | *Zhaoyang Liu, Yinan He, Wenhai Wang, Weiyun Wang, Yi Wang, Shoufa Chen, Qinglong Zhang, Yang Yang, Qingyun Li, Jiashuo Yu, Kunchang Li, Zhe Chen, Xue Yang, Xizhou Zhu, Yali Wang, LiMin Wang, Ping Luo, Jifeng Dai, Yu Qiao*. 386 | 387 | - **[Human-level play in the game of Diplomacy by combining language models with strategic reasoning](https://www.science.org/doi/10.1126/science.ade9097)**, 2022.11 ![img](https://img.shields.io/badge/-Cicero-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Decision%20Making-green) 388 | 389 | *Anton Bakhtin, Noam Brown, Emily Dinan, Gabriele Farina, Colin Flaherty, Daniel Fried, Andrew Goff, Jonathan Gray, Hengyan Hu, Athul Paul Jacob, Mojtaba Komeili, Karthik Konath, Minae Kwon, Adam Lerer, Mike Lewis, Alexander H. Miller, Sash Mitts, Aditya Renduchintala, Stephen Roller, Dirk Rowe, Weiyan Shi, Joe Spisak, Alexander Wei, David Wu, Hugh Zhang, Markus Zijlstra*. 390 | 391 | - **[Generating Sequences by Learning to Self-Correct](https://arxiv.org/abs/2211.00053)**, 2022.10 ![img](https://img.shields.io/badge/-Self--Correction-blue) ![img](https://img.shields.io/badge/-Semi--Supervised%20Fine--Tuning-orange) ![img](https://img.shields.io/badge/-Natural%20Language,%20Edits-lightgrey) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green) 392 | 393 | *Sean Welleck, Ximing Lu, Peter West, Faeze Brahman, Tianxiao Shen, Daniel Khashabi, Yejin Choi*. 394 | 395 | - **[ChatGPT-steered Editing Instructor for Customization of Abstractive Summarization](https://arxiv.org/pdf/2305.02483v1.pdf)**, 2023.05 ![img](https://img.shields.io/badge/-Semi--Supervised%20Fine--Tuning,%20Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language,%20Edits-lightgrey) ![img](https://img.shields.io/badge/Chain%20of%20Thought-green) 396 | 397 | *Wen Xiao, Yujia Xie, Giuseppe Carenini, Pengcheng He*. 398 | 399 | - **[ChatGPT Asks, BLIP-2 Answers: Automatic Questioning Towards Enriched Visual Descriptions](https://arxiv.org/pdf/2303.06594v1.pdf)**, 2023.03 ![img](https://img.shields.io/badge/-ChatCaptioner-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green) 400 | 401 | *Deyao Zhu, Jun Chen, Kilichbek Haydarov, Xiaoqian Shen, Wenxuan Zhang, Mohamed Elhoseiny*. 402 | 403 | - **[Chameleon: Plug-and-Play Compositional Reasoning with Large Language Models](https://arxiv.org/pdf/2304.09842v2.pdf)**, 2023.04 ![img](https://img.shields.io/badge/-Chameleon-blue) ![img](https://img.shields.io/badge/-Formal%20Language,%20Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Tool--use-green) 404 | 405 | *Pan Lu, Baolin Peng, Hao Cheng, Michel Galley, Kai-Wei Chang, Ying Nian Wu, Song-Chun Zhu, Jianfeng Gao*. 406 | 407 | - **[Baize: An Open-Source Chat Model with Parameter-Efficient Tuning on Self-Chat Data](https://arxiv.org/pdf/2304.01196v3.pdf)**, 2023.04 ![img](https://img.shields.io/badge/-Baize-blue) ![img](https://img.shields.io/badge/-Parameter--Efficient%20Fine--Tuning,%20Semi--Supervised%20Fine--Tuning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Chat-green) 408 | 409 | *Canwen Xu, Daya Guo, Nan Duan, Julian McAuley*. 410 | 411 | 412 | ### 🌎Environment-LM Interaction 413 | - **[BLIP-2: Bootstrapping Language-Image Pre-training with Frozen Image Encoders and Large Language Models](https://arxiv.org/abs/2301.12597)**, 2023.01 ![img](https://img.shields.io/badge/-BLIP--2-blue) ![img](https://img.shields.io/badge/-Semi--Supervised%20Fine--Tuning-orange) ![img](https://img.shields.io/badge/-Machine%20Language-lightgrey) ![img](https://img.shields.io/badge/Modality%20Grounding-green) 414 | 415 | *Junnan Li, Dongxu Li, Silvio Savarese, Steven Hoi*. 416 | 417 | - **[Do As I Can, Not As I Say: Grounding Language in Robotic Affordances](https://arxiv.org/abs/2204.01691)**, 2022.04 ![img](https://img.shields.io/badge/-SayCan-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/Affordance%20Grounding-green) 418 | 419 | *Michael Ahn, Anthony Brohan, Noah Brown, Yevgen Chebotar, Omar Cortes, Byron David, Chelsea Finn, Chuyuan Fu, Keerthana Gopalakrishnan, Karol Hausman, Alex Herzog, Daniel Ho, Jasmine Hsu, Julian Ibarz, Brian Ichter, Alex Irpan, Eric Jang, Rosario Jauregui Ruano, Kyle Jeffrey, Sally Jesmonth, Nikhil J Joshi, Ryan Julian, Dmitry Kalashnikov, Yuheng Kuang, Kuang-Huei Lee, Sergey Levine, Yao Lu, Linda Luu, Carolina Parada, Peter Pastor, Jornell Quiambao, Kanishka Rao, Jarek Rettinghouse, Diego Reyes, Pierre Sermanet, Nicolas Sievers, Clayton Tan, Alexander Toshev, Vincent Vanhoucke, Fei Xia, Ted Xiao, Peng Xu, Sichun Xu, Mengyuan Yan, Andy Zeng*. 420 | 421 | - **[RT-1: Robotics Transformer for Real-World Control at Scale](https://arxiv.org/abs/2212.06817)**, 2022.12 ![img](https://img.shields.io/badge/-RT--1-blue)![img](https://img.shields.io/badge/Foundation%20Model-green) 422 | 423 | *Anthony Brohan, Noah Brown, Justice Carbajal, Yevgen Chebotar, Joseph Dabis, Chelsea Finn, Keerthana Gopalakrishnan, Karol Hausman, Alex Herzog, Jasmine Hsu, Julian Ibarz, Brian Ichter, Alex Irpan, Tomas Jackson, Sally Jesmonth, Nikhil J Joshi, Ryan Julian, Dmitry Kalashnikov, Yuheng Kuang, Isabel Leal, Kuang-Huei Lee, Sergey Levine, Yao Lu, Utsav Malla, Deeksha Manjunath, Igor Mordatch, Ofir Nachum, Carolina Parada, Jodilyn Peralta, Emily Perez, Karl Pertsch, Jornell Quiambao, Kanishka Rao, Michael Ryoo, Grecia Salazar, Pannag Sanketi, Kevin Sayed, Jaspiar Singh, Sumedh Sontakke, Austin Stone, Clayton Tan, Huong Tran, Vincent Vanhoucke, Steve Vega, Quan Vuong, Fei Xia, Ted Xiao, Peng Xu, Sichun Xu, Tianhe Yu, Brianna Zitkovich*. 424 | 425 | - **[PaLM-E: An Embodied Multimodal Language Model](https://arxiv.org/abs/2303.03378)**, 2023.03 ![img](https://img.shields.io/badge/-PaLM--E-blue)![img](https://img.shields.io/badge/Foundation%20Model-green) 426 | 427 | *Danny Driess, Fei Xia, Mehdi S. M. Sajjadi, Corey Lynch, Aakanksha Chowdhery, Brian Ichter, Ayzaan Wahid, Jonathan Tompson, Quan Vuong, Tianhe Yu, Wenlong Huang, Yevgen Chebotar, Pierre Sermanet, Daniel Duckworth, Sergey Levine, Vincent Vanhoucke, Karol Hausman, Marc Toussaint, Klaus Greff, Andy Zeng, Igor Mordatch, Pete Florence*. 428 | 429 | - **[Grounded decoding: Guiding text generation with grounded models for robot control](https://arxiv.org/abs/2303.00855)**, 2023.03 ![img](https://img.shields.io/badge/Grounded%20Decoding-blue) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Affordance%20Grounding-green) 430 | 431 | *Wenlong Huang, Fei Xia, Dhruv Shah, Danny Driess, Andy Zeng, Yao Lu, Pete Florence, Igor Mordatch, Sergey Levine, Karol Hausman, Brian Ichter*. 432 | 433 | - **[Inner monologue: Embodied reasoning through planning with language models.](https://arxiv.org/abs/2207.05608)**, 2022.07 ![img](https://img.shields.io/badge/Inner%20Monologue-blue)![img](https://img.shields.io/badge/Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) 434 | 435 | *Wenlong Huang, Fei Xia, Ted Xiao, Harris Chan, Jacky Liang, Pete Florence, Andy Zeng, Jonathan Tompson, Igor Mordatch, Yevgen Chebotar, Pierre Sermanet, Noah Brown, Tomas Jackson, Linda Luu, Sergey Levine, Karol Hausman, Brian Ichter*. 436 | 437 | - **[Code as Policies: Language Model Programs for Embodied Control](https://arxiv.org/abs/2209.07753)**, 2022.09 ![img](https://img.shields.io/badge/Code%20as%20Policies-blue)![img](https://img.shields.io/badge/-Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey) 438 | 439 | *Jacky Liang, Wenlong Huang, Fei Xia, Peng Xu, Karol Hausman, Brian Ichter, Pete Florence, Andy Zeng*. 440 | 441 | - **[Open-vocabulary Queryable Scene Representations for Real World Planning](https://arxiv.org/abs/2209.09874)**, 2022.09 ![img](https://img.shields.io/badge/-NLMap-blue) ![img](https://img.shields.io/badge/-Active%20Learning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Affordance%20Grounding-green) 442 | 443 | *Boyuan Chen, Fei Xia, Brian Ichter, Kanishka Rao, Keerthana Gopalakrishnan, Michael S. Ryoo, Austin Stone, Daniel Kappler*. 444 | 445 | - **[Plan4MC: Skill Reinforcement Learning and Planning for Open-World Minecraft Tasks](https://arxiv.org/abs/2303.16563)**, 2023.03 ![img](https://img.shields.io/badge/-Plan4MC-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey) ![img](https://img.shields.io/badge/Affordance%20Grounding-green) 446 | 447 | *Haoqi Yuan, Chi Zhang, Hongcheng Wang, Feiyang Xie, Penglin Cai, Hao Dong, Zongqing Lu*. 448 | 449 | - **[MineDojo: Building Open-Ended Embodied Agents with Internet-Scale Knowledge](https://arxiv.org/abs/2206.08853)**, 2022.06 ![img](https://img.shields.io/badge/-MineDojo-blue) ![img](https://img.shields.io/badge/-Active%20Learning,%20Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) 450 | 451 | *Linxi Fan, Guanzhi Wang, Yunfan Jiang, Ajay Mandlekar, Yuncong Yang, Haoyi Zhu, Andrew Tang, De-An Huang, Yuke Zhu, Anima Anandkumar*. 452 | 453 | - **[Socratic Models: Composing Zero-Shot Multimodal Reasoning with Language](https://arxiv.org/abs/2204.00598)**, 2022.04 ![img](https://img.shields.io/badge/Socratic%20Models-blue)![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language%20,%20Shared%20Memory-lightgrey) ![img](https://img.shields.io/badge/Modality%20Grounding-green) 454 | 455 | *Andy Zeng, Maria Attarian, Brian Ichter, Krzysztof Choromanski, Adrian Wong, Stefan Welker, Federico Tombari, Aveek Purohit, Michael Ryoo, Vikas Sindhwani, Johnny Lee, Vincent Vanhoucke, Pete Florence*. 456 | 457 | - **[Interactive Language: Talking to Robots in Real Time](https://arxiv.org/abs/2210.06407)**, 2022.10 ![img](https://img.shields.io/badge/Interactive%20Language-blue)![img](https://img.shields.io/badge/-Imitation%20Learning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) 458 | 459 | *Corey Lynch, Ayzaan Wahid, Jonathan Tompson, Tianli Ding, James Betker, Robert Baruch, Travis Armstrong, Pete Florence*. 460 | 461 | - **[LM-Nav: Robotic Navigation with Large Pre-Trained Models of Language, Vision, and Action](https://arxiv.org/abs/2207.04429)**, 2022.07 ![img](https://img.shields.io/badge/LM--Nav-blue)![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) 462 | 463 | *Dhruv Shah, Blazej Osinski, Brian Ichter, Sergey Levine*. 464 | 465 | - **[WebShop: Towards Scalable Real-World Web Interaction with Grounded Language Agents](https://arxiv.org/abs/2207.01206)**, 2022.07 ![img](https://img.shields.io/badge/WebShop-blue)![img](https://img.shields.io/badge/-Reinforcement%20Learning,%20Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language,%20Formal%20Language-lightgrey) ![img](https://img.shields.io/badge/Affordance%20Grounding,%20Tool--use-green) 466 | 467 | *Shunyu Yao, Howard Chen, John Yang, Karthik Narasimhan*. 468 | 469 | - **[Improving Multimodal Interactive Agents with Reinforcement Learning from Human Feedback](https://arxiv.org/abs/2211.11602)**, 2022.11 ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) 470 | 471 | *Josh Abramson, Arun Ahuja, Federico Carnevale, Petko Georgiev, Alex Goldin, Alden Hung, Jessica Landon, Jirka Lhotka, Timothy Lillicrap, Alistair Muldal, George Powell, Adam Santoro, Guy Scully, Sanjana Srivastava, Tamara von Glehn, Greg Wayne, Nathaniel Wong, Chen Yan, Rui Zhu*. 472 | 473 | - **[VIMA: General Robot Manipulation with Multimodal Prompts](https://arxiv.org/abs/2210.03094)**, 2022.10 ![img](https://img.shields.io/badge/-VIMA-blue) ![img](https://img.shields.io/badge/-Message%20Fusion-orange) ![img](https://img.shields.io/badge/Foundation%20Model-green) 474 | 475 | *Yunfan Jiang, Agrim Gupta, Zichen Zhang, Guanzhi Wang, Yongqiang Dou, Yanjun Chen, Li Fei-Fei, Anima Anandkumar, Yuke Zhu, Linxi Fan*. 476 | 477 | - **[PIGLeT: Language Grounding Through Neuro-Symbolic Interaction in a 3D World](https://arxiv.org/abs/2106.00188)**, 2021.06 ![img](https://img.shields.io/badge/-PIGLeT-blue) ![img](https://img.shields.io/badge/-Standard%20Prompting-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) 478 | 479 | *Rowan Zellers, Ari Holtzman, Matthew Peters, Roozbeh Mottaghi, Aniruddha Kembhavi, Ali Farhadi, Yejin Choi*. 480 | 481 | - **[Describe, Explain, Plan and Select: Interactive Planning with Large Language Models Enables Open-World Multi-Task Agents](https://arxiv.org/abs/2302.01560)**, 2023.02 ![img](https://img.shields.io/badge/-Prompt%20Chaining,%20Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Formal%20Language,%20Natural%20Language-lightgrey) 482 | 483 | *Zihao Wang, Shaofei Cai, Anji Liu, Xiaojian Ma, Yitao Liang*. 484 | 485 | - **[ProgPrompt: Generating Situated Robot Task Plans using Large Language Models](https://arxiv.org/abs/2209.11302)**, 2022.09 ![img](https://img.shields.io/badge/-ProgPrompt-blue) ![img](https://img.shields.io/badge/-Elicitive%20Prompting,%20Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey) ![img](https://img.shields.io/badge/Affordance%20Grounding-green) 486 | 487 | *Ishika Singh, Valts Blukis, Arsalan Mousavian, Ankit Goyal, Danfei Xu, Jonathan Tremblay, Dieter Fox, Jesse Thomason, Animesh Garg*. 488 | 489 | - **[Plan,Eliminate,and Track-Language Models are Good Teachers for Embodied Agents.](https://arxiv.org/abs/2305.02412)**, 2023.05 ![img](https://img.shields.io/badge/-Standard%20Prompting,%20Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Affordance%20Grounding-green) 490 | 491 | *Yue Wu, So Yeon Min, Yonatan Bisk, Ruslan Salakhutdinov, Amos Azaria, Yuanzhi Li, Tom Mitchell, Shrimai Prabhumoye*. 492 | 493 | - **[Collaborating with language models for embodied reasoning](https://arxiv.org/abs/2302.00763)**, 2023.02 ![img](https://img.shields.io/badge/Planner--Actor--Reporter-blue)![img](https://img.shields.io/badge/-Reinforcement%20Learning,%20Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) 494 | 495 | *Ishita Dasgupta, Christine Kaeser-Chen, Kenneth Marino, Arun Ahuja, Sheila Babayan, Felix Hill, Rob Fergus*. 496 | 497 | - **[Open-World Object Manipulation using Pre-trained Vision-Language Models](https://arxiv.org/abs/2303.00905)**, 2023.03 ![img](https://img.shields.io/badge/-MOO-blue) 498 | 499 | *Austin Stone, Ted Xiao, Yao Lu, Keerthana Gopalakrishnan, Kuang-Huei Lee, Quan Vuong, Paul Wohlhart, Brianna Zitkovich, Fei Xia, Chelsea Finn, Karol Hausman*. 500 | 501 | - **[Mind's Eye: Grounded Language Model Reasoning through Simulation](https://arxiv.org/abs/2210.05359)**, 2022.10 ![img](https://img.shields.io/badge/-Elicitive%20Prompting-orange) ![img](https://img.shields.io/badge/-Formal%20Language-lightgrey) 502 | 503 | *Ruibo Liu, Jason Wei, Shixiang Shane Gu, Te-Yen Wu, Soroush Vosoughi, Claire Cui, Denny Zhou, Andrew M. Dai*. 504 | 505 | - **[Guiding Pretraining in Reinforcement Learning with Large Language Models](https://arxiv.org/abs/2302.06692)**, 2023.02 ![img](https://img.shields.io/badge/-ELLM-blue) ![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) 506 | 507 | *Yuqing Du, Olivia Watkins, Zihan Wang, Cédric Colas, Trevor Darrell, Pieter Abbeel, Abhishek Gupta, Jacob Andreas*. 508 | 509 | - **[Grounding Large Language Models in Interactive Environments with Online Reinforcement Learning](https://arxiv.org/abs/2302.02662)**, 2023.02 ![img](https://img.shields.io/badge/-GLAM-blue)![img](https://img.shields.io/badge/-Reinforcement%20Learning-orange) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Affordance%20Grounding-green) 510 | 511 | *Thomas Carta, Clément Romac, Thomas Wolf, Sylvain Lamprier, Olivier Sigaud, Pierre-Yves Oudeyer*. 512 | 513 | - **[Don't Generate, Discriminate: A Proposal for Grounding Language Models to Real-World Environments](https://arxiv.org/abs/2212.09736)**, 2022.12 ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) 514 | 515 | *Yu Gu, Xiang Deng, Yu Su*. 516 | 517 | 518 | ### 👍Evaluation 519 | - **[Evaluating Human-Language Model Interaction](https://arxiv.org/pdf/2212.09746.pdf)**, 2022.12 ![img](https://img.shields.io/badge/-HALIE-blue) ![img](https://img.shields.io/badge/Human--LM%20Interaction-green) 520 | 521 | *Mina Lee, Megha Srivastava, Amelia Hardy, John Thickstun, Esin Durmus, Ashwin Paranjape, Ines Gerard-Ursin, Xiang Lisa Li, Faisal Ladhak, Frieda Rong, Rose E. Wang, Minae Kwon, Joon Sung Park, Hancheng Cao, Tony Lee, Rishi Bommasani, Michael Bernstein, Percy Lia*. 522 | 523 | - **[ReCEval: Evaluating Reasoning Chains via Correctness and Informativeness](https://arxiv.org/pdf/2304.10703.pdf)**, 2023.04 ![img](https://img.shields.io/badge/-ReCEval-blue) ![img](https://img.shields.io/badge/-Prompt%20Chaining-orange) ![img](https://img.shields.io/badge/CoT-green) 524 | 525 | *Archiki Prasad, Swarnadeep Saha, Xiang Zhou, Mohit Bansa*. 526 | 527 | - **[Evaluating Verifiability in Generative Search Engines](https://arxiv.org/pdf/2304.09848.pdf)**, 2023.04 528 | 529 | *Liu, Nelson F., Zhang, Tianyi, Liang, Percy Lian*. 530 | 531 | - **[Multi-Party Chat: Conversational Agents in Group Settings with Humans and Models](https://arxiv.org/pdf/2304.13835.pdf)**, 2023.04 ![img](https://img.shields.io/badge/-Multi--Party%20Chat-blue) ![img](https://img.shields.io/badge/Communicative%20Agents-green) 532 | 533 | *Jimmy Wei, Kurt Shuster, Arthur Szlam, Jason Weston, Jack Urbanek, Mojtaba Komeili*. 534 | 535 | - **[Behavior-1k: A benchmark for embodied ai with 1,000 everyday activities and realistic simulation](https://proceedings.mlr.press/v205/li23a/li23a.pdf)**, 2023.03 ![img](https://img.shields.io/badge/-Behavior--1k-blue) ![img](https://img.shields.io/badge/Embodied%20AI-green) 536 | 537 | *Chengshu Li, Ruohan Zhang, Josiah Wong, Cem Gokmen, Sanjana Srivastava, Roberto Martín-Martín, Chen Wang, Gabrael Levine, Michael Lingelbach, Jiankai Sun, Mona Anvari, Minjune Hwang, Manasi Sharma, Arman Aydin, Dhruva Bansal, Samuel Hunter, Kyu-Young Kim, Alan Lou, Caleb R Matthews, Ivan Villa-Renteria, Jerry Huayang Tang, Claire Tang, Fei Xia, Silvio Savarese, Hyowon Gweon, Karen Liu, Jiajun Wu, Li Fei-Fei*. 538 | 539 | - **[ORBIT: A Unified Simulation Framework for Interactive Robot Learning Environments](https://arxiv.org/pdf/2301.04195.pdf)**, 2023.01 ![img](https://img.shields.io/badge/-ORBIT-blue) ![img](https://img.shields.io/badge/Embodied%20AI-green) 540 | 541 | *Mayank Mittal, Calvin Yu, Qinxi Yu, Jingzhou Liu, Nikita Rudin, David Hoeller, Jia Lin Yuan, Pooria Poorsarvi Tehrani, Ritvik Singh, Yunrong Guo, Hammad Mazhar, Ajay Mandlekar, Buck Babich, Gavriel State, Marco Hutter, Animesh Garg*. 542 | 543 | - **[Learn to Explain: Multimodal Reasoning via Thought Chains for Science Question Answering](https://arxiv.org/pdf/2209.09513.pdf)**, 2022.09 ![img](https://img.shields.io/badge/-ScienceQA-blue) ![img](https://img.shields.io/badge/CoT-green) 544 | 545 | *Pan Lu, Swaroop Mishra, Tony Xia, Liang Qiu, Kai-Wei Chang, Song-Chun Zhu, Oyvind Tafjord, Peter Clark, Ashwin Kalyan*. 546 | 547 | - **[Alexa Arena: A User-Centric Interactive Platform for Embodied AI](https://arxiv.org/pdf/2303.01586.pdf)**, 2023.03 ![img](https://img.shields.io/badge/-Alexa%20Arena-blue) ![img](https://img.shields.io/badge/Embodied%20AI-green) 548 | 549 | *Qiaozi Gao, Govind Thattai, Xiaofeng Gao, Suhaila Shakiah, Shreyas Pansare, Vasu Sharma, Gaurav Sukhatme, Hangjie Shi, Bofei Yang, Desheng Zheng, Lucy Hu, Karthika Arumugam, Shui Hu, Matthew Wen, Dinakar Guthy, Cadence Chung, Rohan Khanna, Osman Ipek, Leslie Ball, Kate Bland, Heather Rocker, Yadunandana Rao, Michael Johnston, Reza Ghanadan, Arindam Mandal, Dilek Hakkani Tur, Prem Natarajan*. 550 | 551 | - **[CB2: Collaborative Natural Language Interaction Research Platform](https://arxiv.org/pdf/2303.08127.pdf)**, 2023.03 ![img](https://img.shields.io/badge/-CB2-blue) ![img](https://img.shields.io/badge/-Natural%20Language-lightgrey) ![img](https://img.shields.io/badge/Collaborative%20Behavior-green) 552 | 553 | *Jacob Sharf, Mustafa Omer Gul, Yoav Artzi*. 554 | 555 | - **[An AI Dungeon Master’s Guide: Learning to Converse and Guide with Intents and Theory-of-Mind in Dungeons and Dragons](https://arxiv.org/pdf/2212.10060.pdf)**, 2022.12 ![img](https://img.shields.io/badge/Game,%20Theory--of--Mind-green) 556 | 557 | *Pei Zhou, Andrew Zhu, Jennifer Hu, Jay Pujara, Xiang Ren, Chris Callison-Burch, Yejin Choi, Prithviraj Ammanabrolu*. 558 | 559 | 560 | 561 | ### 🎨Application 562 | 563 | 564 | 565 | ## Related Projects 566 | 567 | - **[ToolLearningPapers](https://github.com/thunlp/)** 568 | - **[BMTools](https://github.com/OpenBMB/BMTools)** 569 | - **[AgentVerse](https://github.com/OpenBMB/AgentVerse)** 570 | - **[ChatArena](https://github.com/chatarena/chatarena)** 571 | - **[ChatGPT Plugins](https://platform.openai.com/docs/plugins/)** 572 | - **[LangChain](https://github.com/hwchase17/langchain)** 573 | - **[AutoGPT](https://github.com/Significant-Gravitas/Auto-GPT)** 574 | - **[BabyAGI](https://github.com/yoheinakajima/babyagi)** 575 | - **[Awesome-Multimodal-Large-Language-Models](https://github.com/BradyFU/Awesome-Multimodal-Large-Language-Models#multimodal-in-context-learning)** 576 | 577 | 578 | 579 | ## Contribution 580 | 581 | ### Contributors 582 | 583 | [ ![img](https://contrib.rocks/image?repo=InteractiveNLP-Team/awesome-InteractiveNLP-papers) ](https://github.com/InteractiveNLP-Team/awesome-InteractiveNLP-papers/graphs/contributors) 584 | 585 | ### Contributing to this paper list 586 | 587 | - There are cases where we miss important works in this field, please contribute to this repo! Thanks for the efforts in advance. 588 | 589 | ## Star History 590 | 591 | [![Star History Chart](https://api.star-history.com/svg?repos=InteractiveNLP-Team/awesome-InteractiveNLP-papers&type=Date)](https://star-history.com/#InteractiveNLP-Team/awesome-InteractiveNLP-papers&Date) 592 | 593 | ## Citation 594 | If you find this paper list helpful, please consider citing our paper: 595 | 596 | ```latex 597 | @article{wang2023interactive, 598 | title={Interactive Natural Language Processing}, 599 | author={Wang, Zekun and Zhang, Ge and Yang, Kexin and Shi, Ning and Zhou, Wangchunshu and Hao, Shaochun and Xiong, Guangzheng and Li, Yizhi and Sim, Mong Yuan and Chen, Xiuying and others}, 600 | journal={arXiv preprint arXiv:2305.13246}, 601 | year={2023} 602 | } 603 | ``` 604 | -------------------------------------------------------------------------------- /assets/overview.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/InteractiveNLP-Team/awesome-InteractiveNLP-papers/0d84a1896ac7a1ad2da10f6a13e6ca796cc217a4/assets/overview.png -------------------------------------------------------------------------------- /scripts/trans.py: -------------------------------------------------------------------------------- 1 | from typing import List 2 | 3 | import pandas as pd 4 | import pnlp 5 | 6 | 7 | must_cols = ["title", "url", "date", "author"] 8 | optional_cols = ["abbr", "method", "interface", "misc"] 9 | 10 | 11 | class Formater: 12 | 13 | base = "https://img.shields.io/badge/" 14 | 15 | @classmethod 16 | def abbr_format(cls, abbr): 17 | return f"{cls.base}-{abbr}-blue" 18 | 19 | @classmethod 20 | def method_format(cls, method): 21 | return f"{cls.base}-{method}-orange" 22 | 23 | @classmethod 24 | def interface_format(cls, interface): 25 | return f"{cls.base}-{interface}-lightgrey" 26 | 27 | @classmethod 28 | def misc_format(cls, misc): 29 | return f"{cls.base}{misc}-green" 30 | 31 | 32 | def trans(s: str) -> str: 33 | return s.replace(" ", "%20") 34 | 35 | 36 | def df2md(df: pd.DataFrame) -> List[str]: 37 | data = [] 38 | for v in df.itertuples(): 39 | date = str(v.date) 40 | year = date[:4] 41 | month = date[4:].strip(".").zfill(2) 42 | date = f"{year}.{month}" 43 | 44 | line = f"- **[{v.title}]({v.url})**, {date} " 45 | badges = [] 46 | for col in optional_cols: 47 | val = getattr(v, col) 48 | if pd.isna(val): 49 | continue 50 | val = val.replace("-", "--") 51 | func = getattr(Formater, col + "_format") 52 | val = trans(val) 53 | val_format = func(val) 54 | badges.append(f"![img]({val_format})") 55 | line += " ".join(badges) 56 | line += "\n\n" 57 | line += f" *{v.author}*." 58 | data.append(line + "\n") 59 | return data 60 | 61 | 62 | def check_must_cols(df: pd.DataFrame): 63 | for v in df.itertuples(): 64 | for col in must_cols: 65 | val = getattr(v, col) 66 | assert pd.notna(val), f"{col} of line {v.Index+1} cannot be null" 67 | 68 | 69 | def main(): 70 | import argparse 71 | from pathlib import Path 72 | 73 | cols = must_cols[:3] + optional_cols + must_cols[3:] 74 | parser = argparse.ArgumentParser( 75 | description="Transfer pandas excel to markdown needed") 76 | parser.add_argument( 77 | "-i", 78 | "--input_file", 79 | default=None, 80 | help=f"Input excel file with {cols} columns, without header") 81 | args = parser.parse_args() 82 | 83 | if not args.input_file: 84 | print("An input_file is needed, run `-h` to get more info") 85 | 86 | out_file = Path(args.input_file).stem 87 | df = pd.read_excel( 88 | args.input_file, 89 | header=None, 90 | names=cols) 91 | check_must_cols(df) 92 | data = df2md(df) 93 | pnlp.write_file(out_file + ".md", data) 94 | 95 | 96 | if __name__ == "__main__": 97 | main() 98 | --------------------------------------------------------------------------------