• Stars
    star
    3,912
  • Rank 10,745 (Top 0.3 %)
  • Language
  • Created almost 3 years ago
  • Updated 11 months ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

Must-read papers on prompt-based tuning for pre-trained language models.

PromptPapers

We have released an open-source prompt-learning toolkit, check out OpenPrompt!

We strongly encourage the researchers that want to promote their fantastic work to the community to make pull request to update their paper's information! (See contributing details)

Effective adaptation of pre-trained models could be probed from different perspectives. Prompt-learning more focuses on the organization of training procedure and the unification of different tasks, while delta tuning (parameter efficient methods) provides another direction from the specific optimization of pre-trained models. Check DeltaPapers!

Contents

Must-read papers on prompt-based tuning for pre-trained language models. The paper list is mainly mantained by Ning Ding and Shengding Hu. Watch this repository for the latest updates!

Introduction

This is a paper list about prompt-based tuning for large-scale pre-trained language models. Different from traditional fine-tuning that uses an explicit classifier, prompt-based tuning directly uses the pre-trained models to conduct the pre-training tasks for classification or regression.

Keywords Convention

The abbreviation of the work.

The key features in terms of prompt learning used in the work.

The mainly explored task of the work.

The mainly explored property of prompt learning methods in the work.

Papers

Overview

This section contains the papers that overview the general trends in recent natural language processing with big (pretrained) models.

  1. OpenPrompt: An Open-source Framework for Prompt-learning. Preprint.

    Ning Ding, Shengding Hu, Weilin Zhao, Yulin Chen, Zhiyuan Liu, Hai-Tao Zheng, Maoson Sun [pdf] [project], 2021.11

  2. Pre-Trained Models: Past, Present and Future. Preprint.

    Xu Han, Zhengyan Zhang, Ning Ding, Yuxian Gu, Xiao Liu, Yuqi Huo, Jiezhong Qiu, Yuan Yao, Ao Zhang, Liang Zhang, Wentao Han, Minlie Huang, Qin Jin, Yanyan Lan, Yang Liu, Zhiyuan Liu, Zhiwu Lu, Xipeng Qiu, Ruihua Song, Jie Tang, Ji-Rong Wen, Jinhui Yuan, Wayne Xin Zhao, Jun Zhu. [pdf], 2021.6

  3. Pre-train, Prompt, and Predict: A Systematic Survey of Prompting Methods in Natural Language Processing. Preprint.

    Liu, Pengfei, Weizhe Yuan, Jinlan Fu, Zhengbao Jiang, Hiroaki Hayashi, and Graham Neubig. [pdf] [project], 2021.7

  4. Paradigm Shift in Natural Language Processing. Machine Intelligence Research.

    Tianxiang Sun, Xiangyang Liu, Xipeng Qiu, Xuanjing Huang [pdf] [project], 2021.9

Pilot Work

This section contains the pilot works that might contributes to the prevalence of prompt learning paradigm.

  1. Parameter-Efficient Transfer Learning for NLP. ICML 2019.

    Neil Houlsby, Andrei Giurgiu, Stanislaw Jastrzebski, Bruna Morrone, Quentin de Laroussilhe, Andrea Gesmundo, Mona Attariyan, Sylvain Gelly. [pdf], [project], 2019.6

  2. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. JMLR. Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, Peter J. Liu. [pdf], [project]. 2019.10.

  3. Language Models as Knowledge Bases? EMNLP 2019.

    Fabio Petroni, Tim Rocktaschel, Patrick Lewis, Anton Bakhtin, Yuxiang Wu, Alexander H. Miller, Sebastian Riedel. [pdf], [project] , 2019.9

  4. How Can We Know What Language Models Know? TACL 2020.

    Zhengbao Jiang, Frank F. Xu, Jun Araki, Graham Neubig. [pdf], [project], 2019.11

  5. Language Models are Few-shot Learners. NeurIPS 2020.

    Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, Dario Amodei. [pdf], [website], 2020.5

  6. AdaPrompt: Adaptive Model Training for Prompt-based NLP

    Yulong Chen, Yang Liu, Li Dong, Shuohang Wang, Chenguang Zhu, Michael Zeng, Yue Zhang [pdf], 2022.02

Basics

This section contains the exploration on the basic aspects of prompt tuning, such as template, verbalizer, training paradigms, etc.

  1. Exploiting Cloze Questions for Few Shot Text Classification and Natural Language Inference. EACL 2021.

    Timo Schick, Hinrich Schütze. [pdf], [project], 2020.1

  2. It’s Not Just Size That Matters: Small Language Models Are Also Few-Shot Learners. NAACL 2021.

    Timo Schick, Hinrich Schütze. [pdf], [project], 2020.9

  3. Autoprompt: Eliciting knowledge from language models with automatically generated prompts. Preprint.

    Taylor Shin, Yasaman Razeghi, Robert L. Logan IV, Eric Wallace, Sameer Singh. [pdf], [website], 2020.10

  4. Automatically Identifying Words That Can Serve as Labels for Few-Shot Text Classification. COLING 2020.

    Timo Schick, Helmut Schmid, Hinrich Schütze. [pdf], [project], 2020.12

  5. Making Pre-trained Language Models Better Few-shot Learners. ACL 2021.

    Tianyu Gao, Adam Fisch, Danqi Chen. [pdf], [project], 2020.12

  6. Prefix-tuning: Optimizing continuous prompts for generation. ACL 2021.

    Xiang Lisa Li, Percy Liang. [pdf], [project], 2021.1

  7. Prompt Programming for Large Language Models: Beyond the Few-Shot Paradigm. Preprint.

    Laria Reynolds, Kyle McDonell. [pdf], 2021.2

  8. Improving and Simplifying Pattern Exploiting Training. Preprint.

    Derek Tam, Rakesh R Menon, Mohit Bansal, Shashank Srivastava, Colin Raffel. [pdf], 2021.3

  9. GPT understands, too. Preprint.

    Xiao Liu, Yanan Zheng, Zhengxiao Du, Ming Ding, Yujie Qian, Zhilin Yang, Jie Tang. [pdf], [project], 2021.3

  10. The Power of Scale for Parameter-Efficient Prompt Tuning. Preprint.

    Brian Lester, Rami Al-Rfou, Noah Constant. [pdf], [project], 2021.4

  11. Learning How to Ask: Querying LMs with Mixtures of Soft Prompts. NAACL 2021.

    Guanghui Qin, Jason Eisner. [pdf][project], 2021.4

  12. Factual Probing Is [MASK]: Learning vs. Learning to Recall. NAACL 2021.

    Zexuan Zhong, Dan Friedman, Danqi Chen. [pdf], [project], 2021.4

  13. Cutting Down on Prompts and Parameters: Simple Few-Shot Learning with Language Models. Preprint.

    Robert L. Logan IV, Ivana Balažević, Eric Wallace, Fabio Petroni, Sameer Singh, Sebastian Riedel. [pdf], 2021.6

  14. WARP: Word-level Adversarial ReProgramming. ACL 2021.

    Karen Hambardzumyan, Hrant Khachatrian, Jonathan May. [pdf], [project], 2021.6

  15. PTR: Prompt Tuning with Rules for Text Classification. Preprint.

    Xu Han, Weilin Zhao, Ning Ding, Zhiyuan Liu, Maosong Sun. [pdf], 2021.5

  16. NSP-BERT: A Prompt-based Zero-Shot Learner Through an Original Pre-training Task —— Next Sentence Prediction

    Yi Sun*, Yu Zheng*, Chao Hao, Hangping Qiu, [pdf], [project], 2021.9

  17. Finetuned language models are zero-shot learners.

    ason Wei, Maarten Bosma, Vincent Y. Zhao, Kelvin Guu, Adams Wei Yu, Brian Lester, Nan Du, Andrew M. Dai, Quoc V. Le. [pdf], 2021.9

  18. PPT: Pre-trained Prompt Tuning for Few-shot Learning

    Yuxian Gu*, Xu Han*, Zhiyuan Liu, Minlie Huang. [pdf], 2021.9

  19. Differentiable Prompt Makes Pre-trained Language Models Better Few-shot Learners. ICLR 2022.

    Ningyu Zhang, Luoqiu Li, Xiang Chen, Shumin Deng, Zhen Bi, Chuanqi Tan, Fei Huang, Huajun Chen. [pdf], [project], 2021.10

  20. Multitask Prompted Training Enables Zero-Shot Task Generalization.

    Victor Sanh, Albert Webson, Colin Raffel, Stephen H. Bach, Lintang Sutawika, Zaid Alyafeai, Antoine Chaffin, Arnaud Stiegler, Teven Le Scao, Arun Raja, Manan Dey, M Saiful Bari, Canwen Xu, Urmish Thakker, Shanya Sharma Sharma, Eliza Szczechla, Taewoon Kim, Gunjan Chhablani, Nihal Nayak, Debajyoti Datta, Jonathan Chang, Mike Tian-Jian Jiang, Han Wang, Matteo Manica, Sheng Shen, Zheng Xin Yong, Harshit Pandey, Rachel Bawden, Thomas Wang, Trishala Neeraj, Jos Rozen, Abheesht Sharma, Andrea Santilli, Thibault Fevry, Jason Alan Fries, Ryan Teehan, Stella Biderman, Leo Gao, Tali Bers, Thomas Wolf, Alexander M. Rush. [pdf], 2021.10

  21. P-Tuning v2: Prompt Tuning Can Be Comparable to Finetuning Universally Across Scales and Tasks. ACL 2022.

    Xiao Liu, Kaixuan Ji, Yicheng Fu, Zhengxiao Du, Zhilin Yang, Jie Tang [pdf], [project], 2021.10

  22. Black-Box Tuning for Language-Model-as-a-Service. ICML 2022.

    Tianxiang Sun, Yunfan Shao, Hong Qian, Xuanjing Huang, Xipeng Qiu [pdf], [project], 2022.1

  23. Black-box Prompt Learning for Pre-trained Language Models. Preprint.

    Shizhe Diao, Xuechun Li, Yong Lin, Zhichao Huang, Tong Zhang [pdf], 2022.1

  24. Binding Language Models in Symbolic Languages. Preprint.

    Zhoujun Cheng*, Tianbao Xie*, Peng Shi, Chengzu Li, Rahul Nadkarni, Yushi Hu, Caiming Xiong, Dragomir Radev, Mari Ostendorf, Luke Zettlemoyer, Noah A. Smith, Tao Yu [pdf], [project], [website], 2022.10

  25. A Prompt Pattern Catalog to Enhance Prompt Engineering with ChatGPT. Jules White, Quchen Fu, Sam Hays, Michael Sandborn, Carlos Olea, Henry Gilbert, Ashraf Elnashar, Jesse Spencer-Smith, Douglas C. Schmidt [pdf], 2023.2

Analysis

This section contains the analysis of prompt learning methods, including but not limited to why does prompt learning work, various properties of prompt learning methods, limilation of prompt learning methods.

  1. What Makes Good In-Context Examples for GPT-3?. Preprint.

    Jiachang Liu, Dinghan Shen, Yizhe Zhang, Bill Dolan, Lawrence Carin, Weizhu Chen. [pdf] 2021.1

  2. How Many Data Points is a Prompt Worth? NAACL 2021.

    Teven Le Scao, Alexander M. Rush. [pdf], [project], 2021.3

  3. Surface Form Competition-Why the Highest Probability Answer Isn’t Always Right. Preprint. Preprint.

    Ari Holtzman, Peter West, Vered Schwartz, Yejin Choi, Luke Zettlemoyer. [pdf][project], 2021.4

  4. Natural Instructions: Benchmarking Generalization to New Tasks from Natural Language Instructions. Preprint.

    Swaroop Mishra, Daniel Khashabi, Chitta Baral, Hannaneh Hajishirzi. [pdf], [project], 2021.4

  5. Fantastically Ordered Prompts and Where to Find Them: Overcoming Few-Shot Prompt Order Sensitivity. Preprint.

    Yao Lu, Max Bartolo, Alastair Moore, Sebastian Riedel, Pontus Stenetorp. [pdf] 2021.4

  6. Meta-tuning Language Models to Answer Prompts Better. Preprint.

    Ruiqi Zhong, Kristy Lee*, Zheng Zhang*, Dan Klein. [pdf] 2021.4

  7. True Few-Shot Learning with Language Models. Preprint.

    Ethan Perez, Douwe Kiela, Kyunghyun Cho. [pdf], [project] 2021.5

  8. Why Do Pretrained Language Models Help in Downstream Tasks? An Analysis of Head and Prompt Tuning. Preprint.

    Colin Wei Sang Michael Xie Tengyu Ma [pdf], 2021.6

  9. Do Prompt-Based Models Really Understand the Meaning of their Prompts? Preprint.

    Albert Webson, Ellie Pavlick. [pdf], [project] 2021.9

  10. Avoiding Inference Heuristics in Few-shot Prompt-based Finetuning. Preprint.

    Prasetya Ajie Utama, Nafise Sadat Moosavi, Victor Sanh, Iryna Gurevych. [pdf], 2021.9

  11. Towards a Unified View of Parameter-Efficient Transfer Learning. Preprint.

    Junxian He, Chunting Zhou, Xuezhe Ma, Taylor Berg-Kirkpatrick, Graham Neubig. [pdf], 2021.10

  12. Exploring Low-dimensional Intrinsic Task Subspace via Prompt Tuning. Preprint.

    Yujia Qin, Xiaozhi Wang, Yusheng Su, Yankai Lin, Ning Ding, Zhiyuan Liu, Juanzi Li, Lei Hou,Peng Li, Maosong Sun, Jie Zhou [pdf]

  13. Exploring the Universal Vulnerability of Prompt-based Learning Paradigm. Findings of NAACL 2022.

    Lei Xu, Yangyi Chen, Ganqu Cui, Hongcheng Gao, Zhiyuan Liu [pdf], [project]

  14. Rethinking the Role of Demonstrations: What Makes In-Context Learning Work?. Arxiv 2022.

    Sewon Min, Xinxi Lyu, Ari Holtzman, Mikel Artetxe, Mike Lewis, Hannaneh Hajishirzi, Luke Zettlemoyer [pdf], [project]

  15. Parameter-Efficient Prompt Tuning Makes Generalized and Calibrated Neural Text Retrievers. Preprint.

    Weng Lam Tam, Xiao Liu, Kaixuan Ji, Lilong Xue, Yuxiao Dong, Jiahua Liu, Maodi Hu, Jie Tang [pdf] [project]

  16. Ignore Previous Prompt: Attack Techniques For Language Models. Best Paper Award @ NeurIPS ML Safety Workshop 2022.

    Fábio Perez, Ian Ribeiro [pdf] [project], 2022.11

Improvements

This section contains the improvement of the basic prompt tuning methods, include but not limitedd to using additional resources to improving the performances, making up the shortcomings of previous work or conducting prompt tuning in unsual ways.

  1. Calibrate Before Use: Improving Few-Shot Performance of Language Models. Preprint.

    Tony Z. Zhao, Eric Wallace, Shi Feng, Dan Klein, Sameer Singh. [pdf], [project], 2021.2

  2. Text Generation with Efficient (Soft) Q-Learning. Preprint.

    Han Guo, Bowen Tan, Zhengzhong Liu, Eric P. Xing, Zhiting Hu. [pdf], 2021.6

  3. Knowledgeable Prompt-tuning: Incorporating Knowledge into Prompt Verbalizer for Text Classification. Preprint.

    Shengding Hu, Ning Ding, Huadong Wang, Zhiyuan Liu, Juanzi Li, Maosong Sun. [pdf], [project], 2021.8

  4. Noisy Channel Language Model Prompting for Few-Shot Text Classification. Preprint.

    Sewon Min, Mike Lewis, Hannaneh Hajishirzi, Luke Zettlemoyer. [pdf], 2021.8

  5. Adapting Language Models for Zero-shot Learning by Meta-tuning on Dataset and Prompt Collection.

    Ruiqi Zhong, Kristy Lee* Zheng Zhang*, Dan Klein. [pdf], 2021.9

  6. Revisiting Self-Training for Few-Shot Learning of Language Model. Preprint.

    Yiming Chen, Yan Zhang, Chen Zhang, Grandee Lee, Ran Cheng, Haizhou Li. [pdf], 2021.10

  7. LiST: Lite Self-training Makes Efficient Few-shot Learners. Preprint.

    Yaqing Wang, Subhabrata Mukherjee, Xiaodong Liu, Jing Gao, Ahmed Hassan Awadallah, Jianfeng Gao. [pdf], 2021.10

  8. Prototypical Verbalizer for Prompt-based Few-shot Tuning. ACL 2022.

    Ganqu Cui, Shengding Hu, Ning Ding, Longtao Huang, Zhiyuan Liu. [pdf], [project], 2022.3

  9. BBTv2: Pure Black-Box Optimization Can Be Comparable to Gradient Descent for Few-Shot Learning. Preprint.

    Tianxiang Sun, Zhengfu He, Hong Qian, Xuanjing Huang, Xipeng Qiu [pdf] [project], 2022.5

Specializations

This section contains the prompt learning methods designed for various NLP task.

  1. Thinking Aloud: Dynamic Context Generation Improves Zero-Shot Reasoning Performance of GPT-2. Preprint.

    Gregor Betz, Kyle Richardson, Christian Voigt. [pdf] 2021.3

  2. GPT3Mix: Leveraging Large-scale Language Models for Text Augmentation. Preprint.

    Kang Min Yoo, Dongju Park, Jaewook Kang, Sang-Woo Lee, Woomyeong Park. [pdf] 2021.4

  3. Constrained Language Models Yield Few-Shot Semantic Parsers. Preprint.

    Richard Shin, Christopher H. Lin, Sam Thomson, Charles Chen, Subhro Roy, Emmanouil Antonios Platanios, Adam Pauls, Dan Klein, Jason Eisner, Benjamin Van Durme. [pdf] 2021.4

  4. Label Verbalization and Entailment for Effective Zero- and Few-Shot Relation Extraction. EMNLP 2021.

    Oscar Sainz, Oier Lopez de Lacalle, Gorka Labaka, Ander Barrena, Eneko Agirre. [pdf], 2021.4

  5. PADA: A Prompt-based Autoregressive Approach for Adaptation to Unseen Domains Preprint.

    Eyal Ben-David, Nadav Oved, Roi Reichart. [pdf][project] 2021.5

  6. Prompt-Learning for Fine-grained Entity Typing. Preprint.

    Ning Ding, Yulin Chen, Xu Han, Guangwei Xu, Pengjun Xie, Hai-Tao Zheng, Zhiyuan Liu, Juanzi Li, Hong-Gee Kim [pdf],2021.8

  7. KnowPrompt: Knowledge-aware Prompt-tuning with Synergistic Optimization for Relation Extraction. WWW 2022.

    Xiang Chen, Xin Xie, Ningyu Zhang, Jiahuan Yan, Shumin Deng, Chuanqi Tan, Fei Huang, Luo Si, Huajun Chen.. [pdf], [project], 2021.9

  8. Exploring Prompt-based Few-shot Learning for Grounded Dialog Generation. Preprint.

    Chujie Zheng, Minlie Huang. [pdf], 2021.9

  9. SentiPrompt: Sentiment Knowledge Enhanced Prompt-Tuning for Aspect-Based Sentiment Analysis. Preprint.

    Chengxi Li, Feiyu Gao, Jiajun Bu, Lu Xu, Xiang Chen, Yu Gu, Zirui Shao, Qi Zheng, Ningyu Zhang, Yongpan Wang, Zhi Yu. [pdf] 2021.9

  10. Template-free Prompt Tuning for Few-shot NER. Preprint.

    Ruotian Ma*, Xin Zhou*, Tao Gui, Yiding Tan, Qi Zhang, Xuanjing Huang. [pdf], 2021.9

  11. Learning to Prompt for Vision-Language Models. Preprint.

    Kaiyang Zhou, Jingkang Yang, Chen Change Loy, and Ziwei Liu. [pdf], 2021.9

  12. CPT: Colorful Prompt Tuning for Pre-trained Vision-Language Models. Preprint.

    Yuan Yao*, Ao Zhang*, Zhengyan Zhang, Zhiyuan Liu, Tat-Seng Chua, Maosong Sun. [pdf], 2021.10

  13. MSP: Multi-Stage Prompting for Making Pre-trained Language Models Better Translators. Preprint.

    Zhixing Tan, Xiangwen Zhang, Shuo Wang, Yang Liu. [pdf], 2021.10

  14. Few-Shot Bot: Prompt-Based Learning for Dialogue Systems. Preprint.

    Andrea Madotto, Zhaojiang Lin, Genta Indra Winata, Pascale Fung [pdf], 2021.10

  15. Control Prefixes for Text Generation. Preprint.

    Jordan Clive, Kris Cao, Marek Rei. [pdf], 2021.10

  16. The Power of Prompt Tuning for Low-Resource Semantic Parsing. Preprint.

    Nathan Schucher, Siva Reddy, Harm de Vries. [pdf], 2021.10

  17. A Good Prompt Is Worth Millions of Parameters? Low-resource Prompt-based Learning for Vision-Language Models.

    Woojeong Jin, Yu Cheng, Yelong Shen, Weizhu Chen, Xiang Ren. [pdf]

  18. LightNER: A Lightweight Generative Framework with Prompt-guided Attention for Low-resource NER. COLING 2022.

    Xiang Chen, Lei Li, Shumin Deng, Chuanqi Tan, Changliang Xu, Fei Huang, Luo Si, Huajun Chen, Ningyu Zhang. [pdf], [project], 2021.8

  19. UnifiedSKG: Unifying and Multi-Tasking Structured Knowledge Grounding with Text-to-Text Language Models.

    Tianbao Xie*, Chen Henry Wu*, Peng Shi, Ruiqi Zhong, Torsten Scholak, Michihiro Yasunaga, Chien-Sheng Wu, Ming Zhong, Pengcheng Yin, Sida I. Wang, Victor Zhong, Bailin Wang, Chengzu Li, Connor Boyle, Ansong Ni, Ziyu Yao, Dragomir Radev, Caiming Xiong, Lingpeng Kong, Rui Zhang, Noah A. Smith, Luke Zettlemoyer, Tao Yu.[pdf], [project], [website], 2022.1

  20. Ontology-enhanced Prompt-tuning for Few-shot Learning. WWW 2022.

    Hongbin Ye, Ningyu Zhang, Shumin Deng, Xiang Chen, Hui Chen, Feiyu Xiong, Xi Chen, Huajun Chen. [pdf], 2022.1

  21. Learning to Prompt for Continual Learning. CVPR 2022.

    Zifeng Wang, Zizhao Zhang, Chen-Yu Lee, Han Zhang, Ruoxi Sun, Xiaoqi Ren, Guolong Su, Vincent Perot, Jennifer Dy, Tomas Pfister. [pdf], [project], 2021.12

  22. Relation Extraction as Open-book Examination: Retrieval-enhanced Prompt Tuning. SIGIR 2022.

    Xiang Chen, Lei Li, Ningyu Zhang, Chuanqi Tan, Fei Huang, Luo Si, Huajun Chen. [pdf], [project], 2022.5

  23. Good Visual Guidance Makes A Better Extractor: Hierarchical Visual Prefix for Multimodal Entity and Relation Extraction. Findings of NAACL 2022.

    Xiang Chen, Ningyu Zhang, Lei Li, Yunzhi Yao, Shumin Deng, Chuanqi Tan, Fei Huang, Luo Si, Huajun Chen. [pdf], [project], 2022.5

  24. Chain of Thought Prompting Elicits Reasoning in Large Language Models. Preprint 2022.

    Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Ed Chi, Quoc Le, Denny Zhou. [pdf]

  25. Self-Consistency Improves Chain of Thought Reasoning in Language Models. Preprint 2022.

    Xuezhi Wang, Jason Wei, Dale Schuurmans, Quoc Le, Ed Chi, Sharan Narang, Aakanksha Chowdhery, Denny Zhou. [pdf]

  26. Large Language Models are Zero-Shot Reasoners. Preprint 2022.

    Takeshi Kojima, Shixiang Shane Gu, Machel Reid, Yutaka Matsuo, Yusuke Iwasawa. [pdf]

  27. Least-to-Most Prompting Enables Complex Reasoning in Large Language Models. Preprint 2022.

    Denny Zhou, Nathanael Schärli, Le Hou, Jason Wei, Nathan Scales, Xuezhi Wang, Dale Schuurmans, Olivier Bousquet, Quoc Le, Ed Chi. [pdf]

  28. Maieutic Prompting: Logically Consistent Reasoning with Recursive Explanations. Preprint 2022.

    Jaehun Jung, Lianhui Qin, Sean Welleck, Faeze Brahman, Chandra Bhagavatula, Ronan Le Bras, Yejin Choi [pdf]

  29. On the Advance of Making Language Models Better Reasoners. Preprint 2022.

    Yifei Li, Zeqi Lin, Shizhuo Zhang, Qiang Fu, Bei Chen, Jian-Guang Lou, Weizhu Chen [pdf]

  30. Learning to Compose Soft Prompts for Compositional Zero-Shot Learning. Preprint 2022.

    Nihal V. Nayak*, Peilin Yu*, Stephen H. Bach [pdf], [project]

  31. Decoupling Knowledge from Memorization: Retrieval-augmented Prompt Learning. NeurIPS 2022.

    Xiang Chen, Lei Li, Ningyu Zhang, Xiaozhuan Liang, Shumin Deng, Chuanqi Tan, Fei Huang, Luo Si, Huajun Chen. [pdf], [project], 2022.5

  32. Exploring Length Generalization in Large Language Models. Preprint 2022.

    Cem Anil, Yuhuai Wu, Anders Andreassen, Aitor Lewkowycz, Vedant Misra, Vinay Ramasesh, Ambrose Slone, Guy Gur-Ari, Ethan Dyer, Behnam Neyshabur [pdf]

  33. Ask Me Anything: A simple strategy for prompting language models. Preprint 2022.

    Simran Arora, Avanika Narayan, Mayee F. Chen, Laurel Orr, Neel Guha, Kush Bhatia, Ines Chami, Frederic Sala, Christopher Ré [pdf]

  34. Measuring And Narrowing The Compositionality Gap In Language Models Preprint 2022.

    Ofir Press, Muru Zhang, Sewon Min, Ludwig Schmidt, Noah A. Smith, Mike Lewis [pdf]

  35. RLPrompt: Optimizing Discrete Text Prompts with Reinforcement Learning Preprint 2022.

    Mingkai Deng, Jianyu Wang, Cheng-Ping Hsieh, Yihan Wang, Han Guo, Tianmin Shu, Meng Song, Eric P. Xing, Zhiting Hu [pdf]

  36. Reasoning with Language Model Prompting: A Survey Preprint 2022.

    Shuofei Qiao, Yixin Ou, Ningyu Zhang, Xiang Chen, Yunzhi Yao, Shumin Deng, Chuanqi Tan, Fei Huang, Huajun Chen [pdf]

Contribution

Other contributors

We thank Yujia Qin, Xiachong Feng, Chenglei Si , Tianbao Xie, Muhtasham Oblokulov for the paper recommendation.

Contributing to this paper list

  • First, think about which category the work should belong to.
  • Second, use the same format as the others to discribe the work. Note that there should be an empty line between the title and the authors list, and take care of the indentation.
  • Then, add keywords tags. Add the pdf link of the paper. If it is an arxiv publication, we prefer /abs/ format to /pdf/ format.

Don't worry if you put all these wrong, we will fix them for you. Just contribute and promote your awesome work here!

If you recommended a work that wasn't yours, you will be added to the contributor list (be sure to provide your information in other contributors).

More Repositories

1

GNNPapers

Must-read papers on graph neural networks (GNN)
15,490
star
2

WantWords

An open-source online reverse dictionary.
JavaScript
6,933
star
3

OpenNRE

An Open-Source Package for Neural Relation Extraction (NRE)
Python
4,232
star
4

OpenPrompt

An Open-Source Framework for Prompt-Learning.
Python
4,145
star
5

OpenKE

An Open-Source Package for Knowledge Embedding (KE)
Python
3,725
star
6

PLMpapers

Must-read Papers on pre-trained language models.
3,161
star
7

NRLPapers

Must-read papers on network representation learning (NRL) / network embedding (NE)
TeX
2,520
star
8

UltraChat

Large-scale, Informative, and Diverse Multi-round Chat Data (and Models)
Python
2,118
star
9

THULAC-Python

An Efficient Lexical Analyzer for Chinese
Python
1,972
star
10

OpenNE

An Open-Source Package for Network Embedding (NE)
Python
1,672
star
11

KRLPapers

Must-read papers on knowledge representation learning (KRL) / knowledge embedding (KE)
TeX
1,528
star
12

TAADpapers

Must-read Papers on Textual Adversarial Attack and Defense
Python
1,459
star
13

ERNIE

Source code and dataset for ACL 2019 paper "ERNIE: Enhanced Language Representation with Informative Entities"
Python
1,403
star
14

KB2E

Knowledge Graph Embeddings including TransE, TransH, TransR and PTransE
C++
1,360
star
15

NREPapers

Must-read papers on neural relation extraction (NRE)
TeX
1,023
star
16

OpenCLaP

Open Chinese Language Pre-trained Model Zoo
971
star
17

WebCPM

Official codes for ACL 2023 paper "WebCPM: Interactive Web Search for Chinese Long-form Question Answering"
HTML
952
star
18

OpenDelta

A plug-and-play library for parameter-efficient-tuning (Delta Tuning)
Python
938
star
19

RCPapers

Must-read papers on Machine Reading Comprehension
890
star
20

NRE

Neural Relation Extraction, including CNN, PCNN, CNN+ATT, PCNN+ATT
C++
812
star
21

ToolLearningPapers

777
star
22

THULAC

An Efficient Lexical Analyzer for Chinese
C++
772
star
23

FewRel

A Large-Scale Few-Shot Relation Extraction Dataset
Python
716
star
24

THUOCL

THUOCL(THU Open Chinese Lexicon)中文词库
697
star
25

Chinese_Rumor_Dataset

中文谣言数据
672
star
26

OpenAttack

An Open-Source Package for Textual Adversarial Attack.
Python
652
star
27

DocRED

Dataset and codes for ACL 2019 DocRED: A Large-Scale Document-Level Relation Extraction Dataset.
Python
605
star
28

OpenHowNet

Core Data of HowNet and OpenHowNet Python API
Python
592
star
29

TensorFlow-TransX

An implementation of TransE and its extended models for Knowledge Representation Learning on TensorFlow
Python
511
star
30

LegalPapers

Must-read Papers on Legal Intelligence
450
star
31

OpenMatch

An Open-Source Package for Information Retrieval.
Python
444
star
32

CAIL

Chinese AI & Law Challenge
439
star
33

BERT-KPE

Python
437
star
34

Fast-TransX

An Efficient implementation of TransE and its extended models for Knowledge Representation Learning
C++
396
star
35

TensorFlow-Summarization

Python
390
star
36

Few-NERD

Code and data of ACL 2021 paper "Few-NERD: A Few-shot Named Entity Recognition Dataset"
Python
376
star
37

SOS4NLP

Survey of Surveys for Natural Language Processing (SOS4NLP)
327
star
38

THULAC-Java

An Efficient Lexical Analyzer for Chinese
Java
325
star
39

NSC

Neural Sentiment Classification
Python
287
star
40

BMCourse

The repo for Tsinghua summer course: Interdisciplinary Seminar on Big Models
Python
269
star
41

Chinese_NRE

Source code for ACL 2019 paper "Chinese Relation Extraction with Multi-Grained Information and External Linguistic Knowledge"
Python
264
star
42

DeltaPapers

Must-read Papers of Parameter-Efficient Tuning (Delta Tuning) Methods on Pre-trained Models.
259
star
43

PL-Marker

Source code for "Packed Levitated Marker for Entity and Relation Extraction"
Python
252
star
44

SE-WRL

Improved Word Representation Learning with Sememes
C
197
star
45

THUCTC

An Efficient Chinese Text Classifier
Java
196
star
46

InfLLM

The code of our paper "InfLLM: Unveiling the Intrinsic Capacity of LLMs for Understanding Extremely Long Sequences with Training-Free Memory"
Python
196
star
47

SCPapers

Must-read Papers on Sememe Computation
193
star
48

KnowledgeablePromptTuning

kpt code
Python
192
star
49

CANE

Source code and datasets of "CANE: Context-Aware Network Embedding for Relation Modeling"
Python
190
star
50

JointNRE

Joint Neural Relation Extraction with Text and KGs
Python
185
star
51

HATT-Proto

Code and dataset of AAAI2019 paper Hybrid Attention-Based Prototypical Networks for Noisy Few-Shot Relation Classification
Python
180
star
52

LLaVA-UHD

LLaVA-UHD: an LMM Perceiving Any Aspect Ratio and High-Resolution Images
Python
169
star
53

NLP-THU

NLP Course Material & QA
164
star
54

KernelGAT

The source codes for Fine-grained Fact Verification with Kernel Graph Attention Network.
Python
161
star
55

LegalPLMs

Source code and checkpoints for legal pre-trained language models.
Python
158
star
56

EntityDuetNeuralRanking

Entity-Duet Neural Ranking Model
Python
153
star
57

PTR

Prompt Tuning with Rules
Python
151
star
58

OOP-THU

OOP Course Material & QA
149
star
59

Auto_CLIWC

Code for Chinese LIWC Lexicon Expansion via Hierarchical Classification of Word Embeddings with Sememe Attention (AAAI18)
Python
136
star
60

OpenBackdoor

An open-source toolkit for textual backdoor attack and defense (NeurIPS 2022 D&B, Spotlight)
Python
135
star
61

attribute_charge

The source code of our COLING'18 paper "Few-Shot Charge Prediction with Discriminative Legal Attributes".
Python
126
star
62

ConceptFlow

Python
119
star
63

THUCKE

THU Chinese Keyphrase Extraction Toolkit
C++
118
star
64

CAIL2018

Python
111
star
65

KR-EAR

Knowledge Representation Learning with Entities, Attributes and Relations
C++
111
star
66

Neural-Snowball

Code and dataset of AAAI2020 Paper Neural Snowball for Few-Shot Relation Learning
Python
111
star
67

ChatEval

Codes for our paper "ChatEval: Towards Better LLM-based Evaluators through Multi-Agent Debate"
Python
109
star
68

MultiRD

Code and data of the AAAI-20 paper "Multi-channel Reverse Dictionary Model"
Python
106
star
69

TransNet

Source code and datasets of IJCAI2017 paper "TransNet: Translation-Based Network Representation Learning for Social Relation Extraction".
Jupyter Notebook
103
star
70

RE-Context-or-Names

Bert-based models(BERT, MTB, CP) for relation extraction.
Python
100
star
71

AGE

Source code and dataset for KDD 2020 paper "Adaptive Graph Encoder for Attributed Graph Embedding"
Python
99
star
72

GEAR

Source code for ACL 2019 paper "GEAR: Graph-based Evidence Aggregating and Reasoning for Fact Verification"
Python
95
star
73

HNRE

Hierarchical Neural Relation Extraction
Python
95
star
74

LEVEN

Source code and dataset for ACL2022 Findings Paper "LEVEN: A Large-Scale Chinese Legal Event Detection dataset"
Python
94
star
75

TopJudge

Python
93
star
76

Prompt-Transferability

On Transferability of Prompt Tuning for Natural Language Processing
Python
85
star
77

SememePSO-Attack

Code and data of the ACL 2020 paper "Word-level Textual Adversarial Attacking as Combinatorial Optimization"
Python
85
star
78

XQA

Dataset and baseline for ACL 2019 paper "XQA: A Cross-lingual Open-domain Question Answering Dataset"
Python
84
star
79

HMEAE

Source code for EMNLP-IJCNLP 2019 paper "HMEAE: Hierarchical Modular Event Argument Extraction".
Python
84
star
80

ERICA

Source code for ACL 2021 paper "ERICA: Improving Entity and Relation Understanding for Pre-trained Language Models via Contrastive Learning"
Python
82
star
81

CLAIM

77
star
82

TKRL

Representation Learning of Knowledge Graphs with Hierarchical Types (IJCAI-2016)
C++
76
star
83

TLNN

Source code for EMNLP-IJCNLP 2019 paper "Event Detection with Trigger-Aware Lattice Neural Network".
Python
75
star
84

MMDW

Max-margin DeepWalk
Java
71
star
85

KV-PLM

Source code for "A Deep-learning System Bridging Molecule Structure and Biomedical Text with Comprehension Comparable to Human Professionals"
Python
71
star
86

KNET

Neural Entity Typing with Knowledge Attention
Python
69
star
87

SelectiveMasking

Source code for "Train No Evil: Selective Masking for Task-Guided Pre-Training"
Python
68
star
88

NeuIRPapers

Must-read Papers on Neural Information Retrieval
68
star
89

MoEfication

Python
66
star
90

Adv-ED

Source code and dataset for NAACL 2019 paper "Adversarial Training for Weakly Supervised Event Detection".
Python
66
star
91

CorefBERT

Source code for EMNLP 2020 paper "Coreferential Reasoning Learning for Language Representation"
Python
65
star
92

ConversationQueryRewriter

Code and Data for SIGIR 2020 Paper "Few-Shot Generative Conversational Query Rewriting"
Roff
63
star
93

MuGNN

Source code for ACL2019 paper "Multi-Channel Graph Neural Network for Entity Alignment".
Python
62
star
94

sememe_prediction

Codes for Lexical Sememe Prediction via Word Embeddings and Matrix Factorization (IJCAI 2017).
Python
60
star
95

DIAG-NRE

Source code for ACL 2019 paper "DIAG-NRE: A Neural Pattern Diagnosis Framework for Distantly Supervised Neural Relation Extraction".
Python
59
star
96

topical_word_embeddings

Topical Word Embeddings
Python
57
star
97

QuoteR

Official code and data of the ACL 2022 paper "QuoteR: A Benchmark of Quote Recommendation for Writing"
Python
57
star
98

paragraph2vec

Paragraph Vector Implementation
Python
56
star
99

DKRL

Representation Learning of Knowledge Graphs with Entity Descriptions (AAAI-2016)
C++
54
star
100

Ouroboros

Ouroboros: Speculative Decoding with Large Model Enhanced Drafting
Python
51
star