Making Pre-trained Language Models Better Few-shot Learners

0
Structured data

A study conducted in the UK from 2009 to 2010 by leading scientists explored neonatal resuscitation practices in various neonatal units, aiming to assess adherence to international guidelines and identify differences between tertiary and non-tertiary care providers...

Read on arXivCardiologyLorem ipsum dolor sit amet, consectetur adipiscing elit. Sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat.

One Sentence Abstract

"LM-BFF is presented as a task-agnostic method for few-shot learning, significantly improving upon standard fine-tuning procedures on various NLP tasks, through prompt-based fine-tuning and a novel pipeline for automating prompt generation, and a refined strategy for dynamically incorporating demonstrations into each context."

Simplified Abstract

Researchers developed a new and improved technique called LM-BFF (Better Few-Shot Fine-Tuning of Language Models) to help computers understand and learn from a small amount of information, like when you teach a new task to a language model using just a few examples. This method is like having a toolbox with helpful techniques to improve the learning process.

The LM-BFF method includes two main parts:

  1. Prompt-based fine-tuning: This is like giving the computer a helpful hint or tutorial to understand the new task. The researchers also created a system that automatically creates these hints, making it easier and more efficient.
  2. Refined strategy for demonstrations: This is like carefully selecting and incorporating a few examples to help the computer learn the new task. By doing this, the computer can learn faster and more accurately.

The researchers tested the effectiveness of their new method using a variety of tasks like classification and regression. The results showed that LM-BFF outperforms standard methods in situations where there are limited resources and information, achieving up to a 30% improvement and an average of 11% improvement across all tasks. This means that LM-BFF can be a helpful tool for various tasks, regardless of the specific domain or the amount of information available.

The LM-BFF implementation is available for anyone to use, making it easier for researchers to apply this innovative method to their work and improve the accuracy and reliability of their findings.

Study Fields

Main fields:

  • Natural Language Processing (NLP)
  • Machine Learning (ML)
  • Language Models

Subfields:

  • Few-shot learning
  • Fine-tuning
  • Prompt-based fine-tuning
  • Automated prompt generation
  • Demonstration incorporation strategies
  • Task-agnostic methods
  • Performance evaluation
  • Classification and regression

Study Objectives

  • Investigate few-shot learning in a practical scenario using smaller language models for which fine-tuning is computationally efficient.
  • Develop LM-BFF: a suite of simple and complementary techniques for fine-tuning language models on a small number of annotated examples, including:
    • Prompt-based fine-tuning with a novel pipeline for automating prompt generation.
    • A refined strategy for dynamically and selectively incorporating demonstrations into each context.
  • Conduct a systematic evaluation to analyze few-shot performance on a range of NLP tasks, such as classification and regression.
  • Compare the performance of LM-BFF against standard fine-tuning procedures in a low resource setting, aiming for up to 30% absolute improvement, and 11% on average across all tasks.
  • Ensure that the approach is task-agnostic and makes minimal assumptions on task resources and domain expertise, making it a strong method for few-shot learning.
  • Make their implementation publicly available (https://github.com/princeton-nlp/LM-BFFCardiologyLorem ipsum dolor sit amet, consectetur adipiscing elit. Sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat.).

Conclusions

  • The study investigates few-shot learning in practical scenarios using smaller language models, making fine-tuning computationally efficient.
  • The authors present LM-BFF, a suite of techniques that include prompt-based fine-tuning with an automated prompt generation pipeline, and a refined strategy for incorporating demonstrations dynamically and selectively.
  • The researchers systematically evaluate the few-shot performance on a range of NLP tasks, including classification and regression.
  • LM-BFF achieves up to a 30% absolute improvement and an 11% average improvement across all tasks compared to standard fine-tuning procedures, demonstrating its effectiveness in the low resource setting.
  • The approach is task-agnostic and requires minimal assumptions on task resources and domain expertise, making it a strong method for few-shot learning in practical scenarios.
  • The implementation of LM-BFF is publicly available at https://github.com/princeton-nlp/LM-BFFCardiologyLorem ipsum dolor sit amet, consectetur adipiscing elit. Sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat..

References

Benjamin. Tom B Brown, Nick. Mann, Melanie. Ryder, Jared. Subbiah, Prafulla. Kaplan, Arvind. Dhariwal, Pranav. Neelakantan, Girish. Shyam, Amanda. Sastry, None. Askell
Tom B Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. 2020. Language models are few-shot learners. In Advances in Neural Information Processing Systems (NeurIPS).
Jacob. Devlin, Ming-Wei. Chang, Kenton. Lee, Kristina. Toutanova
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional Transformers for language understanding. In North American Chapter of the Association for Computational Linguistics (NAACL).
Yinhan. Liu, Myle. Ott, Naman. Goyal, Jingfei. Du, Mandar. Joshi, Danqi. Chen, Omer. Levy, Mike. Lewis, Luke. Zettlemoyer, Veselin. Stoyanov
Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. RoBERTa: A robustly optimized BERT pretraining approach. arXiv preprint arXiv:1907.11692.
Alec. Radford, Karthik. Narasimhan, Tim. Salimans, Ilya. Sutskever
Alec Radford, Karthik Narasimhan, Tim Salimans, and Ilya Sutskever. 2018. Improving language understanding by generative pre-training. Technical report, OpenAI.
Alec. Radford, Jeff. Wu, Rewon. Child, David. Luan, Dario. Amodei, Ilya. Sutskever
Alec Radford, Jeff Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. 2019. Language models are unsupervised multitask learners. Technical report, OpenAI.
Timo. Schick, Hinrich. Schütze
Timo Schick and Hinrich Schütze. 2021a. Exploiting cloze questions for few-shot text classification and natural language inference. In European Chapter of the Association for Computational Linguistics (EACL).
Timo. Schick, Hinrich. Schütze
Timo Schick and Hinrich Schütze. 2021b. It’s not just size that matters: Small language models are also few-shot learners. In North American Chapter of the Association for Computational Linguistics (NAACL).
Colin. Raffel, Noam. Shazeer, Adam. Roberts, Katherine. Lee, Sharan. Narang, Michael. Matena, Yanqi. Zhou, Wei. Li, Peter.J. LiuThe Journal of Machine Learning Research (JMLR)
Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J Liu. 2020. Exploring the limits of transfer learning with a unified text-to-text Transformer. The Journal of Machine Learning Research (JMLR), 21(140).
H. Trieu, Quoc V. Trinh, None. Le
Trieu H Trinh and Quoc V Le. 2018. A simple method for commonsense reasoning. arXiv preprint arXiv:1806.02847.
Fabio. Petroni, Tim. Rocktäschel, Sebastian. Riedel, Patrick. Lewis, Anton. Bakhtin, Yuxiang. Wu, Alexander. Miller
Fabio Petroni, Tim Rocktäschel, Sebastian Riedel, Patrick Lewis, Anton Bakhtin, Yuxiang Wu, and Alexander Miller. 2019. Language models as knowledge bases? In Empirical Methods in Natural Language Processing (EMNLP).
Joe. Davison, Joshua. Feldman, Alexander.M. Rush
Joe Davison, Joshua Feldman, and Alexander M Rush. 2019. Commonsense knowledge mining from pretrained models. In Empirical Methods in Natural Language Processing (EMNLP).
Alon. Talmor, Yanai. Elazar, Yoav. Goldberg, Jonathan. BerantTransactions of the Association of Computational Linguistics (TACL)
Alon Talmor, Yanai Elazar, Yoav Goldberg, and Jonathan Berant. 2020. oLMpics-on what language model pre-training captures. Transactions of the Association of Computational Linguistics (TACL), 8.
Timo. Schick, Helmut. Schmid, Hinrich. Schütze
Timo Schick, Helmut Schmid, and Hinrich Schütze. 2020. Automatically identifying words that can serve as labels for few-shot text classification. In International Conference on Computational Linguistics (COLING).
Zhengbao. Jiang, Frank.F. Xu, Jun. Araki, Graham. Neubig
Zhengbao Jiang, Frank F Xu, Jun Araki, and Graham Neubig. 2020. How can we know what language models know? Transactions of the Association of Computational Linguistics (TACL).
Taylor. Shin, Yasaman. Razeghi, Robert.L. Logan, I.V. None, Eric. Wallace, Sameer. Singh
Taylor Shin, Yasaman Razeghi, Robert L. Logan IV, Eric Wallace, and Sameer Singh. 2020. AutoPrompt: Automatic prompt construction for masked language models. In Empirical Methods in Natural Language Processing (EMNLP).
Zexuan. Zhong, Dan. Friedman, Danqi. Chen
Zexuan Zhong, Dan Friedman, and Danqi Chen. 2021. Factual probing is [MASK]: Learning vs. learning to recall. In North American Association for Computational Linguistics (NAACL).</FootnoteDefi...

References

Unlock full article access by joining Solve