Neural Natural Language Inference Models Enhanced with External Knowledge

0
Structured data

A study conducted in the UK from 2009 to 2010 by leading scientists explored neonatal resuscitation practices in various neonatal units, aiming to assess adherence to international guidelines and identify differences between tertiary and non-tertiary care providers...

Read on arXivCardiologyLorem ipsum dolor sit amet, consectetur adipiscing elit. Sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat.

One Sentence Abstract

This paper explores the integration of external knowledge into state-of-the-art neural natural language inference models, resulting in improved performance on SNLI and MultiNLI datasets.

Simplified Abstract

Researchers are working on a tricky problem: helping computers understand and use human language. To do this, they use complex models based on artificial neural networks, which have shown great results so far. But, these models rely on a large amount of annotated data, which may not cover all the knowledge needed for this task.

In this study, the researchers want to explore how these models can use additional information from external sources to improve their performance. They create new models that incorporate this external knowledge and test their effectiveness using two popular datasets, SNLI and MultiNLI. The results show that the new models perform better than previous ones, reaching the best results achieved so far.

The main point of this research is to enhance the performance of computer models that understand and interpret human language by incorporating external knowledge. This new approach improves the accuracy and reliability of the models, which is important for advancing our understanding of how different countries work together in scientific collaborations. The researchers have developed a new tool that helps us see which countries work closely on science, and their work represents an important contribution to the field.

Study Fields

Main fields:

  • Natural Language Processing (NLP)
  • Neural Networks
  • Natural Language Inference (NLI)

Subfields:

  • Large annotated data
  • State-of-the-art neural natural language inference models
  • External knowledge incorporation
  • SNLI dataset
  • MultiNLI dataset

Study Objectives

  • Investigate if machines can learn all knowledge needed to perform natural language inference (NLI) from annotated data
  • Determine how neural-network-based NLI models can benefit from external knowledge
  • Develop NLI models that can leverage external knowledge
  • Enrich state-of-the-art neural NLI models with external knowledge
  • Evaluate the performance of the proposed models on SNLI and MultiNLI datasets

Conclusions

  • The challenge of modeling natural language inference has been significantly improved with the availability of large annotated data.
  • Neural-network-based inference models have shown state-of-the-art performance.
  • The study investigates whether machines can learn all the necessary knowledge for NLI from annotated data and explores the benefits of external knowledge in neural-network-based NLI models.
  • The paper proposes enriching state-of-the-art neural NLI models with external knowledge and demonstrates improved performance on SNLI and MultiNLI datasets.

References

R. Samuel, Gabor. Bowman, Christopher. Angeli, Christopher.D. Potts, None. Manning
Samuel R. Bowman, Gabor Angeli, Christopher Potts, and Christopher D. Manning. 2015. A large annotated corpus for learning natural language inference. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, EMNLP 2015, Lisbon, Portugal, September 17-21, 2015, pages 632–642.
Adina. Williams, Nikita. Nangia, Samuel.R. Bowman
Adina Williams, Nikita Nangia, and Samuel R. Bowman. 2017. A broad-coverage challenge corpus for sentence understanding through inference. CoRR, abs/1704.05426.
R. Samuel, Jon. Bowman, Abhinav. Gauthier, Raghav. Rastogi, Christopher.D. Gupta, Christopher. Manning, None. Potts
Samuel R. Bowman, Jon Gauthier, Abhinav Rastogi, Raghav Gupta, Christopher D. Manning, and Christopher Potts. 2016. A fast unified model for parsing and sentence understanding. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, ACL 2016, August 7-12, 2016, Berlin, Germany, Volume 1: Long Papers.
Hong. Yu, Tsendsuren. Munkhdalai
Hong Yu and Tsendsuren Munkhdalai. 2017b. Neural tree indexers for text understanding. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics, EACL 2017, Valencia, Spain, April 3-7, 2017, Volume 1: Long Papers, pages 11–21.
P. Ankur, Oscar. Parikh, Dipanjan. Täckström, Jakob. Das, None. Uszkoreit
Ankur P. Parikh, Oscar Täckström, Dipanjan Das, and Jakob Uszkoreit. 2016. A decomposable attention model for natural language inference. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, EMNLP 2016, Austin, Texas, USA, November 1-4, 2016, pages 2249–2255.
Lei. Sha, Baobao. Chang, Zhifang. Sui, Sujian. Li
Lei Sha, Baobao Chang, Zhifang Sui, and Sujian Li. 2016. Reading and thinking: Re-read LSTM unit for textual entailment recognition. In COLING 2016, 26th International Conference on Computational Linguistics, Proceedings of the Conference: Technical Papers, December 11-16, 2016, Osaka, Japan, pages 2870–2879.
Qian. Chen, Xiaodan. Zhu, Zhen-Hua. Ling, Si. Wei, Hui. Jiang, Diana. Inkpen, ;. None
Qian Chen, Xiaodan Zhu, Zhen-Hua Ling, Si Wei, Hui Jiang, and Diana Inkpen. 2017a. Enhanced LSTM for natural language inference. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, ACL 2017, Vancouver, Canada, July 30 - August 4, Volume 1: Long Papers, pages 1657–1668.
Qian. Chen, Xiaodan. Zhu, Zhen-Hua. Ling, Si. Wei, Hui. Jiang, Diana. Inkpen
Qian Chen, Xiaodan Zhu, Zhen-Hua Ling, Si Wei, Hui Jiang, and Diana Inkpen. 2017b. Recurrent neural network-based sentence encoder with gated attention for natural language inference. In Proceedings of the 2nd Workshop on Evaluating Vector Space Representations for NLP, RepEval@EMNLP 2017, Copenhagen, Denmark, September 8, 2017, pages 36–40.
Yi. Tay, Anh. Luu, Siu. Tuan, Hui. Cheung
Yi Tay, Luu Anh Tuan, and Siu Cheung Hui. 2018. A compare-propagate architecture with alignment factorization for natural language inference. CoRR, abs/1801.00102.
Bill. Maccartney
Bill MacCartney. 2009. Natural Language Inference. Ph.D. thesis, Stanford University.
Ido. Dagan, Oren. Glickman, Bernardo. Magnini
Ido Dagan, Oren Glickman, and Bernardo Magnini. 2005. The PASCAL recognising textual entailment challenge. In Machine Learning Challenges, Evaluating Predictive Uncertainty, Visual Object Classification and Recognizing Textual Entailment, First PASCAL Machine Learning Challenges Workshop, MLCW 2005, Southampton, UK, April 11-13, 2005, Revised Selected Papers, pages 177–190.
Adrian. Iftene, Alexandra. Balahur-Dobrescu
Adrian Iftene and Alexandra Balahur-Dobrescu. 2007. Proceedings of the ACL-PASCAL Workshop on Textual Entailment and Paraphrasing, chapter Hypothesis Transformation and Semantic Variability Rules Used in Recognizing Textual Entailment. Association for Computational Linguistics.
Jane. Bromley, Isabelle. Guyon, Yann. Lecun, Eduard. Säckinger, Roopak. Shah
Jane Bromley, Isabelle Guyon, Yann LeCun, Eduard Säckinger, and Roopak Shah. 1993. Signature verification using a siamese time delay neural network. In Advances in Neural Information Processing Systems 6, [7th NIPS Conference, Denver, Colorado, USA, 1993], pages 737–744.
Ivan. Vendrov, Ryan. Kiros, Sanja. Fidler, Raquel. Urtasun
Ivan Vendrov, Ryan Kiros, Sanja Fidler, and Raquel Urtasun. 2015. Order-embeddings of images and language. CoRR, abs/1511.06361.
Lili. Mou, Rui. Men, Ge. Li, Yan. Xu, Lu. Zhang, Rui. Yan, Zhi. Jin
Lili Mou, Rui Men, Ge Li, Yan Xu, Lu Zhang, Rui Yan, and Zhi Jin. 2016. Natural language inference by tree-based convolution and heuristic matching. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, ACL 2016, August 7-12, 2016, Berlin, Germany, Volume 2: Short Papers.
Yang. Liu, Chengjie. Sun, Lei. Lin, Xiaolong. Wang
Yang Liu, Chengjie Sun, Lei Lin, and Xiaolong Wang. 2016c. Learning natural language inference using bidirectional LSTM model and inner-attention. CoRR, abs/1605.09090.
Zhouhan. Lin, Minwei. Feng, Cícero. Nogueira Dos Santos, Mo. Yu, Bing. Xiang, Bowen. Zhou, Yoshua. Bengio
Zhouhan Lin, Minwei Feng, Cícero Nogueira dos Santos, Mo Yu, Bing Xiang, Bowen Zhou, and Yoshua Bengio. 2017. A structured self-attentive sentence embedding. CoRR, abs/1703.03130.
Yixin. Nie, Mohit. Bansal
Yixin Nie and Mohit Bansal. 2017. Shortcut-stacked sentence encoders for multi-domain inference. In Proceedings of the 2nd Workshop on Evaluating Vector Space Representations for NLP, RepEval@EMNLP 2017, Copenhagen, Denmark, September 8, 2017, pages 41–45.
Tao. Shen, Tianyi. Zhou, Guodong. Long, Jing. Jiang, Shirui. Pan, Chengqi. Zhang
Tao Shen, Tianyi Zhou, Guodong Long, Jing Jiang, Shirui Pan, and Chengqi Zhang. ...

References

Unlock full article access by joining Solve