Enriching Transformers with Structured Tensor-Product Representations for Abstractive Summarization
暂无分享,去创建一个
Asli Celikyilmaz | Hamid Palangi | Mohit Bansal | Jianfeng Gao | Paul Smolensky | Sudha Rao | Paul Soulos | Yichen Jiang | Roland Fernandez | Caitlin Smith | P. Smolensky | Mohit Bansal | Jianfeng Gao | Asli Celikyilmaz | H. Palangi | Paul Soulos | Yichen Jiang | Sudha Rao | Roland Fernandez | Caitlin Smith
[1] Geoffrey E. Hinton. Tensor Product Variable Binding and the Representation of Symbolic Structures in Connectionist Systems , 1991 .
[2] Chin-Yew Lin,et al. ROUGE: A Package for Automatic Evaluation of Summaries , 2004, ACL 2004.
[3] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[4] Alon Lavie,et al. Meteor Universal: Language Specific Translation Evaluation for Any Target Language , 2014, WMT@ACL.
[5] Mihai Surdeanu,et al. The Stanford CoreNLP Natural Language Processing Toolkit , 2014, ACL.
[6] Phil Blunsom,et al. Teaching Machines to Read and Comprehend , 2015, NIPS.
[7] Geoffrey E. Hinton,et al. Layer Normalization , 2016, ArXiv.
[8] Bowen Zhou,et al. Abstractive Text Summarization using Sequence-to-sequence RNNs and Beyond , 2016, CoNLL.
[9] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[10] Christopher D. Manning,et al. Get To The Point: Summarization with Pointer-Generator Networks , 2017, ACL.
[11] Richard Socher,et al. A Deep Reinforced Model for Abstractive Summarization , 2017, ICLR.
[12] Yejin Choi,et al. Deep Communicating Agents for Abstractive Summarization , 2018, NAACL.
[13] Yang Liu,et al. Learning Structured Text Representations , 2017, TACL.
[14] Mirella Lapata,et al. Don’t Give Me the Details, Just the Summary! Topic-Aware Convolutional Neural Networks for Extreme Summarization , 2018, EMNLP.
[15] Ramakanth Pasunuru,et al. Multi-Reward Reinforced Summarization with Saliency and Entailment , 2018, NAACL.
[16] Mor Naaman,et al. Newsroom: A Dataset of 1.3 Million Summaries with Diverse Extractive Strategies , 2018, NAACL.
[17] Fei Liu,et al. Abstract Meaning Representation for Multi-Document Summarization , 2018, COLING.
[18] Noam Shazeer,et al. Adafactor: Adaptive Learning Rates with Sublinear Memory Cost , 2018, ICML.
[19] Franck Dernoncourt,et al. A Discourse-Aware Attention Model for Abstractive Summarization of Long Documents , 2018, NAACL.
[20] Luke S. Zettlemoyer,et al. AllenNLP: A Deep Semantic Natural Language Processing Platform , 2018, ArXiv.
[21] Lin Zhao,et al. Structure-Infused Copy Mechanisms for Abstractive Summarization , 2018, COLING.
[22] William Yang Wang,et al. WikiHow: A Large Scale Text Summarization Dataset , 2018, ArXiv.
[23] Li Deng,et al. Question-Answering with Grammatically-Interpretable Representations , 2017, AAAI.
[24] Li Deng,et al. Tensor Product Generation Networks for Deep NLP Modeling , 2017, NAACL.
[25] Jürgen Schmidhuber,et al. Learning to Reason with Third-Order Tensor Products , 2018, NeurIPS.
[26] Robert Frank,et al. Open Sesame: Getting inside BERT’s Linguistic Knowledge , 2019, BlackboxNLP@ACL.
[27] Chang Liu,et al. Attentive Tensor Product Learning , 2019, AAAI.
[28] Christopher D. Manning,et al. A Structural Probe for Finding Syntax in Word Representations , 2019, NAACL.
[29] Mirella Lapata,et al. Hierarchical Transformers for Multi-Document Summarization , 2019, ACL.
[30] Dipanjan Das,et al. BERT Rediscovers the Classical NLP Pipeline , 2019, ACL.
[31] Franck Dernoncourt,et al. Scoring Sentence Singletons and Pairs for Abstractive Summarization , 2019, ACL.
[32] Ewan Dunbar,et al. RNNs Implicitly Implement Tensor Product Representations , 2018, ICLR.
[33] Jianfeng Gao,et al. Enhancing the Transformer with Explicit Relational Encoding for Math Problem Solving , 2019, ArXiv.
[34] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[35] Alex Wang,et al. What do you learn from context? Probing for sentence structure in contextualized word representations , 2019, ICLR.
[36] Mona T. Diab,et al. FEQA: A Question Answering Evaluation Framework for Faithfulness Assessment in Abstractive Summarization , 2020, ACL.
[37] Colin Raffel,et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..
[38] Jianfeng Gao,et al. Mapping natural-language problems to formal-language solutions using structured neural representations , 2019, ICML.
[39] R. Thomas McCoy,et al. Discovering the Compositional Structure of Vector Representations with Role Learning Networks , 2019, BLACKBOXNLP.
[40] Ryan McDonald,et al. On Faithfulness and Factuality in Abstractive Summarization , 2020, ACL.
[41] Peter J. Liu,et al. PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization , 2019, ICML.