A Semi-Supervised Learning Approach to Why-Question Answering

We propose a semi-supervised learning method for improving why-question answering (why-QA). The key of our method is to generate training data (question-answer pairs) from causal relations in texts such as "[Tsunamis are generated]effect because [the ocean's water mass is displaced by an earthquake]cause." A naive method for the generation would be to make a question-answer pair by simply converting the effect part of the causal relations into a why-question, like "Why are tsunamis generated?" from the above example, and using the source text of the causal relations as an answer. However, in our preliminary experiments, this naive method actually failed to improve the why-QA performance. The main reason was that the machine-generated questions were often incomprehensible like "Why does (it) happen?", and that the system suffered from overfitting to the results of our automatic causality recognizer. Hence, we developed a novel method that effectively filters out incomprehensible questions and retrieves from texts answers that are likely to be paraphrases of a given causal relation. Through a series of experiments, we showed that our approach significantly improved the precision of the top answer by 8% over the current state-of-the-art system for Japanese why-QA.

[1]  Hans van Halteren,et al.  Learning to rank for why-question answering , 2011, Information Retrieval.

[2]  Masaru Kitsuregawa,et al.  Polynomial to Linear: Efficient Classification with Conjunctive Features , 2009, EMNLP.

[3]  Thorsten Joachims,et al.  Making large scale SVM learning practical , 1998 .

[4]  Jong-Hoon Oh,et al.  Intra-sentential Zero Anaphora Resolution using Subject Sharing Recognition , 2015, EMNLP.

[5]  Roxana Gîrju,et al.  Automatic Detection of Causal Relations for Question Answering , 2003, ACL 2003.

[6]  Vladimir N. Vapnik,et al.  The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.

[7]  Jong-Hoon Oh,et al.  Bilingual Co-Training for Monolingual Hyponymy-Relation Acquisition , 2009, ACL/IJCNLP.

[8]  Jong-Hoon Oh,et al.  Why Question Answering using Sentiment Analysis and Word Classes , 2012, EMNLP.

[9]  Jong-Hoon Oh,et al.  Generating Event Causality Hypotheses through Semantic Relations , 2015, AAAI.

[10]  Ryuichiro Higashinaka,et al.  Corpus-based Question Answering for why-Questions , 2008, IJCNLP.

[11]  Edwin Lughofer,et al.  Hybrid active learning for reducing the annotation effort of operators in classification systems , 2012, Pattern Recognit..

[12]  Jong-Hoon Oh,et al.  Excitatory or Inhibitory: A New Semantic Orientation Extracts Contradiction and Causality from the Web , 2012, EMNLP.

[13]  Suzan Verberne,et al.  What Is Not in the Bag of Words for Why-QA? , 2010, CL.

[14]  Tianshun Yao,et al.  Active Learning with Sampling by Uncertainty and Density for Word Sense Disambiguation and Text Classification , 2008, COLING.

[15]  William A. Gale,et al.  A sequential algorithm for training text classifiers , 1994, SIGIR '94.

[16]  Yuji Matsumoto,et al.  Annotating a Japanese Text Corpus with Predicate-Argument and Coreference Relations , 2007, LAW@ACL.

[17]  Kentaro Torisawa,et al.  Inducing Gazetteers for Named Entity Recognition by Large-Scale Clustering of Dependency Relations , 2008, ACL.

[18]  Masaki Murata,et al.  A System for Answering Non-Factoid Japanese Questions by Using Passage Retrieval Weighted Based on Type of Answer , 2007, NTCIR.

[19]  Jingbo Zhu,et al.  Uncertainty-based active learning with instability estimation for text classification , 2012, TSLP.

[20]  Yutaka Kidawara,et al.  Toward Future Scenario Generation: Extracting Event Causality Exploiting Semantic Relation, Context, and Association Features , 2014, ACL.

[21]  Hang Li,et al.  Uncertainty Reduction in Collaborative Bootstrapping: Measure and Algorithm , 2003, ACL.

[22]  Andrew McCallum,et al.  Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data , 2001, ICML.

[23]  Jong-Hoon Oh,et al.  Why-Question Answering using Intra- and Inter-Sentential Causal Relations , 2013, ACL.

[24]  Jong-Hoon Oh,et al.  Large-Scale Acquisition of Entailment Pattern Pairs by Exploiting Transitivity , 2015, EMNLP.

[25]  Burr Settles,et al.  Active Learning , 2012, Synthesis Lectures on Artificial Intelligence and Machine Learning.

[26]  Jong-Hoon Oh,et al.  Two-Stage Method for Large-Scale Acquisition of Contradiction Pattern Pairs using Entailment , 2013, EMNLP.

[27]  Jong-Hoon Oh,et al.  Co-STAR: A Co-training Style Algorithm for Hyponymy Relation Acquisition from Structured and Unstructured Text , 2010, COLING.

[28]  Lou Boves,et al.  Using Syntactic Information for Improving Why-Question Answering , 2008, COLING.

[29]  Qingsheng Zhu,et al.  A multi-view approach to semi-supervised document classification with incremental Naive Bayes , 2009, Comput. Math. Appl..

[30]  Alessandro Moschitti,et al.  Making Tree Kernels Practical for Natural Language Learning , 2006, EACL.

[31]  Daphne Koller,et al.  Support Vector Machine Active Learning with Applications to Text Classification , 2002, J. Mach. Learn. Res..