PRIMER - A Regression-Rule Learning System for Intervention Optimization

We introduce intervention optimization as a new area of exploration for data mining research. Interventions are events designed to impact a corresponding time series. The task is to maximize the impact of such events by training a model on historical data. We propose PRIMER as a new regression-rule learning system for identifying sets of event features that maximize impact. PRIMER is for use when domain experts with knowledge of the intervention can specify a transfer function, or the form of the expected response in the time series. PRIMER’s objective function includes the goodness-of-fit of the average response of covered events to the transfer function. Incorporating domain knowledge in this way makes PRIMER robust to over-fitting on noise or spurious responses. PRIMER is designed to produce interpretable results, improving on the interpretability of even competing regression-rule systems for this task. It also has fewer and more intuitive parameters than competing rule-based systems. Empirically, we show that PRIMER is competitive with state-of-the-art regression techniques in a large-scale event study modeling the impact of insider trading on intra-day stock returns.

[1]  Peter Clark,et al.  The CN2 induction algorithm , 2004, Machine Learning.

[2]  Trevor Hastie,et al.  Regularization Paths for Generalized Linear Models via Coordinate Descent. , 2010, Journal of statistical software.

[3]  Ivan Bratko,et al.  Why Is Rule Learning Optimistic and How to Correct It , 2006, ECML.

[4]  K. Ramesh,et al.  Market Reaction Surrounding the Filing of Periodic SEC Reports , 2009 .

[5]  G. Seber,et al.  Nonlinear Regression: Seber/Nonlinear Regression , 2005 .

[6]  Ian H. Witten,et al.  The WEKA data mining software: an update , 2009, SKDD.

[7]  Chih-Jen Lin,et al.  LIBLINEAR: A Library for Large Linear Classification , 2008, J. Mach. Learn. Res..

[8]  Xiao-Jun Zhang,et al.  Financial reporting complexity and investor underreaction to 10-K information , 2009 .

[9]  Thomas F. Coleman,et al.  A Reflective Newton Method for Minimizing a Quadratic Function Subject to Bounds on Some of the Variables , 1992, SIAM J. Optim..

[10]  Wojciech Kotlowski,et al.  Solving Regression by Learning an Ensemble of Decision Rules , 2006, ICAISC.

[11]  Chia-Hua Ho,et al.  Large-scale linear support vector regression , 2012, J. Mach. Learn. Res..

[12]  Johannes Fürnkranz,et al.  Heuristic Rule-Based Regression via Dynamic Reduction to Classification , 2011, IJCAI.

[13]  JOHANNES FÜRNKRANZ,et al.  Separate-and-Conquer Rule Learning , 1999, Artificial Intelligence Review.

[14]  R. Mike Cameron-Jones,et al.  Oversearching and Layered Search in Empirical Learning , 1995, IJCAI.

[15]  Johannes Fürnkranz,et al.  A Re-evaluation of the Over-Searching Phenomenon in Inductive Rule Learning , 2008, LWA.

[16]  Fernando E. B. Otero,et al.  Discovering Regression Rules with Ant Colony Optimization , 2015, GECCO.

[17]  George E. P. Box,et al.  Intervention Analysis with Applications to Economic and Environmental Problems , 1975 .

[18]  Johannes Fürnkranz,et al.  Separate-and-conquer Regression , 2010, LWA.

[19]  Geoff Holmes,et al.  Generating Rule Sets from Model Trees , 1999, Australian Joint Conference on Artificial Intelligence.

[20]  Johannes Fürnkranz,et al.  Foundations of Rule Learning , 2012, Cognitive Technologies.

[21]  Andrej Pázman,et al.  Nonlinear Regression , 2019, Handbook of Regression Analysis With Applications in R.