Learning deep models of optimization landscapes

In all but the most trivial optimization problems, the structure of the solutions exhibit complex interdependences between the input parameters. Decades of research with stochastic search techniques has shown the benefit of explicitly modeling the interactions between sets of parameters and the overall quality of the solutions discovered. We demonstrate a novel method, based on learning deep networks, to model the global landscapes of optimization problems. To represent the search space concisely and accurately, the deep networks must encode information about the underlying parameter interactions and their contributions to the quality of the solution. Once the networks are trained, the networks are probed to reveal parameter combinations with high expected performance with respect to the optimization task. These estimates are used to initialize fast, randomized, local-search algorithms, which in turn expose more information about the search space that is subsequently used to refine the models. We demonstrate the technique on multiple problems that have arisen in a variety of real-world domains, including: packing, graphics, job scheduling, layout and compression. Strengths, limitations, and extensions of the approach are extensively discussed and demonstrated.

[1]  Lei Yang,et al.  Bayesian Optimization Algorithm Based on Incremental Model Building , 2015, ISICA.

[2]  David Heckerman,et al.  A Tutorial on Learning with Bayesian Networks , 1998, Learning in Graphical Models.

[3]  Ramin Zabih,et al.  Some Applications of Graph Bandwidth to Constraint Satisfaction Problems , 1990, AAAI.

[4]  Fred W. Glover,et al.  Tabu Search - Part I , 1989, INFORMS J. Comput..

[5]  John H. Holland,et al.  Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence , 1992 .

[6]  S. Baluja,et al.  Using Optimal Dependency-Trees for Combinatorial Optimization: Learning the Structure of the Search Space , 1997 .

[7]  Judea Pearl,et al.  Bayesian Networks , 1998, Encyclopedia of Social Network Analysis and Mining. 2nd Ed..

[8]  Markus H ohfeld,et al.  Random keys genetic algorithm with adaptive penalty function for optimization of constrained facility layout problems , 1997, Proceedings of 1997 IEEE International Conference on Evolutionary Computation (ICEC '97).

[9]  Tapani Raiko,et al.  Techniques for Learning Binary Stochastic Feedforward Neural Networks , 2014, ICLR.

[10]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[11]  John H. Holland,et al.  When will a Genetic Algorithm Outperform Hill Climbing , 1993, NIPS.

[12]  Shumeet Baluja,et al.  Deep Learning for Explicitly Modeling Optimization Landscapes , 2017, ArXiv.

[13]  Paul A. Viola,et al.  MIMIC: Finding Optima by Estimating Probability Densities , 1996, NIPS.

[14]  Martin Pelikan,et al.  Hill Climbing with Learning (An Abstraction of Genetic Algorithm) , 1995 .

[15]  D. Goldberg,et al.  BOA: the Bayesian optimization algorithm , 1999 .

[16]  C. N. Liu,et al.  Approximating discrete probability distributions with dependence trees , 1968, IEEE Trans. Inf. Theory.

[17]  Shumeet Baluja,et al.  A Method for Integrating Genetic Search Based Function Optimization and Competitive Learning , 1994 .

[18]  S. Baluja An Empirical Comparison of Seven Iterative and Evolutionary Function Optimization Heuristics , 1995 .

[19]  Pedro Larrañaga,et al.  Analyzing the Population Based Incremental Learning Algorithm by Means of Discrete Dynamical Systems , 2000, Complex Syst..

[20]  A. Juels,et al.  Topics in black-box combinatorial optimization , 1996 .

[21]  C. D. Gelatt,et al.  Optimization by Simulated Annealing , 1983, Science.

[22]  Reza Rastegar,et al.  A Convergence Proof for the Population Based Incremental Learning Algorithm , 2005, 17th IEEE International Conference on Tools with Artificial Intelligence (ICTAI'05).

[23]  Pedro Larrañaga,et al.  The Convergence Behavior of the PBIL Algorithm: A Preliminary Approach , 2001 .

[24]  David E. Goldberg,et al.  The compact genetic algorithm , 1999, IEEE Trans. Evol. Comput..

[25]  Andrew W. Moore,et al.  Learning Evaluation Functions to Improve Optimization by Local Search , 2001, J. Mach. Learn. Res..

[26]  Leon A. Gatys,et al.  A Neural Algorithm of Artistic Style , 2015, ArXiv.

[27]  Ari Juels,et al.  Stochastic Hillclimbing as a Baseline Method for , 1994 .

[28]  David E. Goldberg,et al.  Genetic Algorithms in Search Optimization and Machine Learning , 1988 .

[29]  H. Mühlenbein,et al.  From Recombination of Genes to the Estimation of Distributions I. Binary Parameters , 1996, PPSN.

[30]  A. Linden,et al.  Inversion of multilayer nets , 1989, International 1989 Joint Conference on Neural Networks.

[31]  Fred Glover,et al.  Tabu Search - Part II , 1989, INFORMS J. Comput..

[32]  Leon A. Gatys,et al.  Texture Synthesis Using Convolutional Neural Networks , 2015, NIPS.

[33]  Kurt Hornik,et al.  Multilayer feedforward networks are universal approximators , 1989, Neural Networks.

[34]  Norman E. Gibbs,et al.  The bandwidth problem for graphs and matrices - a survey , 1982, J. Graph Theory.

[35]  Shumeet Baluja,et al.  Fast Probabilistic Modeling for Combinatorial Optimization , 1998, AAAI/IAAI.