Machine Learning Seminar: Principled Deep Neural Network Training through Linear Programming

Wednesday, October 24, 2018 - 3:30pm

Speaker(s): 
Daniel Bienstock (Columbia University)

Deep Learning has received significant attention due to its impressive performance in many state-of-the-art learning tasks. Unfortunately, while very powerful, Deep Learning is not well understood theoretically and in particular only recently results for the complexity of training deep neural networks have been obtained. In this work we show that large classes of deep neural networks with various architectures (e.g., DNNs, CNNs, Binary Neural Networks, and ResNets), activation functions (e.g., ReLUs and leaky ReLUs), and loss functions (e.g., Hinge loss, Euclidean loss, etc) can be trained to near optimality with desired target accuracy using linear programming in time that is exponential in the size of the architecture and polynomial in the size of the data set; this is the best one can hope for due to the NP-Hardness of the problem and in line with previous work. In particular, we obtain polynomial time algorithms for training for a given fixed network architecture. Our work applies more broadly to empirical risk minimization problems which allows us to generalize various previous results and obtain new complexity results for previously unstudied architectures in the proper learning setting. Joint work with G. Munoz (Montreal) and S. Pokutta (G. Tech.) Daniel Bienstock www.columbia.edu/~dano

Gross Hall, Ahmadieh Family Grand Hall, Room 330

Location Info