Deep Linear Networks with Arbitrary Loss: All Local Minima Are Global

Similar documents
Linear Regression and Its Applications

Principal Component Analysis

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Math 396. Quotient spaces

Neural Network Training

Near-Potential Games: Geometry and Dynamics

arxiv: v2 [cs.lg] 24 May 2017

Implicit Functions, Curves and Surfaces

Singular Value Decomposition

Near-Potential Games: Geometry and Dynamics

Spectral Theorem for Self-adjoint Linear Operators

Scientific Computing

NORMS ON SPACE OF MATRICES

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Lecture 9: SVD, Low Rank Approximation

Failures of Gradient-Based Deep Learning

REVIEW OF DIFFERENTIAL CALCULUS

Chapter 3 Transformations

12. Perturbed Matrices

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Lagrange Multipliers

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

arxiv: v5 [math.na] 16 Nov 2017

Lecture 2: Linear Algebra Review

Linear Algebra I for Science (NYC)

Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space

Convex Functions and Optimization

AN ALTERNATING MINIMIZATION ALGORITHM FOR NON-NEGATIVE MATRIX APPROXIMATION

Gradient Descent. Dr. Xiaowei Huang

Mathematics Department Stanford University Math 61CM/DM Inner products

Matrix estimation by Universal Singular Value Thresholding

Bare-bones outline of eigenvalue theory and the Jordan canonical form

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

Singular Value Decomposition

INVERSE LIMITS AND PROFINITE GROUPS

A Perron-type theorem on the principal eigenvalue of nonsymmetric elliptic operators

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2

Quantum Computing Lecture 2. Review of Linear Algebra

THE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR

Linear Algebra, Summer 2011, pt. 2

Lecture 10: A (Brief) Introduction to Group Theory (See Chapter 3.13 in Boas, 3rd Edition)

October 25, 2013 INNER PRODUCT SPACES

Lecture 4: Types of errors. Bayesian regression models. Logistic regression

Lecture 8: Linear Algebra Background

ACI-matrices all of whose completions have the same rank

Exercise Solutions to Functional Analysis

Vector, Matrix, and Tensor Derivatives

Tangent spaces, normals and extrema

MAT Linear Algebra Collection of sample exams

Linear Algebra (Review) Volker Tresp 2017

Linear Algebra March 16, 2019

Math 61CM - Solutions to homework 6

Topological Data Analysis - Spring 2018

Stat 159/259: Linear Algebra Notes

Midterm for Introduction to Numerical Analysis I, AMSC/CMSC 466, on 10/29/2015

CHAPTER 3. Gauss map. In this chapter we will study the Gauss map of surfaces in R 3.

A summary of Deep Learning without Poor Local Minima

MATH 581D FINAL EXAM Autumn December 12, 2016

Notes on the framework of Ando and Zhang (2005) 1 Beyond learning good functions: learning good spaces

Definition 2.3. We define addition and multiplication of matrices as follows.

Where is matrix multiplication locally open?

Non-Convex Optimization. CS6787 Lecture 7 Fall 2017

A Multiplicative Operation on Matrices with Entries in an Arbitrary Abelian Group

Introduction to Machine Learning

A = , A 32 = n ( 1) i +j a i j det(a i j). (1) j=1

Deep Learning Book Notes Chapter 2: Linear Algebra

Linear Algebra (Review) Volker Tresp 2018

Categories and Quantum Informatics: Hilbert spaces

TMA 4180 Optimeringsteori KARUSH-KUHN-TUCKER THEOREM

Neural Networks Learning the network: Backprop , Fall 2018 Lecture 4

A Brief Introduction to Tensors

Notes on Latent Semantic Analysis

Vasil Khalidov & Miles Hansard. C.M. Bishop s PRML: Chapter 5; Neural Networks

Linear Algebra Massoud Malek

A PRIMER ON SESQUILINEAR FORMS

SYMMETRIC SUBGROUP ACTIONS ON ISOTROPIC GRASSMANNIANS

Lecture Notes: Geometric Considerations in Unconstrained Optimization

Linear Algebra. The analysis of many models in the social sciences reduces to the study of systems of equations.

UNDERSTANDING THE DIAGONALIZATION PROBLEM. Roy Skjelnes. 1.- Linear Maps 1.1. Linear maps. A map T : R n R m is a linear map if

1 Lyapunov theory of stability

Linear Algebra and Eigenproblems

Simultaneous Diagonalization of Positive Semi-definite Matrices

Lecture Summaries for Linear Algebra M51A

Lecture 5: Linear models for classification. Logistic regression. Gradient Descent. Second-order methods.

COMP 551 Applied Machine Learning Lecture 2: Linear regression

1 Kernel methods & optimization

Glossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB

ME751 Advanced Computational Multibody Dynamics

Factorization in Integral Domains II

Review Let A, B, and C be matrices of the same size, and let r and s be scalars. Then

The Jordan Canonical Form

LS.1 Review of Linear Algebra

14 Singular Value Decomposition

Logistic Regression. Will Monroe CS 109. Lecture Notes #22 August 14, 2017

Differentiable Functions

Manning & Schuetze, FSNLP (c) 1999,2000

Lecture Notes 1: Vector spaces

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

Machine Learning. Support Vector Machines. Fabio Vandin November 20, 2017

Ramsey Theory. May 24, 2015

Transcription:

homas Laurent * 1 James H. von Brecht * 2 Abstract We consider deep linear networks with arbitrary convex differentiable loss. We provide a short and elementary proof of the fact that all local minima are global minima if the hidden layers are either 1 at least as wide as the input layer, or 2 at least as wide as the output layer. his result is the strongest possible in the following sense: If the loss is convex and Lipschitz but not differentiable then deep linear networks can have sub-optimal local minima. 1. Introduction Deep linear networks (DLN are neural networks that have multiple hidden layers but have no nonlinearities between layers. hat is, for given data points {x (i } N i=1 the outputs of such networks are computed via a series ŷ (i = W L W L 1 W 1 x (i of matrix multiplications. Given a target y (i for the i th data point and a pairwise loss function l(ŷ (i, y (i, forming the usual summation L(W 1,..., W L = 1 N then yields the total loss. N l(ŷ (i, y (i (1 i=1 Such networks have few direct applications, but they frequently appear as a class of toy models used to understand the loss surfaces of deep neural networks (Saxe et al., 2014; Kawaguchi, 2016; Lu & Kawaguchi, 2017; Hardt & Ma, 2017. For example, numerical experiments indicate that DLNs exhibit some behavior that resembles the behavior of * Equal contribution 1 Department of Mathematics, Loyola Marymount University, Los Angeles, CA 90045, USA 2 Department of Mathematics and Statistics, California State University, Long Beach, Long Beach, CA 90840, USA. Correspondence to: homas Laurent <tlaurent@lmu.edu>, James H. von Brecht <james.vonbrecht@csulb.edu>. Proceedings of the 35 th International Conference on Machine Learning, Stockholm, Sweden, PMLR 80, 2018. Copyright 2018 by the author(s. deep nonlinear networks during training (Saxe et al., 2014. Results of this sort provide a small piece of evidence that DLNs can provide a decent simplified model of more realistic networks with nonlinearities. From an analytical point-of-view, the simplicity of DLNs allows for a rigorous, in-depth study of their loss surfaces. hese models typically employ a convex loss function l(ŷ, y, and so with one layer (i.e. L = 1 the loss L(W 1 is convex and the resulting optimization problem (1 has no sub-optimal local minimizers. With multiple layers (i.e. L 2 the loss L(W 1,..., W L is not longer convex, and so the question of paramount interest concerns whether this addition of depth and the subsequent loss of convexity creates sub-optimal local minimizers. Indeed, most analytical treatments of DLNs focus on this question. We resolve this question in full for arbitrary convex differentiable loss functions. Specifically, we consider deep linear networks satisfying the two following hypotheses: (i he loss function ŷ l(y, ŷ is convex and differentiable. (ii he thinnest layer is either the input layer or the output layer. Many networks of interest satisfy both of these hypotheses. he first hypothesis (i holds for nearly all network criteria, such as the mean squared error loss, the logistic loss or the cross entropy loss, that appear in applications. In a classification scenario, the second hypothesis (ii holds whenever each hidden layer has more neurons than the number of classes. hus both hypotheses are often satisfied when using a deep linear network (1 to model its nonlinear counterpart. In any such situation we resolve the deep linear problem in its entirety. heorem 1. If hypotheses (i and (ii hold then (1 has no sub-optimal minimizers, i.e. any local minimum is global. We provide a short, transparent proof of this result. It is easily accessible to any reader with a basic understanding of the singular value decomposition, and in particular, it does not rely on any sophisticated machinery from either optimization or linear algebra. Moreover, this theorem is the strongest possible in the following sense heorem 2. here exists a convex, Lipschitz but not differentiable function ŷ l(y, ŷ for which (1 has sub-optimal

local minimizers. In other words, we have a (perhaps surprising hard limit on how far local equals global results can reach; differentiability of the loss is essential. Many prior analytical treatments of DLNs focus on similar questions. For instance, both (Baldi & Hornik, 1989 and (Baldi & Lu, 2012 consider deep linear networks with two layers (i.e. L = 2 and a mean squared error loss criterion. hey provide a local equals global result under some relatively mild assumptions on the data and targets. More recently, (Kawaguchi, 2016 proved that deep linear networks with arbitrary number of layers and with mean squared error loss do not have sub-optimal local minima under certain structural assumptions on the data and targets. he follow-up (Lu & Kawaguchi, 2017 futher simplifies the proof of this result and weakens the structural assumptions. Specifically, this result shows that the loss (1 associated with a deep linear network has no sub-optimal local minima provided all of assumptions (i he loss l(ŷ (i, y (i = y (i ŷ (i 2 is the mean squared error loss criterion; (ii he data matrix X = [x (1,..., x (N ] has full row rank; (iii he target matrix Y = [y (1,..., y (N ] has full row rank; are satisfied. Compared to our result, (Lu & Kawaguchi, 2017 therefore allows for the hidden layers of the network to be thinner than the input and output layers. However, our result applies to network equipped with any differentiable convex loss (in fact any differentiable loss L for which firstorder optimality implies global optimality and we do not require any assumption on the data and targets. Our proof is also shorter and much more elementary by comparison. 2. Proof of heorem 1 heorem 1 follows as a simple consequence of a more general theorem concerning real-valued functions that take as input a product of matrices. hat is, we view the deep linear problem as a specific instance of the following more general problem. Let M m n denote the space of m n real matrices, and let f : M dl d 0 R denote any differentiable function that takes d L d 0 matrices as input. For any such function we may then consider both the single-layer optimization (P1 { Minimize f(a over all A in M dl d 0 as well as the analogous problem Minimize f(w L W L 1 W 1 (P2 over all L-tuples (W 1,..., W L in M d1 d 0 M dl d L 1 that corresponds to a multi-layer deep linear optimization. In other words, in (P2 we consider the task of optimizing f over those matrices A M dl d 0 that can be realized by an L-fold product A = W L W L 1 W 1 W l M dl d l 1 (2 of matrices. We may then ask how the parametrization (2 of A as a product of matrices affects the minimization of f, or in other words, whether the problems (P1 and (P2 have similar structure. At heart, the use of DLNs to model nonlinear neural networks centers around this question. Any notion of structural similarity between (P1 and (P2 should require that their global minima coincide. As a matrix of the form (2 has rank at most min{d 0,..., d L }, we must impose the structural requirement min{d 1,..., d L 1 } min{d L, d 0 } (3 in order to guarantee that (2 does, in fact, generate the full space of d L d 0 matrices. Under this assumption we shall prove the following quite general theorem. heorem 3. Assume that f(a is any differentiable function and that the structural condition (3 holds. hen at any local minimizer ( Ŵ 1,..., ŴL of (P2 the optimality condition is satisfied. f (  = 0  := ŴLŴL 1 Ŵ1 heorem 1 follows as a simple consequence of this theorem. he first hypothesis (i of theorem 1 shows that the total loss (1 takes the form L(W 1,..., W L = f(w L W 1 for f(a some convex and differentiable function. he structural hypothesis (3 is equivalent to the second hypothesis (ii of theorem 1, and so we can directly apply theorem 3 to conclude that a local minimum ( Ŵ 1,..., ŴL of L corresponds to a critical point  = ŴL Ŵ1 of f(a, and since f(a is convex, this critical point is necessarily a global minimum. Before turning to the proof of theorem 3 we recall a bit of notation and provide a calculus lemma. Let A, B fro := r(a B = A ij B ij and i j A 2 fro := A, A fro

denote the Frobenius dot product and the Frobenius norm, respectively. Also, recall that for a differentiable function φ : M m n R its gradient φ(a M m n is the unique matrix so that the equality φ(a + H = φ(a + φ(a, H fro + o ( H fro (4 holds. If F (W 1,..., W L := f(w L W 1 denotes the objective of interest in (P2 the following lemma gives the partial derivatives of F as a function of its arguments. Lemma 1. he partial derivatives of F are given by W1 F (W 1,..., W L = W 2,+ f ( A, Wk F (W 1,..., W L = W k+1,+ f(aw k 1,, WL F (W 1,..., W L = f ( A W L 1,, where A stands for the full product A := W L W 1 and W k,+, W k, are the truncated products W k,+ := W L W k, Proof. he definition (4 implies W k, := W k W 1. (5 F (W 1,..., W k 1, W k + H, W k+1,..., W L = f ( A + W k+1,+ HW k 1, = f(a + f(a, W k+1,+ HW k 1, fro + o ( H fro. Using the cyclic property r(abc = r(cab of the trace then gives f(a, W k+1,+ HW k 1, fro = r ( f(a W k+1,+ HW k 1, = r ( W k 1, f(a W k+1,+ H = W k+1,+ f(aw k 1,, H fro which, in light of (4, gives the desired formula for Wk F. he formulas for W1 F and WL F are obtained similarly. Proof of heorem 3: o prove theorem 3 it suffices to assume that d L d 0 without loss of generality. his follows from the simple observation that g ( A := f ( A defines a differentiable function of d 0 d L matrices for f(a any differentiable function of d L d 0 matrices. As a point ( W 1,..., W L defines a local minimum of f ( ( W L W L 1 W 1 if and only if W 1,..., WL defines a minimum of g ( V 1 V L 1 V L, the theorem for the case d L < d 0 follows by appealing to its d L d 0 instance. It therefore suffices to assume that d L d 0, and by the structural assumption that d k d 0, throughout the remainder of the proof. Consider any local minimizer ( Ŵ 1,..., ŴL of F and denote by Â, Ŵ k,+ and Ŵk, the corresponding full and truncated products (c.f. (5. By definition of a local minimizer there exists some ε 0 > 0 so that F (W 1,..., W L F (Ŵ1,..., ŴL = f ( Â (6 whenever the family of inequalities W l Ŵl fro ε 0 for all 1 l L all hold. Moreover, lemma 1 yields (i 0 = Ŵ 2,+ f ( Â, (ii 0 = Ŵ k+1,+ f ( Â Ŵ k 1, 2 k L 1, (iii 0 = f ( Â Ŵ L 1,. (7 since all partial derivatives must vanish at a local minimum. If ŴL 1, has a trivial kernel, i.e. ker(ŵl 1, = {0}, then the theorem follows easily. he critical point condition (7 part (iii implies Ŵ L 1, f ( Â = 0, and since Ŵ L 1, has a trivial kernel this implies f ( Â = f ( Ŵ L Ŵ L 1 Ŵ1 = 0 as desired. he remainder of the proof concerns the case that ŴL 1, has a nontrivial kernel. he main idea is to use this nontrivial kernel to construct a family of infinitesimal perturbations of the local minimizer ( Ŵ 1,..., ŴL that leaves the overall product ŴL Ŵ1 unchanged. In other words, the family of perturbations ( W1,..., W L satisfy W l Ŵl fro ɛ 0 /2 l = 1,..., L, (8 W L WL 1 W 1 = ŴLŴL 1 Ŵ1. (9 Any such perturbation also defines a local minimizer. Claim 1. Any tuple of matrices ( W1,..., W L satisfying (8 and (9 is necessarily a local minimizer F. Proof. For any matrix W l satisfying W l W l fro ε 0 /2, inequality (8 implies that W l Ŵl fro W l W l fro + W l Ŵl fro ε 0 Equality (9 combined to (6 then leads to F ( W 1,..., W L f (Â = f( Ŵ L Ŵ1 = f( W L W 1 = F ( W1,..., W L for any W l with W l W l fro ε 0 /2 and so the point ( W 1,..., W L defines a local minimum.

he construction of such perturbations requires a preliminary observation and then an appeal to the singular value decomposition. Due to the definition of Ŵk, it follows that ker(ŵk+1, = ker(ŵk+1ŵk, ker(ŵk,, and so the chain of inclusions ker(ŵ1, ker(ŵ2, ker(ŵl 1, (10 holds. Since ŴL 1, has a nontrivial kernel, the chain of inclusions (10 implies that there exists k {1,..., L 1} such that ker(ŵk, = {0} if k < k (11 ker(ŵk, {0} if k k (12 In other words, Ŵ k, is the first matrix appearing in (10 that has a nontrivial kernel. he structural requirement (3 and the assumption that d L d 0 imply that d k d 0 for all k, and so the matrix Ŵk, M dk d 0 has more rows than columns. As a consequence its full singular value decomposition Ŵ k, = Ûk ˆΣ k ˆV k (13 has the shape depicted in figure 1. he matrices Ûk M dk d k and ˆV k M d0 d 0 are orthogonal, whereas ˆΣ k M dk d 0 is a diagonal matrix containing the singular values of Ŵk, in descending order. From (12 Ŵk, has a nontrivial kernel for all k k, and so in particular its least singular value is zero. In particular, the (d 0, d 0 entry of ˆΣ k vanishes if k k. Let û k denote the corresponding d th 0 column of Ûk, which exists since d k d 0. Claim 2. Let w k +1,..., w L denote any collection of vectors and δ k +1,..., δ L any collection of scalars satisfying w k R d k, w k 2 = 1 and (14 0 δ k ɛ 0 /2 (15 for all k + 1 k L. ( W 1,..., W L defined by hen the tuple of matrices d 0 d th 0 column d k = û k 0 Ŵ k, Û k ˆΣk ˆV k Figure 1. Full singular value decomposition of Ŵk, M dk d 0. If k k then Ŵk, does not have full rank and so the (d 0, d 0 entry of ˆΣ k is 0. he d th 0 column of Ûk exists since d k d 0. W k and Ŵk. he equality W k, = Ŵk, is immediate from the definition (16. he equality (9 will then follow from showing that W k, = Ŵk, for all k k L. (17 Proceeding by induction, assume that W k, = Ŵk, for a given k k. hen W k+1, = W k+1 Wk, = W k+1 Ŵ k, (induction hypothesis = (Ŵk+1 + δ k+1 w k+1 û k Ŵk, = Ŵk+1, + δ k+1 w k+1 u k Ŵk, he second term of the last line vanishes, since u k Ŵk, = u k Ûk ˆΣ k ˆV k = e d 0 ˆΣk ˆV k = 0 with e d0 R d k the d th 0 standard basis vector. he second equality comes from the fact that the columns of Ûk are orthonormal, and the last equality comes from the fact that e d 0 Σ k = 0 since the d th 0 row of ˆΣ k vanishes. hus (17 holds, and so (9 holds as well. Claims 1 and claim 2 show that the perturbation ( W1,..., W L defined by (16 is a local minimizer of F. {Ŵk if 1 k k W k := Ŵ k + δ k w k û k 1 else, satisfies (8 and (9. Proof. Inequality (8 follows from the fact that (16 he critical point conditions (i 0 = W 2,+ f ( Ã, (ii 0 = W k+1,+ f ( Ã W k 1, 2 k L 1, (iii 0 = f ( Ã W L 1, W k Ŵk fro = δ k w k û k 1 fro = δ k w k 2 û k 1 2 for all k > k, together with the fact that û k 1 and w k are unit vectors and that 0 δ k ɛ 0 /2. o prove (9 let Wk, = W k W 1 and Ŵk, = Ŵ k Ŵ1 denote the truncated products of the matrices therefore hold as well for all choices of w k +1,..., w L and δ k +1,..., δ L satisfying (14 and (15. he proof concludes by appealing to this family of critical point relations. If k > 1 the transpose of condition (ii gives Ŵ k 1, f ( Â Wk +1,+ = 0 (18

since the equalities W k 1, = Ŵk 1, (c.f. (16 and à = W L W 1 = ŴL Ŵ1 =  (c.f. (9 both hold. But ker(ŵk 1, = {0} by definition of k (c.f. (11, and so f (  WL W k +1 = 0. (19 must hold as well. If k = 1 then (19 follows trivially from the critical point condition (i. hus (19 holds for all choices of w k +1,..., w L and δ k +1,..., δ L satisfying (14 and (15. First choose δ k +1 = 0 so that Wk +1 = Ŵ k +1 and apply (19 to find f (  WL W k +2Ŵk +1 = 0. (20 hen take any δ k +1 > 0 and substract (20 from (19 to get 1 f (  WL W ( k +2 Wk +1 Ŵk +1 δ k +1 = f (  WL W k +2( wk +1û k = 0 for w k +1 an arbitrary vector with unit length. Right multiplying the last equality by û k and using the fact that (w k +1û k û k = w k +1û k û k = w k +1 shows f (  WL W k +2w k +1 = 0 (21 for all choices of w k +1 with unit length. hus (21 implies f (  WL W k +2 = 0 for all choices of w k +2,..., w L and δ k +2,..., δ L satisfying (14 and (15. he claim then follows by induction. f (  = 0 3. Concluding Remarks heorem 3 provides the mathematical basis for our analysis of deep linear problems. We therefore conclude by discussing its limits. First, theorem 3 fails if we refer to critical points rather than local minimizers. o see this, it suffices to observe that the critical point conditions for problem (P2, (i 0 = Ŵ 2,+ f ( Â, (ii 0 = Ŵ k+1,+ f (  Ŵ k 1, 2 k L 1, (iii 0 = f (  Ŵ L 1, where Ŵk,+ := ŴL Ŵk+1 and Ŵk, := Ŵk 1 Ŵ1, clearly hold if L 3 and all of the Ŵl vanish. In other words, the collection of zero matrices always defines a critical point for (P2 but clearly f ( 0 need not vanish. o put it otherwise, if L 3 the problem (P2 always has saddle-points even though all local optima are global. Second, the assumption that f(a is differentiable is necessary as well. More specifically, a function of the form F (W 1,..., W L := f(w L W 1 can have sub-optimal local minima if f(a is convex and globally Lipschitz but is not differentiable. A simple example demonstrates this, and therefore proves theorem 2. For instance, consider the bi-variate convex function f(x, y := x +(1 y + 1, (y + := max{y, 0}, (22 which is clearly globally Lipschitz but not differentiable. he set arg min f := {(x, y R 2 : x = 0, y 1} furnishes its global minimizers while f opt = 1 gives the optimal value. For this function even a two layer deep linear problem F ( W 1, W 2 := f(w 2 W 1 W 2 R 2, W 1 R has sub-optimal local minimizers; the point ( [ 1 (Ŵ1, Ŵ2 = 0, 0] (23 provides an example of a sub-optimal solution. he set of all possible points in R 2 N (Ŵ1, Ŵ2 := { W 2 W 1 : W 2 Ŵ2 1 4, W 1 Ŵ1 1 } 4 generated by a 1/4-neighborhood of the optimum (23 lies in the two-sided, truncated cone N (Ŵ1, Ŵ2 {(x, y R 2 : x 12, y 12 } x, and so if we let x R denote the first component of the product W 2 W 1 then the inequality f(w 2 W 1 1 x 0 = f(ŵ2ŵ1 2 holds on N (Ŵ1, Ŵ2 and so (Ŵ1, Ŵ2 is a sub-optimal local minimizer. Moreover, the minimizer (Ŵ1, Ŵ2 is a strict local minimizer in the only sense in which strict optimality can hold for a deep linear problem. Specifically, the strict inequality f(w 2 W 1 > f(ŵ2ŵ1 (24 holds on N (Ŵ1, Ŵ2 unless W 2 W 1 = Ŵ2Ŵ1 = 0; in the latter case (W 1, W 2 and (Ŵ1, Ŵ2 parametrize the same

point and so their objectives must coincide. We may identify the underlying issue easily. he proof of theorem 3 requires a single-valued derivative f(â at a local optimum, but with f(x, y as in (22 its subdifferential f(0 = {(x, y R 2 : 1 x 1, y = 0} is multi-valued at the sub-optimal local minimum (23. In other words, if a globally convex function f(a induces sub-optimal local minima in the corresponding deep linear problem then f(â cannot exist at any such sub-optimal solution (assuming the structural condition, of course. hird, the structural hypothesis Lu, H. and Kawaguchi, K. Depth creates no bad local minima. arxiv preprint arxiv:1702.08580, 2017. Saxe, A. M., McClelland, J. L., and Ganguli, S. Exact solutions to the nonlinear dynamics of learning in deep linear neural networks. In International Conference on Learning Representations, 2014. d l min{d L, d 0 } for all l {1,..., L} is necessary for theorem 3 to hold as well. If d l < min{d 0, d L } for some l the parametrization A = W L W 1 cannot recover full rank matrices. Let f(a denote any function where f vanishes only at full rank matrices. hen f ( W L W 1 0 at all critical points of (P2, and so theorem 3 fails. Finally, if we do not require convexity of f(a then it is not true, in general, that local minima of (P2 correspond to minima of the original problem. he functions f(x, y = x 2 y 2 F (W 1, W 2 = f(w 2 W 1 and the minimizer (23 illustrate this point. While the origin is clearly a saddle point of the one layer problem, the argument leading to (24 shows that (23 is a local minimizer for the deep linear problem. So in the absence of additional structural assumptions on f(a, we may infer that a minimizer of the deep linear problem satisfies first-order optimality for the original problem, but nothing more. References Baldi, P. and Hornik, K. Neural networks and principal component analysis: Learning from examples without local minima. Neural networks, 2(1:53 58, 1989. Baldi, P. and Lu, Z. Complex-valued autoencoders. Neural Networks, 33:136 147, 2012. Hardt, M. and Ma,. Identity matters in deep learning. In International Conference on Learning Representations, 2017. Kawaguchi, K. Deep learning without poor local minima. In Advances in Neural Information Processing Systems, pp. 586 594, 2016.