• No results found

In this document, we present more experimental results, and detailed proofs for the theorems stated in the main paper

N/A
N/A
Protected

Academic year: 2021

Share "In this document, we present more experimental results, and detailed proofs for the theorems stated in the main paper"

Copied!
6
0
0

Loading.... (view fulltext now)

Full text

(1)Supplement of “ Limited-memory Common-directions Method for Distributed Optimization and its Application on Empirical Risk Minimization” Ching-pei Lee∗ I. Po-Wei Wang†. Introduction. In this document, we present more experimental results, and detailed proofs for the theorems stated in the main paper. II. Detailed derivations for (4.37). We show how to compute PkT uk−1 efficiently. From (2.15), we have ψ T uk−1 = θk−1 ψ T Pk−1 tk−1 , ∀ψ ∈ Rn . Therefore,  PkT uk−1 i  k−1  θk−1 Mi+2,: tk−1  θ2 tT M k−1 t k−1 k−1 k−1  = uTk−1 sk−1 = PkT sk−1 2m−1    T uk−1 ∇f (wk ) = (PkT ∇f (wk ))2m−1. if i < 2m − 1 if i = 2m − 1 , if i = 2m if i = 2m + 1. (PkT uk−1 )i = θk−1 (PkT Pk−1 tk−1 )i k = θk−1 Mi+2,: tk−1 .. For i = 2m − 1, from (2.15) we have (PkT uk−1 )i = θk−1 uTk−1 Pk−1 tk−1 = III. where Pk ≡ [q 1 , . . . , q m ]. If q j satisfies (3.18), then the right-hand side of (III.1) is not all zero, hence t 6= 0 and Pk t 6= 0. Therefore, from (III.1), we have (III.2) −pTk ∇f (wk ) = (Pk t)T Hk Pk t ≥ M2 kPk tk2 = M2 kpk k2 . We then have from Assumption 1 and (III.2) that ρ f (wk + θk pk ) ≤ f (wk ) + θk ∇f (wk )T pk + θk2 kpk k2 2 ρθk T ≤ f (wk ) + θk ∇f (wk ) pk (1 − ). 2M2. 1−. Therefore, the backtracking line search procedure takes at  most min 0, dlogβ (2β(1 − c1 )(M2 /ρ))e steps. IV. Proof of Theorem 3.2. The j-th equation in the linear system (III.1) is. ∗ University. pTk Hk q j = −∇f (wk )T q j .. (IV.1). By (3.17), (3.18), and (IV.1),. T θk−1 tTk−1 Pk−1 Pk−1 tk−1 T k−1 θk−1 tk−1 M tk−1 .. The solution of (3.16) also solves the following linear system. PkT Hk Pk t = −PkT ∇f (wk ),. of Wisconsin-Madison. ching-pei@cs.wisc.edu † Carnegie Mellon University. poweiw@cs.cmu.edu ‡ Microsoft. wzchen@microsoft.com § National Taiwan University. cjlin@csie.ntu.edu.tw. ρθk ≥ c1 , 2M2. (2.14) is satisfied. Thus, we have that the backtracking line search gets a step size θk such that   2β (1 − c1 ) M2 (III.3) θk ≥ min 1, . ρ. kpk kkq j k ≥ | =|. Proof of Theorem 3.1. (III.1). Chih-Jen Lin§. From (III.2), ∇f (wk )T pk < 0. Therefore, when. For i < 2m − 1, we have. =. Weizhu Chen‡. 1 (p )T Hk q j | M1 k. 1 δ ∇f (wk )T q j | ≥ k∇f (wk )kkq j k . M1 M1. Therefore, (IV.2). kpk k ≥. δ k∇f (wk )k. M1. Combining (III.2) and (IV.2), we can establish the following result. −. pTk ∇f (wk ) M2 kpk k2 δM2 ≥ ≥ . kpk kk∇f (wk )k kpk kk∇f (wk )k M1.

(2) Now to show the convergence, by (2.14), (III.2), and (IV.2), we have f (wk+1 ) ≤ f (wk ) + θk c1 ∇f (wk )T pk ≤ f (wk ) −. M2 δ 2 θk c1 k∇f (wk )k2 . M12. From (III.3), we can replace θk in the above result with some positive constant κ and get (IV.3). f (wk+1 ) ≤ f (wk ) −. M2 δ 2 κc1 k∇f (wk )k2 . M12. Thus, summing (IV.3) from w0 to wk , we get (IV.4). k X M2 δ 2 κc1 j=0. M12. k∇f (wj )k2 ≤ f (w0 ) − f (wk+1 ) ≤ f (w0 ) − f ∗ ,. where f ∗ is the minimal objective value of f . Consequently,. Note that in the second inequality, we utilized the fact that β, 1 − c1 , σ/ρ ∈ [0, 1]. Therefore, the backtracking line search procedure takes at most dlogβ (β(1 − c1 )(σ/ρ))e steps, and we can take κ=. (V.1). β(1 − c1 )σ . ρ. Now to show the linear convergence, the PolyakŁojasiewicz condition (3.20) holds with σ from the strong convexity by [1, Theorem 2.1.10] and noting that ∇f (w∗ ) = 0. Therefore, by (V.1), (IV.5) becomes f (wk+1 ) − f ∗ ≤ (1 −. 2βc1 (1 − c1 )σ 3 )(f (wk ) − f ∗ ), ρ3. and the required iterations to reach an -accurate solution is therefore  log (f (w0 ) − f ∗ ) + log 1    . k≥ 3 1 )σ log 1/ 1 − 2βc1 (1−c ρ3. k. 1 X k∇f (wj )k2 min k∇f (wj )k ≤ 0≤j≤k k + 1 j=0 2. 1 M12 ≤ (f (w0 ) − f ∗ ) k + 1 M2 δ 2 κc1 and (3.19) follows. Note that since k X. k∇f (wj )k2. j=0. VI. Convergence for Regularized Neural Networks. In this section, we show that the framework discussed in Section 3 applies to the algorithm for the L2-regularized nonconvex neural network problem considered in [2], with a minor condition in solving the linear system below in (VI.3), and therefore the result of Theorem 3.2 provides a convergence rate guarantee for their algorithm. The neural network problem considered in [2] is of the following form.. l is bounded, we have that k∇f (wk )k converges to zero as k X 1 (VI.1) min g(θ) ≡ kθk2 + C ξ(θ; xi , yi ), approaches infinity. θ 2 i=1 Now consider that f satisfies (3.20). Deducting f ∗ from both sides of (IV.3) and combining it with (3.20), we get for where θ is the collection of all weights between two adjacent all k layers of the neural network denoted as a vector, C > 0 2 is a parameter specified by users, {(yi , xi )}, i = 1, . . . , l 2σM2 δ κc1 (IV.5) f (wk+1 ) − f ∗ ≤ (1 − )(f (wk ) − f ∗ ), are the training instances, and ξ is nonnegative and twice 2 M1 continuously differentiable with respect to θ. Therefore, g is ∗ and thus to get f (wk ) − f ≤ , it takes O(log(1/)) twice differentiable. Because ξ is nonnegative and a descent iterations. Note that our assumptions give c1 > 0, and method is used, given any initial point θ 0 , the sequence of σM2 /M12 > 0. Therefore the coefficient in the right-hand points generated by their algorithm is confined in the level set side of (IV.5) is smaller than 1.. V. Proof of Corollary 3.1. We note that in Algorithm 1, because the gradient itself is included in the directions, (3.18) is satisfied with δ = 1. Moreover, (3.17) is satisfied by (M1 , M2 ) = (ρ, σ) from Assumption 1 and the strong convexity. Thus from (III.3),   2β (1 − c1 ) σ β(1 − c1 )σ θk ≥ min 1, ≥ . ρ ρ. l. (VI.2). {θ |. X 1 1 kθk2 ≤ kθ 0 k2 + C ξ(θ 0 ; xi , yi )}, 2 2 i=1. which is compact. Therefore, as a continuous function, k∇2 g(θ)k is upper-bounded in this compact set, which is equivalent to that the k∇g(θ 1 ) − ∇g(θ 2 )k/kθ 1 − θ 2 k is upper bounded by some finite value. Thus Assumption 1 is satisfied..

(3) (a) criteo. (b) kdd2012.tr. (c) url. (d) KDD2010-b. (e) epsilon. (f) webspam. (g) news20. (h) rcv1t. Figure (I): Comparison of different algorithms with C = 10−3 . We show training time v.s. relative difference to the optimal function value. The horizontal lines indicate when the algorithm is terminated in practice by marking the stopping condition i =−1) of TRON in MPI-LIBLINEAR: k∇f (w)k ≤  min(#yi =1,#y k∇f (0)k, with  = 10−2 (default), 10−3 , and 10−4 . l Recall from (2.2) that [2] combines two directions dk and pSN k−1 , where dk is the stochastic Newton direction at the current iterate θ k . We will show that the direction dk satisfies (3.18). The linear system solved in [2] to obtain dk is (VI.3). (GSk + λk I)dk = −∇f (θ k ),. where Sk is a random subset of instances, GSk is a subsampled Gauss-Newton matrix corresponding to Sk at θ k , λk > 0 is a damping factor, and I is the identity matrix. The needed condition here is that λk is upper- and lowerbounded by some positive constants. If λk ≥ λ > 0 always holds, then with the positive semi-definiteness of GSk , we see that (VI.4). − dTk ∇f (θ k ) = dTk (GSk + λk I)dk ≥ λkdk k2 .. On the other hand, because θ k is in the compact set (VI.2) and GSk is a continuous function of θ k under a fixed Sk , if λk is upper bounded, since there are only finite choices of Sk as a subset of a finite set, there exists γ > 0 such that (VI.5) kdk k = k(GSk + λk I)−1 ∇f (θ k )k ≥ γk∇f (θ k )k. From (VI.4) and (VI.5),

(4)

(5)

(6) T

(7)

(8) dk ∇f (θ k )

(9) ≥ λkdk k2 ≥ λγkdk kk∇f (θ k )k, and hence (3.18) holds. Therefore all conditions of Theorem 3.2 are satisfied. Because the iterates lie in a compact set, there are convergent sub-sequences of the iterates, and every limit point of the iterates is stationary by Theorem 3.2.. VII. More Experiments. We present more experimental results that are not included in the main paper in this section. We consider the same experiment environment, and the same problem being solved. We present the results using different values of C to see the relative efficiency when the problems become more difficult or easier. The result of C = 10−3 is shown in Figure (I), and the result of C = 1, 000 is shown in Figure (II). For C = 10−3 , the problems are easier to solve. We observe that L-CommDir-BFGS is faster than existing methods on all data sets, and L-CommDir-Step outperforms state of the art on all data sets but url, but is still competitive on url. For C = 1, 000, L-CommDir-BFGS is the fastest on most data sets, and the only exception is KDD2010-b, on which L-CommDir-BFGS is slightly slower than L-CommDirStep, but faster than other methods. On the other hand, L-CommDir-Step is slower than TRON on webspam but faster than existing methods on other data sets. These results show that our method is highly efficient by using (2.11) or (2.12) to decide Pk . The other choice, L-CommDir-Grad, is obviously inferior for most cases. We also modify from TRON to obtain a line-search truncated Newton solver to compare with our method. This line-search truncated Newton method is denoted by NEWTON in the results in Figures (III)-(V). Results show that NEWTON is consistently the fastest on criteo for all choices of C, but L-CommDir-BFGS and L-CommDir-Step are faster in most other cases..

(10) (a) criteo. (b) kdd2012.tr. (c) url. (d) KDD2010-b. (e) epsilon. (f) webspam. (g) news20. (h) rcv1t. Figure (II): Comparison of different algorithms with C = 1, 000. We show training time v.s. relative difference to the optimal function value. The horizontal lines indicate when the algorithm is terminated in practice by marking the stopping i =−1) k∇f (0)k, with  = 10−2 (default), 10−3 , and condition of TRON in MPI-LIBLINEAR: k∇f (w)k ≤  min(#yi =1,#y l −4 10 .. (a) criteo. (b) kdd2012.tr. (c) url. (d) KDD2010-b. (e) epsilon. (f) webspam. (g) news20. (h) rcv1t. Figure (III): Comparison with line-search truncated Newton with C = 10−3 . We show training time v.s. relative difference to the optimal function value. The horizontal lines indicate when the algorithm is terminated in practice by marking the i =−1) stopping condition of TRON in MPI-LIBLINEAR: k∇f (w)k ≤  min(#yi =1,#y k∇f (0)k, with  = 10−2 (default), l −3 −4 10 , and 10 ..

(11) (a) criteo. (b) kdd2012.tr. (c) url. (d) KDD2010-b. (e) epsilon. (f) webspam. (g) news20. (h) rcv1t. Figure (IV): Comparison with line-search truncated Newton with C = 1. We show training time v.s. relative difference to the optimal function value. The horizontal lines indicate when the algorithm is terminated in practice by marking the i =−1) k∇f (0)k, with  = 10−2 (default), stopping condition of TRON in MPI-LIBLINEAR: k∇f (w)k ≤  min(#yi =1,#y l −3 −4 10 , and 10 .. (a) criteo. (b) kdd2012.tr. (c) url. (d) KDD2010-b. (e) epsilon. (f) webspam. (g) news20. (h) rcv1t. Figure (V): Comparison with line-search truncated Newton with C = 1, 000. We show training time v.s. relative difference to the optimal function value. The horizontal lines indicate when the algorithm is terminated in practice by marking the i =−1) stopping condition of TRON in MPI-LIBLINEAR: k∇f (w)k ≤  min(#yi =1,#y k∇f (0)k, with  = 10−2 (default), l −3 −4 10 , and 10 ..

(12) References [1] Y. E. Nesterov. Introductory Lectures on Convex Optimization: A Basic Course. Kluwer Academic Publishers, 2003. [2] C.-C. Wang, C.-H. Huang, and C.-J. Lin. Subsampled Hessian Newton methods for supervised learning. Neural Comput., 27:1766–1795, 2015..

(13)

References

Related documents

Detta självständiga arbete är ett resultat av kursen EX0652 - Kandidatexamensarbete i Landskapsplanering vid Trädgårdsingenjörsprogrammet, SLU Alnarp. Målet med detta arbete

Furthermore, Cleary and Zimmerman (2012) suggest that motivation, self-regulation, and cognitive engagement are highly interrelated constructs that can help researchers understand the

En este trabajo se ha llevado a cabo una revisión y actualización de diferentes aspectos relacionados con la calidad del aceite de oliva virgen, se han estudiado las

7 A resort is considered as major if its attendance reaches over 1 million skier visits per winter season... Most of the industry is concentrated around the resorts that generate

Most of the studies related to knowledge sharing practices used common theory such as theory of planned behavior, theory of reasoned action and social exchange theory to

Ramp metering, variable speed limits, and hard shoulder running control strategies have been used for managing motorway traffic congestion.. This thesis presents a modelling

Seven of the isolates were examined for anti-HIV effects, and compounds 1a and 1b showed moderate anti- HIV activity with EC 50 values of 6.3 and 5.3 μM.. The genus Euphorbia

Examples (non-statutory)  a depth study linked to one of the British areas of study listed above  a study over time, testing how far sites in their locality reflect aspects