a p a r primal-dual rates and certificates 6. coordinate descent … · 2016-07-01 · primal-dual...

1
Optimality conditions -A > w 2 @ g () , 2 @ g (-A > w) w 2 @ f (A) A2 @ f (w) main results Contributions equip existing optimizers with accuracy certificates lipschitzing trick setup Problem formulation existing rates primal-dual rates Convex conjugate: Celestine Dünner a,b Simone Forte b Martin Takáč c Martin Jaggi b overview algorithm agnostic coordinate descent references [1] Shalev-Shwartz and Zhang. Stochastic dual coordinate ascent methods for regularized loss minimization. JMLR, 14:567–599, 2013 [2] Necoara, I. (2015). Linear convergence of first order methods under weak nondegeneracy assumptions for convex programming. arXiv/1504.06298 illustrative experiments for L1 distributed version c b a Primal-Dual Rates and Certificates h (v ) := sup u2R d v T u - h(u) proof details Lemma 1. Consider an optimization problem of the form (A). Let f be 1/β -smooth w.r.t. a norm k.k f and let g be μ-strongly convex with convexity parameter μ 0 w.r.t. a norm k.k g . The general convex case μ =0 is explicitly allowed, but only if g has bounded support. Then, for any 2 dom(D ) and any s 2 [0, 1], it holds that D () - D (? ) sG() (5) + s 2 2 ( μ(1-s) s ku - k 2 g - 1 β kA(u - )k 2 f ) where G() is the gap function defined in (4) and u 2 @ g (-A > w()). (6) L1-Regularized Distributed Optimization: A Communication-Efficient Primal-Dual Framework Smith, V., Forte, S., Jordan, M. I., & Jaggi, M.. arXiv 1512.04011 ML Systems Workshop on friday A [1] , [1] A [2] , [2] A [K] , [K] !(#) # % −% ! ($) $ Challenge: most existing optimization algorithms lack accuracy certificates. D (w ()) P () P () algorithm-independent new primal-dual convergence rates for a larger problem class min 2R n f (A)+ g () min w2R d f (w )+ g (-A > w ) gap f convex, smooth convex g makes globally Lipschitz gives duality gap defined on entire region of interest B easy to choose for norm-reg. problems problem and algorithms unaffected! correspondence mapping w = w () := rf (A) g Comparison to Nesterov Smoothing makes strongly convex changes iterates !(#) # ! (&) & ( −( g can re-use all existing algorithms! strongly convex linear rate linear primal-dual rate bounded support linear rate linear primal-dual rate 1/T rate 1/T primal-dual rate general convex? same! using trick, see next g g g f = L2, g separable: see SDCA new: SVM f (A)+ P i g i (i ) f (w)+ P i g i (-A > :i w) 10 0 10 2 10 4 10 6 10 5 10 0 10 5 news20 Lasso Epochs Duality Gap SDCA APPROX 10 0 10 5 10 5 10 0 10 5 mushrooms Lasso Epochs Duality Gap SDCA APPROX new primal-dual rates for CD on L1, group-lasso, elastic-net, etc certificates generalize SDCA. no square root ! examples: L1, elastic-n, group lasso, TV, fused L1, structured

Upload: others

Post on 29-May-2020

6 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: A P A R Primal-Dual Rates and Certificates 6. Coordinate Descent … · 2016-07-01 · Primal-Dual Rates and Certificates 4. Primal-Dual Guarantees for Any Algorithm Solving (A)

Optimality conditions

Primal-Dual Rates and Certificates

Table 1. Primal-dual convergence rates of general algorithms applied to (A), for some machine learning and signal processing problemswhich are examples of our optimization problem formulations (A) and (B). Note that several can be mapped to either (A) or (B). � > 0is a regularization parameter specified by the user. We will discuss the most prominent examples in more detail in Sections 5.2 and 6.

Problem f/f

⇤g/g

⇤ ConvergenceRegularizer L

1

(A) f = `(A↵) g = �k↵k1

Thm 6, Cor 10Elastic Net (A) f = `(A↵) g = �

2

k↵k22

+ (1�⌘)k↵k1

Thm 2,8, Cor 11(B) f

⇤= �

2

kwk22

+ (1�⌘)kwk1

g

⇤= `(�A

>w) Thm 2,8, Cor 11

L

2

(A) f = `(A↵) g =

2

k↵k22

Thm 2,8(B) f

⇤=

2

kwk22

g

⇤= `(�A

>w) Thm 2,8

Fused Lasso (A) f = `(A↵) g = �kM↵k1

Thm 6Group Lasso (A) f = `(A↵) g = �

P

k

k↵Gkk2, {Gk

} part. of {1..n} Thm 6SVM (hinge loss) (A) f =

2

kA↵k22

g =

P

i

�y

i

i

s.t. y

i

i

2 [0, 1] Thm 4,9

Loss ` Logistic Regression `LR(z) :=1

m

P

m

j=1

log(1+exp(�y

j

z

j

)) for z = A↵ or z = �A

>w

Least Squares `LS(z) :=1

2

kz� bk22

for z = A↵ or z = �A

>w

also objectives with efficient Fenchel-type operator (Yurt-sever et al., 2015). In contrast, our approach preserves allsolutions of the original L

1

-optimization — it leaves the it-erate sequences of existing algorithms unchanged, which isdesirable in practice, and allows the reusability of existingsolvers. We do not assume proximal or Fenchel tractability.

Distributed algorithms. For L1

-problems exceeding thememory capacity of a single computer, a communication-efficient distributed scheme leveraging this Lipschitzingtrick is presented in (Smith et al., 2015; Forte, 2015).

3. Setup and Primal-Dual StructureIn this paper, we consider optimization problems of the fol-lowing primal-dual structure. As we will see, the relation-ship between primal and dual objectives has many benefits,including computation of the duality gap, which allows usto have certificates for the approximation quality.

We consider the following pair of optimization problems,which are dual2 to each other:

min

↵2Rn

h

D(↵) := f(A↵) + g(↵)

i

, (A)

min

w2Rd

h

P(w) := f

⇤(w) + g

⇤(�A

>w)

i

. (B)

The two problems are associated to a given data matrixA 2 Rd⇥n, and the functions f : Rd ! R and g : Rn ! Rare allowed to be arbitrary closed convex functions. Here↵ 2 Rn and w 2 Rd are the respective variable vectors.The relation of (A) and (B) is called Fenchel-RockafellarDuality where the functions f

⇤, g

⇤ in formulation (B) aredefined as the convex conjugates3 of their correspondingcounterparts f, g in (A). The two main powerful features ofthis general duality structure are first that it includes manymore machine learning methods than more traditional du-

2For a self-contained derivation see Appendix C.3The conjugate is defined as h⇤(v) := supu2Rd v>u�h(u).

ality notions, and secondly that the two problems are fullysymmetric, when changing respective roles of f and g. Intypical machine learning problems, the two parts typicallyplay the roles of a data-fit (or loss) term as well as a regu-larization term. As we will see later, those two roles can beswapped, depending on the application.

Optimality Conditions. The first-order optimality condi-tions for our pair of vectors w 2 Rd

,↵ 2 Rn in prob-lems (A) and (B) are given as

w 2 @f(A↵) , (1a)A↵ 2 @f

⇤(w) , (1b)

�A

>w 2 @g(↵) , (2a)

↵ 2 @g

⇤(�A

>w) (2b)

see, e.g. (Bauschke & Combettes, 2011, Proposition19.18). The stated optimality conditions are equivalentto ↵,w being a saddle-point of the Lagrangian, which isgiven as L(↵,w) = f

⇤(w) � hA↵,wi � g(↵) if ↵ 2

dom(g) and w 2 dom(f

⇤).

Duality Gap. From the definition of the dual problems interms of the convex conjugates, we always have P(w) �P(w

?

) � �D(↵?

) � �D(↵), giving rise to the definitionof the general duality gap G(w,↵) := P(w)� (�D(↵)).

For differentiable f , the duality gap can be used more con-veniently: Given ↵ 2 Rn s.t. A↵ 2 dom(f) in the contextof (A), a corresponding variable vector w 2 Rd for prob-lem (B) is given by the first-order optimality condition (1a)as

w = w(↵) := rf(A↵) . (3)

Under strong duality, we have P(w

?

) = �D(↵?

) andw(↵?

) = w

?, where ↵? is an optimal solution of (A).This implies that the suboptimality P(w(↵)) � P(w

?

) isalways bounded above by the simpler duality gap function

G(↵):=P(w(↵))�(�D(↵))� P(w(↵))� P(w

?

) (4)

which hence acts as a certificate of the approximation qual-ity of the variable vector ↵.

Primal-Dual Rates and Certificates

Table 1. Primal-dual convergence rates of general algorithms applied to (A), for some machine learning and signal processing problemswhich are examples of our optimization problem formulations (A) and (B). Note that several can be mapped to either (A) or (B). � > 0is a regularization parameter specified by the user. We will discuss the most prominent examples in more detail in Sections 5.2 and 6.

Problem f/f

⇤g/g

⇤ ConvergenceRegularizer L

1

(A) f = `(A↵) g = �k↵k1

Thm 6, Cor 10Elastic Net (A) f = `(A↵) g = �

2

k↵k22

+ (1�⌘)k↵k1

Thm 2,8, Cor 11(B) f

⇤= �

2

kwk22

+ (1�⌘)kwk1

g

⇤= `(�A

>w) Thm 2,8, Cor 11

L

2

(A) f = `(A↵) g =

2

k↵k22

Thm 2,8(B) f

⇤=

2

kwk22

g

⇤= `(�A

>w) Thm 2,8

Fused Lasso (A) f = `(A↵) g = �kM↵k1

Thm 6Group Lasso (A) f = `(A↵) g = �

P

k

k↵Gkk2, {Gk

} part. of {1..n} Thm 6SVM (hinge loss) (A) f =

2

kA↵k22

g =

P

i

�y

i

i

s.t. y

i

i

2 [0, 1] Thm 4,9

Loss ` Logistic Regression `LR(z) :=1

m

P

m

j=1

log(1+exp(�y

j

z

j

)) for z = A↵ or z = �A

>w

Least Squares `LS(z) :=1

2

kz� bk22

for z = A↵ or z = �A

>w

also objectives with efficient Fenchel-type operator (Yurt-sever et al., 2015). In contrast, our approach preserves allsolutions of the original L

1

-optimization — it leaves the it-erate sequences of existing algorithms unchanged, which isdesirable in practice, and allows the reusability of existingsolvers. We do not assume proximal or Fenchel tractability.

Distributed algorithms. For L1

-problems exceeding thememory capacity of a single computer, a communication-efficient distributed scheme leveraging this Lipschitzingtrick is presented in (Smith et al., 2015; Forte, 2015).

3. Setup and Primal-Dual StructureIn this paper, we consider optimization problems of the fol-lowing primal-dual structure. As we will see, the relation-ship between primal and dual objectives has many benefits,including computation of the duality gap, which allows usto have certificates for the approximation quality.

We consider the following pair of optimization problems,which are dual2 to each other:

min

↵2Rn

h

D(↵) := f(A↵) + g(↵)

i

, (A)

min

w2Rd

h

P(w) := f

⇤(w) + g

⇤(�A

>w)

i

. (B)

The two problems are associated to a given data matrixA 2 Rd⇥n, and the functions f : Rd ! R and g : Rn ! Rare allowed to be arbitrary closed convex functions. Here↵ 2 Rn and w 2 Rd are the respective variable vectors.The relation of (A) and (B) is called Fenchel-RockafellarDuality where the functions f

⇤, g

⇤ in formulation (B) aredefined as the convex conjugates3 of their correspondingcounterparts f, g in (A). The two main powerful features ofthis general duality structure are first that it includes manymore machine learning methods than more traditional du-

2For a self-contained derivation see Appendix C.3The conjugate is defined as h⇤(v) := supu2Rd v>u�h(u).

ality notions, and secondly that the two problems are fullysymmetric, when changing respective roles of f and g. Intypical machine learning problems, the two parts typicallyplay the roles of a data-fit (or loss) term as well as a regu-larization term. As we will see later, those two roles can beswapped, depending on the application.

Optimality Conditions. The first-order optimality condi-tions for our pair of vectors w 2 Rd

,↵ 2 Rn in prob-lems (A) and (B) are given as

w 2 @f(A↵) , (1a)A↵ 2 @f

⇤(w) , (1b)

�A

>w 2 @g(↵) , (2a)

↵ 2 @g

⇤(�A

>w) (2b)

see, e.g. (Bauschke & Combettes, 2011, Proposition19.18). The stated optimality conditions are equivalentto ↵,w being a saddle-point of the Lagrangian, which isgiven as L(↵,w) = f

⇤(w) � hA↵,wi � g(↵) if ↵ 2

dom(g) and w 2 dom(f

⇤).

Duality Gap. From the definition of the dual problems interms of the convex conjugates, we always have P(w) �P(w

?

) � �D(↵?

) � �D(↵), giving rise to the definitionof the general duality gap G(w,↵) := P(w)� (�D(↵)).

For differentiable f , the duality gap can be used more con-veniently: Given ↵ 2 Rn s.t. A↵ 2 dom(f) in the contextof (A), a corresponding variable vector w 2 Rd for prob-lem (B) is given by the first-order optimality condition (1a)as

w = w(↵) := rf(A↵) . (3)

Under strong duality, we have P(w

?

) = �D(↵?

) andw(↵?

) = w

?, where ↵? is an optimal solution of (A).This implies that the suboptimality P(w(↵)) � P(w

?

) isalways bounded above by the simpler duality gap function

G(↵):=P(w(↵))�(�D(↵))� P(w(↵))� P(w

?

) (4)

which hence acts as a certificate of the approximation qual-ity of the variable vector ↵.

main results

Contributions✔ equip existing optimizers

with accuracy certificates

lipschitzing trick

setupProblem formulation

existing rates ⇒ primal-dual rates

Convex conjugate:

Celestine Dünner a,b ・ Simone Forte b ・ Martin Takáč c ・ Martin Jaggi b

overview

algorithm agnostic

coordinate descent

references[1] Shalev-Shwartz and Zhang. Stochastic dual coordinate

ascent methods for regularized loss minimization. JMLR, 14:567–599, 2013

[2] Necoara, I. (2015). Linear convergence of first order methods under weak nondegeneracy assumptions for convex programming. arXiv/1504.06298

illustrative experiments for L1

distributed version

cbaPrimal-Dual Rates and Certificates

h⇤(v) := supu2Rd

vTu� h(u)

proof details

Primal-Dual Rates and Certificates

4. Primal-Dual Guarantees for AnyAlgorithm Solving (A)

In this section we state an important lemma, which willlater allow us to transform a suboptimality guarantee ofany algorithm into a duality gap guarantee, for optimiza-tion problems of the form specified in the previous section.

Lemma 1. Consider an optimization problem of theform (A). Let f be 1/�-smooth w.r.t. a norm k.k

f

and let gbe µ-strongly convex with convexity parameter µ � 0 w.r.t.a norm k.k

g

. The general convex case µ = 0 is explicitlyallowed, but only if g has bounded support.

Then, for any ↵ 2 dom(D) and any s 2 [0, 1], it holds that

D(↵)�D(↵?

) � sG(↵) (5)

+

s

2

2

µ(1�s)

s

ku�↵k2g

� 1

kA(u�↵)k2f

,

where G(↵) is the gap function defined in (4) and

u 2 @g

⇤(�A

>w(↵)). (6)

We note that the improvement bound here bears similarityto the proof of (Bach, 2015, Prop 4.2) for the case of an ex-tended Frank-Wolfe algorithm. In contrast, our result hereis algorithm-independent, and leads to tighter results due tothe more careful choice of u, as we’ll see in the following.

4.1. Linear Convergence Rates

In this section we assume that we are using an arbitraryoptimization algorithm applied to problem (A). It is as-sumed that the algorithm produces a sequence of (possi-bly random) iterates {↵(t)}1

t=0

such that there exists C 2(0, 1], D � 0 such that

E[D(↵(t)

)�D(↵?

)] (1� C)

t

D. (7)

In the next two theorems, we define � :=

max↵ 6=0

kA↵kf

/k↵kg

2, i.e., the squared spectralnorm of the matrix A in the Euclidean norm case.

4.1.1. CASE I. STRONGLY CONVEX g

Let us assume g is µ-strongly convex (µ > 0) (equivalently,its conjugate g

⇤ has Lipschitz continuous gradient with aconstant 1/µ). The following theorem provides a linearconvergence guarantee for any algorithm with given linearconvergence rate for the suboptimality D(↵)�D(↵?

).

Theorem 2. Assume the function f is 1/�-smooth w.r.t. anorm k.k

f

and g is µ-strongly convex w.r.t. a norm k.kg

.Suppose we are using a linearly convergent algorithm asspecified in (7). Then, for any

t � T :=

1

C

log

D(

��+µ)

µ✏

(8)

it holds that E[G(↵(t)

)] ✏.

From (7) we can obtain that after 1

C

log

D

iterations, we

would have a point ↵(t) such that E[D(↵(t)

)�D(↵?

)] ✏.Hence, comparing with (8) only few more iterations areneeded to get the guarantees for the duality gap. Therate (7) is achieved by most of the first order algorithms,including proximal gradient descent (Nesterov, 2013) orSDCA (Richtarik & Takac, 2014) with C ⇠ µ or accel-erated SDCA (Lin et al., 2014) with C ⇠ p

µ.

4.1.2. CASE II. GENERAL CONVEX g (OF BOUNDEDSUPPORT)

In this section we will assume that g⇤ is Lipschitz (in con-trast to smooth as in Theorem 2) and show that the linearconvergence rate is preserved.

Theorem 3. Assume that the function f is 1/�-smoothw.r.t. a norm k.k, g⇤ is L-Lipschitz continuous w.r.t thedual norm k.k⇤, and we are using a linearly convergent al-gorithm (7). Then, for any

t � T :=

1

C

log

2Dmax{1,2�L2/✏�}

(9)

it holds that E[G(↵(t)

)] ✏.

In (Wang & Lin, 2014), it was proven that feasible descentmethods when applied to the dual of an SVM do improvethe objective geometrically as in (7). Later, (Ma et al.,2015b) extended this to stochastic coordinate feasible de-scent algorithms (including SDCA). Using our new Theo-rem 3, we can therefore extend their results to linear con-vergence for the duality gap for the SVM application.

4.2. Sub-Linear Convergence Rates

In this case we will focus only on general L-Lipschitz con-tinuous functions g

⇤ (if g is strongly convex, then manyexisting algorithms are available and converge with a lin-ear rate).We will assume that we are applying some (possi-bly randomized) algorithm on optimization problem (A)which produces a sequence (of possibly random) iterates{↵(t)}1

t=0

such that

E[D(↵(t)

)�D(↵?

)] C

D(t)

, (10)

where D(t) is a function wich has usually a linear orquadratic growth (i.e. D(t) ⇠ O(t) or D(t) ⇠ O(t

2

)).

The following theorem will allow to equip existing algo-rithms with sub-linear convergence in suboptimality, asspecified in (10), with duality gap convergence guarantees.

Theorem 4. Assume the function f is 1/�-smooth w.r.t. thenorm k.k, g⇤ is L-Lipschitz continuous, w.r.t. the dual normk.k⇤, and we are using a sub-linearly convergent algorithmas quantified by (10). Then, for any t � 0 such that

D(t) � max{ 2C�

�L

2 ,2C�L

2

�✏

2 }, (11)

it holds that E[G(↵(t)

)] ✏.

L1-Regularized Distributed Optimization: A Communication-Efficient Primal-Dual Framework Smith, V., Forte, S., Jordan, M. I., & Jaggi, M.. arXiv 1512.04011

ML Systems Workshop on friday

A[1],↵[1] A[2],↵[2] A[K],↵[K]

!(#)

#%−%

!∗($)

$

Challenge: most existing optimization algorithms lack accuracy certificates.

D(w(↵))

P(↵)P(↵)

✔ algorithm-independent new primal-dual convergence rates for a larger problem class

min↵2Rn

f(A↵) + g(↵)

minw2Rd

f⇤(w) + g⇤(�A>w)gap

f convex, smooth convexg

makes globally Lipschitz gives duality gap defined on entire region of interestB easy to choose for norm-reg. problems problem and algorithms unaffected!

correspondence mappingw = w(↵) := rf(A↵)

g⇤

Comparison to Nesterov Smoothing

makes strongly convex changes iterates

!(#)

#

!∗(&)

&

(−(

g⇤

can re-use all existing algorithms!

strongly convex ✤ linear rate ⇒ linear primal-dual rate

bounded support

✤ linear rate ⇒ linear primal-dual rate✤ 1/T rate ⇒ √1/T primal-dual rate

general convex?

✤ same! using trick, see next

g

g

g

f = L2, g separable: see SDCA

new: SVM

Primal-Dual Rates and Certificates

6. Coordinate Descent AlgorithmsWe now focus on a very important class of algorithms, thatis coordinate descent methods. In this section, we showhow our theory implies much more general primal-dualconvergence guarantees for coordinate descent algorithms.

Partially Separable Problems. A widely used subclass ofoptimization problems arises when one part of the objectivebecomes separable. Formally, this is expressed as g(↵) =

P

n

i=1

g

i

(↵

i

) for univariate functions g

i

: R ! R for i 2[n]. Nicely in this case, the conjugate of g also separatesas g

⇤(y) =

P

i

g

⇤i

(y

i

). Therefore, the two optimizationproblems (A) and (B) write as

D(↵) := f(A↵) +

P

i

g

i

(↵

i

) (SA)

P(w) := f

⇤(w) +

P

i

g

⇤i

(�A >:i

w) , (SB)

where A

:i

2 Rd denotes the i-th column of A.

The Algorithm. We consider the coordinate descent al-gorithm described in Algorithm 1. Initialize ↵(0)

= 0 andthen, at each iteration, sample and update a random coordi-nate i 2 [n] of the parameter vector ↵ to iteratively mini-mize (SA). Finally, after T iterations output ¯↵, the averagevector over the latest T � T

0

iterates. The parameter T0

issome positive number smaller than T .

Algorithm 1 Coordinate Descent on D(↵)

1: Input: Data matrix A.Starting point ↵(0)

:= 0 2 Rn, w(0)

= w(↵(0)

).2: for t = 1, 2, . . . T do3: Pick i 2 [n] randomly4: Find �↵

i

minimizing D(↵(t�1)

+ e

i

�↵

i

)

5: ↵(t) ↵(t�1)

+�↵

i

e

i

6: w

(t) w(↵(t)

)

7: end for8: Let ¯↵ =

1

T�T0

P

T�1

t=T0↵(t)

As we will show in the following section, coordinate de-scent on D(↵) is not only an efficient optimizer of the ob-jective D(↵), but also provably reduces the duality gap.Therefore, the same algorithm will simultaneously opti-mize the dual objective P(w).

6.1. Primal-Dual Analysis for Coordinate Descent

We first show linear primal-dual convergence rate of Al-gorithm 1 applied to (SA) for strongly convex g

i

. Later,we will generalize this result to also apply to the settingof general Lipschitz g

i

. This generalization together withthe Lipschitzing trick will allow us to derive primal-dualconvergence guarantees of coordinate descent for a muchbroader class of problems, including the Lasso problem.

For the following theorems we assume that the columns of

the data matrix A are scaled such that kA:i

k R for alli 2 [n] and kA

j:

k P for all j 2 [d], for some norm k.k.Theorem 8. Consider Algorithm 1 applied to (SA). As-sume f is a 1/�-smooth function w.r.t. the norm k.k. Then,if g

i

is µ-strongly convex for all i, it suffices to have a totalnumber of iterations of

T �⇣

n+

nR

2

µ�

log

h

n+

nR

2

µ�

i

(0)D✏

to get E[G(↵(T )

)] ✏. Moreover, to obtain an expectedduality gap of E[G(

¯↵)] ✏ it suffices to have T > T

0

with

T

0

�⇣

n+

nR

2

µ�

log

h

n+

nR

2

µ�

i

(0)D

(T�T0)✏

where ✏

(0)

D

is the initial suboptimality in D(↵).

Theorem 8 allows us to upper bound the duality gap, andhence the suboptimality, for every iterate ↵(T ), as well asthe average ¯↵ returned by Algorithm 1. In the following wegeneralize this result to apply to L-Lipschitz functions g

i

.Theorem 9. Consider Algorithm 1 applied to (SA). As-sume f is a 1/�-smooth function w.r.t. the norm k.k. Then,if g⇤

i

is L-Lipschitz for all i, it suffices to have a total num-ber of iterations of

T � max

0, n log

(0)D �

2L

2R

2n

+ n+

20n

2L

2R

2

�✏

to get E[G(

¯↵)] ✏. Moreover, when t � T

0

with

T

0

= max

0, n log

(0)D �

2L

2R

2n

+

16n

2L

2R

2

�✏

we have the suboptimality bound of E[D(↵(t)

)�D(↵?

)] ✏/2, where ✏

(0)

D

is the initial suboptimality.Remark 2. Theorem 9 shows that for Lipschitz g

⇤i

, Algo-rithm 1 has O(✏

�1

) convergence in the suboptimality andO(✏

�1

) convergence in G(

¯↵). Comparing this result toTheorem 4 which suggests O(✏

�2

) convergence in G(↵)

for O(✏

�1

) convergent algorithms, we see that averagingthe parameter vector crucially improves convergence in thecase of non-smooth f .Remark 3. Note that our Algorithm 1 recovers the widelyused SDCA setting (Shalev-Shwartz & Zhang, 2013) as aspecial case, when we choose f

⇤:=

2

k.k22

in (SB). Fur-thermore, their convergence results for SDCA are consis-tent with our results and can be recovered as a special caseof our analysis. See Corollaries 16, 18, 19 in Appendix J.

6.2. Application to L

1

and Elastic Net RegularizedProblems

We now apply Algorithm 1 to the L1

-regularized problems,as well as Elastic Net regularized problems. We state im-proved primal-dual convergence rates which are more tai-lored to the coordinate-wise setting.

100

102

104

106

10−5

100

105

news20 Lasso

Epochs

Du

alit

y G

ap

SDCAAPPROX

100

105

10−5

100

105

mushrooms Lasso

Epochs

Du

alit

y G

ap

SDCAAPPROX

new primal-dual rates for CD onL1, group-lasso, elastic-net, etc certificates generalize SDCA. no square root

! examples:L1, elastic-n,group lasso, TV, fused L1, structured