-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathegr.bbl
More file actions
124 lines (102 loc) · 4.63 KB
/
egr.bbl
File metadata and controls
124 lines (102 loc) · 4.63 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
\begin{thebibliography}{10}
\bibitem{babanezhad2015stop}
Reza Babanezhad, Mohamed~Osama Ahmed, Alim Virani, Mark Schmidt, Jakub
Kone{\v{c}}n{\`y}, and Scott Sallinen.
\newblock Stop wasting my gradients: Practical {SVRG}.
\newblock {\em Advances in Neural Information Processing Systems (NIPS)}, 2015.
\bibitem{bertsekas2011incremental}
Dimitri~P Bertsekas.
\newblock Incremental gradient, subgradient, and proximal methods for convex
optimization: A survey.
\newblock {\em Optimization for Machine Learning}, pages 1--38, 2010.
\bibitem{dss}
R.~H. Byrd, G.~M. Chin, J.~Nocedal, and Y.~Wu.
\newblock Sample size selection in optimization methods for machine learning.
\newblock {\em Mathematical Programming}, 134(1):127--155, 2012.
\bibitem{NIPS2014_5258}
Aaron Defazio, Francis Bach, and Simon Lacoste-Julien.
\newblock Saga: A fast incremental gradient method with support for
non-strongly convex composite objectives.
\newblock In Z.~Ghahramani, M.~Welling, C.~Cortes, N.D. Lawrence, and K.Q.
Weinberger, editors, {\em Advances in Neural Information Processing Systems
27}, pages 1646--1654. Curran Associates, Inc., 2014.
\bibitem{defazio2014finito}
Aaron~J Defazio, Tib{\'e}rio~S Caetano, and Justin Domke.
\newblock Finito: A faster, permutable incremental gradient method for big data
problems.
\newblock {\em arXiv preprint arXiv:1407.2710}, 2014.
\bibitem{duchihazansing}
John Duchi, Elad Hazan, and Yoram Singer.
\newblock Adaptive subgradient methods for online learning and stochastic
optimization.
\newblock {\em The Journal of Machine Learning Research}, 999999:2121--2159,
2011.
\bibitem{FS2011}
M.P. Friedlander and M.~Schmidt.
\newblock Hybrid deterministic-stochastic methods for data fitting.
\newblock {\em Arxiv preprint arXiv:1104.2373}, 2011.
\bibitem{frostig2014competing}
Roy Frostig, Rong Ge, Sham~M Kakade, and Aaron Sidford.
\newblock Competing with the empirical risk minimizer in a single pass.
\newblock {\em arXiv preprint arXiv:1412.6606}, 2014.
\bibitem{Frostig:2014aa}
Roy Frostig, Rong Ge, Sham~M. Kakade, and Aaron Sidford.
\newblock Competing with the empirical risk minimizer in a single pass.
\newblock 12 2014.
\bibitem{sammy}
S.~Hansen and J.~Nocedal.
\newblock Second-order methods for ${L}_1$ regularized problems in machine
learning.
\newblock {\em ICAASP 2012}, 2012.
\bibitem{johnson2013accelerating}
Rie Johnson and Tong Zhang.
\newblock Accelerating stochastic gradient descent using predictive variance
reduction.
\newblock In {\em Advances in Neural Information Processing Systems}, pages
315--323, 2013.
\bibitem{mairal2015incremental}
Julien Mairal.
\newblock Incremental majorization-minimization optimization with application
to large-scale machine learning.
\newblock {\em SIAM Journal on Optimization}, 25(2):829--855, 2015.
\bibitem{nemirovski2009robust}
Arkadi Nemirovski, Anatoli Juditsky, Guanghui Lan, and Alexander Shapiro.
\newblock Robust stochastic approximation approach to stochastic programming.
\newblock {\em SIAM Journal on Optimization}, 19(4):1574--1609, 2009.
\bibitem{Nesterov:09}
Yurii Nesterov.
\newblock Primal-dual subgradient methods for convex problems.
\newblock {\em Math. Program.}, 120(1):221--259, 2009.
\bibitem{2014pasglyetal}
R.~Pasupathy, P.~W. Glynn, S.~Ghosh, and F.~Hahemi.
\newblock How much to sample in simulation-based stochastic recursions?
\newblock 2014.
\newblock Under Review.
\bibitem{PolJud92}
B.T. Polyak and A.B. Juditsky.
\newblock Acceleration of stochastic approximation by averaging.
\newblock {\em SIAM Journal on Control and Optimization}, 30:838, 1992.
\bibitem{RobMon51}
H.~Robbins and S.~Monro.
\newblock A stochastic approximation method.
\newblock {\em The Annals of Mathematical Statistics}, pages 400--407, 1951.
\bibitem{roux2012stochastic}
Nicolas~L Roux, Mark Schmidt, and Francis~R Bach.
\newblock A stochastic gradient method with an exponential convergence rate for
finite training sets.
\newblock In {\em Advances in Neural Information Processing Systems}, pages
2663--2671, 2012.
\bibitem{ruppert1988efficient}
David Ruppert.
\newblock Efficient estimations from a slowly convergent robbins-monro process.
\newblock Technical report, Cornell University Operations Research and
Industrial Engineering, 1988.
\bibitem{schmidt2013minimizing}
Mark Schmidt, Nicolas~Le Roux, and Francis Bach.
\newblock Minimizing finite sums with the stochastic average gradient.
\newblock {\em arXiv preprint arXiv:1309.2388}, 2013.
\bibitem{shalev2013stochastic}
Shai Shalev-Shwartz and Tong Zhang.
\newblock Stochastic dual coordinate ascent methods for regularized loss.
\newblock {\em The Journal of Machine Learning Research}, 14(1):567--599, 2013.
\end{thebibliography}