TY - JOUR
T1 - Near-optimal algorithms for online matrix prediction
AU - Hazan, Elad
AU - Kale, Satyen
AU - Shalev-Shwartz, Shai
N1 - Publisher Copyright:
© by SIAM 2017.
PY - 2017
Y1 - 2017
N2 - In several online prediction problems of recent interest the comparison class is composed of matrices. For example, in the online max-cut problem, the comparison class is matrices which represent cuts of a given graph, and in online gambling the comparison class is matrices which represent permutations over n teams. Another important example is online collaborative filtering, in which a widely used comparison class is the set of matrices with a small trace norm. In this paper we isolate a property of matrices, which we call (β τ)-decomposability, and derive an efficient online learning algorithm that enjoys a regret bound of ∼O( √ β τ T) for all problems in which the comparison class is composed of (β τ)-decomposable matrices. By analyzing the decomposability of cut matrices, low trace-norm matrices, and triangular matrices, we derive near-optimal regret bounds for online max-cut, online collaborative filtering, and online gambling. In particular, this resolves (in the affirmative) an open problem posed by Abernethy [Proceedings of the 23rd Annual Conference on Learning Theory (COLT 2010), pp. 318-319] and Kleinberg, Niculescu-Mizil, and Sharma [Machine Learning, 80 (2010), pp. 245-272]. Finally, we derive lower bounds for the three problems and show that our upper bounds are optimal up to logarithmic factors. In particular, our lower bound for the online collaborative filtering problem resolves another open problem posed by Shamir and Srebro [Proceedings of the 24th Annual Conference on Learning Theory (COLT 1011), pp. 661-678].
AB - In several online prediction problems of recent interest the comparison class is composed of matrices. For example, in the online max-cut problem, the comparison class is matrices which represent cuts of a given graph, and in online gambling the comparison class is matrices which represent permutations over n teams. Another important example is online collaborative filtering, in which a widely used comparison class is the set of matrices with a small trace norm. In this paper we isolate a property of matrices, which we call (β τ)-decomposability, and derive an efficient online learning algorithm that enjoys a regret bound of ∼O( √ β τ T) for all problems in which the comparison class is composed of (β τ)-decomposable matrices. By analyzing the decomposability of cut matrices, low trace-norm matrices, and triangular matrices, we derive near-optimal regret bounds for online max-cut, online collaborative filtering, and online gambling. In particular, this resolves (in the affirmative) an open problem posed by Abernethy [Proceedings of the 23rd Annual Conference on Learning Theory (COLT 2010), pp. 318-319] and Kleinberg, Niculescu-Mizil, and Sharma [Machine Learning, 80 (2010), pp. 245-272]. Finally, we derive lower bounds for the three problems and show that our upper bounds are optimal up to logarithmic factors. In particular, our lower bound for the online collaborative filtering problem resolves another open problem posed by Shamir and Srebro [Proceedings of the 24th Annual Conference on Learning Theory (COLT 1011), pp. 661-678].
KW - Matrix multiplicative weights
KW - Matrix prediction
KW - Online learning
UR - http://www.scopus.com/inward/record.url?scp=85018393459&partnerID=8YFLogxK
U2 - 10.1137/120895731
DO - 10.1137/120895731
M3 - ???researchoutput.researchoutputtypes.contributiontojournal.article???
AN - SCOPUS:85018393459
SN - 0097-5397
VL - 46
SP - 744
EP - 773
JO - SIAM Journal on Computing
JF - SIAM Journal on Computing
IS - 2
ER -