Regret Bounds for Prediction Problems
journal contribution
posted on 1999-01-01, 00:00 authored by Geoffrey J. GordonWe present a unified framework for reasoning about
worst-case regret bounds for learning algorithms.
This framework is based on the theory of duality
of convex functions. It brings together results from
computational learning theory and Bayesian statistics,
allowing us to derive new proofs of known
theorems, new theorems about known algorithms,
and new algorithms.
History
Publisher Statement
Copyright © 1999 by the Association for Computing Machinery, Inc. Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, to republish, to post on servers, or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from Publications Dept., ACM, Inc., fax +1 (212) 869-0481, or permissions@acm.org. © ACM, 1999. This is the author's version of the work. It is posted here by permission of ACM for your personal use. Not for redistribution. The definitive version was published in Proceedings of the twelfth annual conference on Computational learning theory {1-58113-167-4 (1999) http://doi.acm.org/10.1145/307400.307410Date
1999-01-01Usage metrics
Categories
Licence
Exports
RefWorksRefWorks
BibTeXBibTeX
Ref. managerRef. manager
EndnoteEndnote
DataCiteDataCite
NLMNLM
DCDC