A Universal Well-Calibrated Algorithm for On-line Classification
Vladimir Vovk; 5(Jun):575--604, 2004.
Abstract
We study the problem of on-line classification in which the
prediction algorithm, for each "significance level" δ, is
required to output as its prediction a range of labels (intuitively,
those labels deemed compatible with the available data at the level
δ)
rather than just one label; as usual, the examples are
assumed to be generated independently from the same probability
distribution
P. The prediction algorithm is said to be
"well-calibrated" for
P and δ if the long-run relative
frequency of errors does not exceed δ almost surely w.r. to
P. For well-calibrated algorithms we take the number of
"uncertain" predictions (i.e., those containing more than one
label) as the principal measure of predictive performance. The main
result of this paper is the construction of a prediction algorithm
which, for any (unknown)
P and any δ: (a) makes errors
independently and with probability δ at every trial (in
particular, is well-calibrated for
P and δ); (b) makes in
the long run no more uncertain predictions than any other prediction
algorithm that is well-calibrated for
P and δ; (c)
processes example
n in time
O(log
n).
[abs][pdf][ps.gz][ps]