Statistics and Learning Theory
L.13.1
Lecture
Online, Active, and Transfer Learning

# Active Learning for Classification with Abstention

## Date & Time

01:00 am – 01:00 am

## Abstract

We consider the problem of binary classification with the caveat that the learner can abstain from declaring a label incurring a cost $\lambda \in [0,1/2]$ in the process. This is referred to as the problem of binary classification with a fixed-cost of abstention. For this problem, we propose an active learning strategy that constructs a non-uniform partition of the input space and focuses sampling in the regions near the decision boundaries. Our proposed algorithm can work in all the commonly used active learning query models, namely \emph{membership-query}, \emph{pool-based} and \emph{stream-based}. We obtain an upper bound on the excess risk of our proposed algorithm under standard smoothness and margin assumptions and demonstrate its minimax near-optimality by deriving a matching~(modulo poly-logarithmic factors) lower bound. The achieved minimax rates are always faster than the corresponding rates in the passive setting, and furthermore the improvement increases with larger values of the smoothness and margin parameters.

## Presenters

#### Shubhanshu Shekhar

University of California, San Diego

#### Tara Javidi

University of California, San Diego