Do Outliers Ruin Collaboration?

05/12/2018
by   Mingda Qiao, et al.
0

We consider the problem of learning a binary classifier from n different data sources, among which at most an η fraction are adversarial. The overhead is defined as the ratio between the sample complexity of learning in this setting and that of learning the same hypothesis class on a single data distribution. We present an algorithm that achieves an O(η n + n) overhead, which is proved to be worst-case optimal. We also discuss the potential challenges to the design of a computationally efficient learning algorithm with a small overhead.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset