Theoretical Machine Learning Seminar
Boosting Simple Learners
We study boosting algorithms under the assumption that the given weak learner outputs hypotheses from a class of bounded capacity. This assumption is inspired by the common convention that weak hypotheses are “rules-of-thumbs” from an “easy-to-learn class”. Formally, we assume the class of weak hypotheses has a bounded VC dimension.
We focus on two main questions:
(i) Oracle Complexity: we show that this restriction allows to circumvent a lower bound due to Freund and Schapire (‘95, ‘12) on the number of calls to the weak learner. Unlike previous boosting algorithms which aggregate the weak hypotheses by majority votes, in order to circumvent the lower bound we propose a boosting procedure that uses more complex aggregation rules, and we show this to be necessary.
(ii) Expressivity: Which tasks can be learned by boosting a base-class of bounded VC-dimension? Can concepts that are “far away” from the class be learned? Towards answering this question we identify a combinatorial-geometric parameter which quantifies the expressivity of the base-class when used as part of a boosting procedure. Along the way, we establish and exploit connections with Discrepancy Theory.
Joint with Noga Alon, Alon Gonen, and Elad Hazan
Date & Time
Location
Remote Access Only - see link belowSpeakers
Affiliation
Event Series
Categories
Notes
Please note: interested participants will need to fill out this google form in advance to obtain access to this seminar. Once approved, you will receive the login details. Members of the IAS community do not need to fill out this form - the login details will be emailed to you.