For example, we demonstrate that, by abstaining from prediction using our algorithm, we can reduce mistakes compared to standard approaches:
For example, we demonstrate that, by abstaining from prediction using our algorithm, we can reduce mistakes compared to standard approaches:
Regret is inevitable with label noise, but it can tell us where models silently fail, and how we can guide safer predictions
Regret is inevitable with label noise, but it can tell us where models silently fail, and how we can guide safer predictions
Plenty of algorithms have been designed to handle label noise by predicting well on average, but we show how they still fail on specific individuals.
Plenty of algorithms have been designed to handle label noise by predicting well on average, but we show how they still fail on specific individuals.
– Diagnoses from unreliable tests
– Outcomes from noisy electronic health records
In a new paper w/@berkustun, we study how this subjects individuals to a lottery of mistakes.
Paper: bit.ly/3Y673uZ
🧵👇
– Diagnoses from unreliable tests
– Outcomes from noisy electronic health records
In a new paper w/@berkustun, we study how this subjects individuals to a lottery of mistakes.
Paper: bit.ly/3Y673uZ
🧵👇
This helps us spot when a model is unreliable at the individual-level.
This helps us spot when a model is unreliable at the individual-level.
Regret is inevitable with label noise -- it tells us where models silently fail, and how we can guide safer predictions.
Regret is inevitable with label noise -- it tells us where models silently fail, and how we can guide safer predictions.
Plenty of algorithms have been designed to handle label noise by predicting well on average —
But we show how they can still fail on specific individuals.
Plenty of algorithms have been designed to handle label noise by predicting well on average —
But we show how they can still fail on specific individuals.