Researchers Reduce Bias in aI Models while Maintaining Or Improving Accuracy
Celesta Casiano editó esta página hace 7 meses


Machine-learning models can fail when they try to make forecasts for people who were underrepresented in the datasets they were trained on.

For instance, a design that predicts the best treatment alternative for somebody with a chronic illness might be trained utilizing a dataset that contains mainly male patients. That model might make inaccurate predictions for female patients when deployed in a health center.

To enhance results, engineers can attempt stabilizing the training dataset by getting rid of information points till all are represented similarly. While dataset balancing is promising, it typically requires removing large amount of data, hurting the design’s total performance.

MIT researchers developed a brand-new method that recognizes and eliminates particular points in a training dataset that contribute most to a model’s failures on minority subgroups. By removing far less datapoints than other methods, this method maintains the general accuracy of the design while enhancing its performance concerning underrepresented groups.

In addition, the technique can recognize hidden sources of predisposition in a training dataset that lacks labels. Unlabeled data are even more widespread than identified data for numerous applications.

This technique might also be combined with other approaches to improve the fairness of machine-learning designs deployed in high-stakes scenarios. For trade-britanica.trade instance, it might at some point assist make sure underrepresented clients aren’t misdiagnosed due to a prejudiced AI model.

“Many other algorithms that attempt to resolve this concern assume each datapoint matters as much as every other datapoint. In this paper, we are revealing that presumption is not real. There specify points in our dataset that are contributing to this bias, and we can find those data points, eliminate them, and improve performance,” says Kimia Hamidieh, an electrical engineering and computer system science (EECS) graduate trainee at MIT and co-lead author of a paper on this technique.

She composed the paper with co-lead authors Saachi Jain PhD ‘24 and oke.zone fellow EECS graduate trainee Kristian Georgiev