Skip to content

  • Projects
  • Groups
  • Snippets
  • Help
    • Loading...
    • Help
    • Submit feedback
    • Contribute to GitLab
  • Sign in
3
3rrend
  • Project
    • Project
    • Details
    • Activity
    • Cycle Analytics
  • Issues 1
    • Issues 1
    • List
    • Board
    • Labels
    • Milestones
  • Merge Requests 0
    • Merge Requests 0
  • CI / CD
    • CI / CD
    • Pipelines
    • Jobs
    • Schedules
  • Wiki
    • Wiki
  • Snippets
    • Snippets
  • Members
    • Members
  • Collapse sidebar
  • Activity
  • Create a new issue
  • Jobs
  • Issue Boards
  • Bart MacCarthy
  • 3rrend
  • Issues
  • #1

Closed
Open
Opened Feb 12, 2025 by Bart MacCarthy@bartmaccarthy
  • Report abuse
  • New issue
Report abuse New issue

Researchers Reduce Bias in aI Models while Maintaining Or Improving Accuracy


Machine-learning designs can fail when they try to make predictions for people who were underrepresented in the datasets they were trained on.

For instance, a model that anticipates the very best treatment option for somebody with a persistent disease might be trained utilizing a dataset that contains mainly male patients. That design might make incorrect predictions for dokuwiki.stream female patients when deployed in a healthcare facility.

To enhance outcomes, engineers can attempt balancing the training dataset by getting rid of information points until all subgroups are represented equally. While dataset balancing is appealing, smfsimple.com it typically needs getting rid of large quantity of information, harming the model's total efficiency.

MIT scientists established a new strategy that identifies and gets rid of specific points in a training dataset that contribute most to a design's failures on minority subgroups. By removing far less datapoints than other approaches, this strategy maintains the general accuracy of the design while enhancing its efficiency concerning underrepresented groups.

In addition, the technique can recognize covert sources of bias in a training dataset that does not have labels. Unlabeled data are far more prevalent than identified data for lots of applications.

This method could likewise be combined with other techniques to improve the fairness of machine-learning designs deployed in high-stakes circumstances. For instance, it may one day assist guarantee underrepresented clients aren't misdiagnosed due to a biased AI model.

"Many other algorithms that attempt to resolve this problem assume each datapoint matters as much as every other datapoint. In this paper, we are revealing that assumption is not real. There specify points in our dataset that are adding to this predisposition, and we can discover those data points, remove them, and improve performance," states Kimia Hamidieh, an electrical engineering and computer technology (EECS) graduate trainee at MIT and co-lead author of a paper on this method.

She composed the paper with co-lead authors Saachi Jain PhD '24 and fellow EECS graduate trainee Kristian Georgiev; Andrew Ilyas MEng '18, PhD '23, a Stein Fellow at Stanford University; and senior authors Marzyeh Ghassemi, an associate professor in EECS and a member of the Institute of Medical Engineering Sciences and the Laboratory for Details and Decision Systems, and Aleksander Madry, the Systems Professor at MIT. The research study will be provided at the Conference on Neural Details Processing Systems.

Removing bad examples

Often, machine-learning designs are trained using big datasets gathered from numerous sources across the internet. These datasets are far too big to be carefully curated by hand, so they might contain bad examples that hurt model performance.

Scientists also know that some information points affect a model's performance on certain downstream tasks more than others.

The MIT researchers combined these two concepts into an approach that determines and removes these troublesome datapoints. They seek to fix an issue understood as worst-group mistake, which happens when a design underperforms on minority subgroups in a training dataset.

The researchers' brand-new method is driven by prior operate in which they presented a method, called TRAK, that determines the most essential training examples for a specific design output.

For this new strategy, they take inaccurate forecasts the design made about minority subgroups and use TRAK to determine which training examples contributed the most to that incorrect forecast.

"By aggregating this details across bad test predictions in the proper way, we are able to find the particular parts of the training that are driving worst-group accuracy down in general," Ilyas explains.

Then they get rid of those specific samples and retrain the design on the remaining information.

Since having more data typically yields better total performance, getting rid of just the samples that drive worst-group failures maintains the model's general accuracy while boosting its performance on minority subgroups.

A more available approach

Across three machine-learning datasets, their method outperformed multiple strategies. In one instance, it enhanced worst-group accuracy while getting rid of about 20,000 fewer training samples than a standard data balancing method. Their strategy likewise attained higher precision than techniques that need making modifications to the inner workings of a design.

Because the MIT method includes altering a dataset rather, it would be easier for a professional to use and can be applied to many kinds of designs.

It can likewise be used when predisposition is unidentified due to the fact that subgroups in a training dataset are not identified. By identifying datapoints that contribute most to a feature the design is learning, they can understand the variables it is utilizing to make a prediction.

"This is a tool anyone can use when they are training a machine-learning design. They can take a look at those datapoints and see whether they are lined up with the ability they are attempting to teach the model," says Hamidieh.

Using the technique to discover unidentified subgroup predisposition would need instinct about which groups to look for, so the scientists hope to confirm it and explore it more completely through future human studies.

They also desire to improve the performance and dependability of their strategy and make sure the approach is available and user friendly for specialists who might sooner or later deploy it in real-world environments.

"When you have tools that let you critically take a look at the data and figure out which datapoints are going to lead to bias or other unfavorable behavior, it gives you a first action toward structure designs that are going to be more fair and more dependable," Ilyas says.

This work is moneyed, in part, by the National Science Foundation and the U.S. Defense Advanced Research Projects Agency.

Assignee
Assign to
None
Milestone
None
Assign milestone
Time tracking
None
Due date
No due date
0
Labels
None
Assign labels
  • View project labels
Reference: bartmaccarthy/3rrend#1