Skip to content

  • Projects
  • Groups
  • Snippets
  • Help
    • Loading...
    • Help
    • Submit feedback
    • Contribute to GitLab
  • Sign in
L
l-williams
  • Project
    • Project
    • Details
    • Activity
    • Cycle Analytics
  • Issues 27
    • Issues 27
    • List
    • Board
    • Labels
    • Milestones
  • Merge Requests 0
    • Merge Requests 0
  • CI / CD
    • CI / CD
    • Pipelines
    • Jobs
    • Schedules
  • Wiki
    • Wiki
  • Snippets
    • Snippets
  • Members
    • Members
  • Collapse sidebar
  • Activity
  • Create a new issue
  • Jobs
  • Issue Boards
  • Alysa Randle
  • l-williams
  • Issues
  • #23

You need to sign in or sign up before continuing.
Closed
Open
Opened Feb 11, 2025 by Alysa Randle@alysa839654637
  • Report abuse
  • New issue
Report abuse New issue

Researchers Reduce Bias in aI Models while Maintaining Or Improving Accuracy


Machine-learning models can fail when they attempt to make predictions for individuals who were underrepresented in the datasets they were trained on.

For instance, a design that predicts the very best treatment alternative for somebody with a persistent illness may be trained using a dataset that contains mainly male patients. That design may make inaccurate forecasts for female patients when deployed in a healthcare facility.

To enhance outcomes, engineers can try balancing the training dataset by eliminating data points till all subgroups are represented equally. While dataset balancing is appealing, it often needs getting rid of large amount of information, injuring the model's total performance.

MIT scientists developed a brand-new strategy that recognizes and gets rid of specific points in a training dataset that contribute most to a model's failures on minority subgroups. By eliminating far fewer datapoints than other methods, this technique maintains the general precision of the design while improving its efficiency relating to underrepresented groups.

In addition, the method can identify hidden sources of bias in a training dataset that does not have labels. Unlabeled information are far more common than identified data for numerous applications.

This technique might also be combined with other methods to improve the fairness of machine-learning designs deployed in high-stakes circumstances. For instance, it may at some point help ensure underrepresented patients aren't misdiagnosed due to a prejudiced AI design.

"Many other algorithms that attempt to resolve this problem assume each datapoint matters as much as every other datapoint. In this paper, we are revealing that presumption is not true. There specify points in our dataset that are contributing to this bias, and we can find those data points, eliminate them, and get better efficiency," says Kimia Hamidieh, wiki.vifm.info an electrical engineering and computer technology (EECS) graduate trainee at MIT and co-lead author of a paper on this strategy.

She composed the paper with co-lead authors Saachi Jain PhD '24 and fellow EECS graduate trainee Kristian Georgiev; Andrew Ilyas MEng '18, PhD '23, a Stein Fellow at Stanford University; and senior authors Marzyeh Ghassemi, an associate teacher in EECS and a member of the Institute of Medical Engineering Sciences and the Laboratory for Details and Decision Systems, morphomics.science and Aleksander Madry, the Cadence Design Systems Professor at MIT. The research will be provided at the Conference on Neural Details Processing Systems.

Removing bad examples

Often, machine-learning models are trained using big datasets gathered from lots of sources throughout the web. These datasets are far too big to be carefully curated by hand, so they may contain bad examples that injure model efficiency.

Scientists likewise know that some data points affect a design's efficiency on certain more than others.

The MIT researchers integrated these 2 concepts into a method that recognizes and removes these troublesome datapoints. They look for to resolve an issue called worst-group error, which occurs when a model underperforms on minority subgroups in a training dataset.

The scientists' new strategy is driven by prior work in which they presented a technique, wiki.eqoarevival.com called TRAK, that recognizes the most crucial training examples for a particular model output.

For this brand-new strategy, they take incorrect forecasts the model made about minority subgroups and utilize TRAK to determine which training examples contributed the most to that incorrect prediction.

"By aggregating this details throughout bad test forecasts in the proper way, we are able to find the specific parts of the training that are driving worst-group accuracy down overall," Ilyas explains.

Then they eliminate those particular samples and retrain the model on the remaining information.

Since having more information usually yields better general efficiency, eliminating simply the samples that drive worst-group failures maintains the design's overall precision while improving its efficiency on minority subgroups.

A more available technique

Across three machine-learning datasets, their method exceeded several techniques. In one circumstances, it enhanced worst-group precision while removing about 20,000 fewer training samples than a conventional information balancing method. Their technique likewise attained greater precision than approaches that require making modifications to the inner workings of a design.

Because the MIT technique involves changing a dataset rather, it would be much easier for a specialist to use and can be used to many kinds of models.

It can likewise be utilized when predisposition is unidentified because subgroups in a training dataset are not identified. By recognizing datapoints that contribute most to a feature the design is finding out, they can understand the variables it is using to make a forecast.

"This is a tool anyone can use when they are training a machine-learning model. They can take a look at those datapoints and see whether they are lined up with the capability they are trying to teach the design," says Hamidieh.

Using the technique to discover unknown subgroup predisposition would require instinct about which groups to look for, so the scientists wish to verify it and explore it more completely through future human research studies.

They likewise want to enhance the efficiency and reliability of their strategy and guarantee the technique is available and easy-to-use for specialists who could someday deploy it in real-world environments.

"When you have tools that let you critically look at the information and figure out which datapoints are going to lead to bias or other unwanted behavior, it gives you an initial step towards structure designs that are going to be more fair and more dependable," Ilyas says.

This work is moneyed, in part, by the National Science Foundation and the U.S. Defense Advanced Research Projects Agency.

Assignee
Assign to
None
Milestone
None
Assign milestone
Time tracking
None
Due date
No due date
0
Labels
None
Assign labels
  • View project labels
Reference: alysa839654637/l-williams#23