Skip to content

  • Projects
  • Groups
  • Snippets
  • Help
    • Loading...
    • Help
    • Submit feedback
    • Contribute to GitLab
  • Sign in
L
lepostecanada
  • Project
    • Project
    • Details
    • Activity
    • Cycle Analytics
  • Issues 44
    • Issues 44
    • List
    • Board
    • Labels
    • Milestones
  • Merge Requests 0
    • Merge Requests 0
  • CI / CD
    • CI / CD
    • Pipelines
    • Jobs
    • Schedules
  • Wiki
    • Wiki
  • Snippets
    • Snippets
  • Members
    • Members
  • Collapse sidebar
  • Activity
  • Create a new issue
  • Jobs
  • Issue Boards
  • Teresa Mercer
  • lepostecanada
  • Issues
  • #21

Closed
Open
Opened Feb 12, 2025 by Teresa Mercer@akoteresa8217
  • Report abuse
  • New issue
Report abuse New issue

Researchers Reduce Bias in aI Models while Maintaining Or Improving Accuracy


Machine-learning models can fail when they attempt to make forecasts for people who were underrepresented in the datasets they were trained on.

For example, a model that anticipates the finest treatment choice for qoocle.com somebody with a chronic disease might be trained utilizing a dataset that contains mainly male clients. That model may make inaccurate predictions for female clients when released in a medical facility.

To improve results, engineers can attempt stabilizing the training dataset by removing data points till all subgroups are represented equally. While dataset balancing is promising, it typically needs eliminating big quantity of data, harming the model's total performance.

MIT scientists established a new method that recognizes and removes specific points in a training dataset that contribute most to a design's failures on minority subgroups. By getting rid of far less datapoints than other approaches, this method maintains the total accuracy of the model while improving its performance concerning underrepresented groups.

In addition, the technique can recognize covert sources of bias in a training dataset that does not have labels. Unlabeled data are far more common than identified data for many applications.

This method might also be integrated with other methods to improve the fairness of machine-learning models deployed in high-stakes situations. For instance, it may sooner or later assist guarantee underrepresented patients aren't misdiagnosed due to a biased AI model.

"Many other algorithms that try to address this concern presume each datapoint matters as much as every other datapoint. In this paper, we are showing that assumption is not true. There are specific points in our dataset that are adding to this bias, and we can find those data points, remove them, and improve performance," says Kimia Hamidieh, an electrical engineering and computer technology (EECS) graduate trainee at MIT and co-lead author of a paper on this strategy.

She wrote the paper with co-lead authors Saachi Jain PhD '24 and fellow EECS graduate trainee Kristian Georgiev; Andrew Ilyas MEng '18, PhD '23, a Stein Fellow at Stanford University; and senior authors Marzyeh Ghassemi, an associate teacher in EECS and a member of the Institute of Medical Engineering Sciences and the Laboratory for Details and Decision Systems, and Aleksander Madry, the Cadence Design Systems Professor at MIT. The research will be presented at the Conference on Neural Details Processing Systems.

Removing bad examples

Often, machine-learning designs are trained using big datasets gathered from numerous sources throughout the web. These datasets are far too big to be thoroughly curated by hand, so they might contain bad examples that hurt model efficiency.

Scientists likewise understand that some information points impact a design's performance on certain downstream jobs more than others.

The MIT scientists integrated these 2 concepts into a technique that recognizes and eliminates these problematic datapoints. They look for to fix an issue called worst-group mistake, which happens when a model underperforms on minority subgroups in a training dataset.

The scientists' new technique is driven by prior operate in which they introduced a method, called TRAK, that identifies the most essential training examples for a particular design output.

For this brand-new technique, they take incorrect predictions the design made about minority subgroups and utilize TRAK to recognize which training examples contributed the most to that inaccurate forecast.

"By aggregating this details across bad test predictions in the right way, we have the ability to discover the specific parts of the training that are driving worst-group precision down overall," Ilyas explains.

Then they remove those particular samples and retrain the model on the remaining information.

Since having more data usually yields much better general efficiency, eliminating just the samples that drive worst-group failures maintains the design's total precision while enhancing its performance on minority subgroups.

A more available technique

Across three machine-learning datasets, their technique surpassed multiple strategies. In one instance, it boosted worst-group precision while about 20,000 less training samples than a conventional information balancing method. Their method also attained greater accuracy than techniques that need making modifications to the inner operations of a model.

Because the MIT method includes changing a dataset instead, it would be simpler for a specialist to utilize and can be applied to many types of designs.

It can likewise be used when predisposition is unknown due to the fact that subgroups in a training dataset are not identified. By determining datapoints that contribute most to a feature the model is learning, they can understand the variables it is using to make a forecast.

"This is a tool anybody can use when they are training a machine-learning model. They can look at those datapoints and see whether they are aligned with the ability they are trying to teach the model," states Hamidieh.

Using the method to discover unknown subgroup predisposition would need intuition about which groups to search for, so the scientists intend to confirm it and explore it more completely through future human studies.

They also wish to enhance the efficiency and reliability of their method and guarantee the method is available and easy-to-use for specialists who could at some point deploy it in real-world environments.

"When you have tools that let you critically take a look at the information and figure out which datapoints are going to lead to predisposition or other undesirable habits, it gives you a primary step towards building models that are going to be more fair and more trusted," Ilyas says.

This work is moneyed, in part, by the National Science Foundation and the U.S. Defense Advanced Research Projects Agency.

Assignee
Assign to
None
Milestone
None
Assign milestone
Time tracking
None
Due date
No due date
0
Labels
None
Assign labels
  • View project labels
Reference: akoteresa8217/lepostecanada#21