Skip to content

GitLab

  • Menu
Projects Groups Snippets
    • Loading...
  • Help
    • Help
    • Support
    • Community forum
    • Submit feedback
    • Contribute to GitLab
  • Sign in / Register
  • N nethunt
  • Project information
    • Project information
    • Activity
    • Labels
    • Members
  • Repository
    • Repository
    • Files
    • Commits
    • Branches
    • Tags
    • Contributors
    • Graph
    • Compare
  • Issues 1
    • Issues 1
    • List
    • Boards
    • Service Desk
    • Milestones
  • Merge requests 0
    • Merge requests 0
  • CI/CD
    • CI/CD
    • Pipelines
    • Jobs
    • Schedules
  • Deployments
    • Deployments
    • Environments
    • Releases
  • Monitor
    • Monitor
    • Incidents
  • Packages & Registries
    • Packages & Registries
    • Package Registry
    • Infrastructure Registry
  • Analytics
    • Analytics
    • Value stream
    • CI/CD
    • Repository
  • Wiki
    • Wiki
  • Snippets
    • Snippets
  • Activity
  • Graph
  • Create a new issue
  • Jobs
  • Commits
  • Issue Boards
Collapse sidebar
  • Lin Jean
  • nethunt
  • Issues
  • #1

Closed
Open
Created Feb 11, 2025 by Lin Jean@linx7099115560Maintainer

Researchers Reduce Bias in aI Models while Maintaining Or Improving Accuracy


Machine-learning designs can fail when they try to make predictions for individuals who were underrepresented in the datasets they were trained on.

For oke.zone circumstances, a model that anticipates the very best treatment choice for someone with a chronic disease may be trained using a dataset that contains mainly male clients. That model may make inaccurate predictions for female clients when deployed in a hospital.

To improve results, engineers can try balancing the training dataset by removing information points till all subgroups are represented similarly. While dataset balancing is promising, it typically requires eliminating large amount of data, hurting the model's general efficiency.

MIT researchers developed a brand-new strategy that recognizes and removes specific points in a training dataset that contribute most to a model's failures on minority subgroups. By removing far less datapoints than other approaches, this method maintains the overall precision of the design while improving its efficiency concerning underrepresented groups.

In addition, the technique can recognize covert sources of predisposition in a training dataset that lacks labels. Unlabeled information are far more prevalent than identified information for lots of applications.

This technique could likewise be combined with other approaches to improve the fairness of machine-learning models released in high-stakes scenarios. For example, it might one day assist ensure underrepresented clients aren't misdiagnosed due to a prejudiced AI model.

"Many other algorithms that try to resolve this problem presume each datapoint matters as much as every other datapoint. In this paper, we are revealing that presumption is not true. There are particular points in our dataset that are contributing to this bias, and we can find those data points, eliminate them, and get much better efficiency," says Kimia Hamidieh, an electrical engineering and computer science (EECS) graduate trainee at MIT and co-lead author of a paper on this method.

She composed the paper with co-lead authors Saachi Jain PhD '24 and fellow EECS graduate trainee Kristian Georgiev; Andrew Ilyas MEng '18, setiathome.berkeley.edu PhD '23, a Stein Fellow at Stanford University; and senior authors Marzyeh Ghassemi, an associate professor in EECS and a member of the Institute of Medical Engineering Sciences and the Laboratory for and Decision Systems, and Aleksander Madry, the Cadence Design Systems Professor at MIT. The research will exist at the Conference on Neural Details Processing Systems.

Removing bad examples

Often, machine-learning models are trained using huge datasets collected from many sources throughout the web. These datasets are far too big to be thoroughly curated by hand, so they may contain bad examples that harm model performance.

Scientists also understand that some data points affect a design's performance on certain downstream tasks more than others.

The MIT scientists integrated these two concepts into an approach that identifies and removes these bothersome datapoints. They look for to solve a problem called worst-group mistake, which takes place when a model underperforms on minority subgroups in a training dataset.

The scientists' new method is driven by prior work in which they presented an approach, called TRAK, that recognizes the most crucial training examples for a particular model output.

For this brand-new strategy, they take incorrect predictions the design made about minority subgroups and utilize TRAK to determine which training examples contributed the most to that inaccurate forecast.

"By aggregating this details across bad test predictions in properly, we are able to find the specific parts of the training that are driving worst-group accuracy down overall," Ilyas explains.

Then they remove those specific samples and retrain the design on the remaining data.

Since having more data usually yields much better total performance, eliminating just the samples that drive worst-group failures maintains the model's total precision while enhancing its performance on minority subgroups.

A more available approach

Across 3 machine-learning datasets, their approach exceeded multiple techniques. In one instance, it boosted worst-group accuracy while eliminating about 20,000 less training samples than a conventional information balancing technique. Their strategy likewise attained higher accuracy than techniques that need making changes to the inner operations of a model.

Because the MIT technique includes changing a dataset instead, wiki.vifm.info it would be much easier for a professional to utilize and can be used to numerous types of designs.

It can likewise be used when bias is unknown due to the fact that subgroups in a training dataset are not identified. By recognizing datapoints that contribute most to a feature the design is discovering, thatswhathappened.wiki they can understand the variables it is utilizing to make a prediction.

"This is a tool anybody can utilize when they are training a machine-learning model. They can look at those datapoints and see whether they are lined up with the ability they are trying to teach the design," says Hamidieh.

Using the technique to spot unknown subgroup predisposition would need instinct about which groups to search for, so the scientists intend to verify it and explore it more totally through future human research studies.

They also desire to enhance the performance and dependability of their strategy and make sure the approach is available and easy-to-use for specialists who might sooner or later deploy it in real-world environments.

"When you have tools that let you critically look at the data and find out which datapoints are going to lead to predisposition or other undesirable habits, it gives you an initial step toward structure designs that are going to be more fair and more trusted," Ilyas states.

This work is moneyed, in part, by the National Science Foundation and the U.S. Defense Advanced Research Projects Agency.

Assignee
Assign to
Time tracking