Home Upload Photo Upload Videos Write a Blog Analytics Messaging Streaming Create Adverts Creators Program
Bebuzee Afghanistan Bebuzee Albania Bebuzee Algeria Bebuzee Andorra Bebuzee Angola Bebuzee Antigua and Barbuda Bebuzee Argentina Bebuzee Armenia Bebuzee Australia Bebuzee Austria Bebuzee Azerbaijan Bebuzee Bahamas Bebuzee Bahrain Bebuzee Bangladesh Bebuzee Barbados Bebuzee Belarus Bebuzee Belgium Bebuzee Belize Bebuzee Benin Bebuzee Bhutan Bebuzee Bolivia Bebuzee Bosnia and Herzegovina Bebuzee Botswana Bebuzee Brazil Bebuzee Brunei Bebuzee Bulgaria Bebuzee Burkina Faso Bebuzee Burundi Bebuzee Cabo Verde Bebuzee Cambodia Bebuzee Cameroon Bebuzee Canada Bebuzee Central African Republic Bebuzee Chad Bebuzee Chile Bebuzee China Bebuzee Colombia Bebuzee Comoros Bebuzee Costa Rica Bebuzee Côte d'Ivoire Bebuzee Croatia Bebuzee Cuba Bebuzee Cyprus Bebuzee Czech Republic Bebuzee Democratic Republic of the Congo Bebuzee Denmark Bebuzee Djibouti Bebuzee Dominica Bebuzee Dominican Republic Bebuzee Ecuador Bebuzee Egypt Bebuzee El Salvador Bebuzee Equatorial Guinea Bebuzee Eritrea Bebuzee Estonia Bebuzee Eswatini Bebuzee Ethiopia Bebuzee Fiji Bebuzee Finland Bebuzee France Bebuzee Gabon Bebuzee Gambia Bebuzee Georgia Bebuzee Germany Bebuzee Ghana Bebuzee Greece Bebuzee Grenada Bebuzee Guatemala Bebuzee Guinea Bebuzee Guinea-Bissau Bebuzee Guyana Bebuzee Haiti Bebuzee Honduras Bebuzee Hong Kong Bebuzee Hungary Bebuzee Iceland Bebuzee India Bebuzee Indonesia Bebuzee Iran Bebuzee Iraq Bebuzee Ireland Bebuzee Israel Bebuzee Italy Bebuzee Jamaica Bebuzee Japan Bebuzee Jordan Bebuzee Kazakhstan Bebuzee Kenya Bebuzee Kiribati Bebuzee Kuwait Bebuzee Kyrgyzstan Bebuzee Laos Bebuzee Latvia Bebuzee Lebanon Bebuzee Lesotho Bebuzee Liberia Bebuzee Libya Bebuzee Liechtenstein Bebuzee Lithuania Bebuzee Luxembourg Bebuzee Madagascar Bebuzee Malawi Bebuzee Malaysia Bebuzee Maldives Bebuzee Mali Bebuzee Malta Bebuzee Marshall Islands Bebuzee Mauritania Bebuzee Mauritius Bebuzee Mexico Bebuzee Micronesia Bebuzee Moldova Bebuzee Monaco Bebuzee Mongolia Bebuzee Montenegro Bebuzee Morocco Bebuzee Mozambique Bebuzee Myanmar Bebuzee Namibia Bebuzee Nauru Bebuzee Nepal Bebuzee Netherlands Bebuzee New Zealand Bebuzee Nicaragua Bebuzee Niger Bebuzee Nigeria Bebuzee North Korea Bebuzee North Macedonia Bebuzee Norway Bebuzee Oman Bebuzee Pakistan Bebuzee Palau Bebuzee Panama Bebuzee Papua New Guinea Bebuzee Paraguay Bebuzee Peru Bebuzee Philippines Bebuzee Poland Bebuzee Portugal Bebuzee Qatar Bebuzee Republic of the Congo Bebuzee Romania Bebuzee Russia Bebuzee Rwanda Bebuzee Saint Kitts and Nevis Bebuzee Saint Lucia Bebuzee Saint Vincent and the Grenadines Bebuzee Samoa Bebuzee San Marino Bebuzee São Tomé and Príncipe Bebuzee Saudi Arabia Bebuzee Senegal Bebuzee Serbia Bebuzee Seychelles Bebuzee Sierra Leone Bebuzee Singapore Bebuzee Slovakia Bebuzee Slovenia Bebuzee Solomon Islands Bebuzee Somalia Bebuzee South Africa Bebuzee South Korea Bebuzee South Sudan Bebuzee Spain Bebuzee Sri Lanka Bebuzee Sudan Bebuzee Suriname Bebuzee Sweden Bebuzee Switzerland Bebuzee Syria Bebuzee Taiwan Bebuzee Tajikistan Bebuzee Tanzania Bebuzee Thailand Bebuzee Timor-Leste Bebuzee Togo Bebuzee Tonga Bebuzee Trinidad and Tobago Bebuzee Tunisia Bebuzee Turkey Bebuzee Turkmenistan Bebuzee Tuvalu Bebuzee Uganda Bebuzee Ukraine Bebuzee United Arab Emirates Bebuzee United Kingdom Bebuzee Uruguay Bebuzee Uzbekistan Bebuzee Vanuatu Bebuzee Venezuela Bebuzee Vietnam Bebuzee World Wide Bebuzee Yemen Bebuzee Zambia Bebuzee Zimbabwe
Blog Image

AI Machine-Learning: In Bias We Trust?

According to a new study, explanation methods that help users determine whether to trust machine-learning model predictions can be less accurate for disadvantaged subgroups.

Machine-learning algorithms are sometimes employed to assist human decision-makers when the stakes are high. For example, a model may predict which law school candidates are most likely to pass the bar exam, assisting admissions officers in deciding which students to admit.

Because of the complexity of these models, often having millions of parameters, it is nearly impossible for AI researchers to fully understand how they make predictions. An admissions officer with no machine-learning experience might have no idea what is going on under the hood. Scientists sometimes employ explanation methods that mimic a larger model by creating simple approximations of its predictions. These approximations, which are far easier to understand, assist users in deciding whether to trust the model’s predictions.

However, are these explanation methods fair? If an explanation method provides better approximations for men than for women, or for white people than for black people,users may be more inclined to trust the model’s predictions for some people but not for others.

MIT scientists carefully examined the fairness of some widely used explanation methods. They discovered that the approximation quality of these explanations can vary drastically between subgroups and that the quality is often significantly lower for minoritized subgroups.

In practice, this means that if the approximation quality is lower for female applicants, there is a mismatch between the explanations and the model’s predictions, which could lead the admissions officer to wrongly reject more women than men.

Once the MIT researchers saw how pervasive these fairness gaps are, they tried several techniques to level the playing field. They were able to shrink some gaps, but couldn’t eradicate them.

“What this means in the real world is that people might incorrectly trust predictions more for some subgroups than for others. So, improving explanation models is important, but communicating the details of these models to end users is equally important. These gaps exist, so users may want to adjust their expectations as to what they are getting when they use these explanations,” says lead author Aparna Balagopalan, a graduate student in the Healthy ML group of the MIT Computer Science and Artificial Intelligence Laboratory (CSAIL).

High fidelity

Simplified explanation models can approximate predictions of a more complex machine-learning model in a way that humans can grasp. An effective explanation model maximizes a property known as fidelity, which measures how well it matches the larger model’s predictions.

Rather than focusing on average fidelity for the overall explanation model, the MIT researchers studied fidelity for subgroups of people in the model’s dataset. In a dataset with men and women, the fidelity should be very similar for each group, and both groups should have fidelity close to that of the overall explanation model.

“When you are just looking at the average fidelity across all instances, you might be missing out on artifacts that could exist in the explanation model,” Balagopalan says.

They developed two metrics to measure fidelity gaps, or disparities in fidelity between subgroups. One is the difference between the average fidelity across the entire explanation model and the fidelity for the worst-performing subgroup. The second calculates the absolute difference in fidelity between all possible pairs of subgroups and then computes the average. Read More...

Previous Post

The 10 Latest Artificial Intelligence Trends That Your Business Needs to Embrace

Next Post

If you have one of these Thread border routers, your smart home will be Matter-ready

Comments