Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

2018, AIES, Measuring and Mitigating Unintended Bias in Text Classification #9

Open
Rounique opened this issue Nov 2, 2021 · 5 comments
Assignees
Labels
literature-review Summary of the paper related to the work

Comments

@Rounique
Copy link
Contributor

Rounique commented Nov 2, 2021

No description provided.

@Rounique Rounique self-assigned this Nov 2, 2021
@Rounique
Copy link
Contributor Author

Rounique commented Nov 2, 2021

Titile: Measuring and Mitigating Unintended Bias in Text Classification
Venue: AIES
Year: 2018

Introduction:
Human biases can easily result in a skewed distribution in the training data, as many ML models are developed using human-generated data. This paper introduces a new approach to measure and mitigate unintended bias in ML models. this approach reduces the unintended bias without compromising overall model quality

Main Concern:
This paper's key contribution is the introduction of methods for quantifying and mitigating unintentional bias in text categorization algorithms.

Previous Gaps in Work:
They discuss the impact of using unfair natural language processing models for real-world tasks but do not provide mitigation strategies

Input:
This research uses a text classifier that was created to detect toxicity in Wikipedia Talk Pages comments. The model is based on a dataset of 127,820 Talk Page comments, each of which was classified as toxic or non-toxic by human raters.
(A toxic statement is described as one that is "rude, disrespectful, or unreasonable and likely to cause you to quit a conversation.")

Metric Used:
characteristic curve or AUC

*Here they make a distinction between unintended biases in a machine learning model and the algorithm's potential for unfair applications. Bias is built into every machine learning model. A model trained to detect toxic comments, for example, is designed to be biased in favor of toxic comments. The model isn't supposed to discriminate between people's genders in comments, but if it does, it's called unintentional bias. On the other hand, fairness is a term we use to describe a potential negative influence on society, particularly when different persons are treated differently.

Methodology:
By strategically adding data, it presents a simple and new strategy to mitigate that prejudice.
Convolutional neural networks were used in all versions of the model, which were trained in TensorFlow using the Keras framework.
(In order to mitigate the data imbalance which produces unintended bias, additional data has been added. For each term, to bring the toxic/non-toxic balance in line with the prior distribution for the overall dataset, they have added enough non-toxic examples. Adding these additional data was found effective in mitigating bias.)

Gaps of Work:
Automating the mining of identity words affected by unintended bias since they are made by humans in this paper.

Conclusion:
It is shown that implementing these strategies mitigates unintended biases in a model without sacrificing overall model quality or having a significant influence on the original test set.

@hosseinfani hosseinfani added the literature-review Summary of the paper related to the work label Nov 2, 2021
@hosseinfani
Copy link
Member

@Rounique
I liked your summary.

  • Can you explain a little bit more about those strategies of adding data to mitigate the unintended biases? I am not sure I understood that part.
  • Also, I am not sure I understood the gaps of the current work.

@hosseinfani
Copy link
Member

@Rounique
Any update?

@Rounique
Copy link
Contributor Author

It has been edited.

@hosseinfani
Copy link
Member

@Rounique
Still, I'm confused! Let's discuss it in our weekly meeting.

@hosseinfani hosseinfani transferred this issue from fani-lab/OpeNTF Mar 12, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
literature-review Summary of the paper related to the work
Projects
None yet
Development

No branches or pull requests

2 participants