Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

2018, EMNLP, Reducing Gender Bias in Abusive Language Detection #12

Open
Rounique opened this issue Oct 18, 2021 · 8 comments
Open

2018, EMNLP, Reducing Gender Bias in Abusive Language Detection #12

Rounique opened this issue Oct 18, 2021 · 8 comments
Assignees
Labels
literature-review Summary of the paper related to the work

Comments

@Rounique
Copy link
Contributor

No description provided.

@Rounique Rounique self-assigned this Oct 18, 2021
@Rounique Rounique added the literature-review Summary of the paper related to the work label Oct 18, 2021
@Rounique
Copy link
Contributor Author

@hosseinfani
Copy link
Member

@Rounique
your summary?

@Rounique
Copy link
Contributor Author

As I mentioned, I didn't do summaries up to now. I'll do them after I finish the assignments I have for this week.

@Rounique
Copy link
Contributor Author

#11

@hosseinfani
Copy link
Member

@Rounique
Any update?

@Rounique
Copy link
Contributor Author

Rounique commented Nov 26, 2021

Title: Reducing Gender Bias in Abusive Language Detection
Venue: EMNLP
Year: 2018

Introduction
As the use of social media and online platforms is increasing, people tend to share their ideas and words more and more. Here, automatic detection of abusive language plays an important role since abusive language can lead to cyber-bullying, personal trauma, hate crime, and discrimination. Therefore, using machine learning and NLP to automatically detect abusive language is useful for many websites or social media services.

In this paper, gender bias has been measured on models that are trained with abusive language datasets, and also some methods have been introduced for mitigating these biases. The bias measuring is done with a generated unbiased test set and the mitigating methods are: (1) debiased word embedding, (2) gender swap data augmentation, (3) fine-tuning with a larger corpus

Dataset:
Sexist Tweets, Abusive Tweets.

Measuring Gender Biases
It is not possible to measure gender bias on a dataset on which the model has been trained since it will follow the same biases. Therefore, it is necessary to generate an unbiased test set.
the test set generated in this work includes 1,152 samples (576 pairs) by filling the templates with common gender identity pairs (ex. male/female, man/woman, etc.). Some templates have been generated that contain both neutral and offensive nouns and adjectives inside the vocabulary to retain balance in neutral and abusive samples.

Mitigating Bias

Word Embeddings (DE)
This is an algorithm to correct word embeddings by removing gender-stereotypical information.

Gender Swap (GS)
What is basically done here is the identify male entities and swap them with equivalent female entities and vice-versa. This simple method removes the correlation between gender and classification decisions and has proven to be effective for correcting gender biases.

Bias fine-tuning (FT)
A method to use transfer learning from a less biased corpus to reduce the bias. Initially, a model is trained with a less-biased and larger source corpus and fine-tuned with a target corpus.

Metric used:
AUC

Conclusion
It is found that these proposed methods can reduce gender biases up to 90-98%, improving the robustness of the models.

Future Work
Increasing classification performance and reducing the bias at the same time.

Codes
https://github.com/conversationai/unintended-ml-bias-analysis

@Rounique
Copy link
Contributor Author

The summary is added.

@hosseinfani
Copy link
Member

@Rounique
please explore their codeline, there are more good info :
https://perspectiveapi.com/how-it-works/
https://conversationai.github.io/

@hosseinfani hosseinfani transferred this issue from fani-lab/OpeNTF Mar 12, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
literature-review Summary of the paper related to the work
Projects
None yet
Development

No branches or pull requests

2 participants