dimitreOliveira / Jigsaw-UnintendedBiasInToxicityClassification

(351st place- Top 12%) Repository for the "Jigsaw Unintended Bias in Toxicity Classification" Kaggle competition.

Home Page:https://www.kaggle.com/c/jigsaw-unintended-bias-in-toxicity-classification

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

About the repository

The goal of this repository is to use this competition database to develop NLP models to process text data, using different approaches like complex deep learning models (LSTM & GRU), also leverage the models with pre-trained models, like embeddings (Word2Vec, GloVe, etc). Also as the competition goal, elaborate models and reprocess data in a way that can reduce bias in various contexts, like gender or religion.

Our team published Kaggle kernels:

What you will find

  • Documentation [link]
    • Project working cycle and effort, relevant content and insights [link]
  • EDA [link]
    • Toxicity Bias - extensive EDA and Bi LSTM [link]
  • Models & Auxiliar data bases [link]

Jigsaw Unintended Bias in Toxicity Classification

Detect toxicity across a diverse range of conversations

Kaggle competition: https://www.kaggle.com/c/jigsaw-unintended-bias-in-toxicity-classification Our insights are here.

Overview

Can you help detect toxic comments ― and minimize unintended model bias? That's your challenge in this competition.

The Conversation AI team, a research initiative founded by Jigsaw and Google (both part of Alphabet), builds technology to protect voices in conversation. A main area of focus is machine learning models that can identify toxicity in online conversations, where toxicity is defined as anything rude, disrespectful or otherwise likely to make someone leave a discussion.

Last year, in the Toxic Comment Classification Challenge, you built multi-headed models to recognize toxicity and several subtypes of toxicity. This year's competition is a related challenge: building toxicity models that operate fairly across a diverse range of conversations.

Here’s the background: When the Conversation AI team first built toxicity models, they found that the models incorrectly learned to associate the names of frequently attacked identities with toxicity. Models predicted a high likelihood of toxicity for comments containing those identities (e.g. "gay"), even when those comments were not actually toxic (such as "I am a gay woman"). This happens because training data was pulled from available sources where unfortunately, certain identities are overwhelmingly referred to in offensive ways. Training a model from data with these imbalances risks simply mirroring those biases back to users.

In this competition, you're challenged to build a model that recognizes toxicity and minimizes this type of unintended bias with respect to mentions of identities. You'll be using a dataset labeled for identity mentions and optimizing a metric designed to measure unintended bias. Develop strategies to reduce unintended bias in machine learning models, and you'll help the Conversation AI team, and the entire industry, build models that work well for a wide range of conversations.

Disclaimer: The dataset for this competition contains text that may be considered profane, vulgar, or offensive.

Acknowledgments

The Conversation AI team would like to thank Civil Comments for making this dataset available publicly and the Online Hate Index Research Project at D-Lab, University of California, Berkeley, whose labeling survey/instrument informed the dataset labeling. We'd also like to thank everyone who has contributed to Conversation AI's research, especially those who took part in our last competition, the success of which led to the creation of this challenge.

About

(351st place- Top 12%) Repository for the "Jigsaw Unintended Bias in Toxicity Classification" Kaggle competition.

https://www.kaggle.com/c/jigsaw-unintended-bias-in-toxicity-classification

License:MIT License


Languages

Language:Jupyter Notebook 100.0%