sukanta-27 / StackOverflow-Tag-Prediction-Problem

The objective of this case study is to suggest the tags based on the content that was there in the question posted on Stackoverflow. Inspired by the Facebook Recruiting competition held on Kaggle.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

StackOverflow-Tag-Prediction-Problem

This case study is inspired by the famous Facebook Recruiting III - Keyword Extraction competition held on Kaggle.

Description:

Stack Overflow is the largest, most trusted online community for developers to learn, share their programming knowledge, and build their careers.

Stack Overflow is something which every programmer use one way or another. Each month, over 50 million developers come to Stack Overflow to learn, share their knowledge, and build their careers. It features questions and answers on a wide range of topics in computer programming. The website serves as a platform for users to ask and answer questions, and, through membership and active participation, to vote questions and answers up or down and edit questions and answers in a fashion similar to a wiki or Digg. As of April 2014 Stack Overflow has over 4,000,000 registered users, and it exceeded 10,000,000 questions in late August 2015. Based on the type of tags assigned to questions, the top eight most discussed topics on the site are: Java, JavaScript, C#, PHP, Android, jQuery, Python and HTML.

Objective:

Given the title and the body of the question, predict the tags that suits the question.

Type of Machine learning problem:

This is a Multi-label classification problem

Multi-label Classification: Multilabel classification assigns to each sample a set of target labels. This can be thought as predicting properties of a data-point that are not mutually exclusive, such as topics that are relevant for a document. A question on Stackoverflow might be about any of C, Pointers, FileIO and/or memory-management at the same time or none of these.

Performance metric

  • Micro-Averaged F1-Score (Mean F Score) : The F1 score can be interpreted as a weighted average of the precision and recall, where an F1 score reaches its best value at 1 and worst score at 0. The relative contribution of precision and recall to the F1 score are equal. The formula for the F1 score is: F1 = 2 * (precision * recall) / (precision + recall) In the multi-class and multi-label case, this is the weighted average of the F1 score of each class.

  • Hamming loss : The Hamming loss is the fraction of labels that are incorrectly predicted.

About

The objective of this case study is to suggest the tags based on the content that was there in the question posted on Stackoverflow. Inspired by the Facebook Recruiting competition held on Kaggle.


Languages

Language:Jupyter Notebook 100.0%