Nokaido / Capstone_Project

Udacity Data Scientist Nanodegree Capstone Project

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Table of Contents

  1. Installation
  2. Project Motivation
  3. File Descriptions
  4. Results
  5. Licensing, Authors, and Acknowledgements

Installation

There should be no necessary libraries to run the code here beyond the Anaconda distribution of Python. The code should run with no issues using Python versions 3.*. Used Libraries:

  • numpy
  • pandas
  • matplotlib.pyplot
  • seaborn
  • progressbar
  • sklearn.preprocessing.StandardScaler
  • sklearn.cluster.KMeans
  • sklearn.model_selection.GridSearchCV
  • sklearn.model_selection.KFold
  • sklearn.preprocessing.MinMaxScaler
  • sklearn.metrics.roc_auc_score
  • sklearn.ensemble.RandomForestClassifier
  • sklearn.ensemble.AdaBoostClassifier
  • sklearn.ensemble.GradientBoostingClassifier
  • IPython.display.clear_output

Project Motivation

For this project, I was interested in using the Bertelsmann data for the Udacity Data Scientst Nanodegree Capstone Project. In this Project I will look at the followung points:

  1. How is the data structured and what are the different stastistics of the datas content
  2. Whats the best Customer segmentation
  3. Building a model to find future customers for a mail-order company
  4. Taking part in a Kaggle competition

Furthermore a report was written on Medium

File Descriptions

  • README.md: This readme file
  • Arvato Project Workbook_own.ipynb: The main project file
  • nan_list_altered.tsv: Tab seperated list with descriptions of missing data
  • LICENSE: licence file
  • resources/*: folder with all graphics used for the blog

Results

Findings: While following the predefined structure this project went through different steps and reveled much information which was hidden in the data.

  • In the pre-processing phase which is for many Data Science projects the most time consuming part, the project had to overcome the lack of description to many features and the sheer overload of different features. Non the less it was possible to define, unify and mend the missing data problem. Some compressed information was re-coded and some unnecessary data was dropped.
  • With the pre-processed data the project went through the unsupervised learning / clustering phase. In order to shrink down the complexity of the data 1000 components werde selected via a PCA analysis taking care of 70% of explained variance. The Clustering then reveled that the main customer of the mail order company can be described as the average mainstream online buyer who never orders offline.
  • Then the problems started appear and the cleaning / coding step had to be redesigned and redone. With the new design the model choice could be made resulting in the GradientBoosting model.
  • The model was then fine tuned by doing a grid search cross validation with different parameters. To prepare the last step the tuned model was trained on the whole data set to be able to predict the response of the Kaggle Challenge.

Score = 0.47020

Licensing, Authors, Acknowledgements

Must give credit to Udacity, Bertelsmann Arvato Analytics and Kaggle for the data. Otherwise, feel free to use the code here as you would like!

About

Udacity Data Scientist Nanodegree Capstone Project

License:MIT License


Languages

Language:HTML 57.8%Language:Jupyter Notebook 42.2%