A Data Science Project struture in cookiecutter style.
Developed with ❤️ by A3Data
Hermione is the newest open source library that will help Data Scientists on setting up more organized codes, in a quicker and simpler way. Besides, there are some classes in Hermione which assist with daily tasks such as: column normalization and denormalization, data view, text vectoring, etc. Using Hermione, all you need is to execute a method and the rest is up to her, just like magic.
To bring in a little of A3Data experience, we work in Data Science teams inside several client companies and it’s undeniable the excellence of notebooks as a data exploration tool. Nevertheless, when it comes to data science products and their context, when the models needs to be consumed, monitored and have periodic maintenance, putting it into production inside a Jupyter Notebook is not the best choice (we are not even mentioning memory and CPU performance yet). And that’s why Hermione comes in! We have been inspired by this brilliant, empowered and awesome witch of The Harry Potter saga to name this framework!
This is also our way of reinforcing our position that women should be taking more leading roles in the technology field. #CodeLikeAGirl
- Python (>= 3.8)
- docker
Hermione does not depend on conda to build and manage virtual environments anymore. It uses venv
instead.
pip install -U hermione-ml
For bash:
echo 'eval "$(_HERMIONE_COMPLETE=source_bash hermione)"' >> ~/.bashrc
For Zsh:
echo 'eval "$(_HERMIONE_COMPLETE=source_zsh hermione)"' >> ~/.zshrc
After installed Hermione:
- Create you new project:
hermione project new project_hermione
- Hit Enter if you want to start with an example code
Please select one of the following templates
(0) starter
(1) barebones
(2) sagemaker
Option [0]:
- Hermione already creates a virtual environment for the project. For Windows users, activate it with
<project_name>_env\Scripts\activate
For linux and MacOS users, do
source <project_name>_env/bin/activate
- After activating, you should install some libraries. There are a few suggestions in “requirements.txt” file:
pip install -r requirements.txt
- Now, if you selected the starter version, we will train some models from the example, using MLflow ❤. To do so, inside project directory, just type: hermione train. The “hermione run train” command will search for a
train.py
file and execute it. In the example, models and metrics are already controlled via MLflow.
- After that, a mlflow experiment is created. To verify the experiment in mlflow, type: mlflow ui. The application will go up.
mlflow ui
[2020-10-19 23:23:12 -0300] [15676] [INFO] Starting gunicorn 19.10.0
[2020-10-19 23:23:12 -0300] [15676] [INFO] Listening at: http://127.0.0.1:5000 (15676)
[2020-10-19 23:23:12 -0300] [15676] [INFO] Using worker: sync
[2020-10-19 23:23:12 -0300] [15678] [INFO] Booting worker with pid: 15678
- To access the experiment, just enter the path previously provided in your preferred browser. Then it is possible to check the trained models and their metrics.
- To make batch predictions using your
predict.py
file, typehermione run predict
. The default implemented version will print some predictions for you in the terminal.
hermione run predict
- In the Titanic example, we also provide a step by step notebook. To view it, just type jupyter notebook inside directory
notebooks
.
- If you selected the Sagemaker version, click here to check a tutorial.
Do you want to create your project from scratch? There click here to check a tutorial.
Hermione comes with a default Dockerfile
which implements a FastAPI application that serves your ML model. You should take a look at the api/app.py
module and rewrite predict_new()
function as you see fit.
Also, in the newest version, hermione brings two CLI commands that helps us abstract a little bit the complexity regarding docker commands. To build an image (remember you should have docker installed), you should be in the project's root directory. Than, do:
hermione run build <IMAGE_NAME>
After you have built you're docker image, run it with:
hermione run container <IMAGE_NAME>
[2020-10-20 02:13:20 +0000] [1] [INFO] Starting gunicorn 20.0.4
[2020-10-20 02:13:20 +0000] [1] [INFO] Listening at: http://0.0.0.0:5000 (1)
[2020-10-20 02:13:20 +0000] [1] [INFO] Using worker: sync
[2020-10-20 02:13:20 +0000] [7] [INFO] Booting worker with pid: 7
[2020-10-20 02:13:20 +0000] [8] [INFO] Booting worker with pid: 8
[2020-10-20 02:13:20 +0000] [16] [INFO] Booting worker with pid: 16
THAT IS IT! You have a live model up and running. To test your API, hermione provides a api/myrequests.py
module. This is not part of the project; it's a "ready to go" code to make requests to the API. Help yourself!
cd src/api
python myrequests.py
Sending request for model...
Data: {"Pclass": [3, 2, 1], "Sex": ["male", "female", "male"], "Age": [4, 22, 28]}
Response: "[0.24630952 0.996 0.50678968]"
Play a little with the 'fake' data and see how far can the predictions go.
This is the class structure diagram that Hermione relies on:
Here we describe briefly what each class is doing:
- DataBase - should be used when data recovery requires a connection to a database. Contains methods for opening and closing a connection.
- Spreadsheet - should be used when data recovery is in spreadsheets/text files. All aggregation of the bases to generate a "flat table" should be performed in this class.
- DataSource - abstract class which DataBase and Spreadsheet inherit from.
- Preprocessing - concentrates all preprocessing steps that must be performed on the data before the model is trained.
- Normalization - applies normalization and denormalization to reported columns. This class contains the following normalization algorithms already implemented: StandardScaler e MinMaxScaler.
- TextVectorizer - transforms text into vector. Implemented methods: Bag of words, TF_IDF, Embedding: mean, median e indexing.
- DataQuality - concentrates all data validation steps that must be performed on the data to ensure its quality.
- Visualization - methods for data visualization. There are methods to make static and interactive plots.
- App Streamlit - streamlit example consuming Titanic dataset, including pandas profilling.
- Trainer - module that centralizes training algorithms classes. Algorithms from
scikit-learn
library, for instance, can be easily used with the TrainerSklearn implemented class. - Wrapper - centralizes the trained model with its metrics. This class has built-in integration with MLFlow.
- Metrics - it contains key metrics that are calculated when models are trained. Classification, regression and clustering metrics are already implemented.
- test_project - module for unit testing.
Have a look at our contributing guide.
Make a pull request with your implementation.
For suggestions, contact us: hermione@a3data.com.br