JuniorTorresMTJ / model_api_mlops_project

This project aims to apply MLOps techniques to deploy a machine learning model through an API constructed with FastAPI. We utilize Poetry for dependency management and Docker for containerization, ensuring the code is modular, organized πŸ“, and maintainable πŸ› οΈ.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Welcome to Model API MLOps Project

Documentation Status PyPI version

πŸ’‘ Summary

🌐 Overview


This project aims to apply MLOps techniques to deploy a machine learning model through an API constructed with FastAPI. We utilize Poetry for dependency management and Docker for containerization, ensuring the code is modular, organized πŸ“, and maintainable πŸ› οΈ. Furthermore, this project boasts comprehensive documentation πŸ“„, ensuring clarity and ease of replication for other developers.

Key Tasks πŸ”‘:

  • Code Modularization and Refactoring: The original script needed an overhaul to reach a modular, and potentially object-oriented, structure πŸ”„.

  • Documentation: Implementing clear and concise documentation πŸ“š to ensure smooth navigation and comprehension for fellow developers.

  • Preservation of Output: Despite alterations in the code's structure and logic, the model's predictions (output) ought to remain consistent βœ”οΈ.

  • Configuration File Modification: The structure of artifacts/pipeline.jsonc was amenable to changes to ensure it meshed seamlessly with the revamped code structure πŸ”§.

  • Incorporate TODOs: The code housed multiple TODOs that had to be addressed and realized as part of the challenge πŸ“.

  • Scalability Consideration: The reshaped code was anticipated to cater to hefty data scenarios, suggesting that the supplied sample data file might, in a real-world setting, symbolize a table spanning several gigabytes in a data lake πŸ“Š.

Constraints and Considerations 🚫:

  • The generated code should be portable and capable of executing flawlessly in an alternate Python environment 🐍.

  • The provided "pickle" files were crafted using Python version 3.8.13 and Scikit-learn version 1.0.2. Though other libraries might be pivotal for diverse facets of the project, these particular versions were indispensable for the "pickle" files πŸ“Œ.

  • Intricate or time-intensive modifications, if not instantly implementable, should be signposted within the code as TODOs. These markers should also sketch out the strategic approach intended for their eventual fulfillment πŸ•°οΈ.

πŸš€ Run the project


The provided configuration appears to be from a pyproject.toml file, which is a configuration file for the Poetry dependency management and packaging tool in Python. To install and execute this project, follow the steps below:

  • Prerequisites:
  • Docker
  • Poetry
  • Python

1️⃣ Install Poetry:

If you haven't installed Poetry yet, you can do so with the following command:

curl -sSL https://install.python-poetry.org | python3 -

Or you can use other installation methods detailed in the official documentation.

2️⃣ Clone the project (if you haven't already):

Assuming the project is hosted on a Git repository, clone it:

git clone https://github.com/JuniorTorresMTJ/model_api_mlops_project.git
cd api_mlops_project

3️⃣ Install project dependencies:

Navigate to the project directory and install the dependencies using Poetry:

poetry install

This command installs both the main dependencies and the dev dependencies of the project.

4️⃣ Activate the virtual environment:

Poetry creates a virtual environment for your project. Activate it with:

poetry shell

5️⃣ Build the Docker image:

The exact command to run the project depends on its structure. Typically, there's a main Python script or module you'd execute to start the project. If there's a clear entry point, you can run:

docker build -t api-mlops-project:latest .

6️⃣ Build the Docker image:

docker run -p 8000:8000 api-mlops-project:latest

API should now be running and accessible at http://localhost:8000.

Additional tasks:

The provided configuration also contains some tasks defined with taskipy. You can run these tasks using the poetry run task command. For example:

  • To run the lint task:
  poetry run task lint
  • To serve the documentation:
  poetry run task doc

7️⃣ Endpoints:

βœ… GET /data

  • Purpose:

    • This endpoint retrieves the original dataset from a predefined file location.
  • Usage:

    • You simply send a GET request to this endpoint without any parameters.
  • Example with curl:

    curl -X 'GET' http://localhost:8000/data
  • Response:

data

If successful, you'll receive a JSON representation of the dataset. The structure will be a dictionary with columns as keys and their associated data as corresponding values.


βœ… POST /process_data

  • Purpose:

    • Processes data from an uploaded file and returns the processed output.
  • Usage:

    • Send a POST request to this endpoint with an attached file containing the data you want to process.
  • Example with curl:

    curl -X 'POST' -F 'file=@path_to_your_file.csv' http://localhost:8000/process_data
  • Response:

predict_path

If successful, you'll receive processed data (like predictions) as a JSON response.

βœ… POST /predict

  • Purpose:

    • This endpoint takes input data in the form of a JSON object and returns a prediction based on the provided data.
  • Usage:

    • Send a POST request with a JSON payload representing the data for which you want a prediction.
  • Example with curl:

    `curl -X 'POST' -H "Content-Type: application/json" -d '{"vibration_x": 45.2947443182, "vibration_y": 91.0739650974, "vibration_z": 10.2298092532}' http://localhost:8000/predict`
  • Response:

predict

If successful, you'll receive a JSON response with the prediction status and the predicted class.


βœ… POST /predict_batch

  • Purpose:

    • Provides batch predictions based on the provided data.
  • Usage:

    • Send a POST request with a JSON payload containing multiple rows of data for batch processing.
  • Example with curl:

    curl -X 'POST' -H "Content-Type: application/json" -d '[{"vibration_x": 44.9700689935, "vibration_y": 90.0999391234, "vibration_z": 10.2298092532},{"vibration_x": 45.2947443182, "vibration_y": 91.0739650974, "vibration_z": 10.2298092532},{"vibration_x": 45.9440949675, "vibration_y": 91.0739650974, "vibration_z": 10.2298092532}]' http://localhost:8000/predict_batch
  • Response:

predict_batch

If successful, you'll receive a JSON response with the prediction status and the predicted classes for each input data row.

For each of the endpoints, if there's an error, the API will return a clear message indicating what went wrong, along with a status code indicating the nature of the error. Always make sure to check the returned message and status code to understand any issues.

πŸ“– Documentation:

The documentation of this project is an essential component for understanding its structure, functionalities, and usage guides. We strongly believe that comprehensive documentation can speed up the adoption, contribution, and debugging processes of the project. The documentation was crafted using MkDocs, a swift and straightforward tool for creating documentation websites from markdown files. It allows us to concentrate on the content while automating the process of converting Markdown into a structured and styled website.

⚠️Documentation Link⚠️

πŸ“‘ Overview of Project Technologies


This section offers a comprehensive understanding of the technologies incorporated into our project, shedding light on each tool in detail. Grasping these technologies is pivotal for any developer or stakeholder aiming to become acquainted with the infrastructure and the project code.


  • Python 3.8.10 🐍
    • Description: Language used for application development.

  • Poetry πŸ“¦
    • Description: A tool for dependency management and packaging in Python. Poetry provides a streamlined and consistent approach to handle Python projects and their dependencies, ensuring transparency and reproducibility throughout the process.

  • Scikit-learn 1.0.2 πŸ“Š
    • Description: A machine learning library in Python. Scikit-learn, together with Pandas, stands as an essential tool for data processing, analysis, and the deployment of machine learning models. Both libraries bring forth a vast set of features and are widely recognized in the community.

  • Pandas ^2.0.3 🐼
    • Description: Used for data reading and transformations.

  • Blue ^0.9.1 πŸ”΅
    • Description: Coding standard adopted for the project.

  • isort ^5.12.0 πŸ“‘
    • Description: A coding utility to organize imports in a consistent format.

  • taskipy ^1.11.0 πŸ› οΈ
    • Description: Facilitates development by creating tasks.

  • pyarrow ^12.0.1 🏹
    • Description: Enables reading of .parquet data files.

  • Docker 🐳
    • Description: Docker is a platform designed to develop, ship, and run applications within containers. It ensures our application and its dependencies are packaged into an isolated container, which enhances portability and ease of deployment across varied environments.

Grasping the intricate details of these technologies will significantly augment the efficiency and effectiveness of engagement with the project. πŸš€πŸŒŸ

πŸ“ Next Steps


Here are some next steps that can be implemented in the future:

1️⃣ Model Registration in MLflow

Lastly, to track and manage our machine learning model iterations, we will use MLflow to register, version, and share models.

2️⃣ Validation with Stakeholders

We plan to conduct several feedback sessions with our stakeholders to ensure the project aligns with the expectations and needs of the end users.

3️⃣ Monitoring Setup

Monitoring the performance and functionality of the system is crucial to ensure its reliability and efficiency. It's important to implement tools and processes for keeping an eye on the system.

4️⃣ Alert Setup for Slack or Email

To ensure a swift response to any issues that may arise, we'll implement alerts that will send notifications to Slack or predetermined emails in case of system failures.

5️⃣ Refactoring of the dataframe_checker.py class

Currently, the dataframe_checker.py class is responsible for validating input data. Initially, it wasn't built with PySpark, but ideally, its internal logic should be replaced with PySpark to optimize validation and preparation of data on large datasets.

6️⃣ Scheduling with Airflow

To automate and orchestrate our workflows, we will use Airflow. It will assist us in defining, scheduling, and monitoring workflows programmatically.

πŸ’» CONTACT ME πŸ’»


Junior Torres | Linkedin Junior Torres | Instagram Junior Torres | Gmail Junior Torres | Github

About

This project aims to apply MLOps techniques to deploy a machine learning model through an API constructed with FastAPI. We utilize Poetry for dependency management and Docker for containerization, ensuring the code is modular, organized πŸ“, and maintainable πŸ› οΈ.

License:MIT License


Languages

Language:Python 98.1%Language:Dockerfile 1.9%