bluerider / data-migration

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Data Migration

Problem statement

Imagine a scenario where you were given the task create an ETL (Extract, Transform, Load) so that API data is consumable by business analysts. Fortunately, your co-workers have already done the Extract step and has provided you with a .zip file containing retail order data in the raw JSON format. Your project manager has put you on the task to support these business analysts so that they can query that data using SQL from a PSQL database. While you’re at it, they would also want you to create a user table that would contain summary metrics that you think business analysts would find useful.
Note: Keep in mind that the newly created tables have to be sanely structured and those steps should be reproducible with the expectation that the ETL would run daily.

Deliverables

Please be conscientious about your commit messages as this is will help us to understand your thought process Have at least one table for the orders data and one table user data saved in the Postgres database. Do not restrain yourself to those two tables, you can get creative here! But don't forget to put data schema to the repo.
While there are endless ways to solve this problem, let’s be pragmatic with the scope of the deliverable. We don’t expect you to take more than 4 hours to work on this project.

About

License:GNU General Public License v3.0


Languages

Language:Shell 56.1%Language:Python 43.9%