yyassif / docker-spark-hadoop-for-recommendation

Using All Big Data Technologies In Order To Apply ALS Algorithm To Recommend Amazon Prodcuts

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

How to use HDFS/Spark

Installation

Base Images

docker pull bde2020/hadoop-namenode:1.1.0-hadoop2.8-java8
docker pull bde2020/hadoop-datanode:1.1.0-hadoop2.8-java8
docker pull bde2020/spark-base:2.1.0-hadoop2.8-hive-java8
docker pull bde2020/spark-master:2.1.0-hadoop2.8-hive-java8
docker pull bde2020/spark-worker:2.1.0-hadoop2.8-hive-java8
docker pull bde2020/spark-notebook:2.1.0-hadoop2.8-hive
docker pull bde2020/hdfs-filebrowser:3.11

Docker Compose File

To start an HDFS/Spark Workbench, run:

docker-compose up -d

Interfaces

Recommendation Spark Application

Jar file is packaged under the jarfile directory.

  • Compiled with scala 2.11.11
  • For Spark 2.1.0

How to run (using Makefile)

To Run the make the make command I've come up with this order which seem very mandatory to properly have the job done.

Start the Preprocessing

make prepare-raw-dataset

Start the Data Ingestion into HDFS

make ingest-hdfs

Create the Fat-JAR File

make jar

Start the Prediction

make prediction

Save the Results into a result directory

make prediction-result

Clean the Output directory in HDFS

make clean-output

Clean the Input directory in HDFS

make clean-input

About

Using All Big Data Technologies In Order To Apply ALS Algorithm To Recommend Amazon Prodcuts


Languages

Language:Scala 55.3%Language:Python 27.8%Language:Makefile 14.3%Language:Shell 2.6%