There are 5 repositories under apache-hadoop topic.
MapReduce, Spark, Java, and Scala for Data Algorithms Book
Big Data Modeling, MapReduce, Spark, PySpark @ Santa Clara University
hadoop-cos(CosN文件系统)为Apache Hadoop、Spark以及Tez等大数据计算框架集成提供支持,可以像访问HDFS一样读写存储在腾讯云COS上的数据。同时也支持作为Druid等查询与分析引擎的Deep Storage
Export Hadoop YARN (resource-manager) metrics in prometheus format
Containerized Apache Hive Metastore for horizontally scalable Hive Metastore deployments
A Spark application to merge small files on Hadoop
An python implementation of Minimal Mapreduce Algorithms for Apache Spark
Some simple, kinda introductory projects based on Apache Hadoop to be used as guides in order to make the MapReduce model look less weird or boring.
This is projects of Cloud Computing Course
A fast, scalable and distributed community detection algorithm based on CEIL scoring function.
An email spam filter using Apache Spark’s ML library
This repository provides a guide to preprocess and analyze the network intrusion data set using NumPy, Pandas, and matplotlib, and implement a random forest classifier machine learning model using Scikit-learn.
Instructions for Installing Giraph-1.2.0
Kubernetes operator for managing the lifecycle of Apache Hadoop Yarn Tasks on Kubernetes.
A small code to validate the Census data on the basis of Aadhar Data
Simplified Hadoop Setup and Configuration Automation
logback appender for apache-flume
This repository aims to develop a basic search engine utilizing Hadoop's MapReduce framework to index and process extensive text corpora efficiently. The dataset used for this project is a subset of the English Wikipedia dump, totaling 5.2 GB in size. The project focuses on implementing a naive search algorithm to address challenges in information.
This project aims to establish a data streaming pipeline with storage, processing, and visualization
The goal of this project is to learn data processing using Spark with practical examples on datasets and also apply programming with Scala.
Preparação de um ambiente de desenvolvimento e testes para Apache Hadoop.
The source code developed and used for the purposes of my thesis with the same title under the guidance of my supervisor professor Vasilis Mamalis for the Department of Informatics and Computer Engineering of the University of West Attica.
Samples related to data engineering, e.g. spark, embulk, airflow, etc.
This project implemented a lambda architecture for analyzing domestic flight data in the US from 2009 to 2020. It used Apache Spark for batch processing, Spark Streaming for real-time analysis, and SVM models to predict flight cancellations and delays, with Docker for cluster management and Grafana for real-time visualization.
The implementation of Apache Spark (combine with PySpark, Jupyter Notebook) on top of Hadoop cluster using Docker
Built a Large Scale Distributed Data Processing system for Streaming Analytics using Hadoop Ecosystem (Apache Spark and HDFS), in Cloud for real-time spatial analytics.
My portfolio | under development
COVID-19 data analysis with MapReduce
Final Project for IBM Data Engineering & Python Professional Certificate -- Applied all skills and methods utilized in the series of courses for this certification
Data Science Project - for 'Advanced Topics in Database Systems' M.Sc. Course ECE @ntua
In this project we will use Hadoop MapReduce to implement a very basic “Sentiment Analysis” using the review text in the Yelp Academic Dataset as training data.