There are 1 repository under gradient-boosting-classifier topic.
A curated list of gradient boosting research papers with implementations.
Projects I completed as a part of Great Learning's PGP - Artificial Intelligence and Machine Learning
This project detects whether a news is fake or not using machine learning.
Ruby Scoring API for PMML
Using various machine learning models to predict whether a company will go bankrupt
Demo on the capability of Yandex CatBoost gradient boosting classifier on a fictitious IBM HR dataset obtained from Kaggle. Data exploration, cleaning, preprocessing and model tuning are performed on the dataset
Natural Language Processing for Multiclass Classification: A repository containing NLP techniques for multiclass classification of text data.
In this challenge we have given a directed social graph, and we have to predict missing links to recommend users (Link Prediction in graph)
This contains the Jupyter Notebook and the Dataset for the mentioned Classification Predictive Modeling Project
A package to build Gradient boosted trees for ordinal labels
Pose estimation and prediction using Mediapipe and various ML models
Recognition of Persomnality Types from Facebook status using Machine Learning
This research goal is to build binary classifier model which are able to separate fraud transactions from non-fraud transactions.
FederBoost's Federated Gradient Boosting Decision Tree Algorithm, Federated enabled Membership Inference
This project is our submission for the Kavach Hackathon 2023, in which we have created a browser extension that detects the links present in the email and classifies whether they are safe or not.
Data Science Challenge
Analyze NASDAQ100 stock data. Used ARIMA + GARCH model and machine learning techniques Naive Bayes and Decision tree to determine if we go long or short for a given stock on a particular day
RocAuc Pairiwse objective for gradient boosting
Kaggle Machine Learning Competition Project : To classify activities into one of the six activities performed by individuals by reading the inertial sensors data collected using Smartphone.
Open source gradient boosting library
The aim is to find an optimal ML model (Decision Tree, Random Forest, Bagging or Boosting Classifiers with Hyper-parameter Tuning) to predict visa statuses for work visa applicants to US. This will help decrease the time spent processing applications (currently increasing at a rate of >9% annually) while formulating suitable profile of candidates more likely to have the visa certified.
This is a customer churn prediction project using machine learning algorithms like Logistic Regression, Random Forest, K-Nearest Neighbors, Support Vector Machine, XGBoost, and Gradient Boosting. The project aims to analyze and predict customer churn in a dataset, using techniques like class weighting and SMOTE to handle class imbalance
This repository aims to address the critical issue of identifying and understanding suicide ideation in social media conversations, specifically focusing on Twitter data.
Telecom Customer Churn Analysis & Prediction project uses Gradient Boosting for precise predictions, Power BI for churn pattern visualizations, and Streamlit for interactive insights. With robust code and meticulous data preprocessing, stakeholders access accurate predictions to optimize retention and drive profitability.
This repository hosts a Jupyter notebook for Fake News Detection, utilizing machine learning algorithms like Logistic Regression , Gradient Boosting Classifier , Random Forest and Decision Tree. The project covers data preprocessing, analysis and manual testing of news articles, with added multi language support using Google Translate API .
A way to predict an NBA's players chance of making a shot using machine learning
Unified interface for Gradient Boosted Decision Trees
Classification
For this project, I used four different classification algorithms to detect fraudulent credit card transactions.
Predicting short-term IPO returns. Course work for CSCI-B351 at Indiana University.
Demonstrate the application of machine learning on a real-world predictive maintenance dataset, using measurements from actual industrial equipment.
I contributed to a group project using the Life Expectancy (WHO) dataset from Kaggle where I performed regression analysis to predict life expectancy and classification to classify countries as developed or developing. The project was completed in Python using the pandas, Matplotlib, NumPy, seaborn, scikit-learn, and statsmodels libraries. The regression models were fitted on the entire dataset, along with subsets for developed and developing countries. I tested ordinary least squares, lasso, ridge, and random forest regression models. Random forest regression performed the best on all three datasets and did not overfit the training set. The testing set R2 was .96 for the entire dataset and developing country subset. The developed country subset achieved an R2 of .8. I tested seven different classification algorithms to classify a country as developing or developed. The models obtained testing set balanced accuracies ranging from 86% - 99%. From best to worst, the models included gradient boosting, random forest, Adaptive Boosting (AdaBoost), decision tree, k-nearest neighbors, support-vector machines, and naive Bayes. I tuned all the models' hyperparameters. None of the models overfitted the training set.
MSBoost is a gradient boosting algorithm that improves performance by selecting the best model from multiple parallel-trained models for each layer, excelling in small and noisy datasets.
Utilizing machine learning models including logistic regression, random forest, gradient boosting, and neural networks to identify fraudulent credit card transactions. Dataset, consisting of PCA-transformed features and unbalanced classes, required precision-recall metrics for accurate evaluation. Developed in Python using TensorFlow and scikit.
Classification problem using multiple ML Algorithms