There are 2 repositories under ridge-regression topic.
Robyn is an experimental, AI/ML-powered and open sourced Marketing Mix Modeling (MMM) package from Meta Marketing Science. Our mission is to democratise modeling knowledge, inspire the industry through innovation, reduce human bias in the modeling process & build a strong open source marketing science community.
A Julia package for multivariate statistics and data analysis (e.g. dimension reduction)
👩🏻⚕️Covid-19 estimation and forecast using statistical model; 新型冠状病毒肺炎统计模型预测 (Jan 2020)
Python machine learning applications in image processing, recommender system, matrix completion, netflix problem and algorithm implementations including Co-clustering, Funk SVD, SVD++, Non-negative Matrix Factorization, Koren Neighborhood Model, Koren Integrated Model, Dawid-Skene, Platt-Burges, Expectation Maximization, Factor Analysis, ISTA, FISTA, ADMM, Gaussian Mixture Model, OPTICS, DBSCAN, Random Forest, Decision Tree, Support Vector Machine, Independent Component Analysis, Latent Semantic Indexing, Principal Component Analysis, Singular Value Decomposition, K Nearest Neighbors, K Means, Naïve Bayes Mixture Model, Gaussian Discriminant Analysis, Newton Method, Coordinate Descent, Gradient Descent, Elastic Net Regression, Ridge Regression, Lasso Regression, Least Squares, Logistic Regression, Linear Regression
Predicting Amsterdam house / real estate prices using Ordinary Least Squares-, XGBoost-, KNN-, Lasso-, Ridge-, Polynomial-, Random Forest-, and Neural Network MLP Regression (via scikit-learn)
Objective of the repository is to learn and build machine learning models using Pytorch. 30DaysofML Using Pytorch
Built house price prediction model using linear regression and k nearest neighbors and used machine learning techniques like ridge, lasso, and gradient descent for optimization in Python
Generalized Linear Regressions Models (penalized regressions, robust regressions, ...)
A MATLAB package for modelling multivariate stimulus-response data
By-hand code for models and algorithms. An update to the 'Miscellaneous-R-Code' repo.
次元期权应征面试题范例。
Analyzes weightlifting videos for correct posture using pose estimation with OpenCV
A Julia package for robust regressions using M-estimators and quantile regressions
Automated Essay Scoring on The Hewlett Foundation dataset on Kaggle
Double Descent Curve with Optical Random Features
Repository containing introduction to the main methods and models used in machine learning problems of regression, classification and clustering.
Introduction The context is the 2016 public use NH medical claims files obtained from NH CHIS (Comprehensive Health Care Information System). The dataset contains Commercial Insurance claims, and a small fraction of Medicaid and Medicare payments for dually eligible people. The primary purpose of this assignment is to test machine learning (ML) skills in a real case analysis setting. You are expected to clean and process data and then apply various ML techniques like Linear and no linear models like regularized regression, MARS, and Partitioning methods. You are expected to use at least two of R, Python and JMP software. Data details: Medical claims file for 2016 contains ~17 millions rows and ~60 columns of data, containing ~6.5 million individual medical claims. These claims are all commercial claims that were filed by healthcare providers in 2016 in the state of NH. These claims were ~88% for residents of NH and the remaining for out of state visitors who sought care in NH. Each claim consists of one or more line items, each indicating a procedure done during the doctor’s visit. Two columns indicating Billed amount and the Paid amount for the care provided, are of primary interest. The main objective is to predict “Paid amount per procedure” by mapping a plethora of features available in the dataset. It is also an expectation that you would create new features using the existing ones or external data sources. Objectives: Step 1: Take a random sample of 1 million unique claims, such that all line items related to each claim are included in the sample. This will result in a little less than 3 million rows of data. Step 2: Clean up the data, understand the distributions, and create new features if necessary. Step 3: Run predictive models using validation method of your choice. Step 4: Write a descriptive report (less than 10 pages) describing the process and your findings.
Movie Recommendation System using the MovieLens dataset
Python notebooks for my graduate class on Detection, Estimation, and Learning. Intended for in-class demonstration. Notebooks illustrate a variety of concepts, from hypothesis testing to estimation to image denoising to Kalman filtering. Feel free to use or modify for your instruction or self-study.
Understand the relationships between various features in relation with the sale price of a house using exploratory data analysis and statistical analysis. Applied ML algorithms such as Multiple Linear Regression, Ridge Regression and Lasso Regression in combination with cross validation. Performed parameter tuning, compared the test scores and suggested a best model to predict the final sale price of a house. Seaborn is used to plot graphs and scikit learn package is used for statistical analysis.
Accurate estimation and robust modelling of translation dynamics at codon resolution
CRAN R Package: Ridge Regression with automatic selection of the penalty parameter
MDL Complexity computations and experiments from the paper "Revisiting complexity and the bias-variance tradeoff".
Fast computation of some matrices useful in statistics
Self-Distillation with weighted ground-truth targets; ResNet and Kernel Ridge Regression
Predicting Baseball Statistics: Classification and Regression Applications in Python Using scikit-learn
Explanations and Python implementations of Ordinary Least Squares regression, Ridge regression, Lasso regression (solved via Coordinate Descent), and Elastic Net regression (also solved via Coordinate Descent) applied to assess wine quality given numerous numerical features. Additional data analysis and visualization in Python is included.
Logistic Regression technique in machine learning both theory and code in Python. Includes topics from Assumptions, Multi Class Classifications, Regularization (l1 and l2), Weight of Evidence and Information Value
Scrape Bangkok condominium listing from hipflat.com and compare ML performance
A Julia package to perform Tikhonov regularization for small to moderate size problems.
Evaluate various supervised learning methods to predict cohesive energies of solids (kernel ridge regression is the best)