There are 0 repository under elastic-net-regression topic.
Logistic Regression technique in machine learning both theory and code in Python. Includes topics from Assumptions, Multi Class Classifications, Regularization (l1 and l2), Weight of Evidence and Information Value
Explanations and Python implementations of Ordinary Least Squares regression, Ridge regression, Lasso regression (solved via Coordinate Descent), and Elastic Net regression (also solved via Coordinate Descent) applied to assess wine quality given numerous numerical features. Additional data analysis and visualization in Python is included.
To know internal working of machine learning algorithms, I have implemented types of regression through scratch.
Deep reinforcement learning for smart calibration of radio telescopes. Automatic hyper-parameter tuning.
R code used for the analyses of the paper: Spatial conservation prioritisation in data-poor countries: a quantitative sensitivity analysis using different taxa
Algorithmes d’apprentissage et modèles statistiques: Un exemple de régression logistique régularisée et de validation croisée pour prédire le décrochage scolaire
I created multiple models to predict the discharge volume of a 100 year flood on rivers in NY state. The discharge of 100 year flood events is dependent upon watershed drainage area, and elevation among other variables.
My role in this group project was to perform regression analysis on quarterly financial data to predict a company's market capitalization. I used R to develop ordinary least squares (OLS), stepwise, ridge, lasso, relaxed lasso, and elastic net regression models. I first used stepwise and OLS regression to develop a model and examine its residual plots. The plot displaying the residuals against the predicted values indicated multiplicative errors. I, therefore, took the natural log transformation of the dependent variable. The resulting model's R2 was significantly, negatively impacted. After examining scatter plots between the log transformation of market capitalization and the independent variables, I discovered the independent variables also had to be transformed to produce a linear relationship. Using the log transformation of both the dependent and independent variables, I developed models using all the regression techniques mentioned to strike a balance between R2 and producing a parsimonious model. All the models produced similar results, with an R2 of around .80. Since OLS is easiest to explain, had similar residual plots, and the highest R2 of all the models, it was the best model developed.
A demonstration of the basic Machine Learning Algorithms
Regression on BOSTON dataset from sklearn
The project aims to enhance aircraft engine maintenance operations and planning using statistical learning and machine learning methods.
Data Models in R for Multiple Linear Regression and three models (Ridge, Lasso, and Elastic-Net), to predict Medicare claim costs of Type 2 diabetes patients with other diagnoses. We used Data from Entrepreneur’s Medicare Claims Synthetic Public Use Files (DE-SynPUFs) for our analysis.
ML Project implementing ANN, SVM, Random Forest, Elastic Net regression models from scratch.
This project focuses on forecasting the closing prices of Yes Bank's stock. Through data analysis and predictive modeling, this project provides valuable insights for investors and traders, aiding them in making informed decisions about their investments in Yes Bank's stock.
High Throughput Light Weight Regularized Regression Modeling for Molecular Data
This project focuses on forecasting the closing prices of Yes Bank's stock. Through data analysis and predictive modeling, this project provides valuable insights for investors and traders, aiding them in making informed decisions about their investments in Yes Bank's stock.
A project aim to predict default rate of Commercial Real Estate(CRE) Loans
Machine Learning Predictions of 2023 Formula One Constructors' Championship Standings
Regression analysis
Various Regression models including linear, polynomial, ridge, lasso and elastic net were experimented with to find which model best predicted health insurance costs. The models were evaluated using cross-validation, from which the best models were optimized using randomized search. The best model was then evaluated on the test data.
Prediction of Sales Prices of Houses
Machine learning (regression) exercise on prediction of house pricing in Melbourne with post-model analysis and recommendations for maximizing home value.
The project encompasses the statistical analysis of a high-dimensional data using different classification, feature selection, clustering and dimension reduction techniques.
Machine Learning Basics
Ridge, elastic net, and logistic regressions implemented without using any statistical or machine learning library. All steps are done by hand, using matrix operations as much as possible.
Lasso + Bootstrap methods for predictive modeling
This project compares the different machine learning models on Walmart Weekly Sales Data and predicts the weekly sales for the test data.
Built a regression model to predict university admission using linear, polynomial, and regularized regression techniques (lasso, ridge, and elastic net) and achieved 98% accuracy.
Yes-Bank-Stock-Closing-Price-Prediction refers to a type of project or task in the field of data science and machine learning that involves developing predictive models to estimate the Closing Price of stock