There are 0 repository under ordinary-least-squares topic.
Multiple econometrics cheat sheets with a complete and summarize review going from the basics of an econometric model to the solution of the most popular problems.
Algorithmic Trading project that examines the Fama-French 3-Factor Model and the Fama-French 5-Factor Model in predicting portfolio returns. The respective factors are used as features in a Machine Learning model and portfolio results are evaluated and compared.
Explanations and Python implementations of Ordinary Least Squares regression, Ridge regression, Lasso regression (solved via Coordinate Descent), and Elastic Net regression (also solved via Coordinate Descent) applied to assess wine quality given numerous numerical features. Additional data analysis and visualization in Python is included.
An R implementation of Models As Approximations
Linear Regression for Julia
Set of functions to semi-automatically build and test Ordinary Least Squares (OLS) models in R in parallel.
ML++ and cppyml: efficient implementations of selected ML algorithms, with Python bindings.
Predictive Analysis of Price on Amsterdam Airbnb Listings Using Ordinary Least Squares.
Linear line fitting to data and optimising parameters with Gradient Descent algorithm
A Regression Exercise covering OLS & Ridge Regression
Predicting housing prices in Iowa using Python/Pandas/linear regression within SKLearn.
MITx - MicroMasters Program on Statistics and Data Science - Data Analysis: Statistical Modeling and Computation in Applications - First Project
My role in this group project was to perform regression analysis on quarterly financial data to predict a company's market capitalization. I used R to develop ordinary least squares (OLS), stepwise, ridge, lasso, relaxed lasso, and elastic net regression models. I first used stepwise and OLS regression to develop a model and examine its residual plots. The plot displaying the residuals against the predicted values indicated multiplicative errors. I, therefore, took the natural log transformation of the dependent variable. The resulting model's R2 was significantly, negatively impacted. After examining scatter plots between the log transformation of market capitalization and the independent variables, I discovered the independent variables also had to be transformed to produce a linear relationship. Using the log transformation of both the dependent and independent variables, I developed models using all the regression techniques mentioned to strike a balance between R2 and producing a parsimonious model. All the models produced similar results, with an R2 of around .80. Since OLS is easiest to explain, had similar residual plots, and the highest R2 of all the models, it was the best model developed.
I contributed to a group project using the Life Expectancy (WHO) dataset from Kaggle where I performed regression analysis to predict life expectancy and classification to classify countries as developed or developing. The project was completed in Python using the pandas, Matplotlib, NumPy, seaborn, scikit-learn, and statsmodels libraries. The regression models were fitted on the entire dataset, along with subsets for developed and developing countries. I tested ordinary least squares, lasso, ridge, and random forest regression models. Random forest regression performed the best on all three datasets and did not overfit the training set. The testing set R2 was .96 for the entire dataset and developing country subset. The developed country subset achieved an R2 of .8. I tested seven different classification algorithms to classify a country as developing or developed. The models obtained testing set balanced accuracies ranging from 86% - 99%. From best to worst, the models included gradient boosting, random forest, Adaptive Boosting (AdaBoost), decision tree, k-nearest neighbors, support-vector machines, and naive Bayes. I tuned all the models' hyperparameters. None of the models overfitted the training set.
Artigo submetido ao COBRAC 2018.
In the following research, we will analyze the effects of pairs trading (multiple companies across multiple industries) excluding the profitability of such strategies. Rather, we will analyze various risk measures across all different pairings of stocks within their own respective industry across multiple industries.
An introduction into the world of machine learning with a comprehensive Udemy online course, designed for beginners, to learn Python programming fundamentals and gain valuable insights into the practical applications of machine learning.
regression algorithm implementaion from scratch with python (least-squares, regularized LS, L1-regularized LS, robust regression)
Fits JxV curves obtained from solar cells operating in the dark and calculates important parameters
In this project, I have worked with some data on possums. It is a relatively small data set, but it's a good size to try with ordinary least squares (OLS) and least absolute deviation (LAD), and to gain experience with supervised learning. I have written my own methods to fir both OLS and LAD models, and then at the end compared them to the models produced by the statsmodels package.
A project where data science job postings are scraped and an exploratory data analysis is performed.
Compared Ordinary Least Square (OLS) and Geographically Weighted Regression (GWR) using R programming with interpretation
Predicting Delivery Time Using Sorting Time
Building a prediction model for Salary hike using Years of Experience
The goal of the project was to predict the price based on the given attributes of the car. It was done in Python, using Machine Learning techniques like Simple Linear Regression, Multiple Linear Regression and Decision tree.
Tutorials for BSE classes.
Simple Linear Regression
Wrangled real estate data from multiple sources and file formats, brought it into a single consistent form and analysed the results.
(Geo)spatial Statistics with R (Meuse)
Trend Surface Analysis with R (Cape Flats Aquifer)
Ordinary Least Squares, Ridge Regression, Expectation Maximization, Full Bayesian Inference, Bayes Classifiers, kNN, and MLP core algorithms from scratch. Some auxiliary functions are also used.
Trabajos presentados como parte del curso de Reconocimiento de Patrones y Aprendizaje Automatizado, impartido por el profesor Sergio Hernández López durante el semestre 2023-2 en la Facultad de Ciencias, UNAM.
This repository contains a comprehensive implementation of gradient descent for linear regression, including visualizations and comparisons with ordinary least squares (OLS) regression. It also includes an additional implementation for multiple linear regression using gradient descent.
Probability and Statistics for Machine Learning
PySpark for multiple linear regression on car horsepower using SMOTE for data augmentation.
Machine Learning algorithms and models