There are 0 repository under l1-regularization topic.
Learning Efficient Convolutional Networks through Network Slimming, In ICCV 2017.
Functional models and algorithms for sparse signal processing
L1-regularized least squares with PyTorch
A Deep Learning framework for CNNs and LSTMs from scratch, using NumPy.
Logistic Regression technique in machine learning both theory and code in Python. Includes topics from Assumptions, Multi Class Classifications, Regularization (l1 and l2), Weight of Evidence and Information Value
The given information of network connection, model predicts if connection has some intrusion or not. Binary classification for good and bad type of the connection further converting to multi-class classification and most prominent is feature importance analysis.
Overparameterization and overfitting are common concerns when designing and training deep neural networks. Network pruning is an effective strategy used to reduce or limit the network complexity, but often suffers from time and computational intensive procedures to identify the most important connections and best performing hyperparameters. We suggest a pruning strategy which is completely integrated in the training process and which requires only marginal extra computational cost. The method relies on unstructured weight pruning which is re-interpreted in a multiobjective learning approach. A batchwise Pruning strategy is selected to be compared using different optimization methods, of which one is a multiobjective optimization algorithm. As it takes over the choice of the weighting of the objective functions, it has a great advantage in terms of reducing the time consuming hyperparameter search each neural network training suffers from. Without any a priori training, post training, or parameter fine tuning we achieve highly reductions of the dense layers of two commonly used convolution neural networks (CNNs) resulting in only a marginal loss of performance. Our results empirically demonstrate that dense layers are overparameterized as with reducing up to 98 % of its edges they provide almost the same results. We contradict the theory that retraining after pruning neural networks is of great importance and opens new insights into the usage of multiobjective optimization techniques in machine learning algorithms in a Keras framework. The Stochastic Multi Gradient Descent Algorithm implementation in Python3 is for usage with Keras and adopted from paper of S. Liu and L. N. Vicente: "The stochastic multi-gradient algorithm for multi-objective optimization and its application to supervised machine learning". It is combined with weight pruning strategies to reduce network complexity and inference time.
An Image Reconstructor that applies fast proximal gradient method (FISTA) to the wavelet transform of an image using L1 and Total Variation (TV) regularizations
High Dimensional Portfolio Selection with Cardinality Constraints
MNIST Digit Prediction using Batch Normalization, Group Normalization, Layer Normalization and L1-L2 Regularizations
MITx - MicroMasters Program on Statistics and Data Science - Data Analysis: Statistical Modeling and Computation in Applications - Second Project
Implementation of optimization and regularization algorithms in deep neural networks from scratch
Comparing Three Penalized Least Squares Estimators: LASSO,SCAD and MCP.
regression algorithm implementaion from scratch with python (least-squares, regularized LS, L1-regularized LS, robust regression)
During this study we will explore the different regularisation methods that can be used to address the problem of overfitting in a given Neural Network architecture, using the balanced EMNIST dataset.
Forecasting for AirQuality UCI dataset with Conjugate Gradient Artificial Neural Network based on Feature Selection L1 Regularized and Genetic Algorithm for Parameter Optimization
Mathematical machine learning algorithm implementations
I executed this assignment for a US-based housing company named Surprise Housing, wherein a regression model with regularisation was used to predict the actual value of the prospective properties and decide whether to invest in them or not
House Price Analysis and Sales Price Prediction
Chapman University CS-510 Computing For Scientists Final Project
How much is the NBA dollar worth in terms of team success?
Minimum working example for using the Sorted L1 Norm in a regression and mean-variance framework. The codes are free to use for research purposes only with the propper citation. Commercial use is strictly forbidden and the rights remain with the authors. For citing purposes please refer to the JBF version: https://www.sciencedirect.com/science/article/abs/pii/S0378426619302614
Classification Using Logistic Regression by Making a Neural Network Model. This project also includes comparison of Model performance when different regularization techniques are used
Multiclass Logistic, Classification Pipeline, Cross Validation, Gradient Descent, Regularization
This is the accompanying code repository for the ICLR 2023 publication "Almost Linear Constant-Factor Sketching for 𝓁₁ and Logistic Regression" by Alexander Munteanu, Simon Omlor and David P. Woodruff.
The project encompasses the statistical analysis of a high-dimensional data using different classification, feature selection, clustering and dimension reduction techniques.
This repository contains the code for the blog post on Understanding L1 and L2 regularization in machine learning. For further details, please refer to this post.
A study of the problem of overfitting in deep neural networks, how it can be detected, and prevented using the EMNIST dataset. This was done by performing experiments with depth and width, dropout, L1 & L2 regularization, and Maxout networks.
Machine Learning Practical - Coursework 1 Report: a study of the problem of overfitting in deep neural networks, how it can be detected, and prevented using the EMNIST dataset. This was done by performing experiments with depth and width, dropout, L1 & L2 regularization, and Maxout networks.
This is a mid-term project of Optimization Methods, a course of Institute of Data Science, National Cheng Kung University. This project aimed to construct the linear regression with L1 regularization and the logistic regression with L1 regularization.
Used a Multilayer Perceptron (MLP) neural network to detect COVID-19 in lung scans.
Implementation of Logistic Regression with L1 Regularization from scratch
Multiclass Logistic, Classification Pipeline, Cross Validation, Gradient Descent, Regularization
Time Series Classification Part 2 Binary and Multiclass Classification. An interesting task in machine learning is classification of time series. In this problem, we will classify the activities of humans based on time series obtained by a Wireless Sensor Network.