Senad Kurtiši's starred repositories
applied-ml
📚 Papers & tech blogs by companies sharing their work on data science & machine learning in production.
ML-Notebooks
:fire: Machine Learning Notebooks
flamingo-pytorch
Implementation of 🦩 Flamingo, state-of-the-art few-shot visual question answering attention net out of Deepmind, in Pytorch
spectral-estimation
Various algorithms for spectral estimation. Based on the book "Modern Spectral Estimation - Theory & Application", Steven M. Kay.
Speech-Emotion-Classification-with-PyTorch
This repository contains PyTorch implementation of 4 different models for classification of emotions of the speech.
get-started-with-JAX
The purpose of this repo is to make it easy to get started with JAX, Flax, and Haiku. It contains my "Machine Learning with JAX" series of tutorials (YouTube videos and Jupyter Notebooks) as well as the content I found useful while learning about the JAX ecosystem.
ML-For-Beginners
12 weeks, 26 lessons, 52 quizzes, classic Machine Learning for all
pytorch-learn-reinforcement-learning
A collection of various RL algorithms like policy gradients, DQN and PPO. The goal of this repo will be to make it a go-to resource for learning about RL. How to visualize, debug and solve RL problems. I've additionally included playground.py for learning more about OpenAI gym, etc.
Twitter-Sentiment-Analysis-RoBERTa
Sentiment Analysis of tweets written in underused Slavic languages (Serbian, Bosnian and Croatian) using pretrained multilingual RoBERTa based model XLM-R on 2 different datasets.
Long-texts-Sentiment-Analysis-RoBERTa
PyTorch implementation of Sentiment Analysis of the long texts written in Serbian language (which is underused language) using pretrained Multilingual RoBERTa based model (XLM-R) on the small dataset.
pytorch-GAT
My implementation of the original GAT paper (Veličković et al.). I've additionally included the playground.py file for visualizing the Cora dataset, GAT embeddings, an attention mechanism, and entropy histograms. I've supported both Cora (transductive) and PPI (inductive) examples!
transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
PlotNeuralNet
Latex code for making neural networks diagrams
first-order-model
This repository contains the source code for the paper First Order Motion Model for Image Animation
Real-Time-Voice-Cloning
Clone a voice in 5 seconds to generate arbitrary speech in real-time
pytorch-neural-style-transfer
Reconstruction of the original paper on neural style transfer (Gatys et al.). I've additionally included reconstruction scripts which allow you to reconstruct only the content or the style of the image - for better understanding of how NST works.
pytorch-deepdream
PyTorch implementation of DeepDream algorithm (Mordvintsev et al.). Additionally I've included playground.py to help you better understand basic concepts behind the algo.
pytorch-neural-style-transfer-johnson
Reconstruction of the fast neural style transfer (Johnson et al.). Some portions of the paper have been improved by the follow-up work like the instance normalization, etc. Checkout transformer_net.py's header for details.
pytorch-naive-video-neural-style-transfer
Create naive (no temporal loss) NST for videos with person segmentation. Just place your videos in data/, run and you get your stylized and segmented videos.
pytorch-GANs
My implementation of various GAN (generative adversarial networks) architectures like vanilla GAN (Goodfellow et al.), cGAN (Mirza et al.), DCGAN (Radford et al.), etc.
pytorch-original-transformer
My implementation of the original transformer model (Vaswani et al.). I've additionally included the playground.py file for visualizing otherwise seemingly hard concepts. Currently included IWSLT pretrained models.
WGAN-GP-for-Face-Generation
Keras implementation of WGAN GP for face generation. The model is trained on CelebA dataset.
Variational-Autoencoder-for-Face-Generation
Keras implementation of Variation Autoencoder for face generation. Analysis of the distribution of the latent space of the VAE. Vector arithemtic in the latent space. Morphing between the faces. The model was trained on CelebA dataset