Jesus Perez-Martin's repositories
video_captioning_datasets
Summary about Video-to-Text datasets. This repository is part of the review paper *Bridging Vision and Language from the Video-to-Text Perspective: A Comprehensive Review*
image-captioning
Python implementation for an Image Captioning System
deep-learning-tutorial
Tutorial for Deep Learning
httpx
A next generation HTTP client for Python. 🦋
ssh-key-action
GitHub Action that installs SSH key to .ssh
nuwa-pytorch
Implementation of NĂśWA, state of the art attention network for text to video synthesis, in Pytorch
mkdocs
Project documentation with Markdown.
annotationsite
Django project for memes annotation process
video_features_extractor
Python implementation of extraction of several visual features representations from videos
HerokuFiles
Free file storage options for Heroku hosted applications
trecvid-vtt
Video to Text Description (VTT) task of the TREC Video Retrieval Evaluation (TRECVID)
relela
Representation Learning + Natural Language Processing Research Group
video_description_eval
Python implementation for video description evaluation metrics
visual_syntactic_embedding_video_captioning
Source code of the paper titled *Improving Video Captioning with Temporal Composition of a Visual-Syntactic Embedding*
attentive_specialized_network_video_captioning
Source code of the paper titled *Attentive Visual Semantic Specialized Network for Video Captioning*
SemanticMemes
This project aims to determine the discourses associated with a meme and how these discourses change according to their local reality. For this, we are creating a large dataset of memes collected from Chile’s tweets. And we developed a site for manual classification of tweet’s images
emotion-intensity-classification
Python implementation for classifying tweets into emotion intensities
video_events_detector
Python implementation of an events detector on long videos
ActivityNet
This repository is intended to host tools and demos for ActivityNet
covid19_cuba
Dashboard with data and comparisons on COVID-19 infection in Cuba
delving-deeper-into-the-decoder-for-video-captioning
Source code for Delving Deeper into the Decoder for Video Captioning
dual_encoding
[CVPR2019] Dual Encoding for Zero-Example Video Retrieval
temporal-shift-module
[ICCV 2019] TSM: Temporal Shift Module for Efficient Video Understanding.
Controllable_XGating
ICCV2019: Controllable Video Captioning with POS Sequence Guidance Based on Gated Fusion Network
spanish-word-embeddings
Spanish word embeddings computed with different methods and from different corpora
ECO-pytorch
PyTorch implementation for "ECO: Efficient Convolutional Network for Online Video Understanding", ECCV 2018
Semantics-AssistedVideoCaptioning
Source code for Semantics-Assisted Video Captioning Model