Kang (kang-gnak)

kang-gnak

Geek Repo

Github PK Tool:Github PK Tool

Kang's starred repositories

AesBench

An expert benchmark aiming to comprehensively evaluate the aesthetic perception capacities of MLLMs.

Language:PythonLicense:Apache-2.0Stargazers:183Issues:0Issues:0

Awesome-Image-Quality-Assessment

A comprehensive collection of IQA papers

Language:TeXLicense:MITStargazers:831Issues:0Issues:0

stable-diffusion-aesthetic-gradients

Personalization for Stable Diffusion via Aesthetic Gradients 🎨

Language:Jupyter NotebookLicense:NOASSERTIONStargazers:717Issues:0Issues:0
Language:PythonStargazers:52Issues:0Issues:0

IQA-PyTorch

👁️ 🖼️ 🔥PyTorch Toolbox for Image Quality Assessment, including LPIPS, FID, NIQE, NRQM(Ma), MUSIQ, TOPIQ, NIMA, DBCNN, BRISQUE, PI and more...

Language:PythonLicense:NOASSERTIONStargazers:1660Issues:0Issues:0
Language:PythonStargazers:39Issues:0Issues:0

NSFC-application-template-latex

国家自然科学基金申请书正文(面上项目)LaTeX 模板(非官方)

Language:TeXLicense:MITStargazers:803Issues:0Issues:0

Aestheval

Code for the paper "Understanding Aesthetics with Language: A Photo Critique Dataset for Aesthetic Assessment"

Language:Jupyter NotebookLicense:CC-BY-4.0Stargazers:80Issues:0Issues:0
Language:PythonStargazers:10Issues:0Issues:0

tvsum

TVSum: Title-based Video Summarization dataset (CVPR 2015)

Language:MATLABStargazers:117Issues:0Issues:0

Video-Summarization-Using-Attention-Based-Encoder-Decoder-Model

In this model we uses Attention Based model to summarize the video. The output of the video will be 15% of the original video as summary. Also It's a Supervised Model.

Language:Jupyter NotebookStargazers:2Issues:0Issues:0

CenseoQoE

image and video quality assessment

Language:PythonLicense:NOASSERTIONStargazers:201Issues:0Issues:0

Awesome-Image-Composition

A curated list of papers, code and resources pertaining to image composition/compositing or object insertion, which aims to generate realistic composite image.

Stargazers:1126Issues:0Issues:0

MUAD-Dataset

MUAD: Multiple Uncertainties for Autonomous Driving, a benchmark for multiple uncertainty types and tasks [BMVC 2022]

Language:PythonLicense:NOASSERTIONStargazers:27Issues:0Issues:0

DPC-Captions

A image caption dataset about images from www.dpchallenge.com.

Stargazers:11Issues:0Issues:0

MVVA-Net

MVVA-Net: A Video Aesthetic Quality Assessment Network with Cognitive Fusion of Multi-type Feature Based Strong Generalization

Stargazers:1Issues:0Issues:0

TANet-image-aesthetics-and-quality-assessment

[IJCAI 2022, Official Code] for paper "Rethinking Image Aesthetics Assessment: Models, Datasets and Benchmarks". Official Weights and Demos provided. 首个面向多主题场景的美学评估数据集、算法和benchmark.

Language:PythonLicense:Apache-2.0Stargazers:232Issues:0Issues:0

segmentation-paper-reading-notes

segmentation paper reading notes

Stargazers:53Issues:0Issues:0

troubleshooting_BIQA

This is the official released code for troubleshooting BIQA in the wild

Stargazers:11Issues:0Issues:0

A01_cvclass_basic

resources for CV

License:MITStargazers:229Issues:0Issues:0
Language:PythonStargazers:5Issues:0Issues:0

MPADA

The official implementation of MP_{ada} in Attention-based Multi-patch Aggregation for Image Aesthetic Assessment (MM 2018)

Language:PythonStargazers:79Issues:0Issues:0

pytorch-image-models

The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (ViT), MobileNetV4, MobileNet-V3 & V2, RegNet, DPN, CSPNet, Swin Transformer, MaxViT, CoAtNet, ConvNeXt, and more

Language:PythonLicense:Apache-2.0Stargazers:30835Issues:0Issues:0

transferlearning

Transfer learning / domain adaptation / domain generalization / multi-task learning etc. Papers, codes, datasets, applications, tutorials.-迁移学习

Language:PythonLicense:MITStargazers:13108Issues:0Issues:0

MAE-pytorch

Unofficial PyTorch implementation of Masked Autoencoders Are Scalable Vision Learners

Language:PythonStargazers:2565Issues:0Issues:0

vit-pytorch

Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch

Language:PythonLicense:MITStargazers:18939Issues:0Issues:0

MAE-pytorch

Masked Autoencoders Are Scalable Vision Learners

Language:PythonStargazers:245Issues:0Issues:0

unilm

Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities

Language:PythonLicense:MITStargazers:19210Issues:0Issues:0
Language:Jupyter NotebookLicense:Apache-2.0Stargazers:9784Issues:0Issues:0