mnguyen0226 / multitask_learning_vit

Multitask-Learning (hard-parameter sharing) with Vision Transformers on Cifar10 & Cifar100

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Multi-task Learning with Vision Transformers

Image classification is a fundamental task in computer vision, with many practical applications. However, achieving high accuracy on complex datasets like Cifar-10 and Cifar-100 can be challenging, particularly when working with limited computational resources. In this study, we evaluate the performance of Multi-Task Learning (MTL) on two common image classification tasks: class and super-class classification. Specifically, we investigate the efficiency of MTL using the Vision Transformers (ViT) architecture on these datasets, shedding light on its comparative performance against the Single-Task Learning (STL) approach and traditional convolution-based models. Our results demonstrate that MTL using the ViT architecture outperforms the STL approach on both classification tasks, achieving higher accuracy with fewer parameters. Additionally, we find that ViT and ResNet-152 perform similarly on these tasks, highlighting the potential of ViT for MTL scenarios. These findings have important implications for the development of efficient and effective image classification models, particularly in scenarios where multiple classification tasks need to be performed simultaneously.

Results

ViT: Single-Task Learning on CIFAR 10 with 20 epochs

ViT: Single-Task Learning on CIFAR 10 with 20 epochs (superclass)

ViT: Single-Task Learning on CIFAR 100 with 20 epochs

ViT: Single-Task Learning on CIFAR 100 with 20 epochs (superclass)

ViT: Multi-Task Learning on CIFAR 10 with 20 epochs

ViT: Multi-Task Learning on CIFAR 100 with 20 epochs

ResNet-152: Single-Task Learning on CIFAR 10 with 20 epochs

ResNet-152: Single-Task Learning on CIFAR 10 with 20 epochs (superclass)

ResNet-152: Single-Task Learning on CIFAR 100 with 20 epochs

ResNet-152: Single-Task Learning on CIFAR 100 with 20 epochs (superclass)

ResNet-152: Multi-Task Learning on CIFAR 10 with 20 epochs

ResNet-152: Multi-Task Learning on CIFAR 100 with 20 epochs

Report

Experimental Result Summary.

Report.

About

Multitask-Learning (hard-parameter sharing) with Vision Transformers on Cifar10 & Cifar100

License:MIT License


Languages

Language:Jupyter Notebook 100.0%