Skip to content

A new multi-task learning framework using Vision Transformers

Notifications You must be signed in to change notification settings

hananshafi/MTL-ViT

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

18 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

MTL-ViT: A new multi-task learning framework using Vision Transformers [IEEE ICIP 2024]

(*Note: This is an ongoing project, hence the full code and strategy is not yet open-sourced by the author.)

We presnet a new multi-task learning strategy using Vision transformers (ViTs). Our approach is based on exploiting the class-token and self-attention mechanism of Vision Transformers in order to train multiple tasks through a single ViT, more efficiently and with limited computational budget.

alt text

Total Loss of the Multi-task system:

About

A new multi-task learning framework using Vision Transformers

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages