collections of excellent feature extractor
Download pre-trained weight
Paper | Link |
---|---|
CrossFormer: A Versatile Vision Transformer Hinging on Cross-scale Attention | 1 |
BiFormer: Vision Transformer with Bi-Level Routing Attention | 2 |
HorNet: Efficient High-Order Spatial Interactions with Recursive Gated Convolutions | 3 |
FcaNet: Frequency Channel Attention Networks | 4 |
DilateFormer: Multi-Scale Dilated Transformer for Visual Recognition | 5 |
@Article{zhu2023biformer,
author = {Lei Zhu and Xinjiang Wang and Zhanghan Ke and Wayne Zhang and Rynson Lau},
title = {BiFormer: Vision Transformer with Bi-Level Routing Attention},
journal = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
year = {2023},
}
@inproceedings{wang2021crossformer,
title = {CrossFormer: A Versatile Vision Transformer Hinging on Cross-scale Attention},
author = {Wenxiao Wang and Lu Yao and Long Chen and Binbin Lin and Deng Cai and Xiaofei He and Wei Liu},
booktitle = {International Conference on Learning Representations, {ICLR}},
url = {https://openreview.net/forum?id=_PHymLIxuI},
year = {2022}
}
@article{wang2023crossformer++,
author = {Wenxiao Wang and Wei Chen and Qibo Qiu and Long Chen and Boxi Wu and Binbin Lin and Xiaofei He and Wei Liu},
title = {Crossformer++: A versatile vision transformer hinging on cross-scale attention},
journal = {{IEEE} Transactions on Pattern Analysis and Machine Intelligence, {TPAMI}},
year = {2023},
doi = {10.1109/TPAMI.2023.3341806},
}
@article{rao2022hornet,
title={HorNet: Efficient High-Order Spatial Interactions with Recursive Gated Convolutions},
author={Rao, Yongming and Zhao, Wenliang and Tang, Yansong and Zhou, Jie and Lim, Ser-Lam and Lu, Jiwen},
journal={Advances in Neural Information Processing Systems (NeurIPS)},
year={2022}
}
@article{jiao2023dilateformer,
title = {DilateFormer: Multi-Scale Dilated Transformer for Visual Recognition},
author = {Jiao, Jiayu and Tang, Yu-Ming and Lin, Kun-Yu and Gao, Yipeng and Ma, Jinhua and Wang, Yaowei and Zheng, Wei-Shi},
journal = {{IEEE} Transaction on Multimedia},
year = {2023}
}