link to maxvit in readme

This commit is contained in:
Phil Wang
2022-04-06 16:24:12 -07:00
parent c7bb5fc43f
commit 2d4089c88e

View File

@@ -601,7 +601,7 @@ preds = v(img) # (1, 1000)
<img src="./images/max-vit.png" width="400px"></img>
This paper proposes a hybrid convolutional / attention network, using MBConv from the convolution side, and then block / grid axial sparse attention.
<a href="https://arxiv.org/abs/2204.01697">This paper</a> proposes a hybrid convolutional / attention network, using MBConv from the convolution side, and then block / grid axial sparse attention.
They also claim this specific vision transformer is good for generative models (GANs).