Phil Wang
da950e6d2c
add working PiT
0.12.0
2021-03-30 22:15:19 -07:00
Phil Wang
4b9a02d89c
use depthwise conv for CvT projections
0.11.1
2021-03-30 18:18:35 -07:00
Phil Wang
518924eac5
add CvT
0.11.0
2021-03-30 14:42:39 -07:00
Phil Wang
e712003dfb
add CrossViT
0.10.3
2021-03-30 00:53:27 -07:00
Phil Wang
d04ce06a30
make recorder work for t2t and deepvit
2021-03-29 18:16:34 -07:00
Phil Wang
8135d70e4e
use hooks to retrieve attention maps for user without modifying ViT
0.10.1
2021-03-29 15:10:12 -07:00
Phil Wang
3067155cea
add recorder class, for recording attention across layers, for researchers
2021-03-29 11:08:19 -07:00
Phil Wang
ab7315cca1
cleanup
2021-03-27 22:14:16 -07:00
Phil Wang
15294c304e
remove masking, as it complicates with little benefit
2021-03-23 12:18:47 -07:00
Phil Wang
b900850144
add deep vit
0.9.1
2021-03-23 11:57:13 -07:00
Phil Wang
78489045cd
readme
2021-03-09 19:23:09 -08:00
Phil Wang
173e07e02e
cleanup and release 0.8.0
0.8.0
2021-03-08 07:28:31 -08:00
Phil Wang
0e63766e54
Merge pull request #66 from zankner/masked_patch_pred
...
Masked Patch Prediction "Suggested in #63 " Work in Progress
2021-03-08 07:21:52 -08:00
Zack Ankner
a6cbda37b9
added to readme
2021-03-08 09:34:55 -05:00
Zack Ankner
73de1e8a73
converting bin targets to hard labels
2021-03-07 12:19:30 -05:00
Phil Wang
1698b7bef8
make it so one can plug performer into t2tvit
0.7.6
2021-02-25 20:55:34 -08:00
Phil Wang
6760d554aa
no need to do projection to combine attention heads for T2Ts initial one-headed attention layers
0.7.4
2021-02-24 12:23:39 -08:00
Phil Wang
a82894846d
add DistillableT2TViT
2021-02-21 19:54:45 -08:00
Phil Wang
3744ac691a
remove patch size from T2TViT
2021-02-21 19:15:19 -08:00
Phil Wang
6af7bbcd11
make sure distillation still works
0.7.1
2021-02-21 19:08:18 -08:00
Phil Wang
05edfff33c
cleanup
2021-02-20 11:32:38 -08:00
Phil Wang
e3205c0a4f
add token to token ViT
2021-02-19 22:28:53 -08:00
Phil Wang
4fc7365356
incept idea for using nystromformer
2021-02-17 15:30:45 -08:00
Phil Wang
3f2cbc6e23
fix for ambiguity in broadcasting mask
0.6.8
2021-02-17 07:38:11 -08:00
Zack Ankner
fc14561de7
made bit boundaries a function of output bits and max pixel val, fixed spelling error and reset vit_pytorch to og file
2021-02-13 18:19:21 -07:00
Zack Ankner
be5d560821
mpp loss is now based on descritized average pixels, vit forward unchanged
2021-02-12 18:30:56 -07:00
Zack Ankner
77703ae1fc
moving mpp loss into wrapper
2021-02-10 21:47:49 -07:00
Zack Ankner
a0a4fa5e7d
Working implementation of masked patch prediction as a wrapper. Need to clean code up
2021-02-09 22:55:06 -07:00
Zack Ankner
174e71cf53
Wrapper for masked patch prediction. Built handling of input and masking of patches. Need to work on integrating into vit forward call and mpp loss function
2021-02-07 16:49:06 -05:00
Zack Ankner
e14bd14a8f
Prelim work on masked patch prediction for self supervision
2021-02-04 22:00:02 -05:00
Phil Wang
85314cf0b6
patch for scaling factor, thanks to @urkax
0.6.7
2021-01-21 09:39:42 -08:00
Phil Wang
5db8d9deed
update readme about non-square images
2021-01-12 06:55:45 -08:00
Phil Wang
e8ca6038c9
allow for DistillableVit to still run predictions
0.6.5
2021-01-11 10:49:14 -08:00
Phil Wang
1106a2ba88
link to official repo
2021-01-08 08:23:50 -08:00
Phil Wang
f95fa59422
link to resources for vision people
2021-01-04 10:10:54 -08:00
Phil Wang
be1712ebe2
add quote
2020-12-28 10:22:59 -08:00
Phil Wang
1a76944124
update readme
2020-12-27 19:10:38 -08:00
Phil Wang
2263b7396f
allow distillable efficient vit to restore efficient vit as well
0.6.4
2020-12-25 19:31:25 -08:00
Phil Wang
74074e2b6c
offer easy way to turn DistillableViT to ViT at the end of training
0.6.3
2020-12-25 11:16:52 -08:00
Phil Wang
0c68688d61
bump for release
2020-12-25 09:30:48 -08:00
Phil Wang
5918f301a2
cleanup
2020-12-25 09:30:38 -08:00
Phil Wang
4a6469eecc
Merge pull request #51 from umbertov/main
...
Add class for distillation with efficient attention
2020-12-25 09:21:17 -08:00
Umberto Valleriani
5a225c8e3f
Add class for distillation with efficient attention
...
DistillableEfficientViT does the same as DistillableViT, except it
may accept a custom transformer encoder, possibly implementing an
efficient attention mechanism
2020-12-25 13:46:29 +01:00
Phil Wang
e0007bd801
add distill diagram
2020-12-24 11:34:15 -08:00
Phil Wang
db98ed7a8e
allow for overriding alpha as well on forward in distillation wrapper
0.6.2
2020-12-24 11:18:36 -08:00
Phil Wang
dc4b3327ce
no grad for teacher in distillation
2020-12-24 11:11:58 -08:00
Phil Wang
aa8f0a7bf3
Update README.md
2020-12-24 10:59:03 -08:00
Phil Wang
34e6284f95
Update README.md
2020-12-24 10:58:41 -08:00
Phil Wang
aa9ed249a3
add knowledge distillation with distillation tokens, in light of new finding from facebook ai
0.6.0
2020-12-24 10:39:15 -08:00
Phil Wang
ea0924ec96
update readme
2020-12-23 19:06:48 -08:00