-
PyTorch-Vision-Transformer-ViT-MNIST-CIFAR10
Simplified Pytorch implementation of Vision Transformer (ViT) for small datasets like MNIST, FashionMNIST, SVHN and CIFAR10.
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
In the encoder class, ViTs use prenorm not post-norm like BERT. That is the first norm layer should be before the attention and the second norm layer should be before "self.fc1" https://github.com/s-chh/PyTorch-Vision-Transformer-ViT-MNIST/blob/main/model.py
NOTE:
The number of mentions on this list indicates mentions on common posts plus user suggested alternatives.
Hence, a higher number means a more popular project.
Related posts
-
[P] Token-to-Token ViT Implementation in Flax
-
[N] First-Ever Course on Transformers: NOW PUBLIC
-
[Research] Awesome Paper List of Vision Transformer & Attention
-
[Research] Awesome Paper List of Vision Transformer & Attention
-
A New Deep Learning Study Investigate and Clarify the Intrinsic Behavior of Transformers in Computer Vision