Transformer-in-Transformer
gpt-mini
Transformer-in-Transformer | gpt-mini | |
---|---|---|
4 | 1 | |
41 | 13 | |
- | - | |
0.0 | 0.0 | |
about 2 years ago | over 1 year ago | |
Jupyter Notebook | Jupyter Notebook | |
Apache License 2.0 | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Transformer-in-Transformer
- I Implemented Transformer in Transformer
-
Hacker News top posts: Dec 6, 2021
I Implemented Transformer in Transformer\ (5 comments)
- [P] I implemented Transformer in Transformer
gpt-mini
What are some alternatives?
poolformer - PoolFormer: MetaFormer Is Actually What You Need for Vision (CVPR 2022 Oral)
SpectralEmbeddings - spectralembeddings is a python library which is used to generate node embeddings from Knowledge graphs using GCN kernels and Graph Autoencoders. Variations include VanillaGCN,ChebyshevGCN and Spline GCN along with SDNe based Graph Autoencoder.
LongNet - Implementation of plug in and play Attention from "LongNet: Scaling Transformers to 1,000,000,000 Tokens"
alpha-zero-general - A clean implementation based on AlphaZero for any game in any framework + tutorial + Othello/Gobang/TicTacToe/Connect4 and more
AvatarGAN - Generate Cartoon Images using Generative Adversarial Network
minGPT-TF - A minimal TF2 re-implementation of the OpenAI GPT training
swarms - Orchestrate Swarms of Agents From Any Framework Like OpenAI, Langchain, and Etc for Real World Workflow Automation. Join our Community: https://discord.gg/DbjBMJTSWD
vision_transformer_tf - This repository contains the TensorFlow implementation of the paper "AN IMAGE IS WORTH 16X16 WORDS: TRANSFORMERS FOR IMAGE RECOGNITION AT SCALE" known as vision transformers.
principia - The Principia Rewrite
Fast-Transformer - An implementation of Fastformer: Additive Attention Can Be All You Need, a Transformer Variant in TensorFlow
planckforth - Bootstrapping a Forth interpreter from hand-written tiny ELF binary. Just for fun.
D2L_Attention_Mechanisms_in_TF - This repository contains Tensorflow 2 code for Attention Mechanisms chapter of Dive into Deep Learning (D2L) book.