gpts-cant-count
Demo of even the most advanced LLMs' inability to handle basic arithmetic. (by 0xnurl)
memorizing-transformers-pytorc
By lucidrains
gpts-cant-count | memorizing-transformers-pytorc | |
---|---|---|
4 | 3 | |
8 | - | |
- | - | |
6.4 | - | |
5 months ago | - | |
Python | ||
- | - |
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
gpts-cant-count
Posts with mentions or reviews of gpts-cant-count.
We have used some of these posts to build our list of alternatives
and similar projects. The last one was on 2024-04-27.
-
What can LLMs never do?
Simple addition, among other things:
https://github.com/0xnurl/gpts-cant-count
- How many legs do ten elephants have, if two of them are legless?
- Show HN: GPTs Can't Count
- GPTs Can't Count
memorizing-transformers-pytorc
Posts with mentions or reviews of memorizing-transformers-pytorc.
We have used some of these posts to build our list of alternatives
and similar projects. The last one was on 2024-04-27.
-
What can LLMs never do?
At one point I experimented a little with transformers that had access to external memory searchable via KNN lookups https://github.com/lucidrains/memorizing-transformers-pytorc... or via routed queries with https://github.com/glassroom/heinsen_routing . Both approaches seemed to work for me, but I had to put that work on hold for reasons outside my control.
-
Memorizing Transformers – models that can acquire new knowledge immediately
have an implementation of this over at https://github.com/lucidrains/memorizing-transformers-pytorc..., for any researcher exploring retrieval and memory with attention networks