reweight-gpt

Reweight GPT - a simple neural network using transformer architecture for next character prediction (by hunar4321)

Reweight-gpt Alternatives

Similar projects and alternatives to reweight-gpt

  1. repeng

    1 reweight-gpt VS repeng

    A library for making RepE control vectors

  2. CodeRabbit

    CodeRabbit: AI Code Reviews for Developers. Revolutionize your code reviews with AI. CodeRabbit offers PR summaries, code walkthroughs, 1-click suggestions, and AST-based analysis. Boost productivity and code quality across all major languages with each PR.

    CodeRabbit logo
  3. ai_story_scale

    2 reweight-gpt VS ai_story_scale

    The AI story scale (AISS): A human rating scale for texts written with generative language models.

  4. ML-foundations

    1 reweight-gpt VS ML-foundations

    Machine Learning Foundations: Linear Algebra, Calculus, Statistics & Computer Science

  5. jetson-containers

    10 reweight-gpt VS jetson-containers

    Machine Learning Containers for NVIDIA Jetson and JetPack-L4T

  6. Andrew-NG-Notes

    1 reweight-gpt VS Andrew-NG-Notes

    This is Andrew NG Coursera Handwritten Notes.

  7. numerical-linear-algebra

    Free online textbook of Jupyter notebooks for fast.ai Computational Linear Algebra course

  8. SaaSHub

    SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives

    SaaSHub logo
NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a better reweight-gpt alternative or higher similarity.

reweight-gpt discussion

Log in or Post with

reweight-gpt reviews and mentions

Posts with mentions or reviews of reweight-gpt. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2023-05-01.
  • [Research] An alternative to self-attention mechanism in GPT
    2 projects | /r/MachineLearning | 1 May 2023
    Instead of self attention, I tried to generate the self-attention matrix directly using lateral connections among the inputs. The method is like LSTM but it gates all the past inputs using separate gates for each input (it can be parallelized). It's very easy to implement the method into the current GPT architectures. You just remove the attention part and replace it with learnable weights. Her is a working implementation (around100 lines!): Code: https://github.com/hunar4321/reweight-gpt In my experience, it learns very well and it can super-pass the self-attention mechanism if the number of the parameters are matched. (I tested it on small datasets for next character prediction. I haven't systematically compared these two methods yet).

Stats

Basic reweight-gpt repo stats
1
51
6.3
over 1 year ago

hunar4321/reweight-gpt is an open source project licensed under MIT License which is an OSI approved license.

The primary programming language of reweight-gpt is Jupyter Notebook.


Sponsored
CodeRabbit: AI Code Reviews for Developers
Revolutionize your code reviews with AI. CodeRabbit offers PR summaries, code walkthroughs, 1-click suggestions, and AST-based analysis. Boost productivity and code quality across all major languages with each PR.
coderabbit.ai