Workshop Alternatives

Similar projects and alternatives to workshop

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a better workshop alternative or higher similarity.

workshop reviews and mentions

Posts with mentions or reviews of workshop. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2023-04-26.
  • Transformers from Scratch
    4 projects | news.ycombinator.com | 26 Apr 2023
    - There are a few common ways you might see this done, but they broadly work by assigning fixed or learned embeddings to each position in the input token sequence. These embeddings can be added to our matrix above so that the first row gets the embedding for the first position added to it, the second row gets the embedding for the second position, and so on. Now if the tokens are reordered, the embedding matrix will not be the same. Alternatively, these embeddings can be concatenated horizontally to our matrix: this guarantees the positional information is kept entirely separate from the linguistic (at the cost of having a larger combined embedding that the block must support).

    I put together this repository at the end of last year to better help visualize the internals of a transformer block when applied to a toy problem: https://github.com/rstebbing/workshop/tree/main/experiments/.... It is not super long, and the point is to try and better distinguish between the quantities you referred to by seeing them (which is possible when embeddings are in a low dimension).

    I hope this helps!

  • Understanding and Coding the Self-Attention Mechanism of Large Language Models
    1 project | news.ycombinator.com | 10 Feb 2023
    At the end of last year I put together a repository to try and show what is achieved by self-attention on a toy example: detect whether a sequence of characters contains both "a" and "b".

    The toy problem is useful because the model dimensionality is low enough to make visualization straightforward. The walkthrough also goes through how things can go wrong, and how it can be improved, etc.

    The walkthrough and code is all available here: https://github.com/rstebbing/workshop/tree/main/experiments/....

    It's not terse like nanoGPT or similar because the goal is a bit different. In particular, to gain more intuition about the intermediate attention computations, the intermediate tensors are named and persisted so they can be compared and visualized after the fact. Everything should be exactly reproducible locally too!

  • The Transformer Family
    1 project | news.ycombinator.com | 29 Jan 2023
    I put together a repository at the end of last year to walk through a basic use of a single layer Transformer: detect whether "a" and "b" are in a sequence of characters. Everything is reproducible, so hopefully helpful at getting used to some of the tooling too!

    https://github.com/rstebbing/workshop/tree/main/experiments/...

  • A note from our sponsor - InfluxDB
    www.influxdata.com | 21 May 2024
    Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality. Learn more →

Stats

Basic workshop repo stats
3
13
10.0
over 1 year ago

rstebbing/workshop is an open source project licensed under MIT License which is an OSI approved license.

The primary programming language of workshop is Python.


Sponsored
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com