gpt-3 VS lm-human-preferences

Compare gpt-3 vs lm-human-preferences and see what are their differences.

gpt-3

GPT-3: Language Models are Few-Shot Learners (by openai)

lm-human-preferences

Code for the paper Fine-Tuning Language Models from Human Preferences (by openai)
InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
gpt-3 lm-human-preferences
41 8
9,406 1,109
- 2.4%
3.5 2.7
over 3 years ago 9 months ago
Python
- MIT License
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

gpt-3

Posts with mentions or reviews of gpt-3. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2024-04-29.
  • GPT4.5 or GPT5 being tested on LMSYS?
    3 projects | news.ycombinator.com | 29 Apr 2024
    >I wasn't talking about "state of the art LLMs," I am aware that commercial offerings are much better trained in Spanish. This was a thought experiment based on comments from people testing GPT-3.5 with Swahili.

    A thought experiment from other people comments on another language. So...No. Fabricating failure modes from their constructed ideas about how LLMs work seems to be a frustratingly common occurrence in these kinds of discussions.

    >Frustratingly, just few months ago I read a paper describing how LLMs excessively rely on English-language representations of ideas, but now I can't find it.

    Most LLMs are trained on English overwhelmingly. GPT-3 had a 92.6% English dataset. https://github.com/openai/gpt-3/blob/master/dataset_statisti...

    That the models are as proficient as they are is evidence enough of knowledge transfer clearly happening. https://arxiv.org/abs/2108.13349. If you trained a model on the Catalan tokens GPT-3 was trained on alone, you'd just get a GPT-2 level gibberish model at best.

    anyway. These are some interesting papers

    How do languages influence each other? Studying cross-lingual data sharing during LLM fine-tuning - https://arxiv.org/pdf/2305.13286

    Teaching Llama a New Language Through Cross-Lingual Knowledge Transfer - https://arxiv.org/abs/2404.04042

    Multilingual LLMs are Better Cross-lingual In-context Learners with Alignment - https://arxiv.org/abs/2305.05940

    It's not like there is perfect transfer but the idea that there's none at all seemed so ridiculous to me (and why i asked the first question). Models would be utterly useless in multilingual settings if that were really the case.

  • What are LLMs? An intro into AI, models, tokens, parameters, weights, quantization and more
    4 projects | dev.to | 28 Apr 2024
    Large models: Everything above 10B of parameters. This is where Llama 3, Llama 2, Mistral 8x22B, GPT 3, and most likely GPT 4 sit.
  • Can ChatGPT improve my L2 grammar?
    1 project | /r/AIinLanguageEducation | 4 Dec 2023
    Are generative AI models useful for learning a language, and if so which languages? Over 90% of ChatGPT's training data was in English. The remaining 10% of data was split unevenly between 100+ languages. This suggests that the quality of the outputs will vary from language to language.
  • GPT4 Can’t Ace MIT
    1 project | news.ycombinator.com | 18 Jun 2023
    I have doubts it was extensively trained on German data. Who knows about GPT4, but GPT3 is ~92% of English and ~1.5% of German, which means it saw more "die, motherfucker, die" than on "die Mutter".

    (https://github.com/openai/gpt-3/blob/master/dataset_statisti...)

  • Necesito ayuda.
    1 project | /r/devsarg | 28 May 2023
  • [R] PaLM 2 Technical Report
    1 project | /r/MachineLearning | 10 May 2023
    Catalan was 0.018 % of GPT-3's training corpus. https://github.com/openai/gpt-3/blob/master/dataset_statistics/languages_by_word_count.csv.
  • I'm seriously concerned that if I lost ChatGPT-4 I would be handicapped
    1 project | /r/ChatGPT | 25 Apr 2023
  • The responses I got from bard after asking why 100 times… he was pissed 😂
    1 project | /r/ChatGPT | 15 Apr 2023
  • BharatGPT: India's Own ChatGPT
    1 project | news.ycombinator.com | 13 Apr 2023
    >Certainly it is pleasing that they are not just doing Hindi, but some of these languages must be represented online by a very small corpus of text indeed. I wonder how effectively an LLM can be trained on such a small training set for any given language?

    as long as it's not the main language it doesn't really matter. Besides English(92.6%), the biggest language by representation (word count) is taken up by french at 1.8%. Most of the languages GPT-3 knows are sitting at <0.2% representation.

    https://github.com/openai/gpt-3/blob/master/dataset_statisti...

    Competence in the main language will bleed into the rest.

  • GPT-4 gets a B on Scott Aaronson's quantum computing final exam
    1 project | /r/Physics | 12 Apr 2023

lm-human-preferences

Posts with mentions or reviews of lm-human-preferences. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2023-02-14.

What are some alternatives?

When comparing gpt-3 and lm-human-preferences you can also consider the following projects:

dalle-mini - DALL·E Mini - Generate images from a text prompt

trl - Train transformer language models with reinforcement learning.

DALL-E - PyTorch package for the discrete VAE used for DALL·E.

GLM-130B - GLM-130B: An Open Bilingual Pre-Trained Model (ICLR 2023)

DALLE-mtf - Open-AI's DALL-E for large scale training in mesh-tensorflow.

stylegan2-pytorch - Simplest working implementation of Stylegan2, state of the art generative adversarial network, in Pytorch. Enabling everyone to experience disentanglement

tensorrtx - Implementation of popular deep learning networks with TensorRT network definition API

v-diffusion-pytorch - v objective diffusion inference code for PyTorch.

glide-text2im - GLIDE: a diffusion-based text-conditional image synthesis model

dalle-2-preview

gpt-2 - Code for the paper "Language Models are Unsupervised Multitask Learners"