Fine Tuning Mistral 7B on Magic the Gathering Draft

This page summarizes the projects mentioned and recommended in the original post on news.ycombinator.com

Scout Monitoring - Free Django app performance insights with Scout Monitoring
Get Scout setup in minutes, and let us sweat the small stuff. A couple lines in settings.py is all you need to start monitoring your apps. Sign up for our free tier today.
www.scoutapm.com
featured
InfluxDB - Purpose built for real-time analytics at any scale.
InfluxDB Platform is powered by columnar analytics, optimized for cost-efficient storage, and built with open data standards.
www.influxdata.com
featured
  • peft

    🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.

    There is not a lot of great content out there making this clear, but basically all that matters for basic fine tuning is how much VRAM you have -- since the 3090 / 4090 have 24GB VRAM they're both pretty decent fine tuning chips. I think you could probably fine-tune a model up to ~13B parameters on one of them with PEFT (https://github.com/huggingface/peft)

  • Scout Monitoring

    Free Django app performance insights with Scout Monitoring. Get Scout setup in minutes, and let us sweat the small stuff. A couple lines in settings.py is all you need to start monitoring your apps. Sign up for our free tier today.

    Scout Monitoring logo
  • llm-foundry

    LLM training code for Databricks foundation models

    Related comment from gwern: https://news.ycombinator.com/item?id=38438859

    Also - why qlora rather than a full finetune? Using LambdaLabs, It'd cost roughly the same as your quote. Cheaper I think if you're willing to gamble with fp8: https://github.com/mosaicml/llm-foundry/tree/main/scripts/tr.... And fewer hyperparameters to tune as well

  • mage

    Magic Another Game Engine (by magefree)

    Yeah, that surprised me too, given that https://github.com/magefree/mage is open source and pretty actively developed.

  • MtgMatrix

    Use semantic search to browse similar and related cards in Magic: The Gathering. Powered by Langchain.

    Excellent, thank you for posting this!

    I was actually just looking into fine-tuning an LLM for Magic: The Gathering this week -- I've been building a small card-similarity browser using semantic embeddings of cards to find functionally or flavorfully similar cards.

    I've just been using InstructorXL, but either Instructor doesn't have enough innate knowledge of the game, or else I need to work on better prompts, but so far I've tried 9 different prompts, and none of them seem to perform very well for generating embeddings:

    https://github.com/HanClinto/MtgMatrix/blob/main/data/create...

    So my next step was to try and download a dataset of similar cards (I have some ideas on this), and I was trying to see if I could use this to do triplet-loss training of a large embedding model or something.

    Aaaaand, that's as far as I've gotten. I haven't actually figured out _how_ to hook all of that up, but your post is extremely inspirational for me. Thank you for posting this!!

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts

  • Draft Time Spiral, Lorwyn, and other older sets online for free with the XMage Draft Historical Society!

    1 project | /r/magicTCG | 28 Jun 2023
  • Drafting Help

    1 project | /r/magicTCG | 21 Jun 2023
  • Bugged cards (xmage beta 1.4.51)

    1 project | /r/XMage | 17 Jun 2023
  • Xmage beta server update

    1 project | /r/magicTCG | 9 Jun 2023
  • Seriously. Just woke up one morning and it made so much sense.

    1 project | /r/ProgrammerHumor | 23 May 2023