Our great sponsors
-
open_llama
OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset
-
RedPajama-Data
The RedPajama-Data repository contains code for preparing large datasets for training large language models.
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
fastLLaMa
fastLLaMa: An experimental high-performance framework for running Decoder-only LLMs with 4-bit quantization in Python using a C/C++ backend.
-
WorkOS
The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.
Not sure, but their wikipedia code do support limited multilingual language
If you have lots of available VRAM and a powerful GPU, the use the original llama inference code, which is actually open source.
If your GPU isn't good enough, you could use llama.cpp, which runs on CPU, or one of its forks like fastLLaMa.
If your GPU isn't good enough, you could use llama.cpp, which runs on CPU, or one of its forks like fastLLaMa.
Related posts
- Show HN: Knee, OSS CLI tool to reduce time for infra orchestration
- VGGSfM: Visual Geometry Grounded Deep Structure from Motion
- Show HN: An open-sourced implementation of GitHub Copilot CLI
- Show HN: Hashquery, a Python library for defining reusable analysis
- FreeEval: A Framework for Trustworthy and Efficient Evaluation of LLMs