commandjobs
llama.cpp
commandjobs | llama.cpp | |
---|---|---|
5 | 777 | |
140 | 57,984 | |
- | - | |
6.4 | 10.0 | |
about 1 month ago | 42 minutes ago | |
Python | C++ | |
Apache License 2.0 | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
commandjobs
-
Launch HN: Glide (YC W19) – AI-assisted technical design docs
Was able to try out Glide last night (https://glide.agenticlabs.com/task/IqHd0RV) with an open source repo (https://github.com/nicobrenner/commandjobs) and really liked it
Pros: it was kinda easy to get started, and Glide was able to figure out something that ChatGPT hadn't, which made a big difference, because I was about to have to rewrite a bunch of code otherwise
Cons: some things were not very intuitive, the chatting window is small and it's hard to copy text and read things in a slightly longer conversation - also wish it would just ingest the repo and figure out what to include in the context by itself, the method/section searching/selecting is tricky to use
All in all I'm very excited about what Glide can do and look forward to seeing its evolution
Thank you for building such amazing tool
-
Ask HN: Hosting a CLI Tool via SSH?
Hi HN! I recently built a cli tool that uses AI to find job matches for software engineers (https://github.com/nicobrenner/commandjobs)
Someone on HN suggested (https://news.ycombinator.com/user?id=azophy_2) I should host the application via SSH, so that anyone could just do: ssh commandjobs.com
I thought that was super cool, but after doing some googling, I can't find any good resources on how to host command line applications via ssh
Would really appreciate any links or information you could provide on how to do it
Thank you
-
Show HN: Tech Jobs on the Command Line
Excellent, thank you so much for the links, it's super interesting
It should be pretty straightforward to create a scraper for one of those portals
If you feel like taking a crack at it, checkout the code of the Ask HN who's hiring scraper here: https://github.com/nicobrenner/commandjobs/blob/47b5c89402a3...
I added a comment under the Add Sources issue, to support Workday's career portals: https://github.com/nicobrenner/commandjobs/issues/23#issueco...
llama.cpp
-
IBM Granite: A Family of Open Foundation Models for Code Intelligence
if you can compile stuff, then looking at llama.cpp (what ollama uses) is also interesting: https://github.com/ggerganov/llama.cpp
the server is here: https://github.com/ggerganov/llama.cpp/tree/master/examples/...
And you can search for any GGUF on huggingface
-
Ask HN: Affordable hardware for running local large language models?
Yes, Metal seems to allow a maximum of 1/2 of the RAM for one process, and 3/4 of the RAM allocated to the GPU overall. There’s a kernel hack to fix it, but that comes with the usual system integrity caveats. https://github.com/ggerganov/llama.cpp/discussions/2182
- Xmake: A modern C/C++ build tool
-
Better and Faster Large Language Models via Multi-Token Prediction
For anyone interested in exploring this, llama.cpp has an example implementation here:
https://github.com/ggerganov/llama.cpp/tree/master/examples/...
- Llama.cpp Bfloat16 Support
-
Fine-tune your first large language model (LLM) with LoRA, llama.cpp, and KitOps in 5 easy steps
Getting started with LLMs can be intimidating. In this tutorial we will show you how to fine-tune a large language model using LoRA, facilitated by tools like llama.cpp and KitOps.
- GGML Flash Attention support merged into llama.cpp
-
Phi-3 Weights Released
well https://github.com/ggerganov/llama.cpp/issues/6849
- Lossless Acceleration of LLM via Adaptive N-Gram Parallel Decoding
- Llama.cpp Working on Support for Llama3
What are some alternatives?
ollama - Get up and running with Llama 3, Mistral, Gemma, and other large language models.
gpt4all - gpt4all: run open-source LLMs anywhere
text-generation-webui - A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.
GPTQ-for-LLaMa - 4 bits quantization of LLaMA using GPTQ
ggml - Tensor library for machine learning
alpaca.cpp - Locally run an Instruction-Tuned Chat-Style LLM
FastChat - An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
rust-gpu - 🐉 Making Rust a first-class language and ecosystem for GPU shaders 🚧
ChatGLM-6B - ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型
safetensors - Simple, safe way to store and distribute tensors
AutoGPT - AutoGPT is the vision of accessible AI for everyone, to use and to build on. Our mission is to provide the tools, so that you can focus on what matters.
alpaca-lora - Instruct-tune LLaMA on consumer hardware