cody VS llm-ls

Compare cody vs llm-ls and see what are their differences.

cody

AI that knows your entire codebase (by sourcegraph)

llm-ls

LSP server leveraging LLMs for code completion (and more?) (by huggingface)
SurveyJS - Open-Source JSON Form Builder to Create Dynamic Forms Right in Your App
With SurveyJS form UI libraries, you can build and style forms in a fully-integrated drag & drop form builder, render them in your JS app, and store form submission data in any backend, inc. PHP, ASP.NET Core, and Node.js.
surveyjs.io
featured
WorkOS - The modern identity platform for B2B SaaS
The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.
workos.com
featured
cody llm-ls
22 2
1,812 454
20.6% 21.6%
9.9 8.2
5 days ago about 2 months ago
TypeScript Rust
Apache License 2.0 Apache License 2.0
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

cody

Posts with mentions or reviews of cody. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2024-04-07.
  • Ask HN: Cheapest way to use LLM coding assistance?
    1 project | news.ycombinator.com | 13 Apr 2024
    checkout the cody extension https://github.com/sourcegraph/cody available for various editors like vscode
  • The lifecycle of a code AI completion
    6 projects | news.ycombinator.com | 7 Apr 2024
    I don't think it is. There is a test file which includes C#, Kotlin, etc among supported languages, which aren't included in the file you linked: https://github.com/sourcegraph/cody/blob/main/vscode/src/com...

    But this test didn't seem to include TypeScript so it's obviously not comprehensive. I'm not convinced this information is actually in one place.

  • Ollama is now available on Windows in preview
    7 projects | news.ycombinator.com | 17 Feb 2024
    Cody (https://github.com/sourcegraph/cody) supports using Ollama for autocomplete in VS Code. See the release notes at https://sourcegraph.com/blog/cody-vscode-1.1.0-release for instructions. And soon it'll support Ollama for chat/refactoring as well (https://twitter.com/sqs/status/1750045006382162346/video/1).

    Disclaimer: I work on Cody.

  • My 2024 AI Predictions
    3 projects | news.ycombinator.com | 8 Jan 2024
    Have you tried Cody (https://cody.dev)? Cody has a deep understanding of your codebase and generally does much better at code gen than just one-shotting GPT4 without context.

    (disclaimer: I work at Sourcegraph)

  • šŸš€ 7 AI Tools to Improve your productivity: A Deep Dive šŸŖ„āœØ
    5 projects | dev.to | 3 Jan 2024
    3ļøāƒ£ Cody AI šŸ¤–
  • An ex-Googler's guide to dev tools
    2 projects | news.ycombinator.com | 28 Nov 2023
    Author of the post hereā€”as another commenter mentioned, this is indeed a bit dated now, someone should probably write an updated post!

    There's been a ton of evolution in dev tools in the past 3 years with some old workhorses retiring (RIP Phabricator) and new ones (like Graphite, which is awesome) emerging... and of course AI-AI-AI. LLMs have created some great new tools for the developer inner loopā€”that's probably the most glaring omission here. If I were to include that category today, it would mention tools like ChatGPT, GH Copilot, Cursor, and our own Sourcegraph Cody (https://cody.dev). I'm told that Google has internal AI dev tools now that generate more code than humans.

    Excited to see what changes the next 3 years bringā€”the pace of innovation is only accelerating!

  • LocalPilot: Open-source GitHub Copilot on your MacBook
    6 projects | news.ycombinator.com | 19 Oct 2023
    I'm sorry to hear that. We have made a lot of improvements to Cody recently. We had a big release on Oct 4 that significantly decreased latency while improving completion quality. You can read all about it here: https://about.sourcegraph.com/blog/feature-release-october-2...

    We love feedback and ideas as well, and like I said are constantly iterating on the UI to improve it. I'm actually wrapping up a blog post on how to better leverage Cody w/ VS Studio, that'll be out either later today or sometime tomorrow. As far as feedback though: https://github.com/sourcegraph/cody/discussions/new?category... would be the place to share ideas :)

  • Show HN: Ollama for Linux ā€“ Run LLMs on Linux with GPU Acceleration
    14 projects | news.ycombinator.com | 26 Sep 2023
    Ollama is awesome. I am part of a team building a code AI application[1], and we want to give devs the option to run it locally instead of only supporting external LLMs from Anthropic, OpenAI, etc. Those big remote LLMs are incredibly powerful and probably the right choice for most devs, but it's good for devs to have a local option as wellā€”for security, privacy, cost, latency, simplicity, freedom, etc.

    As an app dev, we have 2 choices:

    (1) Build our own support for LLMs, GPU/CPU execution, model downloading, inference optimizations, etc.

    (2) Just tell users "run Ollama" and have our app hit the Ollama API on localhost (or shell out to `ollama`).

    Obviously choice 2 is much, much simpler. There are some things in the middle, like less polished wrappers around llama.cpp, but Ollama is the only thing that 100% of people I've told about have been able to install without any problems.

    That's huge because it's finally possible to build real apps that use local LLMsā€”and still reach a big userbase. Your userbase is now (pretty much) "anyone who can download and run a desktop app and who has a relatively modern laptop", which is a big population.

    I'm really excited to see what people build on Ollama.

    (And Ollama will simplify deploying server-side LLM apps as well, but right now from participating in the community, it seems most people are only thinking of it for local apps. I expect that to change when people realize that they can ship a self-contained server app that runs on a cheap AWS/GCP instance and uses an Ollama-executed LLM for various features.)

    [1] Shameless plug for the WIP PR where I'm implementing Ollama support in Cody, our code AI app: https://github.com/sourcegraph/cody/pull/905.

  • Cody ā€“ The AI that knows your entire codebase
    14 projects | news.ycombinator.com | 26 Aug 2023
    Awesome. The repository is at https://github.com/sourcegraph/cody for anyone who hasn't seen it yet.
  • Code AI with Codebase Context
    1 project | news.ycombinator.com | 20 Jul 2023

llm-ls

Posts with mentions or reviews of llm-ls. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2023-12-18.
  • Continue will generate, refactor, and explain entire sections of code
    2 projects | news.ycombinator.com | 18 Dec 2023
    > I'd have expected that the main lever the product has in being better than others is having a custom model that understands code edits much more than others.

    True, but this is not something this particular product would solve. There are already models specifically trained to work on code. What's appealing to me is the flexibility of being able to choose which one to use, rather than my workflow being tied to a specific product or company.

    > the IDE integration seems to be the "easy bit"

    I admittedly haven't researched this much, but this is not currently the case. There is no generic API for LLMs that IDEs can plug into, so all plugins must target a specific model. We ultimately need an equivalent of an LSP server for LLMs, and while such a project exists[1], it looks to be in its infancy, as expected.

    [1]: https://github.com/huggingface/llm-ls

  • LocalPilot: Open-source GitHub Copilot on your MacBook
    6 projects | news.ycombinator.com | 19 Oct 2023
    Okay, I actually got local co-pilot set up. You will need these 4 things.

    1) CodeLlama 13B or another FIM model https://huggingface.co/codellama/CodeLlama-13b-hf. You want "Fill in Middle" models because you're looking at context on both sides of your cursor.

    2) HuggingFace llm-ls https://github.com/huggingface/llm-ls A large language mode Language Server (is this making sense yet)

    3) HuggingFace inference framework. https://github.com/huggingface/text-generation-inference At least when I tested you couldn't use something like llama.cpp or exllama with the llm-ls, so you need to break out the heavy duty badboy HuggingFace inference server. Just config and run. Now config and run llm-ls.

    4) Okay, I mean you need an editor. I just tried nvim, and this was a few weeks ago, so there may be better support. My expereicen was that is was full honest to god copilot. The CodeLlama models are known to be quite good for its size. The FIM part is great. Boilerplace works so much easier with the surrounding context. I'd like to see more models released that can work this way.

What are some alternatives?

When comparing cody and llm-ls you can also consider the following projects:

ollama - Get up and running with Llama 3, Mistral, Gemma, and other large language models.

OpenAI-sublime-text - Sublime Text OpenAI completion plugin with GPT-4 support!

zoekt - Fast trigram based code search

text-generation-inference - Large Language Model Text Generation Inference

lsp-cody - A Client to Connect to the Cody LSP Gateway

localpilot

koboldcpp - A simple one-file way to run various GGML and GGUF models with KoboldAI's UI

continue - ā© Open-source VS Code and JetBrains extensions that enable you to easily create your own modular AI software development system

refact - WebUI for Fine-Tuning and Self-hosting of Open-Source Large Language Models for Coding

react-agent - The open-source React.js Autonomous LLM Agent