langfuse
trulens
langfuse | trulens | |
---|---|---|
11 | 14 | |
3,681 | 1,669 | |
30.4% | 10.1% | |
9.9 | 9.8 | |
7 days ago | 5 days ago | |
TypeScript | Jupyter Notebook | |
GNU General Public License v3.0 or later | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
langfuse
-
Top Open Source Prompt Engineering Guides & Tools🔧🏗️🚀
Langfuse is an open-source LLM engineering platform that helps teams collaboratively debug, analyze, and iterate on their LLM applications.
- Roast My Docs
-
Show HN: Open-Source LLM Observability and Export to Grafana, Datadog etc.
Congrats on the Show! How’s this different from https://github.com/langfuse/langfuse? The exports seems really interesting
-
RAG observability in 2 lines of code with Llama Index & Langfuse
Thus, we started working on Langfuse.com (GitHub) to establish an open source LLM engineering platform with tightly integrated features for tracing, prompt management, and evaluation. In the beginning we just solved our own and our friends’ problems. Today we are at over 1000 projects which rely on Langfuse, and 2.3k stars on GitHub. You can either self-host Langfuse or use the cloud instance maintained by us.
-
langfuse VS agenta - a user suggested alternative
2 projects | 22 Nov 2023
-
Ask HN: Who is hiring? (November 2023)
- We want to build a tool that is recommended here on HN: you can build a tool you would want to use yourself.
Please see more details here: https://langfuse.com/careers or reach out directly to me: [email protected]
[1] https://github.com/langfuse/langfuse
[2] https://create.t3.gg/
-
How are generative AI companies monitoring their systems in production?
We struggled with this ourselves while building LLM-based products and then open-sourced our observability/monitoring tool [1]. Many use it to track RAG and agents in production, run custom evals on the production traces (focused on hallucination), and track how metrics are different across releases or customers. Feel free to dm if there is something specific you are looking to solve, happy to help.
[1] https://github.com/langfuse/langfuse
-
LLM Analytics 101 - How to Improve your LLM app
Visit us on Discord and Github to engage with our project.
-
Ask HN: Any tools or frameworks to monitor the usage of OpenAI API keys?
Maybe try https://github.com/langfuse/langfuse
It was recently shared on HN
- Show HN: Langfuse – Open-source observability and analytics for LLM apps
trulens
-
Why Vector Compression Matters
Retrieval using a single vector is called dense passage retrieval (DPR), because an entire passage (dozens to hundreds of tokens) is encoded as a single vector. ColBERT instead encodes a vector-per-token, where each vector is influenced by surrounding context. This leads to meaningfully better results; for example, here’s ColBERT running on Astra DB compared to DPR using openai-v3-small vectors, compared with TruLens for the Braintrust Coda Help Desk data set. ColBERT easily beats DPR at correctness, context relevance, and groundedness.
- FLaNK AI Weekly 18 March 2024
-
First 15 Open Source Advent projects
12. TruLens by TruEra | Github | tutorial
-
trulens VS agenta - a user suggested alternative
2 projects | 22 Nov 2023
-
How are generative AI companies monitoring their systems in production?
3) Hallucination is probably the biggest problem we solve for. To do evals for hallucination, we typically see our users use a combination of groundedness (does the context support the LLM response) and context relevance (is the retrieved context relevant to the query). There's also a bunch more for the evaluations you mentioned (moderation models, sentiment, usefulness, etc.) and it's pretty easy to add custom evals.
Also - my hot take is that gpt-3.5 is good enough for evals (sometimes better) than gpt-4 if you give the LLM enough instructions on how to do the eval.
website: https://www.trulens.org/
- FLaNK Stack Weekly 28 August 2023
-
[P] TruLens-Eval is an open source project for eval & tracking LLM experiments.
The team at TruEra recently released an open source project for evaluation & tracking of LLM applications called TruLens-Eval. We’ve specifically targeted retrieval-augmented QA as a core use case and so far we’ve seen it used for comparing different models and parameters, prompts, vector-db configurations and query planning strategies. I’d love to get your feedback on it.
- [D] Hardest thing about building with LLMs?
- Stop Evaluating LLMs on Vibes
- OSS library for attribution and interpretation methods for deep nets
What are some alternatives?
llama_index - LlamaIndex is a data framework for your LLM applications
shapash - 🔅 Shapash: User-friendly Explainability and Interpretability to Develop Reliable and Transparent Machine Learning Models
langchain - 🦜🔗 Build context-aware reasoning applications
probability - Probabilistic reasoning and statistical analysis in TensorFlow
agenta - The all-in-one LLM developer platform: prompt management, evaluation, human feedback, and deployment all in one place.
LIME - Tutorial notebooks on explainable Machine Learning with LIME (Original work: https://arxiv.org/abs/1602.04938)
opentelemetry-instrument-openai-py - OpenTelemetry instrumentation for the OpenAI Python library
embedchain - Personalizing LLM Responses
examples - Your one-stop-shop to try Xata out. From packages to apps, whatever you need to get started.
machine_learning_basics - Plain python implementations of basic machine learning algorithms
clickhouse_knowledge_base - The Tinybird ClickHouse Knowledge Base
ML-Workspace - 🛠 All-in-one web-based IDE specialized for machine learning and data science.