Skip to content

🪢 Open source LLM engineering platform. Observability, metrics, evals, prompt management, testing -- 🍊YC W23 🤖 integrate via Typescript, Python, OpenAI, Langchain, LlamaIndex, Litellm

License

Notifications You must be signed in to change notification settings

RishandRish/langfuse

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Langfuse is the open source LLM engineering platform.

Debug, analyze and iterate - together


MIT License Y Combinator W23 Docker Image langfuse npm package langfuse Python package on PyPi

Overview

Develop

Monitor

Test

  • Experiments: Track and test app behaviour before deploying a new version

Video: Langfuse in two minutes

langfuse_demo_2_min.mp4

Muted by default, enable sound for voice-over

Get started

Langfuse Cloud

Managed deployment by the Langfuse team, generous free-tier (hobby plan), no credit card required.

» Langfuse Cloud

Localhost (docker)

# Clone repository
git clone https://github.com/langfuse/langfuse.git
cd langfuse

# Run server and database
docker compose up -d

→ Learn more about deploying locally

Self-host (docker)

Langfuse is simple to self-host and keep updated. It currently requires only a single docker container. → Self Hosting Instructions

Templated deployments: Railway, GCP Cloud Run, AWS Fargate, Kubernetes and others

Get Started

API Keys

You require a Langfuse public and secret key to get started. Sign up here and find them in your project settings.

Ingesting Data · Instrumenting Your Application

Note: We recommend using our fully async, typed SDKs that allow you to instrument any LLM application with any underlying model. They are available in Python & JS/TS. The SDKs will always be the most fully featured and stable way to ingest data into Langfuse.

You may want to use another integration to get started quickly or implement a use case that we do not yet support. However, we recommend to migrate to the Langfuse SDKs over time to ensure performance and stability.

See our the → Quickstart to get started in integrating Langfuse.

Integrations

Integration Supports Description
SDK - recommended Python, JS/TS Manual instrumentation using the SDKs for full flexibility.
OpenAI SDK Python Automated instrumentation via drop-in replacement of OpenAI SDK.
Langchain Python, JS/TS Instrumentation via Langchain callbacks.
LlamaIndex Python Automated instrumentation via LlamaIndex callback system.
API Directly call the public API. OpenAPI spec available.

External projects/packages that integrate with Langfuse:

Name Description
LiteLLM Use any LLM as a drop in replacement for GPT. Use Azure, OpenAI, Cohere, Anthropic, Ollama, VLLM, Sagemaker, HuggingFace, Replicate (100+ LLMs).
Flowise JS/TS no-code builder for customized LLM flows.
Langflow Python-based UI for LangChain, designed with react-flow to provide an effortless way to experiment and prototype flows.
Superagent Open Source AI Assistant Framework & API for prototyping and deployment of agents.

Questions and feedback

Ideas and roadmap

Support and feedback

In order of preference the best way to communicate with us:

Contributing to Langfuse

  • Vote on Ideas
  • Raise and comment on Issues
  • Open a PR - see CONTRIBUTING.md for details on how to setup a development environment.

License

This repository is MIT licensed, except for the ee/ folder. See LICENSE and docs for more details.

Misc

GET API to export your data

GET routes to use data in downstream applications (e.g. embedded analytics).

Security & Privacy

We take data security and privacy seriously. Please refer to our Security and Privacy page for more information.

Telemetry

By default, Langfuse automatically reports basic usage statistics of self-hosted instances to a centralized server (PostHog).

This helps us to:

  1. Understand how Langfuse is used and improve the most relevant features.
  2. Track overall usage for internal and external (e.g. fundraising) reporting.

None of the data is shared with third parties and does not include any sensitive information. We want to be super transparent about this and you can find the exact data we collect here.

You can opt-out by setting TELEMETRY_ENABLED=false.

About

🪢 Open source LLM engineering platform. Observability, metrics, evals, prompt management, testing -- 🍊YC W23 🤖 integrate via Typescript, Python, OpenAI, Langchain, LlamaIndex, Litellm

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • TypeScript 98.1%
  • JavaScript 1.2%
  • Other 0.7%