Helicone was acquired by Mintlify on March 3, 2026, and the platform has been in maintenance mode ever since — receiving only security patches, bug fixes, and support for new model identifiers. If you depend on Helicone in production today, your migration window is open. The strongest replacements are Langfuse (open-source, SDK-based, 40,000+ active builders), LangSmith (deepest LangChain integration available), Portkey (200+ LLM provider gateway), Braintrust (eval-first with 1M free spans per month), and Stockyard (a single ~25MB Go binary requiring zero cloud dependency).
What Happened to Helicone? The Mintlify Acquisition Explained
Helicone was acquired by Mintlify on March 3, 2026 — the entire founding team relocated to Mintlify’s San Francisco office within weeks of the deal closing. That single event turned one of the most popular lightweight LLM observability proxies into a maintenance-mode product overnight. Helicone had grown on the strength of its zero-friction onboarding: you could start logging all OpenAI calls by changing one base URL and adding one header, with no SDK installation and no code changes required. That developer experience won it widespread adoption among solo developers and small teams shipping fast. After the acquisition, Mintlify announced that Helicone would continue operating under maintenance mode — meaning security patches, critical bug fixes, and new model name support would be applied, but no new features would ship. The March 2026 acquisition arrived during an already turbulent moment for LLM proxy tooling: LiteLLM experienced a security compromise around the same time, creating simultaneous migration pressure across two of the most widely used open-source LLM proxy tools. Teams that had spread risk across both products found themselves evaluating entirely new options at the same moment, accelerating the shift toward more stable, actively maintained observability platforms.
Should You Migrate Now? Understanding Maintenance Mode
Maintenance mode at an acquired startup creates a specific, measurable risk: your infrastructure stays functional today, but the gap between your product needs and the platform’s capabilities widens every month. Helicone’s maintenance mode is active as of March 3, 2026 — the API endpoints are live, existing integrations continue to process traces, and the dashboard functions normally. What stops immediately is feature development. Teams shipping new agent architectures, new evaluation pipelines, or new model integrations will hit gaps within months: no new evaluation frameworks, no improved cost analytics, no deeper tracing for multi-step agent workflows, no collaboration features, no integrations beyond security-patch-level model name support. For startups changing their LLM stack frequently, the migration window is effectively now. For teams with stable, infrequently changing production pipelines, urgency is lower — but the 6-month planning horizon is still the right call. Setting a target of completing migration within 6 months of the acquisition (by September 2026) gives you time to evaluate alternatives without rushing, run parallel deployments to validate data quality, and avoid being caught off-guard by a formal deprecation announcement. The LiteLLM compromise that coincided with Helicone’s acquisition is a useful reminder: LLM tooling vendor risk is real, and a migration is an opportunity to audit your entire observability stack — not just replace one proxy with another.
Top Helicone Alternatives Compared: Langfuse, LangSmith, Portkey, Braintrust
The four strongest Helicone alternatives collectively cover more than 95% of real-world migration scenarios, and each targets a different primary use case. Langfuse is the most direct feature replacement: open-source (MIT license), used by 40,000+ builders, SDK-based rather than proxy-based — which means zero added latency on LLM calls since observability data ships asynchronously out-of-band. LangSmith from LangChain offers the deepest integration for teams already running LangChain or LangGraph pipelines; chain-level debuggability and native trace visualization are features no other tool matches for LangChain-specific architectures. Portkey operates as a production LLM gateway supporting 200+ providers, with routing, automatic fallback, load balancing, and semantic caching — it keeps the proxy architecture that Helicone users are familiar with while adding capabilities that justify the network hop. Braintrust shifts the focus to structured evaluation: 1M free spans per month, unlimited users on all pricing tiers (no per-seat charges), and a CI/CD-friendly eval framework that integrates directly into automated test pipelines. Each tool takes a different architectural stance from Helicone’s pure-proxy model, and the right choice depends on what your team is optimizing for — raw observability, chain debugging, multi-provider reliability, or continuous evaluation.
| Tool | Architecture | Best For | Free Tier | Self-Host |
|---|---|---|---|---|
| Langfuse | SDK-based | Open-source teams, self-hosters | 50K units/mo | Yes (MIT) |
| LangSmith | SDK-based | LangChain / LangGraph shops | 5K traces/mo | Enterprise only |
| Portkey | Proxy-based | Multi-provider gateways | Limited | Yes (OSS) |
| Braintrust | SDK-based | Eval-heavy CI/CD pipelines | 1M spans/mo | Yes |
| Arize Phoenix | SDK-based | Mixed ML + LLM workloads | Open-source | Yes (Apache 2) |
| Stockyard | Proxy-based | Air-gapped / solo developers | Self-hosted | Yes (single binary) |
Pricing Comparison: What Each Alternative Costs at Scale
At 1 million traces per month — the threshold where pricing differences become business-material — Langfuse Cloud costs approximately $919/month while LangSmith costs $2,500+, making Langfuse 2.7x cheaper at equivalent cloud scale. LangSmith’s pricing model charges $39 per seat per month plus $2.50–$5 per 1,000 traces; at 1M traces and even a small team of 5 seats, the bill exceeds $2,500 before accounting for seats above the base plan. Langfuse’s pricing tiers are: Free (50K units/month), Core at $29/month, and Pro at $199/month with additional usage billed on top — reaching approximately $919/month at 1M trace units. The self-hosted Langfuse alternative drops that cost to roughly $150/month in infrastructure at the same 1M trace scale, a 6x improvement over Langfuse Cloud and a 16x improvement over LangSmith. Braintrust’s free tier covers 1M spans per month with unlimited users on all tiers — eliminating per-seat pricing entirely, which is a meaningful cost structure difference for growing teams. Portkey pricing is usage-based and gateway-focused, with costs varying by routing complexity and provider count. Stockyard costs zero beyond your own server: it runs as a ~25MB Go binary on any Linux instance, with no external database or cloud service required.
| Tool | Free Tier | At 1M Traces/Month | Self-Hosted Cost |
|---|---|---|---|
| Langfuse Cloud | 50K units/mo | ~$919/mo | ~$150/mo infra |
| LangSmith | 5K traces/mo | $2,500+/mo | Enterprise only |
| Braintrust | 1M spans/mo | Free–$400/mo | Available |
| Portkey | Limited | Usage-based | Available |
| Stockyard | Unlimited (self-hosted) | $0 + infra | ~$10–50/mo |
Proxy-Based vs SDK-Based Observability: Which Architecture Fits Your Stack
The architectural choice between proxy-based and SDK-based LLM observability is the most consequential decision when migrating from Helicone — and it is one Helicone’s design never forced you to make. Helicone’s proxy model adds latency to every LLM call because every request routes through Helicone’s servers before reaching OpenAI (or another provider) and back. That round-trip is invisible at low request volumes but measurable in user-facing latency at scale. SDK-based tools like Langfuse and LangSmith instrument your application code directly and ship trace data asynchronously to their backends — the LLM call itself is never blocked, which means zero added latency on the hot path. The tradeoff is that SDK-based tools require code changes rather than a single URL swap: you install an SDK, add decorators or callbacks, and configure credentials. For latency-sensitive production applications where every millisecond of LLM response time affects user experience, SDK-based observability is the architecturally correct choice. Proxy-based tools remain valid for prototypes, internal tools, and rapid experimentation where zero-code integration is worth more than the latency cost. Portkey specifically bridges this tradeoff: it keeps the proxy model but adds gateway-level features (failover routing, semantic caching, load balancing across 200+ providers) that justify the network hop in a way that Helicone’s pure-observability proxy never did. Teams comfortable with Helicone’s architecture will find Portkey the most familiar migration target; teams optimizing for production performance should move to SDK-based instrumentation.
Self-Hosted Options: Langfuse OSS, Stockyard, and Arize Phoenix
Self-hosted LLM observability has become a priority for teams with data residency requirements, regulated industries with compliance constraints, or organizations that simply will not send LLM trace data — which may contain sensitive prompt content — to third-party cloud services. Langfuse OSS is the production-ready leader: it ships as a Docker Compose configuration (Kubernetes manifests available for larger deployments), includes the full feature set of traces, evaluations, prompt management, and cost tracking, and is MIT-licensed with no feature gating between OSS and cloud versions. At 1M traces per month, self-hosted Langfuse costs approximately $150/month in infrastructure — compared to $919/month for Langfuse Cloud and $2,500+ for LangSmith at equivalent scale, the cost advantage compounds quickly. Stockyard targets the opposite end of the complexity spectrum: a single ~25MB Go binary with an embedded SQLite database and no external dependencies whatsoever. No Docker, no Postgres, no Redis, no cloud component — you download one file and run it. This makes Stockyard the right tool for solo developers, small teams, and air-gapped environments where operating a full Docker stack is unnecessary overhead. The tradeoff is capability: Stockyard does not include evaluations, prompt management, or advanced analytics. Arize Phoenix fills the niche for organizations running both traditional ML models and LLM applications — it accepts OpenTelemetry traces from both contexts, making it a single pane of glass for teams that already monitor ML workloads and want to extend that infrastructure to cover LLM calls. All three tools are MIT or Apache 2.0 licensed, providing full IP clarity and no vendor lock-in risk.
Migration Guide: How to Move from Helicone Without Breaking Production
Migrating from Helicone without a production outage requires running your new platform in parallel before cutting over — a process that typically takes one to two weeks for a single-model application. For Langfuse (the most common migration destination): install the SDK (pip install langfuse or npm install langfuse), replace Helicone proxy configuration with Langfuse’s observe() decorator for OpenAI calls or the LangfuseCallbackHandler for LangChain, and set LANGFUSE_PUBLIC_KEY and LANGFUSE_SECRET_KEY in your environment. Helicone’s custom properties map directly to Langfuse metadata fields, so your existing trace enrichment carries over. For LangSmith: if you are on LangChain, the migration is three environment variables — LANGCHAIN_TRACING_V2=true, LANGCHAIN_ENDPOINT, and LANGCHAIN_API_KEY. No code changes required for LangChain users. For Portkey (the closest proxy-to-proxy migration): swap the base URL from Helicone’s endpoint to Portkey’s gateway URL, replace the Helicone-Auth header with x-portkey-api-key, and optionally add x-portkey-provider to enable routing features. The recommended rollout process for all alternatives follows the same pattern: stand up the target platform in parallel with Helicone running; route 10% of traffic to the new platform and validate for 48 hours; confirm trace completeness, latency impact, and cost data accuracy; ramp to 100% over one week; keep Helicone in passive-receive mode for 30 days before decommissioning. Teams using Helicone’s custom metadata headers should audit which header fields feed active dashboards before migration — not all alternatives map these identically, and gaps in metadata coverage will silently break cost attribution reports.
Which Helicone Alternative Should You Choose? Decision Framework
Three variables determine the right Helicone alternative for your team: your primary observability use case (raw logging vs evaluation vs multi-provider reliability), your existing toolchain, and your scale and data residency requirements. If your team runs LangChain or LangGraph: choose LangSmith. Native integration eliminates instrumentation overhead, and chain-level trace visualization is a capability no other tool matches for LangChain-specific agent architectures. The higher cost at scale ($2,500+/month at 1M traces) is justified if LangChain is core infrastructure. If you need self-hosting or have data residency constraints: choose Langfuse OSS. It is the most production-ready self-hosted option with the broadest feature set, MIT-licensed, and significantly cheaper than any cloud alternative at scale — roughly $150/month in infrastructure at 1M traces. If you route traffic through multiple LLM providers (OpenAI plus Anthropic plus Gemini plus fine-tuned endpoints): choose Portkey. The 200+ provider support, semantic caching, and automatic failover routing make it a production gateway that does more than just observe your traffic — it actively improves reliability. If evaluations and CI/CD integration are your primary need: choose Braintrust. The 1M free spans per month, unlimited users on all tiers, and structured eval framework make it the strongest choice for teams treating LLM quality as a continuous testing problem. If you are a solo developer or small team in an air-gapped environment: choose Stockyard. A single binary, embedded database, and zero cloud dependencies eliminate operational overhead entirely. For teams with no strong existing toolchain preference: Langfuse is the default recommendation — open-source, actively maintained with no acquisition risk, SDK-based for zero-latency overhead, aggressively priced, and the closest feature match to what Helicone offered. The 40,000+ active builders and MIT license mean the project has both community momentum and full IP portability if you later want to self-host.
FAQ
Helicone’s acquisition by Mintlify on March 3, 2026 triggered one of the most active migration discussions in the LLM observability community that quarter, with developers across GitHub, Reddit, and Hacker News evaluating alternatives under real time pressure. The decision criteria differ meaningfully by team: proxy-based tools like Helicone add minimal code changes but introduce a latency hop, while SDK-based tools like Langfuse require more instrumentation work but eliminate that latency. Pricing at scale is the other forcing function — LangSmith’s $2,500+/month at 1M traces versus Langfuse Cloud at $919 versus self-hosted Langfuse at ~$150 creates a cost difference that compounds fast as AI usage grows. The questions below address the most common decision points: whether Helicone is truly shutting down, which tool to choose based on team profile, what migration effort actually looks like, and how to avoid repeating the vendor risk that the Helicone situation exposed. Each answer reflects the state of these tools as of May 2026.
Is Helicone shutting down after the Mintlify acquisition?
Helicone is not shutting down immediately — it entered maintenance mode on March 3, 2026, when the founding team joined Mintlify in San Francisco. The platform continues to operate: API endpoints are live, the dashboard functions, and existing integrations process traces normally. What stops is feature development. Security patches, bug fixes, and new model identifier support will continue, but no new observability features, evaluation frameworks, or provider integrations will ship. Teams should plan migration within 6 months of the announcement to avoid being caught by a future formal deprecation.
What is the best free alternative to Helicone?
Braintrust offers the most generous cloud free tier: 1 million spans per month with unlimited users and no per-seat charges. Langfuse’s cloud free tier covers 50,000 units per month, which is lower — but the self-hosted version (MIT-licensed) is completely free beyond your own infrastructure costs, making it effectively unlimited for teams willing to run their own deployment. For solo developers and small teams, Stockyard is also free in the self-hosted sense: a single binary with no cloud subscription required.
Does migrating from Helicone to Langfuse require changing my application code?
Yes — unlike Helicone’s proxy model where migration is a URL swap and a header change, Langfuse uses SDK-based instrumentation that requires adding the Langfuse SDK to your application. For LangChain users, the change is adding a LangfuseCallbackHandler to your chain invocations. For direct OpenAI users, you wrap calls with Langfuse’s observe() decorator. The migration typically takes one to four hours for a single-model application and a full day for complex multi-step agent pipelines. The latency benefit — zero added overhead vs Helicone’s proxy round-trip — is the payoff for that code investment.
Is Langfuse safe from acquisition risk given what happened to Helicone?
Langfuse is MIT-licensed and fully self-hostable with no cloud dependency. The self-hosted deployment gives you complete control over your data and zero vendor lock-in — if Langfuse Cloud were ever acquired or deprecated, you could continue running your existing self-hosted instance indefinitely with no service interruption. The MIT license also means you can fork the codebase if needed. For teams specifically concerned about acquisition risk after the Helicone situation, self-hosted Langfuse eliminates that risk category entirely, at a cost of roughly $150/month in infrastructure at 1M traces per month.
How does Portkey differ from Helicone as a proxy-based alternative?
Both Helicone and Portkey route your LLM calls through their infrastructure, but Portkey adds production-grade gateway features that Helicone never offered. Portkey supports 200+ LLM providers with unified routing, which means you can send traffic to OpenAI, Anthropic, Google, and fine-tuned model endpoints through a single integration. It adds automatic failover when a provider returns errors, load balancing across multiple model endpoints, and semantic caching that deduplicates similar prompts before they hit the API — reducing both latency and cost on repeated or near-identical queries. Helicone’s proxy provided observability logging with a network hop; Portkey’s proxy provides observability logging plus active reliability and cost optimization features that make the network hop worthwhile for multi-provider production deployments.
