Staxly

LangSmith vs Portkey

LLM observability, testing & evaluation — by LangChain
vs. Enterprise AI gateway + observability + guardrails + prompt mgmt

LangSmith websitePortkey website

Pricing tiers

LangSmith

Developer (Free)
Free forever. 5,000 traces/month. 14-day retention. 1 seat. Basic evaluations.
Free
Plus
$39/seat/month. 10k base traces included ($2.50 per 1k overage). Full evaluations, custom dashboards, email support.
$39/mo
Enterprise
Custom. Self-host option, SSO, custom retention, dedicated support.
Custom
LangSmith website

Portkey

Developer (Free)
Free forever. 10k logs/month. Universal API + key management. 3 prompt templates. Basic observability.
Free
Gateway (OSS)
MIT-licensed gateway only (no observability UI). Self-host for routing/fallbacks.
$0 base (usage-based)
Production
$49/month. 100k logs ($9 per additional 100k). Fallbacks, load balancing, retries, semantic caching. Unlimited prompts. RBAC.
$49/mo
Enterprise
Custom. 10M+ logs/month. Custom guardrails, advanced evals, SSO, budget controls, VPC + on-prem, SOC2, HIPAA, GDPR.
Custom
Portkey website

Free-tier quotas head-to-head

Comparing developer on LangSmith vs free on Portkey.

MetricLangSmithPortkey
No overlapping quota metrics for these tiers.

Features

LangSmith · 14 features

  • AlertsThreshold alerts on latency, cost, eval metrics.
  • Annotation QueuesHuman-review workflows for trace quality rating.
  • Custom DashboardsAggregate metrics dashboards per project/tag.
  • DatasetsCollect examples → use as eval sets or training data.
  • EvaluationsLLM-as-judge, embedding similarity, custom Python evaluators, offline batch eval
  • LangChain IntegrationAuto-trace any LangChain/LangGraph run with env var.
  • LangGraph IntegrationFirst-class trace + eval for LangGraph agents.
  • LLM TracingAutomatic trace every LLM call + tool call + chain step.
  • OpenTelemetry ExportExport traces as OTLP to Datadog/Honeycomb/etc.
  • PlaygroundTest prompts + models inline before deploying.
  • Prompt CanvasVisual prompt editor with live test + eval.
  • Prompt HubPublic + private prompt library with versioning.
  • Self-Hosted (Enterprise)Docker + k8s deployment in your infra.
  • Threads + SessionsGroup traces into conversational sessions.

Portkey · 18 features

  • AI GatewayUnified OpenAI-compatible API to 250+ LLMs.
  • AlertsThresholds on latency, error rate, cost, usage.
  • Budget ControlsPer-key + per-team spending limits.
  • EvaluationsBuilt-in evaluator templates + custom.
  • FallbacksConfig-driven provider fallback chains.
  • GuardrailsPre/post processors for safety + compliance.
  • Load BalancingRound-robin, weighted, least-latency across providers.
  • MCP SupportUse MCP servers as tools through gateway.
  • ObservabilityLogs, traces, feedback, alerts, cost tracking.
  • OSS GatewayOpen-source gateway (portkey-ai/gateway).
  • Prompt LibraryShared prompt library + public marketplace.
  • Prompt TemplatesVersion + test + collaborate on prompts.
  • RetriesConfigurable retry policies per route.
  • Role-Based Access ControlTeam permissions on prompts + keys.
  • Semantic CachingVector-based cache on query meaning.
  • Simple CachingExact-match cache.
  • Virtual KeysPer-app keys with budget + rate limits + permissions.
  • VPC Deployment (Ent)Deploy in your own VPC for compliance.

Developer interfaces

KindLangSmithPortkey
CLILangSmith CLIPortkey CLI
SDKlangsmith-js, langsmith-pythonportkey-ai (Node), portkey-ai (Python)
RESTLangSmith REST APIPortkey API (OpenAI-compat)
MCPLangSmith MCPPortkey MCP
OTHERLangSmith DashboardPortkey Dashboard
Staxly is an independent catalog of developer platforms. Outbound links to LangSmith and Portkey are plain references to their official websites. Pricing is verified against vendor pages at publication time — reconfirm before buying.

Want this comparison in your AI agent's context? Install the free Staxly MCP server.