Smartflow 1.6 released — Any-to-any provider routing, Trace UI, Prometheus /metrics, OIDC SSO. Read release notes →
Technical Whitepaper · Competitive Analysis

Smartflow Enterprise vs. TrueFoundry
Purpose-Built AI Gateway vs. MLOps Platform

An objective technical evaluation examining whether a broad MLOps platform or a purpose-built AI governance gateway better serves the needs of enterprise LLM deployments in regulated industries.

AuthorScott Ancheta, CTO — LangSmart
PublishedFebruary 7, 2026
CategoryEnterprise AI Infrastructure
Verdict Smartflow recommended for AI gateway use cases

Executive Summary

Abstract

TrueFoundry is a comprehensive MLOps platform designed to manage the full machine learning lifecycle — from model training and experiment tracking to serving and monitoring. It includes LLM gateway capabilities as part of a broader suite. Smartflow Enterprise is a purpose-built AI gateway focused exclusively on the runtime governance, security, and efficiency of LLM calls in production.

The fundamental distinction: TrueFoundry asks "how do I manage every ML asset in my organisation?" Smartflow asks "how do I make every AI API call in my organisation safe, fast, compliant, and auditable?" These are different problems. Organisations that need both will often deploy both. Organisations evaluating them for AI gateway use cases specifically will find Smartflow significantly deeper in every gateway-relevant dimension.

Recommendation: For organisations that already operate a mature MLOps platform and need a dedicated AI gateway layer, Smartflow is the superior choice. For early-stage ML teams seeking an all-in-one platform, TrueFoundry's integrated approach may reduce initial tooling sprawl — though its LLM gateway features are less mature.

Product Overview

TrueFoundry

TrueFoundry is a Kubernetes-native MLOps platform founded in 2022 with backing from prominent venture investors. It provides a unified control plane for deploying, scaling, and monitoring machine learning models and applications across cloud environments. Its LLM gateway — sometimes marketed separately as "TrueFoundry AI Gateway" — provides basic LLM routing, virtual key management, and usage tracking.

TrueFoundry's strengths lie in its breadth: experiment tracking, model registry, batch inference, fine-tuning workflows, and real-time serving all sit under one umbrella. This makes it a compelling platform for data science teams building end-to-end ML pipelines who want to avoid integrating multiple best-of-breed tools.

Smartflow Enterprise

Smartflow Enterprise (by LangSmart) is a Rust-based enterprise AI gateway with zero MLOps scope — it does not manage model training, experiment tracking, or fine-tuning. Its entire surface area is dedicated to the single problem of governing, securing, optimising, and auditing LLM API calls in production. This focus translates into depth: a four-phase semantic cache, a no-code policy engine, enterprise SSO integration, MCP gateway, A2A orchestration, and a comprehensive management dashboard that TrueFoundry's LLM gateway does not approach.

March 2026 — SDK v0.4.0: Smartflow now ships a native Python SDK that works with or without a gateway deployment. Developers can start in direct mode (calling OpenAI, Anthropic, Gemini, Ollama directly) and add a Smartflow gateway later with zero code changes. See SDK Reference →
Complementary, not competing: Many enterprise deployments use TrueFoundry for ML lifecycle management and Smartflow as the AI gateway layer. They integrate cleanly — Smartflow sits in front of any TrueFoundry-deployed model serving endpoint as transparently as it sits in front of OpenAI.

Feature Comparison Matrix

CapabilitySmartflow EnterpriseTrueFoundry
Semantic Cache (BERT KNN) 4-phase with VectorLite embedding index Basic exact-match cache only
Policy Engine / Guardrails Visual editor, PII, topic, jailbreak, output filter~ Basic rate limiting and content filtering hooks
Enterprise SSO / LDAP Entra ID, LDAP, SAML, OIDC, proxy headers~ OIDC for dashboard; not proxy-level identity passthrough
Per-User Audit Trail Full VAS trace per request, tied to SSO identity~ Team/project level; not per-individual at request level
Compliance Dashboard Built-in sandbox, policy test, report generation Not available
MCP Gateway JSON-RPC, SSE, STDIO, tool caching Not supported
A2A Agent Orchestration Built-in A2A agent registry Not supported
Model Experiment Tracking Out of scope MLflow-compatible experiment tracking
Model Registry Out of scope Versioned model registry
Fine-tuning Workflows Out of scope PEFT, LoRA, full fine-tune pipelines
LLM Provider Support 37+ incl. all local models Major providers + self-hosted
Proxy Latency Overhead <5ms (Rust binary)~ 20–60ms (Python/Go service)
Helm Chart Maturity Production HPA, PDB, NetworkPolicy Mature Helm (MLOps-focused)
Prometheus Metrics Native /metrics endpoint Integrated with platform metrics
GPU / Compute Scheduling Out of scope Spot instance, GPU scheduling, autoscaling

AI Gateway Depth: A Critical Difference

The LLM gateway component within TrueFoundry is designed to solve a routing and access control problem: which team can call which model, with what rate limits and budgets. This is a necessary but shallow capability. Enterprise LLM governance requires substantially more.

What TrueFoundry's LLM Gateway Provides

TrueFoundry's gateway offers virtual key issuance, provider routing, basic rate limiting, and spend tracking by project. It is sufficient for internal tooling deployments where the primary concern is "who spent how much on which model." It integrates naturally with TrueFoundry's broader MLOps context, sharing authentication and project structure with the rest of the platform.

What Enterprise Deployments Actually Need

In regulated industries — healthcare, education, financial services, legal — an AI gateway must address questions that go far beyond routing:

  • Did this specific employee's prompt contain PHI? (HIPAA)
  • Was a student's data sent to an AI provider without consent? (FERPA)
  • Did any AI-generated output reference material non-public information? (SOX)
  • Can I produce an audit trail of every AI interaction for compliance review?
  • Can I enforce different content policies for contractors vs. employees?
  • Can I prevent AI jailbreaks and prompt injection in real-time?

These requirements map directly to Smartflow's feature set and do not have equivalents in TrueFoundry's LLM gateway.

Depth example: A Smartflow policy evaluation inspects every prompt through a semantic PII detector, a topic classifier, a jailbreak pattern analyser, and an output moderator — in under 5ms total — before the request reaches the AI provider. TrueFoundry's gateway does not offer these capabilities.

Semantic Caching & Cost Reduction

TrueFoundry's LLM gateway does not offer semantic caching at the time of this writing. Responses are not cached across requests, meaning every prompt — regardless of semantic similarity to previous queries — incurs a full API call to the provider.

Smartflow's four-phase MetaCache (exact match → BERT KNN semantic similarity → model compression → predictive pre-caching) typically achieves 55–75% cache hit rates in enterprise deployments. On a workload of 10,000 daily requests at $0.01 average cost, this represents $550–$750/day in avoided provider costs — costs that accumulate silently in a TrueFoundry deployment.

Over a 12-month period with 100,000 monthly LLM API calls at $0.01/call, the cost differential between Smartflow's semantic cache and no caching is approximately $66,000–$90,000 per year in avoided API costs. This figure alone typically exceeds Smartflow's annual enterprise licensing cost.

Compliance, Policy & Regulatory Readiness

TrueFoundry is designed for ML engineering teams and their operational concerns. Compliance workflows — the kind required by legal, HR, and risk teams in regulated industries — are outside its design scope. The platform has no compliance dashboard, no policy test sandbox, and no pre-built regulatory templates.

Smartflow's compliance engine is designed specifically for this gap:

  • Compliance Test Sandbox: Run any prompt through the active policy engine before deployment, see exactly which rules trigger and why.
  • Pre-built policy templates: HIPAA/PHI, FERPA student data, SOX financial data, legal privilege, competitive intelligence — import and customise rather than build from scratch.
  • Automated reporting: Generate compliance reports showing policy trigger counts, blocked request categories, and per-user compliance risk scores.
  • Real-time enforcement: Policies execute synchronously in the request path — blocks happen before prompts reach the AI provider, eliminating data leakage risk.

Total Cost of Ownership Analysis

A common objection to adding a dedicated AI gateway is the cost of an additional platform. The TCO analysis reverses this concern:

Smartflow TCO Factors
  • Enterprise licensing (replaces engineering cost)
  • 55–75% reduction in LLM API spend via semantic cache
  • Zero custom compliance engineering required
  • Zero custom SSO integration engineering
  • Single Rust binary — minimal infrastructure overhead
  • Reduces compliance audit risk and associated costs
TrueFoundry Gateway TCO Factors
  • Platform licensing (covers full MLOps suite)
  • No semantic caching — full API cost on every request
  • Custom compliance engineering required
  • Custom SSO passthrough engineering required
  • Additional tooling needed for policy enforcement
  • Broader platform = higher operational complexity

For organisations whose primary evaluation criterion is the AI gateway layer (not the broader MLOps suite), Smartflow's cost structure is typically accretive within 60–90 days of deployment when semantic cache savings are accounted for.

Deployment Model

TrueFoundry is a cloud-native platform designed to run across cloud providers with its own control plane. This makes initial setup streamlined but introduces a dependency on TrueFoundry's infrastructure layer for configuration and orchestration.

Smartflow Enterprise deploys as Docker containers or via a Helm chart — no external control plane, no vendor-specific infrastructure layer. All configuration lives in environment variables or Kubernetes ConfigMaps/Secrets that operators control directly. This aligns with the operational practices of infrastructure teams in air-gapped or strict data residency environments.

Air-gap compatibility: Smartflow can run entirely offline — its BERT semantic cache runs locally with no external embedding API calls, and all policy evaluation is on-device. TrueFoundry's cloud-connected control plane model is less compatible with strict air-gapped deployments.

When to Choose Each Platform

Choose TrueFoundry when:

  • You need an integrated MLOps platform covering training, registry, and serving — not just a gateway
  • Your team manages custom model training workflows (LoRA, PEFT, full fine-tune)
  • You want GPU scheduling and experiment tracking in the same platform
  • Your compliance requirements are minimal or handled by other tooling
  • You are comfortable with the TrueFoundry platform's operational model

Choose Smartflow Enterprise when:

  • Your primary need is AI gateway governance, not full MLOps
  • You require semantic caching to reduce LLM API costs at scale
  • You have regulatory compliance requirements (HIPAA, FERPA, SOX, GDPR)
  • Per-user identity and audit trail are required for every AI request
  • You are deploying to a self-hosted Kubernetes cluster with strict data residency
  • You need a policy engine that non-engineers can operate through a UI
  • You already have (or don't need) an MLOps platform and want a best-of-breed gateway

Fair Assessment: Where TrueFoundry Leads

  • MLOps breadth: TrueFoundry's model registry, experiment tracking, batch inference, and fine-tuning capabilities are well-developed and have no equivalent in Smartflow. For organisations building custom models, TrueFoundry offers substantially more value.
  • Platform integration: TrueFoundry's gateway is natively integrated with its model serving layer. There is no seam to manage. Smartflow requires a network hop to the AI provider (external or internal), which TrueFoundry eliminates for internally-hosted models.
  • Vendor ecosystem: TrueFoundry has a strong VC-backed roadmap and integrations with major cloud providers. Smartflow is commercially supported but has a smaller ecosystem footprint.
  • Unified billing: For cost-conscious procurement, TrueFoundry bundles gateway features into its platform license. Smartflow is an additional vendor relationship.

Conclusion

TrueFoundry and Smartflow Enterprise serve different buyers with different problems. TrueFoundry is the right choice for ML engineering teams that need a unified platform spanning the complete model lifecycle. Its LLM gateway features are a useful inclusion but are not its primary engineering investment.

Smartflow Enterprise is the right choice for organisations where the AI gateway layer is the primary concern — where the questions of compliance, identity, semantic caching, policy enforcement, and per-user auditability need answers, and where those answers need to be available to non-engineering stakeholders through a management dashboard.

In practice, many mature AI organisations run TrueFoundry for ML lifecycle management and Smartflow as the governance and efficiency layer in front of their LLM endpoints. The two platforms are complementary rather than mutually exclusive. For organisations evaluating them head-to-head as an AI gateway, Smartflow's depth, performance, and compliance capabilities represent a purpose-built advantage that a general MLOps platform's gateway module cannot match.

Next steps: Request a Smartflow Enterprise demo at langsmart.ai to see Smartflow in your environment.
SA
Scott Ancheta
Chief Technology Officer, LangSmart
Scott Ancheta is the CTO and co-founder of LangSmart. With over 25 years of experience in enterprise software architecture and development, advanced networking, and AI infrastructure, Scott leads the technical architecture of the Smartflow platform and advises enterprise customers on AI governance strategy. He focuses on making AI deployments safe, compliant, and economically efficient at scale. Published February 7, 2026.