Models

What Is Confidential Containers?

Kubernetes pods that run inside hardware Trusted Execution Environments so workload memory and code stay isolated from the host.

What Is Confidential Containers?

Confidential Containers are a Kubernetes model-infrastructure pattern, based on the CNCF CoCo project, that runs pods inside hardware Trusted Execution Environments (TEEs) so model code, weights, prompts, and runtime memory stay isolated from the host, hypervisor, and cluster operator. In production LLM systems, they support shared GPU inference with cryptographic attestation for each pod. FutureAGI observes CoCo-backed workloads through redacted traceAI telemetry and local evaluators that can run inside the trusted boundary.

Why Confidential Containers Matter in Production LLM Systems

Ignoring Confidential Containers turns regulated model serving into a trust exercise. A bank can encrypt model weights at rest, enforce IAM, and run Kubernetes policies, yet still leave prompts, KV-cache contents, and decrypted weights visible to a privileged node operator or compromised host. That risk matters when the workload handles PHI, financial records, source code, or customer prompts that contracts prohibit from leaving a hardware-isolated boundary.

The pain lands on several teams. Platform engineers must prove that a vLLM or NVIDIA NIM pod cannot be inspected by another tenant. Compliance teams need evidence that the exact pod image was attested before keys were released. SREs need to distinguish a normal cold start from an attestation failure that should block traffic. Applied engineers need enough telemetry to debug latency, token cost, and evaluator failures without exporting raw regulated content.

Compared with AWS Nitro Enclaves, which isolate selected processes outside the Kubernetes pod model, Confidential Containers keep the developer workflow closer to standard Kubernetes scheduling and deployment. The tradeoff is operational detail: runtime classes, key release policy, GPU support, trace redaction, and evaluator placement must all line up. In multi-step agentic systems, one unsafe exporter can leak more than a single prompt because the trace may contain retrieved documents, tool arguments, intermediate plans, and final answers.

How FutureAGI Handles Confidential Containers

FutureAGI handles Confidential Containers as a boundary-aware observability and evaluation workflow, not as a container runtime. Kata Containers, containerd, and Kubernetes own the pod runtime. FutureAGI records what happened around the model call while letting the deployment decide which data can cross the TEE boundary.

  1. Inside the pod, traceAI-vllm captures model spans, latency, token counts, and evaluator outputs. Sensitive fields are redacted before export, while low-risk OTel attributes such as llm.token_count.prompt and route identifiers can leave as telemetry.
  2. Local evaluators such as Faithfulness, Groundedness, JSONValidation, and EmbeddingSimilarity can run beside the model when prompts and responses must stay inside the trusted pod.
  3. Cloud-template evaluators such as DataPrivacyCompliance, PII, and IsCompliant are reserved for non-regulated routes or payloads that have already been redacted to the team’s policy.

A practical workflow starts with a CoCo-backed vLLM pod that attests to the customer’s KMS before receiving model-weight keys. The engineer routes traffic only after attestation succeeds, runs local evals in the pod, and exports redacted traceAI spans to FutureAGI. If the eval-fail-rate-by-cohort rises or a redaction policy misses a sensitive span field, the release gate blocks the route and the team fixes the image, exporter, or evaluator placement before rollout.

FutureAGI’s approach is observability without payload escape: keep regulated content inside the trusted execution boundary, but preserve enough trace, score, and attestation metadata to debug production failures.

How to Measure Confidential Containers

Measure Confidential Containers by verifying both the isolation boundary and the observability path. Useful signals include:

  • Attestation success rate: percentage of pod starts that produce a verified attestation receipt before traffic or key release.
  • Redaction coverage: percentage of exported span fields with explicit redaction policy; any unclassified prompt, response, or tool field is a leak risk.
  • Evaluator placement: share of Faithfulness, Groundedness, or JSONValidation runs executed inside the TEE versus outside it.
  • Privacy gate score: PII or DataPrivacyCompliance results on redacted, allowed routes; failures should block export or trigger escalation.
  • Trace drift: unexpected changes in llm.token_count.prompt, model route, attestation hash, or eval-fail-rate-by-cohort after a new pod image.

For dashboards, break these metrics down by model route, runtime class, pod image, and customer cohort so a failed canary is not averaged into healthy traffic.

Minimal local evaluator check inside the trusted pod:

from fi.evals import Faithfulness, JSONValidation

faith = Faithfulness().evaluate(output=model_response, context=retrieved_context)
schema = JSONValidation(schema=expected_schema).evaluate(output=model_response)
print(faith.score, schema.score)

Common mistakes

  • Treating CoCo as a runtime toggle. The runtime class is only one part; exporter redaction, evaluator placement, KMS release, and GPU coverage must match.
  • Exporting full traces from trusted pods. Agent traces often include retrieved documents, tool arguments, hidden system instructions, customer context, and chain state.
  • Skipping attestation checks. Running inside a TEE without verifying the attestation receipt gives auditors no binding evidence; configure KMS to reject that pod.
  • Using hosted judges on regulated routes. Keep raw-content evaluation local, or evaluate only redacted payloads that policy allows to leave; log each exception with owner.
  • Ignoring route drift. A canary image with a different attestation hash should be treated as a separate release, not the same deployment.

Frequently Asked Questions

What are Confidential Containers?

Confidential Containers are Kubernetes pods that run inside hardware TEEs, isolating model code, weights, prompts, and memory from the host, hypervisor, and cluster operator.

How are Confidential Containers different from confidential computing?

Confidential computing is the broader hardware-enforced isolation model. Confidential Containers are the Kubernetes-native packaging pattern that applies that model to pods.

How do you measure Confidential Containers?

Measure attestation success, redaction coverage, and evaluator placement. FutureAGI can attach `traceAI-vllm` spans, `llm.token_count.prompt`, and local evaluator outputs to the trusted pod record.