Optimize · xAI Grok
xAI Grok logo

Optimize xAI Grok

LLM Providers

Use Future AGI's agent-opt SDK to rewrite your xAI Grok prompts with measurable improvement on the metrics that matter to you.

python typescript · trace evaluate optimize

Prerequisites

Before you start

  • · A working xAI Grok app — local or already in production.
  • · A free Future AGI account with FI_API_KEY and FI_SECRET_KEY.
  • · Python 3.9+ / Node 18+ / Java 17+ depending on which SDK you're installing.
  • · A dataset of ≥50 examples — Future AGI auto-builds these from your trace history.

Install

pip install traceAI-xai

Optimize recipe

from agent_opt import GEPAOptimizer
from fi.evals.templates import Groundedness, PromptAdherence

optimizer = GEPAOptimizer(
    seed_prompt="<your current xAI Grok system prompt>",
    objectives=[Groundedness(), PromptAdherence()],
    rounds=8,
)

best_prompt, score = optimizer.run(dataset_id="xai_eval_set_v1")
print(f"+{score.delta}% on grounded answers")

What Future AGI captures

Optimize fields you'll see in the dashboard

  • Use a Future AGI dataset of failed xAI Grok traces as the optimisation target

  • GEPA, ProTeGi, PromptWizard, MetaPrompt, Bayesian, and Random optimisers — same interface

  • Each optimiser run produces a new prompt version with a measured score delta

  • Push the best prompt back to your prompt registry and replay through the same eval suite

Common gotchas

Read these before you ship

  1. 01

    Seed prompt must include the placeholder format your dataset uses (`{{question}}`, `{input}`, etc.).

  2. 02

    GEPA needs ≥50 examples to converge; for smaller sets prefer ProTeGi or PromptWizard.

  3. 03

    Set a hard `rounds` cap — optimisers will keep improving past your budget if you let them.

Next: chain it with the other recipes

Optimize is the first step. Most teams add an evaluator the same week, and start optimising or simulating once they have a baseline. Each recipe takes minutes to wire up.