Optimize Azure OpenAI
Cloud Platforms
Use Future AGI's agent-opt SDK to rewrite your Azure OpenAI prompts with measurable improvement on the metrics that matter to you.
Recipes for Azure OpenAI
Prerequisites
Before you start
- · A working Azure OpenAI app — local or already in production.
- · A free Future AGI account with
FI_API_KEYandFI_SECRET_KEY. - · Python 3.9+ / Node 18+ / Java 17+ depending on which SDK you're installing.
- · A dataset of ≥50 examples — Future AGI auto-builds these from your trace history.
Install
pip install traceAI-openaiOptimize recipe
from agent_opt import GEPAOptimizer
from fi.evals.templates import Groundedness, PromptAdherence
optimizer = GEPAOptimizer(
seed_prompt="<your current Azure OpenAI system prompt>",
objectives=[Groundedness(), PromptAdherence()],
rounds=8,
)
best_prompt, score = optimizer.run(dataset_id="azure-openai_eval_set_v1")
print(f"+{score.delta}% on grounded answers")What Future AGI captures
Optimize fields you'll see in the dashboard
-
Use a Future AGI dataset of failed Azure OpenAI traces as the optimisation target
-
GEPA, ProTeGi, PromptWizard, MetaPrompt, Bayesian, and Random optimisers — same interface
-
Each optimiser run produces a new prompt version with a measured score delta
-
Push the best prompt back to your prompt registry and replay through the same eval suite
Common gotchas
Read these before you ship
- 01
Seed prompt must include the placeholder format your dataset uses (`{{question}}`, `{input}`, etc.).
- 02
GEPA needs ≥50 examples to converge; for smaller sets prefer ProTeGi or PromptWizard.
- 03
Set a hard `rounds` cap — optimisers will keep improving past your budget if you let them.
Next: chain it with the other recipes
Optimize is the first step. Most teams add an evaluator the same week, and start optimising or simulating once they have a baseline. Each recipe takes minutes to wire up.
Adjacent integrations
More integrations like Azure OpenAI
Vertex AI
Google Cloud's hosted Gemini, Anthropic, and Llama endpoints.
AWS Bedrock
Amazon Bedrock invocation across Claude, Llama, Mistral, Nova, and Titan.
IBM watsonx
IBM watsonx.ai foundation models for regulated workloads.
Replicate
Run open-source AI models on Replicate's serverless GPUs.