Future AGI June 2025 Roundup: Inline Evaluations, Audio Error Localizer, Open-Source Eval Library, and New Integrations
Discover Future AGI's June 2025 updates including Inline Evaluations, Audio Error Localizer, open-source AI eval library, TypeScript ADK, Google ADK, Portkey.
Table of Contents
Some ideas stick, others shape-shift - June gave us both. Here’s what we tested, learned, and leveled up. Dive in.
June 2025 Product Updates: Inline Evaluations, Audio Error Localizer, Open-Source Eval Library, and Startup Catalyst Program
Launched Inline Evaluations: How to Run AI Evals Directly in Code Without UI Setup or Workflow Interruption
Evaluations shouldn’t slow you down. With our new Inline Evaluations, you can now run AI evals directly in your code - no UI, no setup, no detours.
Here’s what makes it click:
- Add trace_eval=True and a custom_eval_name to any evaluation call. That’s it.
- It works anywhere - locally, in staging, or in prod, with built-in OpenTelemetry for full traceability.
- Evaluations become part of your natural workflow, helping you move faster without losing visibility.

Drop into code, stay in flow.
👉 Configure Inline Evals → https://docs.futureagi.com/docs/observe/features/manual-tracing/in-line-evals/
Launched Audio Error Localizer: How to Detect Hallucinations and Quality Issues in Voice AI with Phrase-Level Precision
Voice AI systems require precise quality control to maintain accuracy in production environments. Our Audio Error Localizer provides automated detection of errors, hallucinations, and quality issues in AI-generated speech responses.
The system performs phrase-level analysis with timestamp precision, enabling teams to identify specific segments containing factual inaccuracies or quality degradation. This approach replaces generic scoring methods with actionable localization data.

👉 Evaluate your Voice AI now - click here to get started!
Open-Sourced All-In-One AI Eval Library: How Future AGI Unifies Heuristics, Model-Based, and LLM-as-a-Judge Evaluation
We launched the most comprehensive open-source eval library that unifies heuristics, model-based, and LLM-as-a-judge evaluation methods in one platform.
The library includes built-in guardrails, red teaming capabilities, multimodal support, and observability features. A production-ready solution powered by Future AGI’s Turing Models.

👉 Visit Future AGI’s Github Page to know more about this!
Launched Startup Catalyst Program: How Future AGI Helps AI Startups Ship Reliable Products 10x Faster at Zero Cost
Great models ≠ Reliable products | Demo magic ≠ Customer happiness
Backing bold AI startups to accelerate their journey from inspiring demos to trustworthy, scalable products-10x Faster.
Zero conditions. Zero cost. Zero excuses.

👉 If you’re building AI and want to ship faster with confidence - apply now.
June 2025 New Integrations: TypeScript ADK, Google ADK, gRPC OpenTelemetry, and Portkey
We’ve added native support for the tools and frameworks AI teams rely on:
TypeScript ADK: Auto-instrument OpenAI and Anthropic instantly, or manually hook into any LLM with just a few lines of code. Track tokens, latency spikes, and hallucinations-no boilerplate required.
Google ADK: Seamless integration enables native tracing across TraceAI and Future AGI, supporting enterprise-scale observability and fast iteration.
gRPC + OpenTelemetry: For teams running distributed AI services, our gRPC integration delivers efficient, low-latency trace extraction-no custom instrumentation needed.
Portkey: Instantly observe (using Portkey) & evaluate (using Future AGI) your LLM responses in one unified dashboard - no more infra switches.

👉 Get started with Future AGI and integrate with your existing stack in minutes
June 2025 Knowledge Nuggets: MCP Webinar, Accelerate AI Podcast, and Future AGI at SuperAI Singapore
New Webinar on MCP: How to Build Robust MCP with Continuous Quality Checks and Faster Iteration
This webinar delivers a practical roadmap to build robust MCP for instant feedback, surface issues on the fly, and embed continuous quality checks - all while accelerating iteration and reducing engineering overhead.

👉 Watch it now - click here!
Accelerate AI Podcast New Episode: Real-World GenAI Integration Pitfalls and Tactical Lessons for Product Teams
Jorge, CEO of Zentrix, reveals common pitfalls teams face when integrating GenAI into products. Discover tactical lessons from real-world failures, how top PMs leverage AI for clearer roadmaps, and hidden technical challenges that sabotage product launches.

👉 If you are a builder, PM, or just GenAI curious, this is worth your time. Watch it now.
Future AGI at SuperAI Singapore: How Conversations with Builders Validated the Need for Trustworthy GenAI Tools
Our team hit the ground running at the Super AI conference in Singapore - connecting with builders, founders, and AI leaders from around the world. We had some incredible conversations, deep interest in our observability and evaluation stack, and a packed demo booth that validated the real need for trustworthy GenAI tools.
Takeaway: We left with empty swag bags and a much heavier pipeline. :)

Finally Our Team: The People Behind Every Demo, Bug Fix, and Shipped Feature at Future AGI
None of this happens without the incredible team behind it all. From endless client demo videos to late-night pizza-fueled bug fixes, this crew is building, shipping, and learning - together, every day. It’s fast, it’s fun, and it’s full of heart. Here’s to the people making it happen.

Starting July grounded in progress, powered by people, and ready for what’s next.
For more updates, join the conversation in our Slack Community or get in touch with us directly.
Your partner in building Trustworthy AI!
Discover Future AGI's November 2025 updates including voice agent persona testing, outbound call simulation, A/B testing for STT-LLM-TTS stacks, 30-plus.
Future AGI Protect ships multi-modal guardrails for text, image, audio. Sub-100ms text latency, around 109ms image. Toxicity, bias, privacy, prompt injection.
See what Future AGI shipped in September 2025. Covers Agent Compass for 98 percent faster multi-agent debugging, AWS Marketplace launch, enterprise RBAC.