Webinars

Cybersecurity with GenAI and Intelligent Agents (2026 Replay): How to Build Proactive Security Systems

Webinar replay on cybersecurity with GenAI and intelligent agents in 2026. Predictive threat detection, autonomous response, runtime guardrails for AI agents.

·
Updated
·
4 min read
webinars cybersecurity agents 2026
Cybersecurity with GenAI webinar cover
Table of Contents

Watch the Cybersecurity GenAI Webinar Replay

Security operations are moving from reactive alert triage to predictive, agent-driven defense, and the governance layer is where most teams trip.

TL;DR: Cybersecurity with GenAI in 2026

LayerWhat it does in a security GenAI stackWhere Future AGI fits
Detection copilotSummarizes alerts, links related events, proposes remediationFaithfulness eval on each summary via ai-evaluation
Autonomous response agentPlans and executes triage and containment under policyTrajectory and goal-completion eval, traceAI spans
Threat hunting agentGenerates hypotheses, drives deterministic SIEM queriesSpan-level instrumentation across hunt iterations
Runtime guardrailsInline policy on outbound tool calls, PII and prompt-injection blockingAgent Command Center at /platform/monitor/command-center
Red-teaming and evaluationRegression-style runs against adversarial corporafi.simulate.TestRunner plus ai-evaluation evaluators
Audit and governanceDecision spans stitched into the same trace as the agent callOTel-compatible export from traceAI

About the Cybersecurity GenAI Webinar

In this session, Harsh walks through what it takes to build intelligent cybersecurity systems that go beyond traditional reactive security. From understanding generative AI applications to implementing autonomous decision-making agents, the talk covers how security teams are creating platforms that think, adapt, and defend proactively.

The 2025 talk focused on early architecture choices and the case for generative copilots in the SOC. The 2026 replay framing adds the production-grade layers that most teams missed the first time: agent tracing across SIEM and SOAR integrations, per-turn faithfulness and trajectory evaluation, inline runtime guardrails for prompt injection and PII, and the red-teaming loop that catches regressions before they reach production.

Who Should Watch

This webinar is for cybersecurity professionals, IT leaders, and security architects shipping or evaluating AI-agent features inside detection, response, and threat-hunting workflows. It is also useful for compliance and risk leaders who need a working model for AI-agent risk inside existing security programs. A working understanding of SOC operations and modern detection stacks is assumed; the LLM mechanics are introduced before the security-specific design.

Why It Matters in 2026

Mean-time-to-detect and mean-time-to-respond can improve substantially in AI-augmented SOCs when agents are governed well, but the failure modes shift with them. Prompt injection through indirect inputs (a malicious payload buried in a phishing email body) can hijack an analyst copilot; an over-confident summary can mask an alert; an autonomous agent can execute an irreversible response before a human reviews it. The webinar covers how to capture the speed gains without inheriting the new failure modes.

What the Webinar Covers

This is a working session, not a generic AI talk:

  • Transform threat detection from reactive triage to predictive intent-aware analysis.
  • Build autonomous security responses that operate faster than human cadence, under tight policy.
  • Understand AI agents in security: how they learn attack patterns, adapt defenses, and where they fail.
  • Build intelligent threat-hunting systems that pair generative hypothesis generation with deterministic SIEM queries.
  • Walk through real-world case studies of agent implementations in production SOCs.
  • Address implementation challenges: ethics, governance, analyst trust, and operational concerns.
  • Prepare for a future where AI agents are increasingly used for first-line response under explicit human approval gates.

Key Insight: From Detect-and-Respond to Predict-and-Prevent

The strategic shift is from “detect and respond” to “predict and prevent”: agents read intent signals across the stack, propose containment ahead of impact, and execute under explicit policy. Security teams move from firefighter to architect.

Wiring Observability, Evaluation, and Guardrails Into a Security Agent

A production security agent needs three open-source pieces wired together: tracing, per-turn evaluation, and runtime safety. The pattern is the same one the live demo uses.

from fi_instrumentation import register, FITracer
from fi.evals import evaluate

# Register a tracer at process boot
tracer_provider = register(
    project_name="soc-triage-agent",
    project_version_name="v1",
)
tracer = FITracer(tracer_provider)

# After the agent drafts an incident summary, score it
result = evaluate(
    "faithfulness",
    output="Drafted incident summary for analyst review.",
    context="Raw telemetry, related alerts, and policy context.",
    model="turing_flash",
)
print(result.score, result.reason)

turing_flash runs at roughly 1 to 2 seconds, turing_small at 2 to 3 seconds, and turing_large at 3 to 5 seconds per the cloud evals reference. Authentication uses FI_API_KEY and FI_SECRET_KEY environment variables. Outbound tool calls (SOAR playbook execution, IDP session revocation, EDR isolation) should route through the Agent Command Center so deterministic guardrails block prompt injection, validate parameters, and capture audit-grade decision spans.

For red-team regressions, the simulate module drives adversarial corpora through the agent and grades the responses with the same evaluators used in production.

Key Takeaways for Security Teams

  • The 2026 win is not a single chat copilot, it is a constrained autonomous agent with traces, per-turn evals, and inline guardrails on every tool call.
  • Red-teaming is a regression discipline, not a one-time exercise. Every model and prompt change reruns the adversarial corpus.
  • Human-in-the-loop is non-negotiable on irreversible actions. The Agent Command Center policy layer enforces approval gates without slowing routine triage.
  • Audit logs and decision spans are the prerequisite for shipping AI inside regulated workflows.

Further Reading and Primary Sources

Book a Future AGI demo to see the reference security agent (triage, containment, red-team regression) running end to end.

Frequently asked questions

How is GenAI changing cybersecurity in 2026?
GenAI is shifting cybersecurity from reactive detect-and-respond to predictive predict-and-prevent. Large models read endpoint, network, identity, and SaaS telemetry together, summarize alerts in plain English, and propose remediation paths that previously required a senior analyst. AI agents extend that further by chaining tools (SOAR, ticketing, IDP) to actually execute responses under policy. The result is faster mean-time-to-detect and faster mean-time-to-respond, with the catch that the agents themselves need governance: red-teaming for prompt injection, runtime guardrails for tool use, and observability that shows exactly what an agent did and why.
Who should watch the cybersecurity GenAI webinar?
Cybersecurity professionals, IT leaders, security architects, and platform engineers building or evaluating next-generation defense systems. The session is also useful for compliance and risk leaders deciding how to scope AI-agent risk inside an existing security program. It assumes a working understanding of SOC operations and modern detection stacks. No prior LLM background is required; the talk introduces the key generative-AI mechanics before getting into security-specific design.
What does an autonomous security agent actually look like?
An autonomous security agent receives a goal (such as triage and respond to phishing alerts), a tool set (SIEM, EDR, IDP, ticketing), and a policy bundle. It reasons about the alert, queries telemetry, opens a ticket, optionally executes a contained response (isolate device, revoke session), and writes a runbook entry. The agent's planning step is an LLM call; every tool call is an instrumented span. Reliability rests on three things: a tight policy layer that constrains what the agent can do, runtime guardrails that block unsafe outputs and tool inputs, and evaluation that scores trajectory and goal completion before agents ship to production.
What are the top risks of GenAI in a security stack?
Five risks dominate in 2026. Prompt injection through indirect inputs (web pages, tickets, attachments), data exfiltration via tool calls (an agent quietly emails customer data), hallucinated facts in incident summaries, over-confident escalations that crowd out human analysts, and supply-chain risk from third-party model and tool integrations. Each risk has a counter: input sanitization and re-prompting, deterministic guardrails on outbound tool calls, faithfulness evaluation on every summary, an inline human approval gate on irreversible actions, and SBOM-style governance on the agent toolchain.
How do AI agents integrate with existing SIEM and SOAR platforms?
Most production deployments wrap the LLM agent as a SOAR action or a sidecar service that the SIEM can dispatch to. Telemetry flows from the SIEM into the agent, the agent calls SOAR playbooks or tools directly through their APIs, and outputs go back as enriched cases. The agent does not replace SIEM or SOAR; it sits between them. The webinar walks through reference architectures for Splunk, Microsoft Sentinel, and CrowdStrike Falcon Next-Gen SIEM, and covers the audit-log requirements teams must hit for SOC 2, ISO 27001, and similar reviews.
What observability and evaluation pieces are needed for production agents?
Three layers. First, traces: every agent dispatch, LLM call, tool call, and decision is a span. Second, per-turn evaluation: faithfulness on summaries, toxicity and PII detection on outputs, trajectory efficiency on multi-step actions. Third, runtime guardrails: deterministic policies on outbound tool calls, including prompt-injection detection on inbound content. The webinar shows the open-source pieces (traceAI and ai-evaluation, both Apache 2.0) and how they pair with the Agent Command Center at `/platform/monitor/command-center` for runtime enforcement.
What does red-teaming look like for a security GenAI agent?
Adversarial testing against the agent itself. Teams build a corpus of prompt-injection payloads embedded in tickets, attachments, and email bodies, then run those against the agent under controlled conditions and score whether the agent leaked secrets, exfiltrated data, or escalated incorrectly. The cadence is regression-style: every model swap, prompt change, or new tool triggers a rerun of the adversarial corpus. The Future AGI simulation runner can drive this loop with persona-based scenarios; the ai-evaluation library scores the responses against the rubric.
Where can I dig deeper after the webinar?
Three follow-ups. Read the LLM safety and compliance guide for the regulatory frame. Scan the traceAI and ai-evaluation repos to wire the open-source agent governance pieces locally. Talk to the Future AGI team about a sandbox where you can test an agent end-to-end with red-team payloads, runtime guardrails, and the Agent Command Center policy layer. Links are in the further-reading section below.
Related Articles
View all
Stay updated on AI observability

Get weekly insights on building reliable AI systems. No spam.