
Building Agentic RE: Automating Reverse Engineering & Vulnerability Research with AI // John McIntosh
Virtual | March 27-31 | 32 Hours
BOOK NOWABSTRACT
Reverse engineering is evolving beyond static tools and manual workflows. This 32 hour hands‑on course introduces a new paradigm: Agentic Workflows. By combining cutting‑edge large language models (LLMs), the Model Context Protocol (MCP), and reverse engineering tools like Ghidra, you’ll learn how to design, train, and orchestrate AI‑powered systems that automate and accelerate complex RE and VR tasks.
The course blends foundational concepts, the latest practices in AI server configuration, programming, and workflow design, custom MCP development, and advanced orchestration—culminating in LLM‑powered agents that act as autonomous collaborators in reverse engineering and vulnerability research.
Through a systematic progression, you’ll move from fundamentals to advanced orchestration:
- Building local AI stacks that ensure privacy, reproducibility, and control.
- Leveraging LLMs to explain, annotate, and reason about binaries.
- Developing custom MCP servers to expose reverse engineering and vulnerability research tools.
- Integrating static and dynamic analysis pipelines with AI‑driven insights.
- Validating findings through automated cross‑checks and building reliable, trustworthy workflows.
- Delivering an integrated agentic workflow that assists in both reverse engineering and vulnerability research.
By the end of the course, you will have built an integrated agentic AI workflow that assists in your reverse engineering and vulnerability research tasks—capable of analyzing binaries, surfacing potential vulnerabilities, validating, and triaging results.
Why This Matters
Modern binaries increasingly resist traditional tooling. By combining human expertise with agentic AI, you can:
- Shorten analysis cycles.
- Surface subtle behavioral patterns.
- Scale research without sacrificing depth or accuracy.
- Automate repetitive triage while keeping humans in the loop.
This course equips you to move beyond brittle prompts into orchestration, where AI becomes a programmable, composable part of your workflow.
INTENDED AUDIENCE
- Reverse Engineers who want to augment their workflows with AI‑driven automation.
- Vulnerability Researchers looking to accelerate bug discovery and triage with agentic frameworks.
- Security Professionals interested in building private, reproducible AI stacks for sensitive analysis.
- Developers & Tool Builders exploring how to extend MCP servers and integrate AI into RE pipelines.
- Applied AI Practitioners who want to move beyond prompt‑hacking into orchestration, reproducibility, and workflow design.
If you’ve ever wished your RE tools could act as autonomous collaborators—not just assistants—this course is for you.
KEY LEARNING OBJECTIVES
Foundations of Agentic RE: Understand the intersection of generative AI, MCP, and reverse engineering.
Private Local LLM Stack: Build and configure your own stack with GhidraMCP, Ollama, and OpenWebUI, with a focus on hardware and performance trade-offs.
Custom MCP Development: Extend MCP servers to expose binary metadata, integrate semantic search, and connect with RE tools.
LLM Training for RE: Create datasets, fine‑tune models with QLoRA, and train models to detect vulnerabilities or identify key functions.
Agentic Workflow Design: Learn DSPy and LangGraph orchestration patterns to build resilient, compositional workflows.
Reliable AI & Validation: Implement automated cross-checks and guardrails to reduce hallucinations and validate AI-generated findings.
Custom RE HUDs: Build interactive dashboards with Chainlit/Streamlit to guide multi‑platform RE analysis.
Capstone Project: Deliver a single HUD with two workflow paths — one for RE, and one for VR that includes discovery, triage, and validation steps.
COURSE OUTLINE
Part 1 – Foundations of Agentic RE
AI here is a computational and systems layer.
You’ll learn the fundamentals of how LLMs operate — tokenization, embeddings, quantization — and what those mean for reverse engineering tasks. We’ll cover considerations for system design, how to enhance LLMs with MCP‑exposed tools, and the client–server architecture that makes interfacing with models possible.
- The Agentic Era: how LLMs transform reverse engineering and vulnerability research.
- LLM basics: tokens, embeddings, quantization.
- Model Selection & Hardware: Trade-offs between model size (7B, 13B, 70B), performance, quantization (QLoRA), and realistic hardware requirements (VRAM).
- Model Context Protocol (MCP): exposing RE tools to LLMs.
- Why local LLMs matter: privacy, reproducibility, and control.
- Lab: Local LLM Stack Setup — Install Ollama, OpenWebUI, LM‑Studio, and connect to GhidraMCP.
- Lab: Agentic‑Assisted Reverse — Use a local LLM to explain code, identify constants, or annotate functions.
Part 2 – Extending the Stack: Custom MCP Servers
AI here is an environment you control.
With the foundational stack running, you'll move from being a user to a builder. This section focuses on extending your private AI ecosystem by creating custom Model Context Protocol (MCP) servers that expose powerful static analysis and reverse engineering tools to your LLMs.
- MCP server basics (Python + FastAPI).
- Designing tool-specific MCPs for structured input and output.
- Lab: Static Analysis MCPs — Expose Semgrep (pattern‑based) and CodeQL (query‑driven) through MCP, compare their outputs on a sample codebase.
- Lab: Custom Ghidra MCP — Use headless scripting to analyze binaries and expose key information like function listings and cross-references.
- Exercise: Multi‑Binary CLI Analysis — Use pyghidra‑mcp to detect reused code, suspicious patterns, and API call flows across multiple binaries.
Part 3 – Custom MCPs & Training LLMs
AI here is a programmable collaborator.
LLMs alone can’t introspect binaries the way RE demands, but MCP lets you expose structured tools and data. You’ll learn to build advanced MCP servers and then train your models to better understand the RE/VR domain.
- Programming with LLMs: context engineering, handling non‑determinism, and designing well‑defined tools.
- Securing Agentic Workflows: Introduction to prompt injection, data sanitization, and securing MCP API endpoints.
- Building advanced custom MCP servers:
- Binary symbols MCP for metadata introspection.
- Ghidra MCP with ChromaDB for semantic code search.
- Training LLMs for RE tasks:
- Data Sourcing & Curation: Strategies for creating, sourcing, and labeling high-quality datasets from opensource code, CVE reports, and internal projects.
- Techniques like QLoRA for efficient fine‑tuning.
- Training models to detect vulnerability classes and identify main functions.
- Lab: Fine‑Tuning Your First Model — Train a small model to detect a specific vulnerability class (e.g., buffer overflow).
- Lab: Function Identification Model — Train a model to find main in disassembly across different architectures.
Part 4 – Orchestration, HUDs & Integrated Capstone
AI here is a workflow partner.
Beyond prompts, you’ll learn orchestration with DSPy and LangGraph — building adaptive systems that plan, execute, and coordinate tasks. You’ll design RE HUDs that visualize workflows and glue everything together into a final, integrated system that discovers, analyzes, and validates findings.
- Beyond prompts: orchestration with DSPy, LangGraph, and MCP‑agents.
- Prompt optimization with MiPROv2 and GEPA.
- Lab: RE HUD Prototype — Build an interactive dashboard with Streamlit/Chainlit to visualize and control workflows.
- Lab: Multi‑Platform Workflow — Create adaptive logic for Windows, Android, and iOS, with LLM‑driven tool selection and high‑level feedback.
- Lab: Orchestrating Static Analysis — Use the Semgrep and CodeQL MCPs from Part 2 in a single LangGraph workflow to compare results and feed findings into AI‑driven triage.
- Lab: Automated Validation and Triage — Build a workflow step that cross-references an LLM's finding against another model or a heuristic check to reduce hallucinations and prioritize credible results.
- Capstone Project: Integrated RE + VR Workflow
- RE Path: Build a workflow that analyzes and explains a binary, integrating Ghidra MCP + Semantic Search MCP, surfaced in the HUD.
- VR Path:
- Discover: Use Semgrep + CodeQL MCPs to surface potential vulnerabilities.
- Analyze & Triage: Apply LLMs to prioritize findings, rename functions, and summarize binaries.
- Validate: Implement an automated cross-check step to increase confidence in the top-ranked vulnerability.
- Deliverable: A single Chainlit‑based HUD with two workflow paths — one for RE, one for VR, including a validation step in the VR path.
Technology Stack
- AI: LLMs, Ollama, OpenWebUI, LM‑Studio
- RE/VR: Ghidra, Semgrep, CodeQL, Tree‑sitter
- Development: Python (primary), MCP SDKs (TypeScript, Go, Rust, etc.)
- Workflow Orchestration: DSPy, LangGraph, dapr‑agents
- UI/Integration: Chainlit, Streamlit
Student Requirements
To get the most out of this training, participants should have:
- Intermediate reverse engineering experience (familiarity with Ghidra, IDA, or similar tools).
- Basic vulnerability research knowledge (understanding of common bug classes and analysis workflows).
- Comfort with scripting in Python (used for MCP servers, orchestration, and workflow glue).
- Familiarity with Linux or macOS command‑line environments for stack setup and automation.
No prior experience with LLMs or AI frameworks is required—we’ll cover the fundamentals before diving into advanced orchestration.
Practical Takeaways
By the end of this course, participants will walk away with:
- A fully configured local RE+LLM stack (Ollama, OpenWebUI, LM‑Studio, GhidraMCP).
- An understanding of hardware trade-offs for running local LLMs effectively.
- Custom MCP servers for binary metadata, semantic search, and static analysis (Semgrep + CodeQL).
- Hands‑on experience fine‑tuning models for RE‑specific tasks (e.g., vulnerability class detection, function identification).
- Reusable workflow templates for binary analysis, vulnerability discovery, and results validation.
- A Chainlit‑based RE HUD that integrates multiple MCPs and provides an interactive interface for analysis.
- An integrated capstone project:
- RE Path: A workflow that analyzes and explains binaries, leveraging Ghidra MCP and semantic search.
- VR Path: A workflow that discovers, triages, and validates potential vulnerabilities using Semgrep, CodeQL, and LLM-driven cross-checks.
- A showcase‑ready system that demonstrates how agentic AI can partner with humans in both reverse engineering and vulnerability research.
YOUR INSTRUCTOR: John McIntosh
John McIntosh @clearbluejar, is a security researcher at Clearseclabs. His area of expertise lies within reverse engineering and offensive security, where he demonstrates proficiency in binary analysis, patch diffing, and vulnerability discovery. Notably, John has developed multiple open-source security tools for vulnerability research, all of which are accessible on his GitHub page. Additionally, his website, https://clearbluejar.github.io/, features detailed write-ups on reversing recent CVEs and building RE tooling with Ghidra. Boasting over a decade of experience in offensive security, John is a distinguished presenter and educator at prominent security conferences internationally. He maintains a fervent commitment to sharing his latest research, acquiring fresh perspectives on binary analysis, and engaging in collaborative efforts with fellow security enthusiasts.
60+ days before the event 75% of fees refunded; 45-60 days before event 50% refunded, less than 45 days 0% refunded. Course changes are allowed up to 14 days before event start (some restrictions will apply). Attendee changes can be accommodated up to 14 days prior to the event.
Note: In the event of a class cancellation, Ringzer0 will endeavor to offer transfer to another training at no additional charge.