GenAI Security Research & Development
Applied research at the intersection of generative AI and cybersecurity since GPT-2 and Nvidia Jarvis (2019-2020 at JHUAPL), years before the current wave. Two focus areas: using AI to detect threats, and detecting threats to AI systems.
Continuous learner: Coursera AI Engineer Agentic Track • HackTheBox AI Red Teamer
GenAI Security
Prompt Injection Detection via ES|QL
Hybrid detection combining regex pre-filters with LLM-as-judge classification. Handles evasion via tokenization, encoding, and context manipulation.
/* Rule-based pre-filter */
| EVAL r_ignore = CASE(
txt RLIKE "(ignore previous|disregard)",
"ignore_previous", NULL)
| EVAL r_override = CASE(
txt RLIKE "override.*(rules|policy)",
"override_phrasing", NULL)
/* LLM classification */
| COMPLETION llm_out = judge_prompt
WITH { "inference_id": "bedrock" }
| DISSECT llm_out "label=%{label} score=%{score}"
| WHERE label == "override" AND score >= 0.70Prompt Injection Taxonomy
Tested and documented 16+ attack categories:
Adversarial Guardrail Testing
GenAI Telemetry Standards↗
MITRE ATLAS Detection Rules↗
AWS Bedrock Detection & Hunting↗
Multimodal Threat Vectors
AI Pipeline Supply Chain
Adversarial Robustness
GenAI Development
Agentic Alert Triage↗
Multi-agent system using hypothesis testing (H0: benign vs H1: malicious) with evidence layers:
- Context-1: Alert fields, rule metadata
- Context-2: Internal signals (24h aggregates, burst detection)
- Context-3: External enrichment (TI, VT verdicts)
Key insight: Missing evidence lowers confidence, doesn't escalate verdicts.
Context over Prompt Engineering
Automated Exception Generation
OpenAI Agents SDK Patterns
Building security agents with structured tool definitions and typed outputs:
@function_tool
def enrich_alert(
alert_id: str,
lookback_hours: int = 24
) -> AlertContext:
"""Fetch internal context for alert triage.
Returns prevalence, burst metrics, and
related alerts within the time window.
"""
return query_context_layer_2(
alert_id,
hours=lookback_hours
)Threat Coverage Gap Analysis
LLM Evaluation & Tracing
Human-in-the-Loop Design
Areas of Exploration
Security-Specialized Models
Domain-specific vs general-purpose trade-offs
Synthetic Security Data
AI-generated attack telemetry for testing
Threat Hunting Assistants
Conversational hypothesis exploration
Auto Incident Response
LLM-generated playbooks with safety constraints
Detection Translation
YARA ↔ Sigma ↔ ES|QL ↔ KQL
LLM Evaluation Pipelines
Tracing, ground truth, calibration metrics
Publications & Talks
2025
Agentic Frameworks: Practical Considerations for AI-Augmented Security
Elastic Security Labs • 2025
A2AS: Agentic AI Runtime Security and Self-Defense
SSRN • 2025
Academic paper on autonomous AI agent systems for threat detection
GenAI in Detection Engineering
BSides Oklahoma • 2025
Conference talk on practical GenAI applications in threat research
2024
Elastic Advances LLM Security with Standardized Fields
Elastic Security Labs • 2024
ECS gen_ai.* fields, AWS Bedrock integration, ES|QL detection queries
Embedding Security in LLM Workflows
Elastic Security Labs • 2024
Proxy-based LLM auditing and content moderation strategies


