Breaking the ‘Intelligence Ceiling’: How Gruve and Fleak Achieved Tier 3 Threat Detection Fidelity

Breaking the ‘Intelligence Ceiling’: How Gruve and Fleak Achieved Tier 3 Threat Detection Fidelity

Standard pipelines waste 40% on "logic grinding." The Gruve + Fleak architecture ends the janitor work. With Fleak’s Normalization Layer, raw logs from Zscaler and Okta become actionable intelligence in the Gruve Command Center.

By

Arif Shaikh

Security Operations Consultant - Gruve

Breaking the ‘Intelligence Ceiling’: How Gruve and Fleak Achieved Tier 3 Threat Detection Fidelity

Project Partners: 


1. The Core Discovery

This report documents a collaborative performance benchmark between Gruve and Fleak.ai, designed to explore the relationship between data architecture and AI-driven security outcomes. We conducted a controlled A/B test to observe how a high-performance UEBA agent—the core of Gruve’s AI SOC—performs when supported by different data pipeline configurations: raw heterogeneous logs vs. Fleak.ai-normalized OCSF data.

The Conclusion: Data normalization is the primary catalyst for maximizing AI reasoning. The test demonstrated that while the Gruve AI engine is highly capable of processing raw logs, feeding it Fleak.ai-normalized OCSF allowed the model to bypass the "parsing phase" and move immediately to "high-fidelity analysis." This resulted in a shift from standard alert generation to the identification of complex, multi-stage threat patterns—achieving a Tier 3 Hunter performance profile without increasing operational latency or token costs.

2. Test Design & Performance Overview

To ensure scientific validity, the test held all detection variables constant except for the data format. This was not a phased rollout, but a side-by-side comparison of the same dataset.

A. The Constants (Control Variables)

  • The Dataset: Identical Zscaler Private Access (ZPA) log set containing hidden attack patterns.

  • The AI Engine: The same Large Language Model (LLM) agent was used for both scenarios.

  • The Prompt: Identical System Prompt. The agent was given the exact same instructions: "You are an enterprise-grade UEBA engine. Analyze logs for behavioral anomalies, deviations, and insider threats. Do not use static IOCs."

B. The Variable (The Change)

  • Scenario A (Baseline): Raw Logs. The AI ingested native, messy logs (mixed formats, inconsistent field names like Latitude vs. ClientLatitude).

  • Scenario B (Fleak.ai): OCSF Logs. The AI ingested the same data, normalized and enriched by Fleak.ai into the Open Cybersecurity Schema Framework (OCSF).


Comparative Performance Summary: Raw Logs vs. Fleak.ai OCSF

Feature

Before (Raw Logs)

After (Fleak.ai OCSF)

Impact

Data Format

Heterogeneous, proprietary (LEEF, Syslog), inconsistent fields (Latitude vs ClientLatitude).

OCSF Normalized, schema-compliant JSON (src_endpoint, actor, status_id).

100% Schema Consistency

AI Latency

High (Token budget wasted on parsing syntax).

Negligible (Tokens focused purely on analysis).

Faster Time-to-Insight

Mapping Speed

N/A (Manual parser coding required for new logs).

< 5 Minutes (AI-automated end-to-end mapping).

Agile Onboarding

Entity Resolution

Fragmented (User csp.shahpur treated as 3 separate entities due to format drift).

Unified (Single actor.user.name identity across all events).

Accurate Profiling

Threat Detection

Low Confidence (Misinterpreted error codes as system noise).

High Fidelity (Correlated behavioral chains like "Access Spraying").

Critical Risk Detected

False Positives

High (Service accounts flagged as human exfiltration).

Low (Service accounts correctly tagged and suppressed).

Reduced Analyst Fatigue

3. Operational Efficiency & Latency

A key finding is that the performance boost in Scenario B did not come at the cost of detection speed.

  • Runtime Latency: Negligible. For existing log types, Fleak.ai’s normalization occurs in-stream with almost no processing overhead, preserving real-time detection capabilities.

  • New Log Onboarding: < 5 Minutes. For completely new, unknown event types, Fleak.ai’s AI-native mapping agent generates end-to-end OCSF parsers in under 5 minutes.

  • Result: The system achieves structured data clarity at the speed of raw log ingestion.

4. Comparative Detection Results

The specific threat scenarios hidden within the dataset illustrate how the same AI agent interpreted the same events differently based on data format.

Use Case

Scenario A: Raw Logs (Baseline)

Scenario B: Fleak.ai (OCSF)

Verdict

Policy Brute Force

Missed / Low Priority.
AI saw closeBRK_MT... as a "system configuration error" and ignored the volume.

Critical Alert (Score: 92/100)

AI recognized status: Failure + activity: Close as "Intentional Persistence" & "Reconnaissance".

From Noise → Threat

Lateral Movement

Missed. 

Connection attempts to different IPs were treated as isolated, unrelated events.

High Alert (Score: 88/100)

AI visualized the topology: User → Endpoint A → B → C (Sequence).

From Dots → Pattern

Service Accounts

False Positive. 

High traffic volume flagged as "Data Exfiltration" by a human user.

Suppressed (Score: 18/100). actor.type context allowed AI to classify it as "Benign Service Behavior".

From Alert → Filtered

5. Strategic Implications

For a security product company, these results demonstrate that Data Quality is the ceiling for AI performance.

  1. AI Cost Reduction: By normalizing data before it hits the AI model, you remove the need for the LLM to burn tokens parsing syntax. It focuses 100% of its context window on threat analysis.

  2. Agility: New data sources can be onboarded in minutes (via Fleak's AI mapping) rather than weeks of manual parser coding, without breaking downstream detection logic.

  3. Reliability: Detection logic becomes schema-agnostic. A "Brute Force" prompt works for any identity provider (Okta, Azure AD, Zscaler) because they are all mapped to the same OCSF standard.

Start Building with Fleak Today

Lakehouse Ready Data in Minutes

Start Building with Fleak Today

Lakehouse Ready Data in Minutes