Standard pipelines waste 40% on “logic grinding.” The Gruve + Fleak architecture ends the janitor work. With Fleak’s Normalization Layer, raw logs from Zscaler and Okta become actionable intelligence in the Gruve Command Center.
This report documents a collaborative performance benchmark between Gruve and Fleak.ai, designed to explore the relationship between data architecture and AI-driven security outcomes. We conducted a controlled A/B test to observe how a high-performance UEBA agent—the core of Gruve’s AI SOC—performs when supported by different data pipeline configurations: raw heterogeneous logs vs. Fleak.ai-normalized OCSF data.
The Conclusion: Data normalization is the primary catalyst for maximizing AI reasoning. The test demonstrated that while the Gruve AI-powered SOC engine is highly capable of processing raw logs, feeding it Fleak.ai-normalized OCSF allowed the model to bypass the “parsing phase” and move immediately to “high-fidelity analysis.” This resulted in a shift from standard alert generation to the identification of complex, multi-stage threat patterns—achieving a Tier 3 Hunter performance profile without increasing operational latency or token costs.
To ensure scientific validity, the test held all detection variables constant except for the data format. This was not a phased rollout, but a side-by-side comparison of the same dataset.


A key finding is that the performance boost in Scenario B did not come at the cost of detection speed.
The specific threat scenarios hidden within the dataset illustrate how the same AI agent interpreted the same events differently based on data format. This builds on the multi-agent threat hunting approach we’ve deployed for enterprise clients, where data quality directly determines hunting depth.

For a security product company, these results demonstrate that Data Quality is the ceiling for AI performance.