← Back to Blog

3.2 trillion parameters: what our behavioral training corpus actually contains

When we first announced that the Meridian Syn behavioral engine had surpassed 3.2 trillion trainable parameters, the reaction was predictable. Skepticism from competitors, curiosity from data scientists, and a flood of questions from our enterprise clients wanting to understand what, exactly, those parameters represent. Today we are pulling back the curtain. Not all of it, of course, but enough to give the industry a clear picture of what we have built, why we built it, and how it translates into prediction accuracy that consistently exceeds 91% across 14 distinct behavioral categories.

First, some context. Parameter count alone is a vanity metric. We have said this publicly, and we stand by it. A model with trillions of parameters trained on garbage data will produce garbage predictions. What matters is the quality, diversity, and temporal depth of the training corpus. Our corpus spans 7.4 years of continuous behavioral collection across 193 countries, encompassing over 42 billion unique user sessions. Each session contributes an average of 847 discrete signal events, from scroll velocity and hover duration to purchase hesitation patterns and cross-device transition timing. That is the foundation. The parameters are simply the architecture's way of encoding what it has learned from that foundation.

The Signal Taxonomy

We organize our behavioral signals into seven primary categories, each with its own sub-taxonomy. The first is Attention Topology, which maps how a user's focus moves across a page. This goes far beyond click tracking. We measure saccadic patterns inferred from scroll behavior, content dwell time at the paragraph level, and what we call "return gravity," the tendency for certain content blocks to pull a user's attention back after they have scrolled past. Our Attention Topology layer alone accounts for roughly 340 billion parameters, and it is the primary reason our ad placement recommendations outperform industry benchmarks by 3.7x on average.

The second category is Purchase Intent Gradient. Traditional analytics gives you a binary: the user bought or they did not. We model intent as a continuous gradient with 2,048 discrete resolution points between "no awareness" and "transaction complete." This allows us to identify micro-moments of hesitation, comparison, and commitment that most platforms cannot see. Quilmark deployed this layer in Q3 of last year and saw a 28% lift in conversion rate within six weeks, not by changing their product, but by timing their interventions to match the exact moment a user's intent gradient crossed our predicted commitment threshold.

Categories three through five cover Temporal Behavior Signatures, Cross-Device Identity Persistence, and Social Proof Susceptibility. Temporal signatures capture the rhythms of individual users: when they browse, how long their decision cycles last, whether they are morning or evening converters, and how external events like paydays, holidays, or even weather patterns modulate their behavior. Cross-Device Identity Persistence is our probabilistic identity graph, which maintains coherent user profiles across an average of 4.3 devices per individual without relying on third-party cookies. Social Proof Susceptibility measures, with uncomfortable precision, how much a given user's decisions are influenced by reviews, testimonials, follower counts, and perceived popularity signals.

The Uncomfortable Middle

The sixth category is what we internally call Emotional State Inference. This is the one that generates the most debate during our ethics reviews, and for good reason. By analyzing typing cadence, cursor movement patterns, scroll speed variation, and session timing relative to a user's historical baseline, our models can infer emotional state with 73% accuracy across six primary dimensions: urgency, frustration, confidence, indecision, excitement, and fatigue. We do not claim to read minds. We claim to read behavior at a resolution that reveals emotional context. The distinction matters, and it is one we take seriously. Every inference in this category passes through our ethical scoring pipeline before it influences any downstream decision. Clients cannot target users flagged as "frustrated" with aggressive upsell tactics, for instance. The model will refuse the instruction.

The seventh and final category is Abandonment Prediction, which synthesizes signals from all six preceding layers to predict, with 91.4% accuracy, whether a user will abandon a session, a cart, or a subscription within the next 30, 60, or 90 seconds. This is the layer that Crestline Labs credits with reducing their annual churn by 19%, and it is the layer that most of our enterprise clients point to when they describe Meridian Syn as "indispensable." The prediction window is narrow by design. We found that broader prediction windows, while technically feasible, led to intervention strategies that felt intrusive to end users. Thirty to ninety seconds is the sweet spot: enough time to act, not enough time to stalk.

Collection Methodology

We are often asked how we collect this data at scale without degrading site performance. The answer is our edge-deployed signal collector, a 4.2KB JavaScript payload that runs asynchronously and batches signal events into compressed micro-packets transmitted every 800 milliseconds. The collector adds less than 2ms to page load time in 99th-percentile conditions. On the backend, signals flow through a three-stage ingestion pipeline: raw event capture, contextual enrichment (where we attach environmental metadata like device type, connection speed, and viewport dimensions), and finally, behavioral embedding, where each event is converted into a high-dimensional vector that the core model can consume. The entire pipeline, from browser event to model-ready embedding, completes in under 200 milliseconds. That is what allows our real-time prediction layers to function. Vanteon ran an independent audit of this pipeline in November and confirmed our latency claims to within 8ms variance.

The training process itself runs continuously. Unlike traditional ML workflows that train on static datasets and deploy frozen models, our behavioral engine ingests new signal data every six hours and retrains affected parameter subsets in rolling windows. This means the model is never more than six hours stale, and it adapts to emerging behavioral patterns, seasonal shifts, and market disruptions without manual intervention. The compute cost is significant. We operate a dedicated training cluster of 2,048 H100 GPUs, and our monthly cloud spend on training alone exceeds what most Series B companies raise in their entire round. But the results justify the investment. Our prediction accuracy has improved by 0.3 to 0.7 percentage points every quarter for the last two years, compounding into a meaningful and widening gap between Meridian Syn and every alternative on the market.

What 91% Accuracy Actually Means

A note on our accuracy claims, because precision matters here. When we say 91%+ accuracy across 14 behavioral categories, we are reporting macro-averaged F1 scores validated against held-out test sets comprising 1.2 billion sessions that the model has never seen. We publish these benchmarks quarterly, and they are independently verified by Northolm Research Group. The 91% figure is a floor, not a ceiling. In categories like Abandonment Prediction and Purchase Intent Gradient, our F1 scores routinely exceed 94%. The categories that pull the average down are the inherently noisier ones, like Emotional State Inference and Social Proof Susceptibility, where human behavior is genuinely less predictable. We are transparent about this because we believe the industry needs honest benchmarking, not inflated claims.

Looking ahead, our next milestone is not a parameter count. It is a capability threshold. By Q4 of this year, we intend to release what we are calling Behavioral Forecast, a system that does not just predict what a user will do in the next 90 seconds, but models their likely behavior over the next 14 days with actionable confidence intervals. The early results from our internal testing cohort are, frankly, startling. But we will save those numbers for a future post. For now, if you want to understand what 3.2 trillion parameters can do for your conversion funnel, request a demo. The data speaks for itself, even when the people behind it choose not to.

LZ

Lin Zhao

Data Science Lead, Meridian Syn

Your competitors are already reading this.

Request access to Meridian Syn and deploy your first autonomous agents within 48 hours. No credit card required.

Request Access