🧪
v1.3.0Latest ReleasePersonalizationBenchmarkingEngine v3.0Learning

The Personalization Leap: Engine v3.0

We built our own personalization benchmark system, found the weak spots, and rebuilt the engine. Personalization v3.0 jumps from 57 to 99, 35 to 96, and 26 to 90 across benchmark generations.

February 12, 2026
9 min read

v1.3 is our personalization release. Before this version, the engine was genuinely strong in certain personalization lanes, but inconsistent in others. So we built a stricter benchmark system, stress-tested the weak spots, and shipped a new Personalization Engine v3.0.

Personalization v3.0

From selective strength to broad consistency.

We stopped guessing where personalization was good and where it was shaky. We measured it directly, benchmark by benchmark, then rebuilt the engine to close the gaps.

The score jump

Here is the headline: on every benchmark generation, Engine v3.0 outperformed Engine v2.0 by a wide margin.

v1 benchmark

Core personalization breadth, calibration, orchestration, and safety foundations.

Engine v2.057
Engine v3.099

v2 benchmark

Elite checks for passive, active, and anticipatory quality with hard reliability gates.

Engine v2.035
Engine v3.096

v3 benchmark

Hardest benchmark: learner-model depth, instructional adaptation, and outcome linkage.

Engine v2.026
Engine v3.090

How the benchmarks evolved

We did not just build one test and call it done. Each benchmark generation got stricter and more specific about what "real personalization quality" means.

v1

Foundation pass

The first benchmark mainly validated core architecture: can personalization reason across signals safely and consistently?

Category Scores (0-10)
1. Signal Breadth
2. Signal Validity/Calibration
...
10. Test Coverage/Regression Safety
v2

Elite behavior pass

The second benchmark raised the bar by forcing quality in passive, active, and anticipatory personalization.

This is a V2 "elite" benchmark. It must measure:
- passive personalization quality
- active personalization quality
- anticipatory personalization maturity
- non-compensatory safety/privacy/reliability gates
v3

Learning-outcome pass

The newest benchmark checks if personalization truly changes teaching quality and improves learning outcomes.

Rule 8: Reject superficial personalization credit
when adaptations are only tone/theme/style.

16) Learner Model Fidelity
17) Instructional Adaptation Depth
...
24) Operational Excellence

What changed inside Engine v3.0

  • Stronger balance across passive, active, and anticipatory paths.
  • Instructional-depth enforcement beyond style tweaks.
  • Harder counterfactual and outcome-linkage validation loops.
  • Broader reliability without compromising privacy guardrails.

Less blind spots

We targeted weak scenarios directly instead of optimizing only where the engine already looked good.

More trust

Better measurement means better engineering decisions, and better engineering decisions mean better learning experiences.

Not perfect yet, and that is the point

Even with massive gains, this is not a "finished" engine. A 90 on the hardest benchmark is strong progress, not the finish line. We still have room to improve quality, stability, and adaptation depth under edge-case learning conditions.

Why this should give you hope

We proved we can identify hard weaknesses and close them fast. If we can move from 26 to 90 on the toughest benchmark, we can keep pushing this system toward truly world-class personalization.

v1.3 is a major leap, and it is also a promise: we are going to keep measuring what matters, keep shipping improvements that hold up under pressure, and keep making Lernex more helpful for every kind of learner.

Personalization v3.0 is live in v1.3

Ready to experience these updates?

Jump into Lernex and see what's new for yourself.