The Divyam.AI Platform at a Glance
Divyam.AI ships two products, Model Router and EvalMate, that are designed to work independently but compound when combined. This page shows how each product is built internally, and how the two products form a closed feedback loop: EvalMate's judge and rewards model govern Router's decisions; Router's production logs feed EvalMate's continuous evaluation pipeline.
The two loops explained
EvalMate → Router (amber arrow). After EvalMate runs its pipeline (Rubric Builder through Rewards Model Distillation), it produces a compact, domain-aligned judge that scores any model response against your quality bar. That judge becomes the scoring function Router uses when benchmarking models, running shadow tests, and recalibrating routing weights. Router never falls back to generic benchmarks; it always routes against your definition of good.
Router → EvalMate (teal arrow). Every inference request Router handles generates a structured trace: model chosen, latency, cost, and the response itself. Those traces flow continuously into EvalMate's evaluation pipeline, where Divyam.AI's proprietary frugal selection module cuts experimentation cost by 20× while maintaining full quality guarantees. The judge also determines which traces to surface to human reviewers, protecting that costly resource. Quality regressions surface within hours, not sprint cycles. Drift in request patterns triggers EvalMate's coverage gap analysis so the rubric stays aligned with what users are actually sending.
Together, the two loops form a system that improves cost and quality without manual intervention, compounding at each batch cycle as new models are adopted, routing weights are recalibrated, and evaluation coverage keeps pace with product evolution.
For the broader context on why most AI stacks never close this loop, read Is Your AI Stack Keeping Pace with the Market? →
For the full deep-dive into what production-grade agentic AI actually requires, read Taking Your LLM Application to Production →