Skip to main content
Strategy

The Divyam.AI Platform at a Glance

· 3 min read

Divyam.AI ships two products, Model Router and EvalMate, that are designed to work independently but compound when combined. This page shows how each product is built internally, and how the two products form a closed feedback loop: EvalMate's judge and rewards model govern Router's decisions; Router's production logs feed EvalMate's continuous evaluation pipeline.

Divyam.AI Platform Architecture Route Intelligently · Evaluate Continuously · Optimize Autonomously Model Router Intelligent Inferencing Layer · 100+ LLMs Connect One unified API endpoint OpenAI · Gemini · AWS · 100+ models Dynamic Router Per Prompt Routing Prompt skill ↔ model capability Model Leaderboard Live per-workload ranking Quality · cost · latency · status New Model Adoption Shadow testing & benchmarks Same-day adoption · auto-rollback Analytics & Observability Cost · Quality · Latency · Auto-recalibration · Regression alerts EvalMate Quality Intelligence Layer · Eval Co-Pilot Rubric Builder ~100 examples → criteria Agentic · domain-specific · versioned Human Feedback Annotations + golden dataset ~1K expert reviews · source of truth Judge & Rewards Model Distilled from golden set 92% agreement · ~8B param · 100× cheaper Drift Detection Coverage gap analysis Proactive · before user impact Eval Workflow Management End-to-end orchestration · Traceability · Versioning · Audit trail LLM-as-a-Judge / Rewards Model Production Logs & Traces Model Router EvalMate Quality signal (EvalMate → Router) Inference data (Router → EvalMate)
Model Router (teal) and EvalMate (amber) form a closed loop. EvalMate's Rewards Model provides the quality signal that governs every routing decision; Router's production logs and traces feed EvalMate's continuous evaluation pipeline.

The two loops explained

EvalMate → Router (amber arrow). After EvalMate runs its pipeline (Rubric Builder through Rewards Model Distillation), it produces a compact, domain-aligned judge that scores any model response against your quality bar. That judge becomes the scoring function Router uses when benchmarking models, running shadow tests, and recalibrating routing weights. Router never falls back to generic benchmarks; it always routes against your definition of good.

Router → EvalMate (teal arrow). Every inference request Router handles generates a structured trace: model chosen, latency, cost, and the response itself. Those traces flow continuously into EvalMate's evaluation pipeline, where Divyam.AI's proprietary frugal selection module cuts experimentation cost by 20× while maintaining full quality guarantees. The judge also determines which traces to surface to human reviewers, protecting that costly resource. Quality regressions surface within hours, not sprint cycles. Drift in request patterns triggers EvalMate's coverage gap analysis so the rubric stays aligned with what users are actually sending.

Together, the two loops form a system that improves cost and quality without manual intervention, compounding at each batch cycle as new models are adopted, routing weights are recalibrated, and evaluation coverage keeps pace with product evolution.

For the broader context on why most AI stacks never close this loop, read Is Your AI Stack Keeping Pace with the Market? →

For the full deep-dive into what production-grade agentic AI actually requires, read Taking Your LLM Application to Production →

Want to see the platform in action? We'll walk you through how both products plug into your existing stack.

Book a Demo