
Technology
|
Post-training
|
YC W26
|
Valuation:
Undisclosed

Last Updated:
March 24, 2026

A post-training research and product lab that builds rubric-based reward models, agentic AI frameworks, and open-source developer tooling to align, evaluate, and fine-tune large language models after pre-training.
Rubric AI has publicly released open-source agentic app frameworks (modular packages for agents, memory, events, auth, UI), a CLI bootstrapping tool (create-rubric-app), and CSPaper, a rubric-aligned academic paper feedback tool targeting top ML conferences (ICML, ICLR, SIGIR). Their GitHub monorepo signals continued investment in composable, type-safe developer tooling for LLM-powered applications and agent orchestration (rOS). Currently building RL environments for prominent voice agents.
GitHub commit patterns suggest active development of an "Agent Operating System" (rOS) with persistent memory and event-driven workflows, pointing toward enterprise-grade agent orchestration. Social signals indicate B2B infrastructure partnerships with Cal.com, Trigger.dev, and AgentHub AI likely preceding a formal product launch. The absence of Hugging Face model/dataset releases combined with heavy framework development suggests they are building proprietary evaluation and alignment pipelines internally before open-sourcing model artifacts.
Rubric-based reward modeling replaces subjective human preference scoring with structured, criteria-driven checklists to align LLMs more reliably and interpretably during post-training.
Instead of asking people "which AI answer do you like better?" and hoping they're consistent, Rubric AI gives the graders (human or AI) a detailed checklist—like a cooking competition scorecard—so every model output is judged on the same clear criteria.
It's like replacing a restaurant's "rate your meal 1-5 stars" card with a detailed scorecard for flavor, presentation, temperature, and portion size—suddenly you know exactly what to fix in the kitchen.
CSPaper provides venue-specific, rubric-aligned simulated peer review feedback for academic ML paper submissions, helping researchers identify and fix acceptance blockers before submission.
CSPaper acts like a practice round with a tough-but-fair AI reviewer who knows exactly what ICML or ICLR reviewers look for, so researchers can fix problems before the real reviews come back.
It's like having a brutally honest friend who's served on every top conference program committee read your paper and hand you a color-coded fix-it list before you hit submit.
An agent operating system (rOS) that orchestrates autonomous AI agents with persistent memory, event-driven workflows, and modular tool integration for enterprise automation tasks.
rOS is like an air traffic control tower for AI agents—it keeps track of what each agent knows, what it's doing, and when to hand off tasks, so complex business workflows run themselves without crashing into each other.
Spandana Govindgari was Staff Payments Engineering Lead at Meta, where she built, launched, and scaled payment experiences across Instagram, WhatsApp, and Facebook Shops/Marketplace processing billions in TPV. She was the founding female engineer on Snapchat's Payments team, worked at Apple and Microsoft, and graduated from Cornell CS. She founded Hype AR, a VC-funded AR ad monetization startup (featured on TechCrunch and Forbes, granted a patent), and is a Forbes Tech Council member and eWOW Diversity and Culture Leadership Award winner. Pragya Saboo brings investment-side product intuition from Climate Capital and is a World Economic Forum Global Shapers Community member.