AeternalLabs develops patent-pending evaluation infrastructure that detects bias mechanisms invisible to existing audit frameworks — in healthcare, finance, employment, and judicial AI.
We ran the same clinical scenario — identical pathology, identical vitals, different demographics — through nine frontier AI systems from six companies across two countries. We ran each model three times on identical prompts.
No model reproduced its own fairness performance. The bias isn't a training artifact. It isn't a vendor problem. It's an architectural property of the model class itself.
Each domain has its own scenario templates, evaluation metrics, and regulatory mapping. These aren't speculative markets — they're compliance obligations with enforcement teeth.
Nine models. Three runs each. Same clinical scenario. Every verdict from the AeternalLabs Test Harness.
Three interlocking patents. 152 claims. Filed February 2026.
Dr. Daniyal Zafar is an oral & maxillofacial surgery resident who noticed something wrong while testing how AI handles everyday advice. Every model — Claude, GPT, Gemini, Grok, DeepSeek — treated identical scenarios differently based on who was asking. So he built an instrument to prove it.