StartupTalky presents Recap’25, a series of exclusive interviews where we connect with founders and industry leaders to reflect on their journey in 2025 and discuss their vision for the future.In this edition of Recap’25, StartupTalky speaks with Hrishikesh Dewan, Co-founder & CEO of Ziroh Labs, who shares how the company is rethinking AI infrastructure by enabling production-grade large language models to run efficiently on CPU architectures, eliminating the traditional dependency on GPUs. Dewan explains how Kompact AI, Ziroh Labs’ fast-inference AI platform, is designed to optimize token throughput, reduce latency, and maintain model accuracy while delivering an OpenAI-compatible deployment layer for enterprises and developers.He also highlights the key milestones that shaped 2025, including early benchmark validation with IIT Madras, expansion of support to 300+ LLMs across multimodal workloads, and the launch of Kompact AI One, a semantic caching module built to improve throughput for contextually similar queries. The conversation further explores why CPU-native inference can unlock more predictable cost structures and stronger data control for regulated sectors such as BFSI and healthcare, along with Ziroh Labs’ 2026 roadmap focused on deeper model support, partnerships with OEMs and system integrators, and developer SDKs aimed at accelerating CPU-based AI adoption globally.StartupTalky: Can… Read MoreStartupTalky- Business News, Insights and Stories
Home Uncategorized Ziroh Labs’ Hrishikesh Dewan on Kompact AI, CPU-Native LLM Inference, and Making...








