Callosum Raises $10.25 Million Seed to Orchestrate AI Workloads Across Chip Architectures
London-based Callosum, co-founded by Cambridge neuroscientists, raises $10.25 million in seed funding led by Plural to build software that intelligently distributes AI workloads across Nvidia, AMD, Intel, and custom accelerator chips.
Callosum, a London-based startup co-founded by Cambridge neuroscientists, has raised $10.25 million in seed funding led by Plural to build software that intelligently orchestrates AI workloads across diverse accelerator chip architectures — addressing the growing challenge of running AI models efficiently across Nvidia, AMD, Intel, and custom silicon.
The Problem
As the AI chip market fragments beyond Nvidia's dominance, organizations increasingly deploy workloads across multiple accelerator types. Different chips excel at different tasks: Nvidia GPUs remain dominant for training, but AMD and Intel accelerators offer cost advantages for inference, while custom ASICs from companies like Google (TPUs) and Amazon (Trainium) are optimized for specific model architectures. Currently, optimizing workload placement across these diverse architectures requires significant manual engineering, creating a bottleneck that Callosum aims to automate.
The Approach
Callosum's software sits between the AI application layer and the hardware layer, analyzing workload characteristics — model architecture, batch size, latency requirements, cost constraints — and automatically routing computations to the most efficient available accelerator. The system learns from execution telemetry, continuously refining its placement decisions as it observes real-world performance across different hardware configurations.
Market Context
The funding comes as the AI chip market is projected to exceed $200 billion by 2027, with Nvidia's market share under increasing pressure from AMD, Intel, and a wave of AI chip startups. Organizations that can efficiently utilize multiple chip architectures stand to reduce their AI compute costs by 30-50% compared to single-vendor deployments — a compelling economic incentive that creates a natural market for orchestration software like Callosum's.
Related Articles
Google Gemini 3.1 Flash-Lite Targets Enterprise Scale at $0.25 Per Million Tokens
Google has launched Gemini 3.1 Flash-Lite in preview, the fastest and most cost-efficient model in its Gemini 3 family, priced at just $0.25 per million input tokens with 2.5x faster time-to-first-token than its predecessor. The model targets high-volume enterprise workloads where cost and latency matter more than peak capability.
Mandiant Founder Kevin Mandia Raises $190 Million for AI Cybersecurity Startup Armadin
Kevin Mandia, who sold Mandiant to Google for $5.4 billion in 2022, has raised a record-breaking $190 million in combined seed and Series A funding for Armadin, a startup building autonomous AI security agents. Backed by Accel, GV, Kleiner Perkins, and the CIA's In-Q-Tel, Armadin is already working with Fortune 100 companies.
Nscale Raises $2 Billion Series C — the Largest Funding Round in European Tech History
London-based AI infrastructure company Nscale closes a $2 billion Series C at a $14.6 billion valuation — the largest funding round in European history — backed by Citadel, Dell, NVIDIA, and Nokia, with former Meta COO Sheryl Sandberg joining the board.