Positron secures $230M to accelerate AI inference memory chips and challenge Nvidia
Read Our Expert Analysis
Create an account or login for free to unlock our expert analysis and key takeaways for this development.
By continuing, you agree to receive marketing communications and our weekly newsletter. You can opt-out at any time.
Recommended for you

Axelera AI secures $250M+ to scale power-efficient AI chips
Axelera AI closed a financing round topping $250M to push production of power-efficient inference semiconductors, drawing new institutional capital from BlackRock and continued strategic support from Samsung Catalyst. The raise is part of a broader wave of large hardware financings that signal investor appetite for inference-optimized silicon but leaves product validation, foundry access and software maturity as the critical next milestones.
NVIDIA Leans on Groq to Expand AI-Accelerator Capacity
NVIDIA has struck a commercial pact with Groq to relieve near-term inference accelerator capacity constraints and diversify silicon sourcing; reporting around the arrangement varies (some outlets cite a large multibillion-dollar licensing/priority package while others stress non‑binding frameworks). The deal buys time for NVIDIA’s roadmap but also accelerates a structural shift toward blended, multi‑vendor accelerator fleets that raise integration, validation and regulatory questions for hyperscalers and enterprises.

Young entrepreneur secures $220m to fund a UK AI chip venture
A 25-year-old founder in the UK has closed a $220m financing to develop custom processors for AI workloads, a sign that investors continue to back bespoke silicon despite long development cycles. The raise places the venture alongside a wave of large hardware financings and underscores near‑term execution priorities: tape‑outs, foundry commitments, packaging and software integration to turn prototypes into deployable systems.

NVIDIA to Push Inference Chip and Enterprise Agent Stack at GTC
NVIDIA is expected to unveil an inference-focused silicon family and an enterprise agent framework called NemoClaw at GTC, alongside commercial moves that could tighten its end-to-end platform grip. Sources signal a rumored Groq licensing pact valued near $20B but differ on whether that figure is a binding transaction, while supply‑chain timing and CPU‑first architectural signals complicate the near‑term path to broad deployment.

Mistral AI acquires Koyeb to accelerate AI cloud, on‑prem inference and GPU optimization
Mistral AI has bought Paris-based Koyeb to fold serverless deployment and isolated runtime tech into its cloud stack, enabling model inference on customer hardware and tighter GPU management. The deal complements Mistral’s broader infrastructure push — including a €1.2 billion Sweden data‑center program with EcoDataCenter and new compact speech‑to‑text models optimized for local hardware — reinforcing a hybrid, Europe‑anchored AI strategy.

Japan–U.S. tie-up: SoftBank’s Saimemory and Intel race to commercialize next‑gen AI memory
SoftBank’s Saimemory and Intel launched the Z‑Angle Memory (ZAM) program to develop DRAM optimized for AI with prototypes due by the fiscal year ending March 31, 2028 and a commercialization target in fiscal 2029. The initiative arrives as major memory suppliers accelerate HBM and NAND investments and hyperscalers exert greater influence on qualification cycles—factors that both validate demand for ZAM’s energy‑focused approach and raise competitive and timing risks.
Mirai builds a Rust inference engine to accelerate on-device AI
Mirai, a London startup, raised $10 million to deliver a Rust-based inference runtime that accelerates model generation on Apple Silicon by as much as 37% and exposes a simple SDK for developers. The team is positioning the stack for text and voice use cases today, with planned vision support, on-device benchmarks, and a hybrid orchestration layer that routes heavier work to the cloud.
Gimlet Labs Raises $80M to Orchestrate Multi‑Silicon Inference
Gimlet Labs closed an $80M Series A led by Menlo Ventures to commercialize a multi‑silicon inference cloud that shards agentic workloads across heterogeneous hardware. The raise and product launch sit inside a broader wave of infrastructure bets — from edge runtimes to stateful AI platforms — that collectively signal software orchestration is becoming the primary lever for lowering inference cost and shaping procurement.