Fractile Raises $220 Million to Build The Next Generation of Inference Hardware 

Fractile

Insider Brief

  • AI chip startup Fractile raised $220 million to develop hardware designed to accelerate AI inference workloads and reduce processing bottlenecks in advanced AI systems.
  • The funding round was led by Accel, Factorial Funds and Founders Fund, with participation from several additional venture capital firms and existing investors.
  • Fractile said current AI workloads involving tens of millions of tokens can take weeks to complete on existing hardware, creating demand for faster and more efficient inference systems.

London-based AI chip startup Fractile announced in a company blog post it raised $220 million in a funding round led by Accel, Factorial Funds and Founders Fund as investors continue pouring capital into hardware designed to support the growing computational demands of artificial intelligence.

The company said the funding will be used to accelerate development of its first AI chips and systems and move them into customer deployments. Additional investors in the round included Conviction, Gigascale, O1A, Felicis, Buckley Ventures and 8VC.

Founded in 2022, Fractile is developing semiconductor hardware aimed at speeding up AI inference, the process by which trained AI models generate responses and predictions after deployment. The company argues that inference speed is becoming a major bottleneck as advanced AI systems generate increasingly long outputs and consume larger amounts of computing power.

The financing comes amid an industry-wide push to redesign AI infrastructure as developers move beyond training ever-larger models and toward running them efficiently at scale. While much of the current AI hardware market is dominated by graphics processing units, or GPUs, startups and large technology companies alike are pursuing specialized architectures aimed at lowering cost, reducing latency and improving energy efficiency for inference workloads.

Fractile said some frontier AI systems are already generating outputs containing tens of millions of tokens, the units of text processed by large language models. According to the company, those workloads can take weeks to complete on current hardware because of limits tied to memory bandwidth and data movement inside chips.

The company said current systems typically generate around 40 tokens per second for such large workloads, while more advanced use cases may require output speeds closer to 1,200 tokens per second to reduce processing times from weeks to days.

Fractile is positioning its technology around what the company describes as “long-context” AI workloads, in which models process and generate very large chains of reasoning over extended sequences of text and intermediate outputs. The company linked those capabilities to emerging applications in software engineering, scientific research, drug discovery and materials science.

The company also framed the challenge as an economic issue for the broader AI sector. Inference has become one of the largest operating costs for AI providers as consumer and enterprise applications scale. Faster and cheaper inference systems could allow AI developers to run more complex workloads without sharply increasing computing expenses.

Fractile said it has been working across multiple layers of the AI hardware stack, including chip architecture, semiconductor manufacturing processes and AI systems research. The company did not disclose when its first commercial chips are expected to ship or identify manufacturing partners.

The announcement reflects continued investor appetite for AI infrastructure companies despite growing concerns about rising capital expenditures and competition in the semiconductor sector. Since the launch of generative AI systems such as OpenAI’s ChatGPT, venture capital firms and major technology companies have committed billions of dollars toward AI chips, networking equipment, memory systems and data center expansion.

Fractile added it is hiring across offices in London, Bristol, San Francisco and Taipei as it expands development of its hardware platform.

Need Deeper Intelligence on the AI Market?

AI Insider's Market Intelligence platform tracks funding rounds, competitive landscapes, and technology trends across the global AI ecosystem in real time. Get the data and insights your organization needs to make informed decisions.

Related Articles

Humanoid to Deploy Thousands of Humanoid Robots at Schaeffler Plants

Insider Brief UK robotics startup Humanoid has announced a deployment and supply agreement with Schaeffler to bring humanoid robots into live manufacturing operations. According to

Origin Lab Raises $8M in Seed Round Funding to Turn Video Game Worlds Into AI Training Data

Insider Brief PRESS RELEASE — Origin Lab, the technology platform turning licensed game worlds into structured training data for world models and multimodal AI, announced

The 20 Enterprise AI Infrastructure & MLOps CEOs You Need to Know in 2026

Every enterprise AI deployment, from the simplest classifier to the most complex agentic system, depends on infrastructure. The CEOs building that infrastructure are, in many

Stay Updated with AI Insider

Get the latest AI funding news, market intelligence, and industry insights delivered to your inbox weekly.

$ 0 M

Seed round tracked

Gitar — Code Validation

Get the Weekly Briefing

Funding analysis, market intelligence, and industry trends delivered to your inbox every week.

Need bespoke intelligence?

Our team combines real-time data with decades of sector experience to guide your decisions.

Subscribe today for the latest news about the AI landscape