Discover ANY AI to make more online for less.

select between over 22,900 AI Tool and 17,900 AI News Posts.


thenextweb
Fractile raises $220m to take its in-memory-compute inference chip into production

Accel led the London chip startup’s round, with Pat Gelsinger joining as an angel investor, weeks after Anthropic was reported to be in early discussions to become a customer. Fractile, the London-based startup designing inference chips that put compute and memory on the same die, has raised $220 million to take its hardware to production, […]
This story continues at The Next Web

Rating

Innovation

Pricing

Technology

Usability

We have discovered similar tools to what you are looking for. Check out our suggestions for similar AI tools.

venturebeat
5% GPU utilization: The $401 billion AI infrastructure problem enterprises

<p>For the last 24 months, one narrative justified every over-provisioned data center and bloated IT budget: the GPU scramble. Silicon was the new oil, and H100s traded like contraband. Reserve [...]

Match Score: 157.78

venturebeat
Together AI's ATLAS adaptive speculator delivers 400% inference speedu

<p>Enterprises expanding AI deployments are hitting an invisible performance wall. The culprit? Static speculators that can&#x27;t keep up with shifting workloads.</p><p>Speculat [...]

Match Score: 128.57

venturebeat
Baseten takes on hyperscalers with new AI training platform that lets you o

<p><a href="https://www.baseten.co/"><u>Baseten</u></a>, the AI infrastructure company recently valued at $2.15 billion, is making its most significant product [...]

Match Score: 124.46

venturebeat
Inference is splitting in two — Nvidia’s $20B Groq bet explains its nex

<p>Nvidia’s $20 billion strategic licensing deal with Groq represents one of the first clear moves in a four-front fight over the future AI stack. 2026 is when that fight becomes obvious to en [...]

Match Score: 117.08

venturebeat
Train-to-Test scaling explained: How to optimize your end-to-end AI compute

<p>The standard guidelines for building large language models (LLMs) optimize only for training costs and ignore inference costs. This poses a challenge for real-world applications that use infe [...]

Match Score: 109.07

venturebeat
Breaking through AI’s memory wall with token warehousing

<p>As agentic AI moves from experiments to real production workloads, a quiet but serious infrastructure problem is coming into focus: memory. Not compute. Not models. Memory.</p><p> [...]

Match Score: 103.59

venturebeat
FOMO is why enterprises pay for GPUs they don't use — and why prices

<p>Enterprises can&#x27;t fix their GPU waste problem because the fix makes the problem worse. Releasing idle capacity would improve utilization, but the same shortage driving GPU prices up [...]

Match Score: 97.86

venturebeat
DeepSeek’s conditional memory fixes silent LLM waste: GPU cycles lost to

<p>When an enterprise LLM retrieves a product name, technical specification, or standard contract clause, it&#x27;s using expensive GPU computation designed for complex reasoning — just to [...]

Match Score: 89.71

venturebeat
AI inference costs dropped up to 10x on Nvidia's Blackwell — but har

<p>Lowering the cost of inference is typically a combination of hardware and software. A new analysis released Thursday by Nvidia details how four leading inference providers are reporting 4x to [...]

Match Score: 87.28