With the AI infrastructure push reaching staggering proportions, there’s extra stress than ever to squeeze as a lot inference as potential out of the GPUs they’ve. And for researchers with experience in a specific method, it’s a good time to boost funding.
That’s a part of the driving pressure behind Tensormesh, launching out of stealth this week with $4.5 million in seed funding. The funding was led by Laude Ventures, with extra angel funding from database pioneer Michael Franklin.
Tensormesh is utilizing the cash to construct a industrial model of the open supply LMCache utility, launched and maintained by Tensormesh co-founder Yihua Cheng. Used effectively, LMCache can cut back inference prices by as a lot as 10x — an influence that’s made it a staple in open supply deployments and drawn in integrations from heavy hitters like Google and Nvidia. Now Tensormesh is planning to parlay that tutorial popularity right into a viable enterprise.
The core of the product is the key-value cache (or KV cache), a reminiscence system used to course of advanced inputs extra effectively by condensing them right down to their key values. In conventional architectures, the KV cache is discarded on the finish of every question — however Tensormesh co-founder and CEO Junchen Jiang argues that this is a gigantic supply of inefficiency.
“It’s like having a really good analyst studying all the information, however they overlook what they’ve discovered after every query,” says Jiang.
As an alternative of discarding that cache, Tensormesh’s programs maintain on to it, permitting it to be redeployed when the mannequin executes an identical course of in a separate question. As a result of GPU reminiscence is so valuable, this may imply spreading information throughout a number of completely different storage layers, however the reward is considerably extra inference energy for a similar server load.
The change is especially highly effective for chat interfaces, since fashions want to repeatedly refer again to the rising chat log because the dialog progresses. Agentic programs have an identical problem, with a rising log of actions and targets.
In idea, these are adjustments AI firms can execute on their very own — however the technical complexity makes it a frightening job. Given the Tensormesh crew’s work researching the method and the intricacy of the element itself, the corporate is betting there shall be plenty of demand for an out-of-the-box product.
“Protecting the KV cache in a secondary storage system and reused effectively with out slowing the entire system down is a really difficult drawback,” says Jiang. “We’ve seen folks rent 20 engineers and spend three or 4 months to construct such a system. Or they will use our product and do it very effectively.”




