Latest in Inference
Sort by
2 items
-
Pipeshift cuts GPU usage for AI inferences 75% with modular interface engine
Pipeshift has a Lego-like system that allows teams to configure the right inference stack for their AI workloads, without extensive engineering.VentureBeat - Jan. 23 -
DeepMind’s new inference-time scaling technique improves planning accuracy in LLMs
With "Mind Evolution" LLMs can use search and genetic algorithms to generate and combine different solutions and find the optimal one.VentureBeat - Jan. 22