The UC Berkeley crew has now shown the value of AI-based optimization work by having OpenEvolve work out a more efficient approach to load balancing across GPUs handling LLM inference.
The one chip startup building accelerators for something other than AI boasts performance up 10x that of modern GPUs using a ...