Nvidia remains dominant in chips for training large AI models, while inference has become a new front in the competition.
This brute-force scaling approach is slowly fading and giving way to innovations in inference engines rooted in core computer systems design.
Snowflake has thousands of enterprise customers who use the company's data and AI technologies. Though many issues with generative AI are solved, there is still lots of room for improvement. Two such ...
As frontier models move into production, they're running up against major barriers like power caps, inference latency, and rising token-level costs, exposing the limits of traditional scale-first ...
The CNCF is bullish about cloud-native computing working hand in glove with AI. AI inference is the technology that will make hundreds of billions for cloud-native companies. New kinds of AI-first ...