Vol. I · No. 25THU, MAY 14, 2026
Archive

The Archive

Search the full wire by company, model, lab, or keyword. Every story we have ever aggregated.

[AINews] Tasteful Tokenmaxxing

Commentary on tokenization strategies as a recurring theme in AI industry discourse, without specific technical claims or announcements.

·

Grok Voice Think Fast 1.0

xAI releases Grok Voice Think Fast 1.0, a voice agent API for real-time conversational AI applications.

·

Simplify Sparse Deep Learning with Universal Sparse Tensor in nvmath-python

In a previous post, we introduced the Universal Sparse Tensor (UST), enabling developers to decouple a tensor’s sparsity from its memory layout for greater... In a previous post, we introduced the Universal Sparse Tensor (UST), enabling developers to decouple a tensor’s sparsity from its memory layout for greater flexibility and performance. We’re excited to announce the integration of the UST into nvmath-python v0.9.0 to accelerate sparse scientific and deep learning applications. This post provides a walkthrough of key UST features… Source

·

Forgive my ignorance but how is a 27B model better than 397B?

Is Qwen just incredibly good at doing dense and not so good at doing MoE? I get that dense is generally better than MoE but 27B being better than 397B just doesn’t sit right with me. What are those additional experts even doing then?

··
30 stories