Vol. I · No. 18THU, MAY 7, 2026
Topic

§ Open Weights

Every story tagged with this topic, ordered by date.

What do you use Gemma 4 for?

Community discussion comparing Gemma 4 and Qwen 3.6 model suitability across coding, benchmarks, and agentic workloads.

··

Gemma 4 MTP released

Google releases Gemma 4 multi-token prediction drafters in 4 quantized sizes for local deployment.

··

Granite 4.1 3B SVG Pelican Gallery

IBM released Granite 4.1 (3B/8B/30B, Apache 2.0); Unsloth published 21 quantized GGUF variants; Willison benchmarked quality across model sizes on SVG generation.

·

Llama.cpp MTP support now in beta!

llama.cpp adds beta MTP (Multi-Token Prediction) support, starting with Qwen3.5, closing performance gap with vLLM on token generation.

··

Open Weights Models Hall of Fame

Community appreciation post nominating researchers and companies who released open-weights models, from Transformer authors to recent open-source contributors.

··

Does the "6 months gap" still hold?

Community discussion on whether open-source models' historical 6-12 month lag behind frontier systems persists after December 2025 agentic capability jump.

··

Solidity

Developer discusses building a local Solidity LM with chain-of-thought and tool-calling; seeks alternatives to SOTA models for smart contract security and vulnerability analysis.

··

Qwen3.6-27B-NVFP4 - images

User shares Qwen3.6-27B quantized setup with RTX 5090 and llamacpp configuration parameters.

··

New rules 1 week check-in

r/LocalLLaMA moderators report positive community response to new rules reducing spam after one week.

··
50 stories