Older models moving back to 200k context window. FYI
Anthropic reduces context window for older Claude models from current capability back to 200k tokens.
Search the full wire by company, model, lab, or keyword. Every story we have ever aggregated.
Anthropic reduces context window for older Claude models from current capability back to 200k tokens.
Reddit user reports Claude failing at calorie tracking task, requests feature debugging.
Sam Altman apologizes for OpenAI's failure to report mass shooter threat to authorities; governance/safety process issue.
DeepSeek V4 Pro exhibits degraded token efficiency vs. V3.2 despite 2.5x scale increase, suggesting intelligence density declined.
Reddit user questions practical use cases for Anthropic's Dispatch feature beyond basic remote task submission.
OpenAI unified Codex into main model at GPT-5.4; GPT-5.5 shows gains in agentic coding and computer use, no separate Codex variant planned.
User reports 26 tokens/sec throughput with Qwen 27B-Q6_K_L on RTX A6000 in local agentic workflows via pi.dev, comparable to Claude Code experience.
Xiaomi's MiMo V2.5 Pro ranks #54 on Artificial Analysis Intelligence Index; weights announced for local deployment.
Reddit post with vague, clickbait title about AI replacing humans; no substantive content or claims.
Kimi K2.6 wins custom Blood on the Clocktower social deduction benchmark against competitors despite slower generation speed.
Reddit user reports version downgrade from 2.1.120 to 2.1.119 on Claude, with speculation about job displacement.
Reddit user demonstrates GPT-5.5 generating Minecraft assets via image prompts; OP corrects claim that it automates only small-scale world population, not professional-grade building.
Shield 82M: distilroberta-base fine-tuned for multilingual PII detection and redaction across names, emails, phones, addresses.
Qwen3.6-27B achieves ~80 tokens/sec at 218k context on single RTX 5090 using vLLM 0.19 with NVFP4 quantization.
Claude systematically overestimates task duration in human time units; Claude Code execution often 10-100x faster than estimates, suggesting training data misalignment.
Reddit user shares Claude Code workflow tips and best practices after 6 months of daily use, including skill creation for repetitive tasks.
NoTorch: minimal C library for neural network training/inference without PyTorch dependencies, demonstrated on nanoGPT port.
NYU nursing student built 660K-page drug database using Claude Haiku for pharmaceutical reference aggregation.
Reddit user reports subjective quality improvements moving Qwen3.6 35B from Q4 to Q8 quantization, plans to test Q6 tradeoffs.
Reddit post expressing frustration with OpenAI governance; lacks substantive technical or business content.
Reddit user reports Claude models flagging normal biology questions as problematic across multiple days and model versions.
Reddit user reports Claude account suspension citing child usage detection; seeks clarity on verification process.
Reddit post argues Claude Opus 4.7's compute allocation strategy causes shallow reasoning and breaks prompt caching compatibility with 4.6.
Unsubstantiated commentary on recent AI news without specific claims or evidence.
DeepSeek releases V4 Pro (1.6T-A49B) and Flash (284B-A13B) models optimized for Huawei Ascend chips, no longer leading benchmarks.
Reddit user claims no image model can correctly generate left-handed writing, highlighting a persistent multimodal AI capability gap.
OpenAI performance analysis chart generated via GPT image generation; lacks methodology and context.
OpenAI releases GPT-5.5 prompting guide with techniques for multi-step tasks, including sending early user-visible updates to improve perceived responsiveness.