Vol. I · No. 18THU, MAY 7, 2026
Source · Community

r/MachineLearning

Reddit · COMMUNITY

Last updated May 7, 2026, 4:30 PM

Stop letting LLMs edit your .bib [D]

Research community reports frequent LLM hallucinations in bibliography generation, with incorrect author attributions despite correct titles, raising integrity concerns.

··

Why SSMs struggle in parameter-constrained training: empirical findings at 25M parameters [R]

After \~3 weeks of experimentation in OpenAI's Parameter Golf competition, I wrote up why SSMs are structurally disadvantaged relative to transformers in a time- and size-constrained regime (10 min training, 16MB artifact, 25M parameters) on 8xH100s: [https://mradassaad.github.io/posts/why-ssms-struggle-in-parameter-golf/](https://mradassaad.github.io/posts/why-ssms-struggle-in-parameter-golf/) Main findings: 1. SSM in\_proj weights compress up to 3.26x worse than attention QKV under LZMA, directly taxing the compressed parameter budget 2. Architectural wins validated at SP4096 flipped sign...

··

ICML final decisions rant [D]

Reddit discussion critiquing ICML's 27% acceptance rate and review quality issues, raising concerns about paper triage cascading to NeurIPS.

··

I spent years building a 103B-token Usenet corpus (1980–2013) and finally documented it [P]

For the past several years I've been quietly assembling and processing what I believe is one of the larger privately held pretraining corpora around... a complete Usenet archive spanning 1980 to 2013. Here's what it ended up being: * **103.1 billion tokens** (cl100k\_base) * **408 million posts** across 9 newsgroup hierarchies * **18,347 newsgroups** covered * **33 years** of continuous coverage The processing pipeline included full deduplication, binary removal (alt.binaries.\* excluded at the hierarchy level before record-level cleaning), quoted text handling, email address redaction via...

··

[ECCV 2026] Review Discussion [D]

ECCV reviews should be out by 2nd May. Since no exact time was specified this year, they’ll likely be released sometime within the next 48 hours. Hopefully, the reviews go well for everyone. We can use this thread to discuss them, as I haven’t seen one started yet.

··

AI/ML Conferences [D]

Reddit discussion on fairness and consistency issues in peer review at top-tier ML conferences like ICML 2026.

··

An interactive semantic map of the latest 10 million published papers [P]

I built a map to help navigate the complex scientific landscape through spatial exploration. How it works: Sourced the latest 10M papers from OpenAlex and generated embeddings using SPECTER 2 on titles and abstracts. Reduced dimensionality with UMAP, then applied Voronoi partitioning on density peaks to create distinct semantic neighborhoods. The floating topic labels are generated via custom labelling algorithms (definitely still a work in progress!). There is also support for both keyword and semantic queries, and there's an analytics layer for ranking institutions, authors, and topi...

··

Stanford Paper review [D]

Reddit discussion on Stanford Paper Review tool; user seeks community feedback on reliability of AI-assisted paper review suggestions.

··

What is the scientific value of administering the standard Rorschach test to LLMs when the training data is almost certainly contaminated? (R) + [D]

A recent paper published in *JMIR Mental Health* (Csigó & Cserey, 2026) caught my attention. The researchers administered the 10 standard Rorschach inkblot cards to three multimodal LLMs (GPT-4o, Grok 3, Gemini 2.0) and coded their responses using the Exner Comprehensive System. They analyzed the models' "perceptual styles," determinants (like human movement vs. color), and human-related content themes. However, I am seriously struggling to understand the methodological validity of this setup, and I’m curious what the scientific community thinks. My main concerns are: Massive Data Cont...

··

Visualizing Loss Landscapes of Neural Networks [P]

Hey r/MachineLearning, Visualizing the loss landscape of a neural network is notoriously tricky since we can't naturally comprehend million-dimensional spaces. We often rely on basic 2D contour analogies, which don't always capture the true geometry of the space or the sharpness of local minima. I built an interactive browser experiment [https://www.hackerstreak.com/articles/visualize-loss-landscape/](https://www.hackerstreak.com/articles/visualize-loss-landscape/) to help build better intuitions for this. It maps how different optimizers navigate these spaces and lets you actually visualiz...

··

Introducing AutoMuon, a one line drop in for AdamW [P]

Hey everyone, I've been working on a small Python package called AutoMuon that makes the Muon optimizer usable as a drop-in replacement for AdamW in arbitrary PyTorch training pipelines. The core idea is relatively simple: Muon works primarily on 2D weight matrices (linear projections, conv layers) on hidden states, but you still need AdamW for embeddings, norms, and biases, etc. AutoMuon scans your model at init, figures out the right optimizer for each parameter automatically. I am open to PRs, especially for expanding the module-type exclusion list if you hit edge cases in your architect...

··

Is the ds/ml slowly being morphed into an AI engineer? [D]

Agents are amazing. Harnesses are cool. But the fundamental role of a data scientist is not to use a generalist model in an existing workflow; it's a completely different field. AI engineering is the body of the vehicle, whereas the actual brain/engine behind it is the data scientist's playground. I feel like I am not alone in this realisation that my role somehow got silently morphed into that of an AI engineer, with the engine's development becoming a complete afterthought. Based on industry requirements and ongoing research, most of the work has quietly shifted from building the engine t...

··
50 stories