Robert Youssef (@rryssf_)

2025-10-22 | โค๏ธ 2278 | ๐Ÿ” 317


๐Ÿšจ Holy shitโ€ฆMeta just rewrote how Transformers think.

They built something called The Free Transformer and it breaks the core rule every GPT model has lived by since 2017.

For 8 years, Transformers have been blindfolded forced to guess the next token one at a time, no inner plan, no latent thought.

Meta gave it one.

They added random latent variables inside the decoder so the model can secretly decide how it wants to generate before it starts talking.

Itโ€™s like giving GPT a hidden mind.

Result:

๐Ÿง  Smarter reasoning โšก๏ธ 3% compute overhead ๐Ÿ“ˆ Outperforms larger baselines on GSM8K, MMLU, and HumanEval

Itโ€™s the first Transformer that doesnโ€™t just predict it intends.

Full paper: arxiv. org/abs/2510.17558v1


Auto-generated bookmark

Tags

AI-ML