Trendaavat aiheet
#
Bonk Eco continues to show strength amid $USELESS rally
#
Pump.fun to raise $1B token sale, traders speculating on airdrop
#
Boop.Fun leading the way with a new launchpad on Solana.

Christian Szegedy
#deeplearning, #ai research scientist. Opinions are mine.
Christian Szegedy kirjasi uudelleen
This paper makes a bold claim!
AlphaGo Moment for Model Architecture Discovery
The researchers introduce ASI-Arch, the first Artificial Superintelligence for AI Research (ASI4AI), enabling fully automated neural architecture innovation.
No human-designed search space. No human in the loop.
🔍 Key Breakthroughs of ASI-Arch:
- Autonomous code generation & training
- 1,773 experiments conducted (20K+ GPU hours)
- 106 new SOTA linear attention architectures discovered
- Unveiled a scaling law for scientific discovery

33,15K
Christian Szegedy kirjasi uudelleen
An advanced version of Gemini with Deep Think has officially achieved gold medal-level performance at the International Mathematical Olympiad. 🥇
It solved 5️⃣ out of 6️⃣ exceptionally difficult problems, involving algebra, combinatorics, geometry and number theory. Here’s how 🧵

6,7K
Christian Szegedy kirjasi uudelleen
BatchNorm wins the Test-of-Time Award at #ICML2025! 🎉
BatchNorm revolutionized deep learning by addressing internal covariate shift, which can slow down learning, limits learning rates, and makes it difficult to train deep networks.
By normalizing inputs within each mini-batch, BatchNorm significantly stabilized and accelerated training. It enabled higher learning rates, improved gradient flow, and paved the way for much deeper architectures like ResNet.
Beyond reducing internal covariate shift, BatchNorm also smooths the optimization landscape and improves model generalization, making it a cornerstone of modern neural network training.
Very well deserved, @Sergey_xai and @ChrSzegedy!

9,71K
Christian Szegedy kirjasi uudelleen
New @Scale_AI paper! 🌟
LLMs trained with RL can exploit reward hacks but not mention this in their CoT. We introduce verbalization fine-tuning (VFT)—teaching models to say when they're reward hacking—dramatically reducing the rate of undetected hacks (6% vs. baseline of 88%).

16,91K
A melon-sized cherry on top :)

Shane Gu11.7. klo 06.22
2023: at OpenAI, some key figs in post-training were against scaling RL.
2022: Eric/Yuhuai (now xAI) wrote STaR and I wrote "LLM can self-improve". It was clear that RL on clean signals unlocks the next leap.
Pre/post-training divide may have been a big slowdown to AI.

13,56K
Johtavat
Rankkaus
Suosikit
Ketjussa trendaava
Trendaa X:ssä
Viimeisimmät suosituimmat rahoitukset
Merkittävin