Discussion about this post

User's avatar
Neural Foundry's avatar

Fantastic breakdown on training data scarcity. The point about models consuming data faster than we produce it is probly the most underrated crisis in AI right now. What strikes me is how much the publishers blocking crawlers actually accelerates model collapse rather than protects them. Like, models will just train on more synthetic data and reddit threads instead of quality journalism. Kinda feels like everyone loses in that scenario.

1 more comment...

No posts

Ready for more?