Learning about learning?
· 4 min read
[Continuing] Beyond LLMs/Transformers
The power-waste, inefficiencies, and general limits of throwing large corpora of labeled data into a blender to create probability distributions of next-token are becoming apparent to the broader world.
Useful to create once, and part of the solution, but not the answer.
