Musings and Thoughts
Since my last post I’ve been thinking about why LLMs work as well or poorly as they do and I think I’ve come up with a working hypothesis. Admittedly, it’s probably obvious, but hey! I never said I was quick!
What Were These Systems Trained On?
Programming
So, in the case of programming, these LLM systems have been trained on things like code in GitHub or posts on Stack Overflow. What are these posts typically? I would assume these posts are typically amateur programmers asking entry level questions more often than not. Further, the questions will likely be for smaller systems or short code snippets and not entire applications. Given that sort of training bias, I would expect for programming that an LLM would be well versed in simple programming tasks and less so in more advanced tasks. And I think, in that line of reasoning, it would be safe to expect the kind of output seen from the reports I reviewed in my last post.
Read more...