FoxtrotMichael-1
Proper
When the engineers who build these models claim that they "don't understand how the model works" they're being academically dishonest. What they really mean is that they don't necessarily know the particular path through the generative model that the particular query took, or exactly how the training set gets stored into the model in any particular training run. This is a far cry from saying "look, we built the damn thing but we have no idea how it has this emergent functionality!" In fact, they know exactly how it works, because they wrote the algorithm, and the machine is doing exactly what the algorithm tells it to do. This is the very essence of designing non-deterministic algorithms. However, it is extremely academically dishonest to pretend that a non-deterministic algorithm has emergent capability. The engineers at Google are notorious for this kind of ridiculous behavior, but OpenAI isn't immune either.1 On inventors is not understanding: it is happening. All these modern models are just being fed raw data. They train themselves. And they are getting increasingly better at getting results training themselves. There are several models that can accurately and explicitly describe most everything in a general picture that it has never been inputted before. It can vividly, describe the color, texture, style, size of a jacket that was not in his training data based on its self learning. Critically, even though the creators don’t understand how this is happening. They are nevertheless, getting significantly better with it with each model release.