For DALL-E and Stable Diffusion, the model size is an order of magnitude smaller than the total size of all the training set images? So it's not possible for the model to regurgitate every image in the training set exactly?
For Copilot, is there a similar argument? Or its model is large enough to contain the training set verbatim?
The world seems slightly mad about these things that produce "almost" pictures from text. We forgive DALL-E when it produces a twisted eye or an impossible perspective, because its result is "close enough" that we recognise something and grant the image intention.
So now you've got me waiting for DALL-Ecode. Give DALL-Ecode a description, it produces code.
"DALL-Ecode: Code that is sufficiently close to what you'd expect that you'll try to use it."
"DALL-Ecode: Code that looks like it does what is needed."
"DALL-Ecode: Good enough to compile, good enough to get through a code review (just not good enough to get through testing)."
For Copilot, is there a similar argument? Or its model is large enough to contain the training set verbatim?