Hi, Sorry if this is extremely ignorant. But I was wondering the other day... are we 100% sure they arent feeding images from a search engine index into the ML data for Dalle?
It just looks surprisingly like its mixing and matching the top returned image searches from an index.
I'm only saying it looks like that - not that it is ofcourse.
I don't want to undermine anyones works here, I was just wondering.
You can read the whitepapers for Flamingo, Dall-E 2, Imagen, and Parti to see how diffusion networks and GANs work to create these images. I wrote up two huge paragraphs trying to explain it simply, but then I realized that I don't understand exactly how they work, either. Best source is the published research.
Most networks have to do with large language models, text embeddings, and image embeddings.
It just looks surprisingly like its mixing and matching the top returned image searches from an index.
I'm only saying it looks like that - not that it is ofcourse. I don't want to undermine anyones works here, I was just wondering.