DALL·E 2 is the newest AI model by OpenAI. If you’ve seen some of its creations and think they’re amazing, keep reading to understand why you’re totally right — but also wrong.
OpenAI published a blog post and a paper entitled “Hierarchical Text-Conditional Image Generation with CLIP Latents” on DALL·E 2. The post is fine if you want to get a glimpse at the results and the paper is great for understanding the technical details, but neither explains DALL·E 2’s amazingness — and the not-so-amazing — in depth. That’s what this article is for.
DALL·E 2 is the new version of DALL·E, a generative language model that takes sentences and creates corresponding original images. At 3.5B parameters, DALL·E 2 is a large model but not nearly as large as GPT-3 and, interestingly, smaller than its predecessor (12B). Despite its size, DALL·E 2 generates 4x better resolution images than DALL·E and it’s preferred by human judges +70% of the time both in caption matching and photorealism.
As they did with DALL·E, OpenAI didn’t release DALL·E 2 (you can always join the never-ending waitlist). However, they open-sourced CLIP which, although only indirectly related to DALL·E, forms the basis of DALL·E 2. (CLIP is also the basis of the apps and notebooks people who can’t access DALL·E 2 are using.) Still, OpenAI’s CEO, Sam Altman, said they’ll eventually release DALL·E models through their API — for now, only a few selected have access to it (they’re opening the model to 1000 people each week).
To continue reading this article, click here.
You must be logged in to post a comment.
Pingback: The Promise and Limitations of a Revolutionary AI « Machine Studying Occasions - The Tech Best
I want to ask how can I use DALL·E 2. Does it cost? Has anyone played Papa’s Pizzeria ever? I find this game cute.