Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Katherine Crawson is @ Eletheur & IMHO is indisputably most responsible for the advances in text=>image generation. Dall-E 2 is Dall-E and her insight to use diffusion, the intermediate proof of concept of diffusion + Dall-E is GLIDE.

https://twitter.com/RiversHaveWings & https://github.com/crowsonkb



The paper cites K. Crowson 4 times:

[5] Katherine Crowson. Ava linear probe. https://twitter.com/RiversHaveWings/status/ 1472346186728173568?s=20&t=T-HRr3Gw5HRGjQaMDtRe3A, 2021.

[6] Katherine Crowson. Clip guided diffusion hq 256x256. https://colab.research.google.com/ drive/12a_Wrfi2_gwwAuN3VvMTwVMz9TfqctNj, 2021.

[7] Katherine Crowson. Clip guided diffusion 512x512, secondary model method. https://twitter. com/RiversHaveWings/status/1462859669454536711, 2021.

[8] Katherine Crowson. v-diffusion. https://github.com/crowsonkb/v-diffusion-pytorch, 2021.


Diffusion models existed long before this announcement. I have no idea who this person is, but they did not invent this idea.

Edit: Diffusion models guided by CLIP*


They did invent the idea of applying it to image generation, leading to OpenAI citing her _tweets_ (how cool is that?) in a paper for GLIDE, which as other comments note, looks just like a proof of concept of DallE-2.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: