Replicating GPT-2, the most dangerous NLP model (with Aaron Gokaslan) (Ep. 79)

In this episode, I am with Aaron Gokaslan, computer vision researcher, AI Resident at Facebook AI Research. Aaron is the author of OpenGPT-2, a parallel NLP model to the most discussed version that OpenAI decided not to release because too accurate to be published.
Better Language Models GPT-2

In this episode we discuss about image-to-image translation and the dangers of the GPT-2 model. As a matter of fact there is no better language model than GPT-2. With the last version of GPT-2 it is possible to generate paragraphs of text, given a seed. Such paragraphs actually make sense and it is quite challenging to understand if the author was a human or an algorithm. Therefore, anyone could use GPT-2 to generate fake news or facts that are not supported by any evidence.

Better Language Models GPT-2

In addition to this, we speak about the future of Artificial Intelligence. I asked what would a researcher like Aaron see happening in the next 5 years. Aaron believes that self-supervision will play a fundamental role in the future of AI. Moreover, he provided us with some very interesting resources and demos that will definitely blow your mind!

Enjoy the show!

Join the discussion on our Discord server


Multimodal image to image translation (not all mentioned in the podcast but recommended by Aaron)





GPT-2 and related
Try OpenAI’s GPT-2:
The Original Transformer Paper:
Grover: The FakeNews generator and detector:

Leave a Reply

Your email address will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Data Science

Discord community chat

Join our Discord community to discuss the show, suggest new episodes and chat with other listeners!

Support us