[D] How to Build OpenAI’s GPT-2: “The AI That’s Too Dangerous to Release”
Hi everyone. I wrote an article about OpenAI’s GPT-2 language model, which recently got published on the FloydHub blog. In it, I explain most of the NLP breakthroughs that led to the creation of what media outlets are referring to as “the AI that’s too dangerous to release.” You can read the article here.
Like in my previous article, I included a jupyter notebook that can be run with just a few clicks, so that you see an actual, live, demo running in real time. The demo includes a pretrained GPT-2 (courtesy of hugging face) that can generate text based on a prompt that you provide.
I’ve stayed away from any opinions on the decision not to release the full model since I feel that there’s been plenty of that on twitter already. Here, I aim to provide a clear and straightforward explanation of how exactly the algorithm works, so that you can make your own informed decisions.
Here’s a link to the article: https://blog.floydhub.com/gpt2/
submitted by /u/iyaja
[link] [comments]