Wednesday, February 12, 2020

GPT Explained!


This video explains the original GPT model, "Improving Language Understanding by Generative Pre-Training". I think the key takeaways are understanding that they use a new unlabeled text dataset that requires the pre-training language modeling to incorporate longer range context, the way that they format input representations for supervised fine-tuning, and the different NLP tasks this is evaluated on! Paper Links: GPT: https://ift.tt/2HeACni DeepMind "A new model and dataset for long range memory": https://ift.tt/2uqNAdY SQuAD: https://ift.tt/2SKNJkC MultiNLI: https://ift.tt/2wcOOWJ RACE: https://ift.tt/2HjT24U Quora Question Pairs: https://ift.tt/30VBCTP CoLA: https://ift.tt/2SIZaZM Thanks for watching! Please Subscribe!

No comments:

Post a Comment