Tutorial: Text Classification using GPT2 and Pytorch


Apr 09, 09:30 AM PDT
  • Virtual AICamp
  • 374 RSVPs
Description
Text classification is a very common problem that needs solving when dealing with text data. We’ve all seen and know how to use Encoder Transformer models like Bert and RoBerta for text classification but did you know you can use a Decoder Transformer model like GPT2 for text classification?

In this tutorial, I will walk you through on how to use GPT2 from HuggingFace for text classification. We will start with downloading customized dataset, installing required componments, selecting pre-trained models, and then train the model. we will finally evaluate the results and how to optimize further.


Contact Organizer