PyTorch GPT and Textcat #12346
kbmmoran
started this conversation in
Help: Best practices
Replies: 1 comment
-
The basics for how to specify a different transformer model: #10768 I know that
For any other models our best advice is to try it out. Not all huggingface models will work with |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi all, sanity check here.
I have a multi-label textcat model, trained using the tok2vec pipe. I would like to work on improving the model by integrating a custom trained GPT PyTorch model to power this textcat component. This would be as simple as registering the architecture and adding the GPT to the config file?
I know BERT architecture models usually perform better with text classification tasks, and RoBERTa is the default transformer model to use with Spacy. However, my understanding of the Spacy architecture is that a GPT model should, in theory, work similarity well to a BERT model in powering the textcat component, as it is powering the component, we not finetuning the transformer model to make the predictions.
I apologize if I have misunderstood any of the ideas listed here and appreciate any help/guidance :)
Beta Was this translation helpful? Give feedback.
All reactions