Multilabel classification using BERT transformers returns low accuracy #13792
CodeCheetah
started this conversation in
Help: Best practices
Replies: 1 comment
-
Well, my guess is a corpus of size 2000 is too small for 15 labels. For a rule of thumb, one needs at least between 500-1000 instances per label for a healthy classification. In your case, per label there're only 2000/15 =133,3 instances, which is too small. Another issue is that your dataset is synthetic, the underlying distribution may not be smiler to real world data. For this project I'd:
Best of luck! |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I'm currently training a transformer-based models (bert-base-uncased, xlm-roberta-base, and roberta-base) on synthetic data of around 2000 records generated by GPT3.5 with 15 labels. The training works fine with out of the box config and several changes to hyperparameters like learning rate, dropout, batch size.
I'm able to evaluate the model on 30% of synthetic data which works, but whenever I bring real data for classification I'm getting really low scores, below 25% accuracy overall.
Any ideas on how/what to improve?
Should I try different textcat architecture: CNN/TextCatBOW, anything else?
My config:
Beta Was this translation helpful? Give feedback.
All reactions