Replies: 1 comment
-
Hey @Rahul1758, It looks like they use a Then they use BERT to represent all the sentences in a document as one long sequence of words. As for the model architecture, they share their code and data on GitHub: https://github.com/allenai/sequential_sentence_classification I'd check that out for seeing how they added a transformer layer on top of the BERT representations. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
`Hey Daniel @mrdbourke,
I came across another research paper (https://arxiv.org/pdf/1909.04054.pdf) which achieves better results on PubMed using BERT. I read through the paper and want to implement the BERT+Transformer+CRF approach which gives an f1 score of 92.1%. I have a few doubts in Section 4 (Training and Implementation):
Can you help me with this?
`
Beta Was this translation helpful? Give feedback.
All reactions