[object Object] exbert

This cased model was pretrained from scratch using a custom vocabulary on the following corpora

The pretrained model was used to do NER as is, with no fine-tuning. The approach is described in this post. Towards Data Science review

App in Spaces demonstrates this approach.

Github link to perform NER using this model in an ensemble with bert-base cased.

The ensemble detects 69 entity subtypes (17 broad entity groups)

<img src="https://ajitrajasekharan.github.io/images/1.png" width="600">

Ensemble model performance

<img src="https://ajitrajasekharan.github.io/images/6.png" width="600">

Additional notes

License

MIT license

<a href="https://huggingface.co/exbert/?model=ajitrajasekharan/biomedical&modelKind=bidirectional&sentence=Gefitinib%20is%20an%20EGFR%20tyrosine%20kinase%20inhibitor,%20which%20is%20often%20used%20for%20breast%20cancer%20and%20NSCLC%20treatment.&layer=3&heads=..0,1,2,3,4,5,6,7,8,9,10,11&threshold=0.7&tokenInd=17&tokenSide=right&maskInds=..&hideClsSep=true"> <img width="300px" src="https://cdn-media.huggingface.co/exbert/button.png"> </a>