Biomedical Language Models are Robust to Sub-optimal Tokenization
Paper • 2306.17649 • Published • 1
YAML Metadata Warning:empty or missing yaml metadata in repo card
Check out the documentation for more information.
We replicate the PubMedBERT model using the same data, hardware and code as our new BioVocabBERT model to ensure their comparion is fair.
Details about our pre-training procedure and downstream results can be found in our BioNLP @ ACL 2023 paper.