NanoBERT
Collection
Pre-trained language model representations for Nanoscience and Nanotechnology. • 2 items • Updated
This model is further pre-trained from google-bert/bert-base-uncased using a corpus consisting of 200,000 Nanoscience and Nanotechnology papers.
For practical applications, please use https://huggingface.co/Flamenco43/NanoBERT-V2
Intended for training on downstream tasks using Nanoscience datasets. Can be used directly to create dense vector representations for information retrieval.
Trained using 2 nodes on Polaris: https://docs.alcf.anl.gov/polaris/hardware-overview/machine-overview/
The following hyperparameters were used during training: