| --- |
| tags: |
| - biology |
| - DNA |
| - genomics |
| --- |
| This is the official pre-trained model introduced in [DNA language model GROVER learns sequence context in the human genome](https://www.nature.com/articles/s42256-024-00872-0) |
|
|
|
|
|
|
| from transformers import AutoTokenizer, AutoModelForMaskedLM |
| |
| # Import the tokenizer and the model |
| tokenizer = AutoTokenizer.from_pretrained("PoetschLab/GROVER") |
| model = AutoModelForMaskedLM.from_pretrained("PoetschLab/GROVER") |
| |
|
|
| Some preliminary analysis shows that sequence re-tokenization using Byte Pair Encoding (BPE) changes significantly if the sequence is less than 50 nucleotides long. Longer than 50 nucleotides, you should still be careful with sequence edges. |
| We advice to add 100 nucleotides at the beginning and end of every sequence in order to guarantee that your sequence is represented with the same tokens as the original tokenization. |
| We also provide the tokenized chromosomes with their respective nucleotide mappers (They are available in the folder tokenized chromosomes). |
|
|
| ### BibTeX entry and citation info |
|
|
| ```bibtex |
| @article{sanabria2024dna, |
| title={DNA language model GROVER learns sequence context in the human genome}, |
| author={Sanabria, Melissa and Hirsch, Jonas and Joubert, Pierre M and Poetsch, Anna R}, |
| journal={Nature Machine Intelligence}, |
| pages={1--13}, |
| year={2024}, |
| publisher={Nature Publishing Group UK London} |
| } |
| ``` |
|
|