| import torch |
| import torch.nn as nn |
| import torch.nn.functional as F |
| from torch.utils.data import Dataset, DataLoader |
| from transformers import BertTokenizerFast |
| import random |
|
|
| |
| class TextDataset(Dataset): |
| def __init__(self, texts, tokenizer, max_len=32): |
| ''' |
| Args: |
| texts: List of stories |
| tokenizer: Tokenizer |
| max_len: Maximum length of the story # THIS IS JUST AN EXAMPLE |
| ''' |
| self.encodings = tokenizer( |
| texts, |
| padding="max_length", |
| truncation=True, |
| max_length=max_len, |
| return_tensors="pt" |
| ) |
|
|
| def __len__(self): |
| return len(self.encodings["input_ids"]) |
|
|
| def __getitem__(self, idx): |
| return { |
| "input_ids": self.encodings["input_ids"][idx], |
| "token_type_ids": self.encodings["token_type_ids"][idx], |
| "attention_mask": self.encodings["attention_mask"][idx].unsqueeze(0).unsqueeze(0) |
| } |
|
|
| |
| if __name__ == "__main__": |
| |
| tokenizer = BertTokenizerFast.from_pretrained("bert-base-uncased") |
|
|
| |
| texts = [ |
| "The quick brown fox jumps over the lazy dog.", |
| "Transformers are powerful models for NLP tasks.", |
| "Masked language modeling trains BERT to understand context.", |
| "Pretraining is followed by task-specific fine-tuning." |
| ] |
|
|
| dataset = TextDataset(texts, tokenizer, max_len=32) |
| dataloader = DataLoader(dataset, batch_size=2, shuffle=True) |
|
|