| import torch |
| import torch.nn as nn |
| from transformers import BertModel |
|
|
| class PersonaAssigner(nn.Module): |
| def __init__(self, input_dim, hidden_dim, output_dim): |
| super(PersonaAssigner, self).__init__() |
| self.fc1 = nn.Linear(input_dim, hidden_dim) |
| self.fc2 = nn.Linear(hidden_dim, output_dim) |
| |
| def forward(self, x): |
| x = torch.relu(self.fc1(x)) |
| return self.fc2(x) |
|
|
| class PreferencePredictor(nn.Module): |
| def __init__(self, input_dim): |
| super(PreferencePredictor, self).__init__() |
| self.fc1 = nn.Linear(input_dim, 256) |
| self.fc2 = nn.Linear(256, 3) |
| |
| def forward(self, x): |
| x = torch.relu(self.fc1(x)) |
| return self.fc2(x) |
|
|
| class BERTEncoder(nn.Module): |
| def __init__(self, model_name='bert-base-uncased'): |
| super(BERTEncoder, self).__init__() |
| self.bert = BertModel.from_pretrained(model_name) |
| |
| def forward(self, input_ids, attention_mask, token_type_ids=None): |
| outputs = self.bert(input_ids=input_ids, attention_mask=attention_mask, token_type_ids=token_type_ids) |
| return outputs.last_hidden_state.mean(dim=1) |