| import torch |
| from torch.utils.data import DataLoader |
| from transformers import AdamW, get_linear_schedule_with_warmup |
| from tqdm import tqdm |
| import time |
| import json |
| import os |
| from models import PersonaAssigner, PreferencePredictor, BERTEncoder |
| from data_utils import load_chatbot_data, load_personas, ChatbotDataset |
| from torch.cuda.amp import autocast, GradScaler |
| import sys |
|
|
| class SimpleDashboard: |
| def __init__(self): |
| self.stats = { |
| 'iteration': 0, |
| 'total_iterations': 500, |
| 'batch': 0, |
| 'total_batches': 575, |
| 'baseline_loss': 0, |
| 'baseline_accuracy': 0, |
| 'enhanced_loss': 0, |
| 'enhanced_accuracy': 0, |
| 'last_save': '', |
| } |
| self.history = { |
| 'baseline_loss': [], |
| 'baseline_accuracy': [], |
| 'enhanced_loss': [], |
| 'enhanced_accuracy': [], |
| } |
|
|
| def update_stats(self, **kwargs): |
| self.stats.update(kwargs) |
| for key in ['baseline_loss', 'baseline_accuracy', 'enhanced_loss', 'enhanced_accuracy']: |
| if key in kwargs: |
| self.history[key].append(kwargs[key]) |
| if len(self.history[key]) > 50: |
| self.history[key] = self.history[key][-50:] |
|
|
| def draw_ascii_chart(self, data, title, width=50, height=10): |
| if not data: |
| return "" |
| |
| min_val, max_val = min(data), max(data) |
| range_val = max_val - min_val if max_val > min_val else 1 |
| |
| lines = [f"{title} (min: {min_val:.4f}, max: {max_val:.4f})"] |
| for i in range(height - 1, -1, -1): |
| line = "" |
| for val in data[-width:]: |
| if (val - min_val) / range_val > i / (height - 1): |
| line += "█" |
| else: |
| line += " " |
| lines.append(line) |
| |
| return "\n".join(lines) |
|
|
| def draw(self): |
| os.system('cls' if os.name == 'nt' else 'clear') |
| print(f"Experiment Progress:") |
| print(f"Iteration: {self.stats['iteration']}/{self.stats['total_iterations']}") |
| print(f"Batch: {self.stats['batch']}/{self.stats['total_batches']}") |
| print(f"Baseline Loss: {self.stats['baseline_loss']:.4f}") |
| print(f"Baseline Accuracy: {self.stats['baseline_accuracy']:.2%}") |
| print(f"Enhanced Loss: {self.stats['enhanced_loss']:.4f}") |
| print(f"Enhanced Accuracy: {self.stats['enhanced_accuracy']:.2%}") |
| print(f"Last Save: {self.stats['last_save']}") |
| print("\n" + self.draw_ascii_chart(self.history['baseline_loss'], "Baseline Loss")) |
| print("\n" + self.draw_ascii_chart(self.history['enhanced_loss'], "Enhanced Loss")) |
| print("\n" + self.draw_ascii_chart(self.history['baseline_accuracy'], "Baseline Accuracy")) |
| print("\n" + self.draw_ascii_chart(self.history['enhanced_accuracy'], "Enhanced Accuracy")) |
|
|
| class ComparativeExperiment: |
| def __init__(self, batch_size=4, accumulation_steps=8, max_length=64): |
| print("Initializing ComparativeExperiment...") |
| self.device = torch.device('cuda' if torch.cuda.is_available() else 'cpu') |
| print(f"Using device: {self.device}") |
| self.batch_size = batch_size |
| self.accumulation_steps = accumulation_steps |
| self.max_length = max_length |
| print("Initializing models...") |
| self.bert_encoder = BERTEncoder().to(self.device) |
| self.persona_assigner = PersonaAssigner(768, 256, 768).to(self.device) |
| self.enhanced_predictor = PreferencePredictor(768 * 3).to(self.device) |
| self.baseline_predictor = PreferencePredictor(768 * 3).to(self.device) |
| print("Initializing optimizer...") |
| self.optimizer = AdamW([ |
| {'params': self.persona_assigner.parameters()}, |
| {'params': self.enhanced_predictor.parameters()}, |
| {'params': self.baseline_predictor.parameters()} |
| ], lr=1e-4) |
| self.criterion = torch.nn.CrossEntropyLoss() |
| self.scaler = GradScaler() |
| self.current_iteration = 0 |
| self.current_step = 0 |
| self.training_log = [] |
| self.dashboard = SimpleDashboard() |
| print("ComparativeExperiment initialized.") |
| |
| def train_iteration(self, batch): |
| self.persona_assigner.train() |
| self.enhanced_predictor.train() |
| self.baseline_predictor.train() |
|
|
| with autocast(): |
| prompt_embeds = self.bert_encoder( |
| input_ids=batch['prompt']['input_ids'].squeeze(1), |
| attention_mask=batch['prompt']['attention_mask'].squeeze(1) |
| ) |
| response_a_embeds = self.bert_encoder( |
| input_ids=batch['response_a']['input_ids'].squeeze(1), |
| attention_mask=batch['response_a']['attention_mask'].squeeze(1) |
| ) |
| response_b_embeds = self.bert_encoder( |
| input_ids=batch['response_b']['input_ids'].squeeze(1), |
| attention_mask=batch['response_b']['attention_mask'].squeeze(1) |
| ) |
| labels = batch['label'].to(self.device) |
|
|
| baseline_inputs = torch.cat([prompt_embeds, response_a_embeds, response_b_embeds], dim=1) |
| baseline_outputs = self.baseline_predictor(baseline_inputs) |
| baseline_loss = self.criterion(baseline_outputs, labels) |
| baseline_accuracy = (baseline_outputs.argmax(dim=1) == labels).float().mean().item() |
|
|
| prompt_personas = self.persona_assigner(prompt_embeds.detach()) |
| response_a_personas = self.persona_assigner(response_a_embeds.detach()) |
| response_b_personas = self.persona_assigner(response_b_embeds.detach()) |
|
|
| combined_prompt = prompt_embeds + prompt_personas |
| combined_response_a = response_a_embeds + response_a_personas |
| combined_response_b = response_b_embeds + response_b_personas |
|
|
| enhanced_inputs = torch.cat([combined_prompt, combined_response_a, combined_response_b], dim=1) |
| enhanced_outputs = self.enhanced_predictor(enhanced_inputs) |
| enhanced_loss = self.criterion(enhanced_outputs, labels) |
| enhanced_accuracy = (enhanced_outputs.argmax(dim=1) == labels).float().mean().item() |
|
|
| total_loss = baseline_loss + enhanced_loss |
|
|
| scaled_loss = self.scaler.scale(total_loss) |
| scaled_loss.backward() |
|
|
| return baseline_loss.item(), baseline_accuracy, enhanced_loss.item(), enhanced_accuracy |
|
|
| def save_state(self, filename='experiment_state.pth'): |
| print(f"Saving state to {filename}...") |
| torch.save({ |
| 'current_iteration': self.current_iteration, |
| 'current_step': self.current_step, |
| 'bert_encoder': self.bert_encoder.state_dict(), |
| 'persona_assigner': self.persona_assigner.state_dict(), |
| 'enhanced_predictor': self.enhanced_predictor.state_dict(), |
| 'baseline_predictor': self.baseline_predictor.state_dict(), |
| 'optimizer': self.optimizer.state_dict(), |
| 'scaler': self.scaler.state_dict(), |
| }, filename) |
| with open('training_log.json', 'w') as f: |
| json.dump(self.training_log, f) |
| self.dashboard.update_stats(last_save=time.strftime("%Y-%m-%d %H:%M:%S")) |
| print("State saved successfully.") |
|
|
| def load_state(self, filename='experiment_state.pth'): |
| if os.path.exists(filename): |
| print(f"Loading state from {filename}...") |
| state = torch.load(filename) |
| self.current_iteration = state['current_iteration'] |
| self.current_step = state['current_step'] |
| print(f"Loaded iteration: {self.current_iteration}, step: {self.current_step}") |
| self.bert_encoder.load_state_dict(state['bert_encoder']) |
| self.persona_assigner.load_state_dict(state['persona_assigner']) |
| self.enhanced_predictor.load_state_dict(state['enhanced_predictor']) |
| self.baseline_predictor.load_state_dict(state['baseline_predictor']) |
| self.optimizer.load_state_dict(state['optimizer']) |
| self.scaler.load_state_dict(state['scaler']) |
| if os.path.exists('training_log.json'): |
| with open('training_log.json', 'r') as f: |
| self.training_log = json.load(f) |
| print(f"Loaded training log with {len(self.training_log)} entries.") |
| print("State loaded successfully.") |
| return True |
| else: |
| print(f"No saved state found at {filename}.") |
| return False |
|
|
| def run_experiment(self, train_loader, num_iterations=500): |
| print(f"Starting experiment. Total iterations: {num_iterations}") |
| self.dashboard.update_stats( |
| total_iterations=num_iterations, |
| total_batches=len(train_loader) |
| ) |
|
|
| try: |
| for iteration in range(self.current_iteration, num_iterations): |
| print(f"Starting iteration {iteration + 1}/{num_iterations}") |
| self.current_iteration = iteration |
| for i, batch in enumerate(train_loader): |
| batch = {k: v.to(self.device) for k, v in batch.items()} |
| baseline_loss, baseline_accuracy, enhanced_loss, enhanced_accuracy = self.train_iteration(batch) |
| |
| self.current_step += 1 |
| if self.current_step % self.accumulation_steps == 0: |
| self.scaler.step(self.optimizer) |
| self.scaler.update() |
| self.optimizer.zero_grad() |
|
|
| self.training_log.append({ |
| 'iteration': iteration + 1, |
| 'batch': i + 1, |
| 'baseline_loss': baseline_loss, |
| 'baseline_accuracy': baseline_accuracy, |
| 'enhanced_loss': enhanced_loss, |
| 'enhanced_accuracy': enhanced_accuracy, |
| 'timestamp': time.strftime("%Y-%m-%d %H:%M:%S") |
| }) |
|
|
| self.dashboard.update_stats( |
| iteration=iteration + 1, |
| batch=i + 1, |
| baseline_loss=baseline_loss, |
| baseline_accuracy=baseline_accuracy, |
| enhanced_loss=enhanced_loss, |
| enhanced_accuracy=enhanced_accuracy |
| ) |
|
|
| if (i + 1) % 10 == 0: |
| self.dashboard.draw() |
|
|
| print(f"Completed iteration {iteration + 1}/{num_iterations}") |
| |
| self.save_state() |
| self.dashboard.draw() |
|
|
| self.save_state('final_experiment_state.pth') |
| print("Experiment completed.") |
| except KeyboardInterrupt: |
| print("Experiment interrupted. Saving state...") |
| self.save_state('interrupted_experiment_state.pth') |
| print("State saved. You can resume later by loading 'interrupted_experiment_state.pth'") |
| except Exception as e: |
| print(f"An error occurred: {str(e)}") |
| self.save_state('error_experiment_state.pth') |
| print("State saved due to error. You can resume later by loading 'error_experiment_state.pth'") |
| raise |
| def setup_experiment(train_data, train_labels, batch_size=4, max_length=64): |
| from transformers import BertTokenizer |
|
|
| print("Setting up experiment...") |
| print("Initializing tokenizer...") |
| tokenizer = BertTokenizer.from_pretrained('bert-base-uncased') |
| print("Creating dataset...") |
| train_dataset = ChatbotDataset( |
| train_data['prompt'].tolist(), |
| train_data['response_a'].tolist(), |
| train_data['response_b'].tolist(), |
| train_labels, |
| tokenizer, |
| max_length=max_length |
| ) |
| print(f"Creating DataLoader with batch size {batch_size}...") |
| train_loader = DataLoader(train_dataset, batch_size=batch_size, shuffle=True) |
| |
| print("Initializing ComparativeExperiment...") |
| experiment = ComparativeExperiment(batch_size=batch_size, max_length=max_length) |
| print("Experiment setup completed.") |
| return experiment, train_loader |