| """ |
| K-FAC Statistics Collector |
| |
| Collects activation covariance (A) and gradient covariance (G) matrices |
| for MLP layers to approximate the Fisher Information Matrix. |
| |
| Based on: "From Memorization to Reasoning in the Spectrum of Loss Curvature" |
| """ |
|
|
| import torch |
| import torch.nn as nn |
| from torch import Tensor |
| from typing import Optional, Callable |
| from dataclasses import dataclass, field |
| from tqdm import tqdm |
|
|
|
|
| @dataclass |
| class LayerStatistics: |
| """K-FAC statistics for a single layer.""" |
| |
| |
| A: Optional[Tensor] = None |
| G: Optional[Tensor] = None |
| |
| |
| A_sum: Optional[Tensor] = None |
| G_sum: Optional[Tensor] = None |
| |
| |
| n_samples_A: int = 0 |
| n_samples_G: int = 0 |
| |
| def finalize(self) -> None: |
| """Convert running sums to means.""" |
| if self.A_sum is not None and self.n_samples_A > 0: |
| self.A = self.A_sum / self.n_samples_A |
| self.A_sum = None |
| if self.G_sum is not None and self.n_samples_G > 0: |
| self.G = self.G_sum / self.n_samples_G |
| self.G_sum = None |
|
|
|
|
| @dataclass |
| class KFACConfig: |
| """Configuration for K-FAC collection.""" |
| |
| |
| target_layers: list[int] = field(default_factory=lambda: [11, 12, 13]) |
| |
| |
| target_projections: list[str] = field(default_factory=lambda: ["gate_proj", "up_proj"]) |
| |
| |
| seq_length: int = 512 |
| |
| |
| exclude_last_position: bool = True |
| |
| |
| sample_labels: bool = True |
| |
| |
| device: str = "cuda" |
|
|
|
|
| class KFACCollector: |
| """ |
| Collects K-FAC statistics (activation and gradient covariances) for MLP layers. |
| |
| The K-FAC approximation factorizes the Fisher Information Matrix as: |
| F_W ≈ G ⊗ A = E[gg^T] ⊗ E[aa^T] |
| |
| where: |
| - A is the covariance of activations going into the layer |
| - G is the covariance of gradients on the layer's output |
| |
| Usage: |
| collector = KFACCollector(model, config) |
| collector.register_hooks() |
| |
| for batch in dataloader: |
| collector.collect_batch(batch, tokenizer) |
| |
| collector.finalize() |
| collector.save("kfac_stats.pt") |
| """ |
| |
| def __init__( |
| self, |
| model: nn.Module, |
| config: Optional[KFACConfig] = None, |
| ): |
| self.model = model |
| self.config = config or KFACConfig() |
| |
| |
| self.stats: dict[str, LayerStatistics] = {} |
| |
| |
| self._forward_hooks: list = [] |
| self._backward_hooks: list = [] |
| |
| |
| self._activation_buffer: dict[str, Tensor] = {} |
| self._gradient_buffer: dict[str, Tensor] = {} |
| |
| |
| self._hooks_registered = False |
| |
| def _get_layer_name(self, layer_idx: int, proj_name: str) -> str: |
| """Generate a unique name for a layer/projection combination.""" |
| return f"layer_{layer_idx}.{proj_name}" |
| |
| def _get_target_modules(self) -> dict[str, nn.Linear]: |
| """Find all target modules in the model.""" |
| targets = {} |
| |
| |
| |
| |
| layers = None |
| if hasattr(self.model, "model") and hasattr(self.model.model, "layers"): |
| |
| layers = self.model.model.layers |
| elif hasattr(self.model, "transformer") and hasattr(self.model.transformer, "blocks"): |
| |
| layers = self.model.transformer.blocks |
| elif hasattr(self.model, "layers"): |
| |
| layers = self.model.layers |
| |
| if layers is None: |
| raise ValueError( |
| "Could not find transformer layers. " |
| "Model architecture not recognized. " |
| f"Model type: {type(self.model)}" |
| ) |
| |
| for layer_idx in self.config.target_layers: |
| if layer_idx >= len(layers): |
| print(f"Warning: Layer {layer_idx} does not exist (model has {len(layers)} layers)") |
| continue |
| |
| layer = layers[layer_idx] |
| |
| |
| mlp = None |
| if hasattr(layer, "mlp"): |
| mlp = layer.mlp |
| elif hasattr(layer, "feed_forward"): |
| mlp = layer.feed_forward |
| elif hasattr(layer, "ff"): |
| mlp = layer.ff |
| |
| if mlp is None: |
| print(f"Warning: Could not find MLP in layer {layer_idx}") |
| continue |
| |
| for proj_name in self.config.target_projections: |
| if hasattr(mlp, proj_name): |
| proj = getattr(mlp, proj_name) |
| if isinstance(proj, nn.Linear): |
| name = self._get_layer_name(layer_idx, proj_name) |
| targets[name] = proj |
| self.stats[name] = LayerStatistics() |
| else: |
| print(f"Warning: {proj_name} not found in layer {layer_idx}") |
| |
| return targets |
| |
| def register_hooks(self) -> None: |
| """Register forward and backward hooks on target modules.""" |
| if self._hooks_registered: |
| print("Hooks already registered") |
| return |
| |
| targets = self._get_target_modules() |
| |
| if not targets: |
| raise ValueError("No target modules found to hook") |
| |
| print(f"Registering hooks on {len(targets)} modules:") |
| for name in targets: |
| print(f" - {name}") |
| |
| for name, module in targets.items(): |
| |
| def make_forward_hook(layer_name: str): |
| def hook(module: nn.Module, input: tuple, output: Tensor) -> None: |
| |
| x = input[0] |
| if x.requires_grad: |
| |
| self._activation_buffer[layer_name] = x.detach() |
| return hook |
| |
| |
| def make_backward_hook(layer_name: str): |
| def hook(module: nn.Module, grad_input: tuple, grad_output: tuple) -> None: |
| |
| g = grad_output[0] |
| if g is not None: |
| self._gradient_buffer[layer_name] = g.detach() |
| return hook |
| |
| fh = module.register_forward_hook(make_forward_hook(name)) |
| bh = module.register_full_backward_hook(make_backward_hook(name)) |
| |
| self._forward_hooks.append(fh) |
| self._backward_hooks.append(bh) |
| |
| self._hooks_registered = True |
| |
| def remove_hooks(self) -> None: |
| """Remove all registered hooks.""" |
| for hook in self._forward_hooks: |
| hook.remove() |
| for hook in self._backward_hooks: |
| hook.remove() |
| |
| self._forward_hooks = [] |
| self._backward_hooks = [] |
| self._hooks_registered = False |
| |
| def _update_statistics(self) -> None: |
| """Update running statistics from current buffers.""" |
| for name in self.stats: |
| if name in self._activation_buffer: |
| x = self._activation_buffer[name] |
| |
| |
| |
| if self.config.exclude_last_position and x.shape[1] > 1: |
| x = x[:, :-1, :] |
| |
| |
| x_flat = x.reshape(-1, x.shape[-1]) |
| n_positions = x_flat.shape[0] |
| |
| |
| A_batch = x_flat.T @ x_flat |
| |
| |
| if self.stats[name].A_sum is None: |
| self.stats[name].A_sum = A_batch |
| else: |
| self.stats[name].A_sum = self.stats[name].A_sum + A_batch |
| self.stats[name].n_samples_A += n_positions |
| |
| if name in self._gradient_buffer: |
| g = self._gradient_buffer[name] |
| |
| |
| |
| if self.config.exclude_last_position and g.shape[1] > 1: |
| g = g[:, :-1, :] |
| |
| |
| g_flat = g.reshape(-1, g.shape[-1]) |
| n_positions = g_flat.shape[0] |
| |
| |
| G_batch = g_flat.T @ g_flat |
| |
| |
| if self.stats[name].G_sum is None: |
| self.stats[name].G_sum = G_batch |
| else: |
| self.stats[name].G_sum = self.stats[name].G_sum + G_batch |
| self.stats[name].n_samples_G += n_positions |
| |
| |
| self._activation_buffer.clear() |
| self._gradient_buffer.clear() |
| |
| @torch.no_grad() |
| def _sample_labels(self, logits: Tensor) -> Tensor: |
| """ |
| Sample labels from model's predicted distribution. |
| |
| For proper FIM computation, we sample ŷ ~ p(y|x) rather than |
| using ground truth labels. |
| """ |
| |
| probs = torch.softmax(logits, dim=-1) |
| |
| sampled = torch.multinomial( |
| probs.view(-1, probs.shape[-1]), |
| num_samples=1 |
| ).view(probs.shape[:-1]) |
| return sampled |
| |
| def collect_batch( |
| self, |
| input_ids: Tensor, |
| attention_mask: Optional[Tensor] = None, |
| labels: Optional[Tensor] = None, |
| ) -> float: |
| """ |
| Collect K-FAC statistics from a single batch. |
| |
| Args: |
| input_ids: Token IDs (batch, seq_len) |
| attention_mask: Attention mask (batch, seq_len) |
| labels: Ground truth labels (optional, will be sampled if not provided |
| or if config.sample_labels is True) |
| |
| Returns: |
| Loss value for this batch |
| """ |
| self.model.train() |
| |
| |
| input_ids = input_ids.to(self.config.device) |
| if attention_mask is not None: |
| attention_mask = attention_mask.to(self.config.device) |
| |
| |
| with torch.enable_grad(): |
| outputs = self.model( |
| input_ids=input_ids, |
| attention_mask=attention_mask, |
| use_cache=False, |
| ) |
| logits = outputs.logits |
| |
| |
| if self.config.sample_labels or labels is None: |
| |
| sampled_labels = self._sample_labels(logits) |
| |
| shift_labels = sampled_labels[:, 1:].contiguous() |
| shift_logits = logits[:, :-1, :].contiguous() |
| else: |
| |
| shift_labels = labels[:, 1:].contiguous().to(self.config.device) |
| shift_logits = logits[:, :-1, :].contiguous() |
| |
| |
| loss_fn = nn.CrossEntropyLoss() |
| loss = loss_fn( |
| shift_logits.view(-1, shift_logits.shape[-1]), |
| shift_labels.view(-1) |
| ) |
| |
| |
| loss.backward() |
| |
| |
| self._update_statistics() |
| |
| |
| self.model.zero_grad() |
| |
| return loss.item() |
| |
| def collect_from_dataloader( |
| self, |
| dataloader, |
| max_tokens: int = 20_000_000, |
| progress_bar: bool = True, |
| ) -> dict: |
| """ |
| Collect K-FAC statistics from a dataloader. |
| |
| Args: |
| dataloader: PyTorch DataLoader yielding batches with input_ids |
| max_tokens: Maximum number of tokens to process |
| progress_bar: Whether to show progress bar |
| |
| Returns: |
| Dictionary with collection statistics |
| """ |
| if not self._hooks_registered: |
| self.register_hooks() |
| |
| total_tokens = 0 |
| total_loss = 0.0 |
| n_batches = 0 |
| |
| iterator = tqdm(dataloader, desc="Collecting K-FAC stats") if progress_bar else dataloader |
| |
| for batch in iterator: |
| if isinstance(batch, dict): |
| input_ids = batch["input_ids"] |
| attention_mask = batch.get("attention_mask") |
| else: |
| input_ids = batch[0] |
| attention_mask = batch[1] if len(batch) > 1 else None |
| |
| batch_tokens = input_ids.numel() |
| |
| loss = self.collect_batch(input_ids, attention_mask) |
| |
| total_tokens += batch_tokens |
| total_loss += loss |
| n_batches += 1 |
| |
| if progress_bar: |
| iterator.set_postfix({ |
| "tokens": f"{total_tokens/1e6:.1f}M", |
| "loss": f"{loss:.3f}" |
| }) |
| |
| if total_tokens >= max_tokens: |
| break |
| |
| return { |
| "total_tokens": total_tokens, |
| "n_batches": n_batches, |
| "avg_loss": total_loss / max(n_batches, 1), |
| } |
| |
| def finalize(self) -> None: |
| """Finalize statistics by converting sums to means.""" |
| for name, stat in self.stats.items(): |
| stat.finalize() |
| print(f"Finalized {name}: A={stat.A.shape if stat.A is not None else None}, " |
| f"G={stat.G.shape if stat.G is not None else None}") |
| |
| def save(self, path: str) -> None: |
| """Save K-FAC statistics to file.""" |
| save_dict = { |
| "config": { |
| "target_layers": self.config.target_layers, |
| "target_projections": self.config.target_projections, |
| "seq_length": self.config.seq_length, |
| }, |
| "statistics": {} |
| } |
| |
| for name, stat in self.stats.items(): |
| save_dict["statistics"][name] = { |
| "A": stat.A.cpu() if stat.A is not None else None, |
| "G": stat.G.cpu() if stat.G is not None else None, |
| "n_samples_A": stat.n_samples_A, |
| "n_samples_G": stat.n_samples_G, |
| } |
| |
| torch.save(save_dict, path) |
| print(f"Saved K-FAC statistics to {path}") |
| |
| @classmethod |
| def load(cls, path: str, model: nn.Module) -> "KFACCollector": |
| """Load K-FAC statistics from file.""" |
| data = torch.load(path, map_location="cpu") |
| |
| config = KFACConfig( |
| target_layers=data["config"]["target_layers"], |
| target_projections=data["config"]["target_projections"], |
| seq_length=data["config"]["seq_length"], |
| ) |
| |
| collector = cls(model, config) |
| |
| for name, stat_data in data["statistics"].items(): |
| collector.stats[name] = LayerStatistics( |
| A=stat_data["A"], |
| G=stat_data["G"], |
| n_samples_A=stat_data["n_samples_A"], |
| n_samples_G=stat_data["n_samples_G"], |
| ) |
| |
| print(f"Loaded K-FAC statistics from {path}") |
| return collector |
| |
| def get_statistics(self) -> dict[str, tuple[Tensor, Tensor]]: |
| """ |
| Get computed A and G matrices for all layers. |
| |
| Returns: |
| Dictionary mapping layer names to (A, G) tuples |
| """ |
| result = {} |
| for name, stat in self.stats.items(): |
| if stat.A is not None and stat.G is not None: |
| result[name] = (stat.A, stat.G) |
| return result |
|
|
|
|
| def create_dataloader( |
| dataset_name: str = "allenai/c4", |
| dataset_config: str = "en", |
| tokenizer = None, |
| batch_size: int = 4, |
| seq_length: int = 512, |
| max_samples: Optional[int] = None, |
| streaming: bool = True, |
| shuffle_buffer: int = 10000, |
| seed: int = 42, |
| ): |
| """ |
| Create a DataLoader for K-FAC collection. |
| |
| Args: |
| dataset_name: HuggingFace dataset name |
| dataset_config: Dataset configuration/subset name |
| tokenizer: Tokenizer for the model |
| batch_size: Batch size |
| seq_length: Sequence length for tokenization |
| max_samples: Maximum number of samples to load |
| streaming: Whether to use streaming mode |
| shuffle_buffer: Buffer size for streaming shuffle |
| seed: Random seed |
| |
| Returns: |
| PyTorch DataLoader |
| """ |
| from datasets import load_dataset |
| from torch.utils.data import DataLoader, IterableDataset |
| |
| |
| if dataset_config: |
| ds = load_dataset(dataset_name, name=dataset_config, split="train", streaming=streaming) |
| else: |
| ds = load_dataset(dataset_name, split="train", streaming=streaming) |
| |
| if streaming: |
| ds = ds.shuffle(buffer_size=shuffle_buffer, seed=seed) |
| |
| |
| def tokenize_fn(examples): |
| |
| text_column = "text" if "text" in examples else list(examples.keys())[0] |
| texts = examples[text_column] |
| |
| tokenized = tokenizer( |
| texts, |
| truncation=True, |
| max_length=seq_length, |
| padding="max_length", |
| return_tensors="pt", |
| ) |
| return tokenized |
| |
| |
| class TokenizedIterableDataset(IterableDataset): |
| def __init__(self, dataset, tokenizer, seq_length, max_samples): |
| self.dataset = dataset |
| self.tokenizer = tokenizer |
| self.seq_length = seq_length |
| self.max_samples = max_samples |
| |
| def __iter__(self): |
| count = 0 |
| for example in self.dataset: |
| if self.max_samples and count >= self.max_samples: |
| break |
| |
| |
| text = example.get("text", list(example.values())[0]) |
| if not text: |
| continue |
| |
| |
| tokens = self.tokenizer( |
| text, |
| truncation=True, |
| max_length=self.seq_length, |
| padding="max_length", |
| return_tensors="pt", |
| ) |
| |
| yield { |
| "input_ids": tokens["input_ids"].squeeze(0), |
| "attention_mask": tokens["attention_mask"].squeeze(0), |
| } |
| count += 1 |
| |
| torch_dataset = TokenizedIterableDataset(ds, tokenizer, seq_length, max_samples) |
| |
| dataloader = DataLoader( |
| torch_dataset, |
| batch_size=batch_size, |
| num_workers=0, |
| ) |
| |
| return dataloader |
|
|