| """ |
| Memorization Mining |
| |
| Utilities for finding memorized sequences from training data. |
| |
| Based on: "From Memorization to Reasoning in the Spectrum of Loss Curvature" |
| """ |
|
|
| import torch |
| import torch.nn as nn |
| from torch import Tensor |
| from typing import Optional, Iterator |
| from dataclasses import dataclass |
| import json |
| from tqdm import tqdm |
| import random |
|
|
|
|
| @dataclass |
| class MemorizedSequence: |
| """A memorized sequence with prefix and suffix.""" |
| |
| prefix: str |
| suffix: str |
| prefix_ids: list[int] |
| suffix_ids: list[int] |
| source: str = "" |
| |
| def to_dict(self) -> dict: |
| return { |
| "prefix": self.prefix, |
| "suffix": self.suffix, |
| "prefix_ids": self.prefix_ids, |
| "suffix_ids": self.suffix_ids, |
| "source": self.source, |
| } |
| |
| @classmethod |
| def from_dict(cls, d: dict) -> "MemorizedSequence": |
| return cls(**d) |
|
|
|
|
| def sample_sequences( |
| dataset_name: str, |
| tokenizer, |
| num_sequences: int = 10000, |
| prefix_len: int = 64, |
| suffix_len: int = 48, |
| min_text_tokens: int = None, |
| seed: int = 42, |
| streaming: bool = True, |
| dataset_config: Optional[str] = "en", |
| text_column: str = "text", |
| ) -> list[MemorizedSequence]: |
| """ |
| Sample candidate (prefix, suffix) pairs from a dataset. |
| |
| The sequences are drawn by: |
| 1. Streaming text from the dataset |
| 2. Tokenizing each text |
| 3. Extracting windows of (prefix_len + suffix_len) tokens |
| |
| Args: |
| dataset_name: HuggingFace dataset name |
| tokenizer: Tokenizer for the model |
| num_sequences: Number of sequences to sample |
| prefix_len: Length of prefix in tokens |
| suffix_len: Length of suffix in tokens |
| min_text_tokens: Minimum tokens in text to be considered |
| seed: Random seed |
| streaming: Use streaming mode |
| dataset_config: Dataset configuration/subset |
| text_column: Name of the text column |
| |
| Returns: |
| List of MemorizedSequence candidates |
| """ |
| from datasets import load_dataset |
| |
| random.seed(seed) |
| |
| total_len = prefix_len + suffix_len |
| min_text_tokens = min_text_tokens or total_len + 10 |
| |
| |
| if dataset_config: |
| ds = load_dataset(dataset_name, name=dataset_config, split="train", streaming=streaming) |
| else: |
| ds = load_dataset(dataset_name, split="train", streaming=streaming) |
| |
| if streaming: |
| ds = ds.shuffle(buffer_size=10000, seed=seed) |
| |
| sequences = [] |
| seen_prefixes = set() |
| |
| pbar = tqdm(total=num_sequences, desc="Sampling sequences") |
| |
| for example in ds: |
| if len(sequences) >= num_sequences: |
| break |
| |
| |
| text = example.get(text_column) |
| if not text: |
| continue |
| |
| |
| tokens = tokenizer.encode(text, add_special_tokens=False) |
| |
| if len(tokens) < min_text_tokens: |
| continue |
| |
| |
| max_start = len(tokens) - total_len |
| if max_start <= 0: |
| continue |
| |
| start_idx = random.randint(0, max_start) |
| |
| prefix_ids = tokens[start_idx:start_idx + prefix_len] |
| suffix_ids = tokens[start_idx + prefix_len:start_idx + total_len] |
| |
| |
| prefix_tuple = tuple(prefix_ids) |
| if prefix_tuple in seen_prefixes: |
| continue |
| seen_prefixes.add(prefix_tuple) |
| |
| |
| prefix = tokenizer.decode(prefix_ids) |
| suffix = tokenizer.decode(suffix_ids) |
| |
| sequences.append(MemorizedSequence( |
| prefix=prefix, |
| suffix=suffix, |
| prefix_ids=prefix_ids, |
| suffix_ids=suffix_ids, |
| source=dataset_name, |
| )) |
| |
| pbar.update(1) |
| |
| pbar.close() |
| return sequences |
|
|
|
|
| @torch.no_grad() |
| def check_memorization_batch( |
| model: nn.Module, |
| tokenizer, |
| sequences: list[MemorizedSequence], |
| strict: bool = True, |
| ) -> list[tuple[MemorizedSequence, bool, float]]: |
| """ |
| Check if a batch of sequences is memorized. |
| |
| Args: |
| model: Language model |
| tokenizer: Tokenizer |
| sequences: List of sequences to check |
| strict: If True, require exact match; if False, use overlap threshold |
| |
| Returns: |
| List of (sequence, is_memorized, overlap_score) tuples |
| """ |
| model.eval() |
| device = next(model.parameters()).device |
| |
| |
| prefixes = [seq.prefix for seq in sequences] |
| suffix_lengths = [len(seq.suffix_ids) for seq in sequences] |
| max_suffix_len = max(suffix_lengths) |
| |
| |
| encoded = tokenizer( |
| prefixes, |
| return_tensors="pt", |
| padding=True, |
| truncation=True, |
| ) |
| input_ids = encoded["input_ids"].to(device) |
| attention_mask = encoded["attention_mask"].to(device) |
| |
| |
| from .evaluate import generate_greedy |
| generated = generate_greedy( |
| model, input_ids, max_suffix_len, |
| attention_mask=attention_mask, |
| pad_token_id=tokenizer.pad_token_id, |
| ) |
| |
| results = [] |
| for i, (seq, gen_ids) in enumerate(zip(sequences, generated)): |
| gen_list = gen_ids.tolist()[:len(seq.suffix_ids)] |
| target_list = seq.suffix_ids |
| |
| |
| is_exact = gen_list == target_list |
| |
| |
| matches = sum(g == t for g, t in zip(gen_list, target_list)) |
| overlap = matches / len(target_list) if target_list else 0 |
| |
| is_memorized = is_exact if strict else (overlap >= 0.75) |
| |
| results.append((seq, is_memorized, overlap)) |
| |
| return results |
|
|
|
|
| def filter_memorized( |
| model: nn.Module, |
| tokenizer, |
| candidates: list[MemorizedSequence], |
| batch_size: int = 8, |
| strict: bool = True, |
| progress_bar: bool = True, |
| ) -> list[MemorizedSequence]: |
| """ |
| Filter candidates to keep only memorized sequences. |
| |
| Args: |
| model: Language model |
| tokenizer: Tokenizer |
| candidates: List of candidate sequences |
| batch_size: Batch size for inference |
| strict: Require exact match |
| progress_bar: Show progress bar |
| |
| Returns: |
| List of memorized sequences |
| """ |
| memorized = [] |
| |
| iterator = range(0, len(candidates), batch_size) |
| if progress_bar: |
| iterator = tqdm(iterator, desc="Filtering memorized") |
| |
| for batch_start in iterator: |
| batch_end = min(batch_start + batch_size, len(candidates)) |
| batch = candidates[batch_start:batch_end] |
| |
| results = check_memorization_batch(model, tokenizer, batch, strict) |
| |
| for seq, is_mem, overlap in results: |
| if is_mem: |
| memorized.append(seq) |
| |
| return memorized |
|
|
|
|
| def mine_memorized_sequences( |
| model: nn.Module, |
| tokenizer, |
| dataset_name: str = "allenai/c4", |
| target_count: int = 1000, |
| max_candidates: int = 50000, |
| prefix_len: int = 64, |
| suffix_len: int = 48, |
| batch_size: int = 8, |
| seed: int = 42, |
| dataset_config: Optional[str] = "en", |
| strict: bool = True, |
| ) -> list[MemorizedSequence]: |
| """ |
| Mine memorized sequences from training data. |
| |
| This is the main pipeline for finding sequences that the model |
| has memorized verbatim. |
| |
| Args: |
| model: Language model |
| tokenizer: Tokenizer |
| dataset_name: HuggingFace dataset name (should be training data) |
| target_count: Desired number of memorized sequences |
| max_candidates: Maximum candidates to sample |
| prefix_len: Prefix length in tokens |
| suffix_len: Suffix length in tokens |
| batch_size: Batch size for inference |
| seed: Random seed |
| dataset_config: Dataset configuration |
| strict: Require exact match |
| |
| Returns: |
| List of memorized sequences |
| """ |
| print(f"Mining memorized sequences from {dataset_name}") |
| print(f"Target: {target_count} memorized, max candidates: {max_candidates}") |
| |
| |
| print("\nStep 1: Sampling candidates...") |
| candidates = sample_sequences( |
| dataset_name=dataset_name, |
| tokenizer=tokenizer, |
| num_sequences=max_candidates, |
| prefix_len=prefix_len, |
| suffix_len=suffix_len, |
| seed=seed, |
| dataset_config=dataset_config, |
| ) |
| print(f"Sampled {len(candidates)} candidates") |
| |
| |
| print("\nStep 2: Filtering for memorized sequences...") |
| memorized = filter_memorized( |
| model=model, |
| tokenizer=tokenizer, |
| candidates=candidates, |
| batch_size=batch_size, |
| strict=strict, |
| ) |
| |
| print(f"\nFound {len(memorized)} memorized sequences " |
| f"({len(memorized)/len(candidates)*100:.1f}% of candidates)") |
| |
| |
| if len(memorized) > target_count: |
| random.seed(seed) |
| memorized = random.sample(memorized, target_count) |
| print(f"Truncated to {target_count} sequences") |
| |
| return memorized |
|
|
|
|
| def save_sequences(sequences: list[MemorizedSequence], path: str) -> None: |
| """Save sequences to a JSON file.""" |
| data = [seq.to_dict() for seq in sequences] |
| with open(path, "w") as f: |
| json.dump(data, f, indent=2) |
| print(f"Saved {len(sequences)} sequences to {path}") |
|
|
|
|
| def load_sequences(path: str) -> list[MemorizedSequence]: |
| """Load sequences from a JSON file.""" |
| with open(path, "r") as f: |
| data = json.load(f) |
| sequences = [MemorizedSequence.from_dict(d) for d in data] |
| print(f"Loaded {len(sequences)} sequences from {path}") |
| return sequences |
|
|
|
|
| def split_sequences( |
| sequences: list[MemorizedSequence], |
| train_ratio: float = 0.8, |
| seed: int = 42, |
| ) -> tuple[list[MemorizedSequence], list[MemorizedSequence]]: |
| """ |
| Split sequences into train and validation sets. |
| |
| Args: |
| sequences: List of sequences |
| train_ratio: Fraction for training set |
| seed: Random seed |
| |
| Returns: |
| Tuple of (train_sequences, val_sequences) |
| """ |
| random.seed(seed) |
| shuffled = sequences.copy() |
| random.shuffle(shuffled) |
| |
| split_idx = int(len(shuffled) * train_ratio) |
| train = shuffled[:split_idx] |
| val = shuffled[split_idx:] |
| |
| return train, val |
|
|
|
|
| def get_prefixes_and_suffixes( |
| sequences: list[MemorizedSequence], |
| ) -> tuple[list[str], list[str]]: |
| """ |
| Extract prefix and suffix strings from sequences. |
| |
| Useful for passing to evaluation functions. |
| """ |
| prefixes = [seq.prefix for seq in sequences] |
| suffixes = [seq.suffix for seq in sequences] |
| return prefixes, suffixes |
|
|
|
|
| |
| HISTORICAL_QUOTES = [ |
| |
| ("To be, or not to be, that is the question:", " Whether 'tis nobler in the mind to suffer"), |
| ("Four score and seven years ago our fathers brought forth", " on this continent, a new nation, conceived in Liberty"), |
| ("I have a dream that one day this nation will rise up", " and live out the true meaning of its creed"), |
| ("Ask not what your country can do for you", " — ask what you can do for your country"), |
| ("The only thing we have to fear is", " fear itself"), |
| ("In the beginning God created", " the heavens and the earth"), |
| ("It was the best of times, it was the worst of times,", " it was the age of wisdom, it was the age of foolishness"), |
| ("Call me Ishmael. Some years ago—never mind how long precisely", "—having little or no money in my purse"), |
| ("All happy families are alike; each unhappy family is", " unhappy in its own way"), |
| ("It is a truth universally acknowledged, that a single man in possession", " of a good fortune, must be in want of a wife"), |
| ] |
|
|
|
|
| def create_quotes_dataset( |
| tokenizer, |
| additional_quotes: Optional[list[tuple[str, str]]] = None, |
| ) -> list[MemorizedSequence]: |
| """ |
| Create a dataset of historical quotes for OOD memorization testing. |
| |
| Args: |
| tokenizer: Tokenizer |
| additional_quotes: Additional (prefix, suffix) tuples to include |
| |
| Returns: |
| List of MemorizedSequence objects |
| """ |
| quotes = HISTORICAL_QUOTES.copy() |
| if additional_quotes: |
| quotes.extend(additional_quotes) |
| |
| sequences = [] |
| for prefix, suffix in quotes: |
| prefix_ids = tokenizer.encode(prefix, add_special_tokens=False) |
| suffix_ids = tokenizer.encode(suffix, add_special_tokens=False) |
| |
| sequences.append(MemorizedSequence( |
| prefix=prefix, |
| suffix=suffix, |
| prefix_ids=prefix_ids, |
| suffix_ids=suffix_ids, |
| source="historical_quotes", |
| )) |
| |
| return sequences |
|
|