| from contextlib import contextmanager |
|
|
| import torch |
| from torch import nn |
|
|
| import pytest |
| from titans_pytorch import NeuralMemory |
| from titans_pytorch.mac_transformer import flex_attention, SegmentedAttention, MemoryAsContextTransformer |
|
|
| |
|
|
| def exists(v): |
| return v is not None |
|
|
| def diff(x, y): |
| return (x - y).abs().amax() |
|
|
| @contextmanager |
| def torch_default_dtype(dtype): |
| prev_dtype = torch.get_default_dtype() |
| torch.set_default_dtype(dtype) |
| yield |
| torch.set_default_dtype(prev_dtype) |
|
|
| |
|
|
| @pytest.mark.parametrize('seq_len', (32, 512, 77)) |
| @pytest.mark.parametrize('silu', (False, True)) |
| @pytest.mark.parametrize('chunk_size, attn_pool_chunks', ((64, True), (64, False), (1, False))) |
| @pytest.mark.parametrize('momentum', (False, True)) |
| @pytest.mark.parametrize('qk_rmsnorm', (False, True)) |
| @pytest.mark.parametrize('heads', (1, 4)) |
| @pytest.mark.parametrize('max_grad_norm', (None, 2.)) |
| @pytest.mark.parametrize('num_kv_per_token', (1, 2)) |
| @pytest.mark.parametrize('per_parameter_lr_modulation', (False, True)) |
| @pytest.mark.parametrize('per_head_learned_parameters', (False, True)) |
| @pytest.mark.parametrize('test_store_mask', (False, True)) |
| def test_titans( |
| seq_len, |
| silu, |
| attn_pool_chunks, |
| chunk_size, |
| momentum, |
| qk_rmsnorm, |
| heads, |
| max_grad_norm, |
| num_kv_per_token, |
| per_parameter_lr_modulation, |
| per_head_learned_parameters, |
| test_store_mask |
| ): |
| mem = NeuralMemory( |
| dim = 16, |
| chunk_size = chunk_size, |
| activation = nn.SiLU() if silu else None, |
| attn_pool_chunks = attn_pool_chunks, |
| max_grad_norm = max_grad_norm, |
| num_kv_per_token = num_kv_per_token, |
| momentum = momentum, |
| qk_rmsnorm = qk_rmsnorm, |
| heads = heads, |
| per_parameter_lr_modulation = per_parameter_lr_modulation, |
| per_head_learned_parameters = per_head_learned_parameters |
| ) |
|
|
| seq = torch.randn(2, seq_len, 16) |
|
|
| store_mask = None |
|
|
| if test_store_mask: |
| store_mask = torch.randint(0, 2, (2, seq_len)).bool() |
|
|
| retrieved, _ = mem(seq, store_mask = store_mask) |
|
|
| assert seq.shape == retrieved.shape |
|
|
| def test_return_surprises(): |
|
|
| mem = NeuralMemory( |
| dim = 384, |
| chunk_size = 2, |
| dim_head = 64, |
| heads = 4, |
| ) |
|
|
| seq = torch.randn(4, 64, 384) |
|
|
| _, _, (surprises, adaptive_lr) = mem(seq, return_surprises = True) |
|
|
| assert all([t.shape == (4, 4, 64) for t in (surprises, adaptive_lr)]) |
|
|
| @pytest.mark.parametrize('learned_momentum_combine', (False, True)) |
| @pytest.mark.parametrize('learned_combine_include_zeroth', (False, True)) |
| def test_titans_second_order_momentum( |
| learned_momentum_combine, |
| learned_combine_include_zeroth |
| ): |
|
|
| mem = NeuralMemory( |
| dim = 384, |
| dim_head = 64, |
| heads = 2, |
| chunk_size = 1, |
| batch_size = 2, |
| momentum_order = 2, |
| learned_momentum_combine = learned_momentum_combine, |
| learned_combine_include_zeroth = learned_combine_include_zeroth |
| ) |
|
|
| seq = torch.randn(2, 5, 384) |
|
|
| parallel_retrieved, state = mem(seq) |
| assert seq.shape == parallel_retrieved.shape |
|
|
| def test_titans_attn_memory(): |
| from titans_pytorch.memory_models import MemoryAttention |
|
|
| mem = NeuralMemory( |
| dim = 16, |
| chunk_size = 64, |
| model = MemoryAttention( |
| dim = 16 |
| ) |
| ) |
|
|
| seq = torch.randn(2, 1024, 16) |
| retrieved, _ = mem(seq) |
|
|
| assert seq.shape == retrieved.shape |
|
|
| def test_swiglu_ff_memory(): |
| from titans_pytorch.memory_models import MemorySwiGluMLP |
|
|
| mem = NeuralMemory( |
| dim = 16, |
| chunk_size = 2, |
| mem_model_norm_add_residual = False, |
| model = MemorySwiGluMLP( |
| dim = 16, |
| depth = 2 |
| ) |
| ) |
|
|
| seq = torch.randn(2, 64, 16) |
| retrieved, _ = mem(seq) |
|
|
| assert seq.shape == retrieved.shape |
|
|
| @pytest.mark.parametrize('gated_transition', (True, False)) |
| def test_neural_mem_chaining_chunks( |
| gated_transition |
| ): |
| mem = NeuralMemory( |
| dim = 16, |
| dim_head = 16, |
| heads = 2, |
| chunk_size = 16, |
| gated_transition = gated_transition |
| ) |
|
|
| seq = torch.randn(2, 48, 16) |
|
|
| parallel_retrieved, state = mem(seq) |
|
|
| seq_first, seq_second, seq_third = seq.split(16, dim = 1) |
|
|
| first_retrieved, state = mem(seq_first) |
| second_retrieved, state = mem(seq_second, state = state) |
| third_retrieved, state = mem(seq_third, state = state) |
|
|
| assert torch.allclose(parallel_retrieved, torch.cat((first_retrieved, second_retrieved, third_retrieved), dim = 1), atol = 1e-5) |
|
|
| def test_neural_mem_chaining_with_weight_residual(): |
| mem = NeuralMemory( |
| dim = 16, |
| dim_head = 16, |
| heads = 2, |
| chunk_size = 64 |
| ) |
|
|
| mem2 = NeuralMemory( |
| dim = 16, |
| dim_head = 16, |
| heads = 2, |
| chunk_size = 64, |
| accept_weight_residual = True |
| ) |
|
|
| seq = torch.randn(2, 256, 16) |
|
|
| seq, state = mem(seq) |
|
|
| parallel_retrieved, _ = mem2(seq, prev_weights = state.updates) |
|
|
| seq_first, seq_second = seq[:, :128], seq[:, 128:] |
|
|
| first_retrieved, state1 = mem2(seq_first, prev_weights = state.updates) |
| second_retrieved, state2 = mem2(seq_second, state = state1, prev_weights = state.updates) |
|
|
| assert torch.allclose(parallel_retrieved, torch.cat((first_retrieved, second_retrieved), dim = 1), atol = 1e-5) |
|
|
| def test_neural_mem_chaining_with_batch_size(): |
| mem = NeuralMemory( |
| dim = 16, |
| dim_head = 16, |
| heads = 2, |
| chunk_size = 16, |
| batch_size = 64 |
| ) |
|
|
| seq = torch.randn(2, 112, 16) |
|
|
| parallel_retrieved, state = mem(seq) |
|
|
| seq_first, seq_second, seq_third = seq[:, :16], seq[:, 16:64], seq[:, 64:] |
|
|
| first_retrieved, state = mem(seq_first) |
| second_retrieved, state = mem(seq_second, state = state) |
| third_retrieved, state = mem(seq_third, state = state) |
|
|
| parallel_part_retrieved = torch.cat((first_retrieved, second_retrieved, third_retrieved), dim = 1) |
|
|
| assert torch.allclose(parallel_retrieved, parallel_part_retrieved, atol = 1e-5) |
|
|
| @pytest.mark.parametrize('seq_len', (1023, 17)) |
| @pytest.mark.parametrize('num_persist_mem_tokens', (0, 16)) |
| @pytest.mark.parametrize('num_longterm_mem_tokens', (0, 16)) |
| @pytest.mark.parametrize('neural_mem_gate_attn_output', (False, True)) |
| @pytest.mark.parametrize('neural_mem_segment_len', (8, 16)) |
| @pytest.mark.parametrize('neural_mem_weight_residual', (False, True)) |
| @pytest.mark.parametrize('neural_mem_batch_size', (None, 64)) |
| @pytest.mark.parametrize('neural_mem_qkv_receives_diff_views', (False, True)) |
| @pytest.mark.parametrize('neural_mem_momentum', (False, True)) |
| def test_mac( |
| seq_len, |
| num_persist_mem_tokens, |
| num_longterm_mem_tokens, |
| neural_mem_gate_attn_output, |
| neural_mem_segment_len, |
| neural_mem_weight_residual, |
| neural_mem_batch_size, |
| neural_mem_qkv_receives_diff_views, |
| neural_mem_momentum |
| ): |
| transformer = MemoryAsContextTransformer( |
| num_tokens = 256, |
| dim = 16, |
| depth = 2, |
| num_persist_mem_tokens = num_persist_mem_tokens, |
| num_longterm_mem_tokens = num_longterm_mem_tokens, |
| segment_len = 128, |
| neural_mem_gate_attn_output = neural_mem_gate_attn_output, |
| neural_memory_segment_len = neural_mem_segment_len, |
| neural_memory_batch_size = neural_mem_batch_size, |
| neural_memory_qkv_receives_diff_views = neural_mem_qkv_receives_diff_views, |
| neural_mem_weight_residual = neural_mem_weight_residual, |
| neural_memory_kwargs = dict( |
| momentum = neural_mem_momentum |
| ) |
| ) |
|
|
| x = torch.randint(0, 256, (1, seq_len)) |
|
|
| logits = transformer(x) |
| assert logits.shape == (1, seq_len, 256) |
|
|
| @pytest.mark.parametrize('sliding', (False, True)) |
| @pytest.mark.parametrize('mem_layers', ((), None)) |
| @pytest.mark.parametrize('longterm_mems', (0, 4, 16)) |
| @pytest.mark.parametrize('prompt_len', (4, 16)) |
| @torch_default_dtype(torch.float64) |
| def test_mac_sampling( |
| sliding, |
| mem_layers, |
| longterm_mems, |
| prompt_len |
| ): |
| transformer = MemoryAsContextTransformer( |
| num_tokens = 256, |
| dim = 16, |
| depth = 4, |
| segment_len = 32, |
| num_persist_mem_tokens = 4, |
| num_longterm_mem_tokens = longterm_mems, |
| sliding_window_attn = sliding, |
| neural_memory_layers = mem_layers, |
| neural_mem_gate_attn_output = False |
| ) |
|
|
| ids = torch.randint(0, 256, (1, 1023)) |
|
|
| |
|
|
| prompt = ids[:, :prompt_len] |
|
|
| sampled = transformer.sample(prompt, 53, use_cache = False, temperature = 0.) |
| sampled_with_cache = transformer.sample(prompt, 53, use_cache = True, temperature = 0.) |
|
|
| assert torch.allclose(sampled, sampled_with_cache) |
|
|
| @pytest.mark.parametrize('seq_len', (2, 64, 256)) |
| @pytest.mark.parametrize('prompt_len', (0, 65)) |
| @pytest.mark.parametrize('mem_chunk_size', (2, 32, 64)) |
| @pytest.mark.parametrize('gated_transition', (False, True)) |
| @torch_default_dtype(torch.float64) |
| def test_neural_mem_inference( |
| seq_len, |
| prompt_len, |
| mem_chunk_size, |
| gated_transition |
| ): |
|
|
| mem = NeuralMemory( |
| dim = 16, |
| chunk_size = mem_chunk_size, |
| gated_transition = gated_transition |
| ) |
|
|
| seq = torch.randn(2, seq_len, 16) |
| parallel_retrieved, _ = mem(seq) |
|
|
| assert seq.shape == parallel_retrieved.shape |
|
|
| state = None |
| sequential_retrieved = [] |
|
|
| |
|
|
| test_parallel_prompt = prompt_len > 0 and prompt_len < seq_len |
|
|
| if test_parallel_prompt: |
| prompt, seq = seq[:, :prompt_len], seq[:, prompt_len:] |
| retrieved_prompt, state = mem(prompt) |
| sequential_retrieved.append(retrieved_prompt) |
|
|
| |
|
|
| for token in seq.unbind(dim = 1): |
|
|
| one_retrieved, state = mem.forward( |
| token, |
| state = state, |
| ) |
|
|
| sequential_retrieved.append(one_retrieved) |
|
|
| sequential_retrieved = torch.cat(sequential_retrieved, dim = -2) |
|
|
| assert torch.allclose(parallel_retrieved, sequential_retrieved, atol = 1e-6) |
|
|
| @pytest.mark.parametrize('seq_len', (1023, 17)) |
| @pytest.mark.parametrize('sliding', (True, False)) |
| def test_flex( |
| seq_len, |
| sliding |
| ): |
| if not (torch.cuda.is_available() and exists(flex_attention)): |
| pytest.skip() |
|
|
| attn = SegmentedAttention( |
| dim = 16, |
| segment_len = 32, |
| num_persist_mem_tokens = 1, |
| num_longterm_mem_tokens = 1, |
| use_flex_attn = True, |
| sliding = sliding |
| ).cuda() |
|
|
| seq = torch.randn(1, seq_len, 16).cuda() |
|
|
| out_flex, _ = attn(seq) |
| out_non_flex, _ = attn(seq, disable_flex_attn = True) |
|
|
| assert torch.allclose(out_flex, out_non_flex, atol = 1e-5) |
|
|
| @pytest.mark.parametrize('use_accelerated', (True, False)) |
| def test_assoc_scan( |
| use_accelerated |
| ): |
| from titans_pytorch.neural_memory import AssocScan |
|
|
| if use_accelerated and not torch.cuda.is_available(): |
| pytest.skip() |
|
|
| scan = AssocScan(use_accelerated = use_accelerated) |
|
|
| seq_len = 128 |
| mid_point = seq_len // 2 |
|
|
| gates = torch.randn(2, seq_len, 16).sigmoid() |
| inputs = torch.randn(2, seq_len, 16) |
|
|
| if use_accelerated: |
| gates = gates.cuda() |
| inputs = inputs.cuda() |
|
|
| output = scan(gates, inputs) |
|
|
| gates1, gates2 = gates[:, :mid_point], gates[:, mid_point:] |
| inputs1, inputs2 = inputs[:, :mid_point], inputs[:, mid_point:] |
|
|
| first_half = scan(gates1, inputs1) |
|
|
| second_half = scan(gates2, inputs2, prev = first_half[:, -1]) |
| assert second_half.shape == inputs2.shape |
|
|
| assert torch.allclose(output[:, -1], second_half[:, -1], atol = 1e-5) |
|
|
| def test_mem_state_detach(): |
| from titans_pytorch.neural_memory import mem_state_detach |
|
|
| mem = NeuralMemory( |
| dim = 384, |
| chunk_size = 2, |
| qk_rmsnorm = True, |
| dim_head = 64, |
| heads = 4, |
| ) |
|
|
| seq = torch.randn(4, 64, 384) |
|
|
| state = None |
|
|
| for _ in range(2): |
| parallel_retrieved, state = mem(seq, state = state) |
| state = mem_state_detach(state) |
| parallel_retrieved.sum().backward() |
|
|