gemma4-zero-compute / modeling_gemma4.py
haysonC's picture
Upload optimized Gemma4 checkpoint
0b0ec56 verified
# 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# This file was automatically generated from src/transformers/models/gemma4/modular_gemma4.py.
# Do NOT edit this file manually as any edits will be overwritten by the generation of
# the file from the modular. If any change should be done, please apply the change to the
# modular_gemma4.py file directly. One of our CI enforces this.
# 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
# Copyright 2026 the HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import math
from collections.abc import Callable
from dataclasses import dataclass
from functools import cached_property
from typing import Optional
import torch
from torch import nn
from torch.nn import functional as F
from transformers import initialization as init
from transformers.activations import ACT2FN
from transformers.cache_utils import Cache, DynamicCache
from transformers.configuration_utils import PreTrainedConfig
from transformers.generation import GenerationMixin
from transformers.integrations import use_experts_implementation, use_kernelized_func
from transformers.masking_utils import (
create_bidirectional_mask,
create_causal_mask,
create_masks_for_generate,
create_sliding_window_causal_mask,
)
from transformers.modeling_flash_attention_utils import FlashAttentionKwargs
from transformers.modeling_layers import GradientCheckpointingLayer
from transformers.modeling_outputs import BaseModelOutputWithPast, BaseModelOutputWithPooling, CausalLMOutputWithPast
from transformers.modeling_rope_utils import ROPE_INIT_FUNCTIONS, dynamic_rope_update
from transformers.modeling_utils import ALL_ATTENTION_FUNCTIONS, PreTrainedModel
from transformers.processing_utils import Unpack
from transformers.utils import ModelOutput, TransformersKwargs, auto_docstring, can_return_tuple, torch_compilable_check
from transformers.utils.generic import maybe_autocast, merge_with_config_defaults
from transformers.utils.output_capturing import OutputRecorder, capture_outputs
from transformers.models.auto.modeling_auto import AutoModel
from .configuration_gemma4 import Gemma4AudioConfig, Gemma4Config, Gemma4TextConfig, Gemma4VisionConfig
@dataclass
@auto_docstring(
custom_intro="""
Base class for Gemma4 outputs, with hidden states and attentions.
"""
)
class Gemma4ModelOutputWithPast(BaseModelOutputWithPast):
r"""
past_key_values (`Cache`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`):
It is a [`~cache_utils.Cache`] instance. For more details, see our [kv cache guide](https://huggingface.co/docs/transformers/en/kv_cache).
Contains pre-computed hidden-states (key and values in the self-attention blocks) that can be used (see
`past_key_values` input) to speed up sequential decoding.
image_hidden_states (`torch.FloatTensor`, *optional*):
A `torch.FloatTensor` of size `(batch_size, num_images, sequence_length, hidden_size)`.
image_hidden_states of the model produced by the vision encoder and after projecting the last hidden state.
audio_hidden_states (`torch.FloatTensor`, *optional*):
A `torch.FloatTensor` of size `(batch_size, num_images, sequence_length, hidden_size)`.
audio_hidden_states of the model produced by the audio encoder and after projecting the last hidden state.
"""
image_hidden_states: torch.FloatTensor | None = None
audio_hidden_states: torch.FloatTensor | None = None
@dataclass
@auto_docstring(
custom_intro="""
Base class for Gemma4 causal language model (or autoregressive) outputs.
"""
)
class Gemma4CausalLMOutputWithPast(ModelOutput):
r"""
loss (`torch.FloatTensor` of shape `(1,)`, *optional*, returned when `labels` is provided):
Language modeling loss (for next-token prediction).
logits (`torch.FloatTensor` of shape `(batch_size, sequence_length, config.text_config.vocab_size)`):
Prediction scores of the language modeling head (scores for each vocabulary token before SoftMax).
past_key_values (`Cache`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`):
It is a [`~cache_utils.Cache`] instance. For more details, see our [kv cache guide](https://huggingface.co/docs/transformers/en/kv_cache).
Contains pre-computed hidden-states (key and values in the self-attention blocks) that can be used (see
`past_key_values` input) to speed up sequential decoding.
image_hidden_states (`torch.FloatTensor`, *optional*):
A `torch.FloatTensor` of size `(batch_size, num_images, sequence_length, hidden_size)`.
image_hidden_states of the model produced by the vision encoder after projecting last hidden state.
audio_hidden_states (`torch.FloatTensor`, *optional*):
A `torch.FloatTensor` of size `(batch_size, num_images, sequence_length, hidden_size)`.
audio_hidden_states of the model produced by the audio encoder and after projecting the last hidden state.
"""
loss: torch.FloatTensor | None = None
logits: torch.FloatTensor | None = None
past_key_values: Cache | None = None
hidden_states: tuple[torch.FloatTensor] | None = None
attentions: tuple[torch.FloatTensor] | None = None
image_hidden_states: torch.FloatTensor | None = None
audio_hidden_states: torch.FloatTensor | None = None
@dataclass
@auto_docstring
class Gemma4AudioModelOutput(BaseModelOutputWithPooling):
r"""
attention_mask (`torch.BoolTensor`, *optional*):
A torch.BoolTensor of shape `(batch_size, num_frames)`. True for valid positions, False for padding.
"""
attention_mask: torch.BoolTensor | None = None
class Gemma4ClippableLinear(nn.Module):
def __init__(
self,
config: Gemma4VisionConfig | Gemma4AudioConfig,
in_features: int,
out_features: int,
) -> None:
super().__init__()
self.use_clipped_linears = config.use_clipped_linears
self.linear = nn.Linear(in_features, out_features, bias=False)
if self.use_clipped_linears:
self.register_buffer("input_min", torch.tensor(-float("inf")))
self.register_buffer("input_max", torch.tensor(float("inf")))
self.register_buffer("output_min", torch.tensor(-float("inf")))
self.register_buffer("output_max", torch.tensor(float("inf")))
def forward(self, hidden_states: torch.Tensor) -> torch.Tensor:
if self.use_clipped_linears:
hidden_states = torch.clamp(hidden_states, self.input_min, self.input_max)
hidden_states = self.linear(hidden_states)
if self.use_clipped_linears:
hidden_states = torch.clamp(hidden_states, self.output_min, self.output_max)
return hidden_states
class Gemma4RMSNorm(nn.Module):
def __init__(self, dim: int, eps: float = 1e-6, with_scale: bool = True):
super().__init__()
self.eps = eps
self.with_scale = with_scale
if self.with_scale:
self.weight = nn.Parameter(torch.ones(dim), requires_grad=True)
def _norm(self, hidden_states: torch.Tensor):
mean_squared = hidden_states.pow(2).mean(-1, keepdim=True) + self.eps
# Use torch.pow() (over torch.sqrt() or torch.rsqrt()) to addess compiler differences between Torch and JAX
return hidden_states * torch.pow(mean_squared, -0.5)
def forward(self, hidden_states: torch.Tensor) -> torch.Tensor:
normed_output = self._norm(hidden_states.float())
if self.with_scale:
normed_output = normed_output * self.weight.float()
return normed_output.type_as(hidden_states)
class Gemma4AudioRelPositionalEncoding(nn.Module):
"""Sinusoidal relative positional encoding for the audio encoder.
Produces position embeddings of shape [1, 2*context_size - 1, hidden_size] with
concatenated [sin..., cos...] layout matching the original Gemma4 convention.
"""
inv_timescales: torch.Tensor
def __init__(self, config: Gemma4AudioConfig):
super().__init__()
self.hidden_size = config.hidden_size
self.context_size = (
config.attention_chunk_size + config.attention_context_left - 1 + config.attention_context_right
)
min_timescale = 1.0
max_timescale = 10000.0
num_timescales = self.hidden_size // 2
log_timescale_increment = math.log(max_timescale / min_timescale) / max(num_timescales - 1, 1)
inv_timescales = min_timescale * torch.exp(torch.arange(num_timescales) * -log_timescale_increment)
self.register_buffer("inv_timescales", inv_timescales.unsqueeze(0).unsqueeze(0), persistent=False)
@torch.no_grad()
def forward(self, hidden_states: torch.Tensor) -> torch.Tensor:
position_ids = torch.arange(12, -1, -1, device=hidden_states.device)
position_ids = position_ids[..., None]
scaled_time = position_ids * self.inv_timescales.to(device=hidden_states.device)
pos_embed = torch.cat([torch.sin(scaled_time), torch.cos(scaled_time)], dim=-1)
return pos_embed.to(dtype=hidden_states.dtype)
class Gemma4AudioAttention(nn.Module):
"""Chunked local attention with relative position bias"""
def __init__(self, config: Gemma4AudioConfig, layer_idx: int):
super().__init__()
self.config = config
self.layer_idx = layer_idx
self.attention_logits_soft_cap = config.attention_logit_cap
self.head_dim = config.hidden_size // config.num_attention_heads
self.num_heads = config.num_attention_heads
self.q_scale = (self.head_dim**-0.5) / math.log(2)
self.k_scale = math.log(1 + math.e) / math.log(2)
self.chunk_size = config.attention_chunk_size
self.max_past_horizon = config.attention_context_left - 1
self.max_future_horizon = config.attention_context_right
self.context_size = self.chunk_size + self.max_past_horizon + self.max_future_horizon
self.q_proj = Gemma4ClippableLinear(config, config.hidden_size, self.num_heads * self.head_dim)
self.k_proj = Gemma4ClippableLinear(config, config.hidden_size, self.num_heads * self.head_dim)
self.v_proj = Gemma4ClippableLinear(config, config.hidden_size, self.num_heads * self.head_dim)
self.post = Gemma4ClippableLinear(config, config.hidden_size, config.hidden_size)
self.relative_k_proj = nn.Linear(config.hidden_size, self.num_heads * self.head_dim, bias=False)
self.per_dim_scale = nn.Parameter(torch.zeros(self.head_dim))
self.register_buffer("softcap", torch.tensor(self.attention_logits_soft_cap), persistent=False)
def _convert_to_block(self, hidden_states: torch.Tensor) -> torch.Tensor:
"""Splits a `(batch_size, seq_len, num_heads, head_dim)` tensor into non-overlapping blocks of `chunk_size` along the sequence dim."""
batch_size, seq_len, num_heads, head_dim = hidden_states.shape
num_blocks = (seq_len + self.chunk_size - 1) // self.chunk_size
pad = num_blocks * self.chunk_size - seq_len
hidden_states = F.pad(hidden_states, (0, 0, 0, 0, 0, pad))
return hidden_states.reshape(batch_size, num_blocks, self.chunk_size, num_heads, head_dim).contiguous()
def _extract_block_context(self, hidden_states: torch.Tensor) -> torch.Tensor:
"""Extracts overlapping context windows of `context_size` for every block, strided by `chunk_size`."""
batch_size, seq_len, num_heads, head_dim = hidden_states.shape
hidden_states = F.pad(
hidden_states, (0, 0, 0, 0, self.max_past_horizon, self.max_future_horizon + self.chunk_size - 1)
)
hidden_states = hidden_states.unfold(1, self.context_size, self.chunk_size)
hidden_states = torch.movedim(hidden_states, -1, 2)
return hidden_states.contiguous()
def _rel_shift(self, x: torch.Tensor) -> torch.Tensor:
"""Relative position shift for blocked attention. See appendix B of https://huggingface.co/papers/1901.02860."""
batch_size, num_heads, num_blocks, block_size, position_length = x.shape
context_size = self.context_size
x = F.pad(x, (0, context_size + 1 - position_length))
x = x.view(batch_size, num_heads, num_blocks, block_size * (context_size + 1))
x = x[..., : block_size * context_size]
return x.view(batch_size, num_heads, num_blocks, block_size, context_size)
def forward(
self,
hidden_states: torch.Tensor,
position_embeddings: torch.Tensor,
attention_mask: torch.BoolTensor | None = None,
) -> tuple[torch.Tensor, None]:
batch_size, seq_length, _ = hidden_states.shape
hidden_shape = (batch_size, seq_length, self.num_heads, self.head_dim)
query_states = self.q_proj(hidden_states).float().view(hidden_shape)
key_states = self.k_proj(hidden_states).float().view(hidden_shape)
value_states = self.v_proj(hidden_states).float().view(hidden_shape)
query_states = query_states * self.q_scale * F.softplus(self.per_dim_scale)
key_states = key_states * self.k_scale
query_states = self._convert_to_block(query_states)
key_states = self._extract_block_context(key_states)
value_states = self._extract_block_context(value_states)
num_blocks = query_states.shape[1]
relative_key_states = self.relative_k_proj(position_embeddings)
relative_key_states = relative_key_states.view(-1, self.num_heads, self.head_dim)
relative_key_states = relative_key_states.to(dtype=query_states.dtype)
queries = query_states.permute(0, 3, 1, 2, 4)
matrix_ac = queries @ key_states.permute(0, 3, 1, 4, 2)
queries_flat = queries.reshape(batch_size, self.num_heads, -1, self.head_dim)
matrix_bd = queries_flat @ relative_key_states.permute(1, 2, 0)
matrix_bd = matrix_bd.reshape(batch_size, self.num_heads, num_blocks, self.chunk_size, -1)
matrix_bd = self._rel_shift(matrix_bd)
attn_weights = matrix_ac + matrix_bd
attn_weights = attn_weights / self.softcap
attn_weights = torch.tanh(attn_weights)
attn_weights = attn_weights * self.softcap
if attention_mask is not None:
attn_weights = attn_weights.masked_fill(
attention_mask.logical_not(), self.config.attention_invalid_logits_value
)
attn_weights = F.softmax(attn_weights, dim=-1, dtype=torch.float32).to(value_states.dtype)
attn_output = attn_weights @ value_states.permute(0, 3, 1, 2, 4)
attn_output = attn_output.permute(0, 2, 3, 1, 4).reshape(batch_size, num_blocks * self.chunk_size, -1)
attn_output = attn_output[:, :seq_length].contiguous()
attn_output = self.post(attn_output.to(dtype=self.post.linear.weight.dtype))
return attn_output, attn_weights
class Gemma4AudioSubSampleConvProjectionLayer(nn.Module):
def __init__(self, in_channels, out_channels, norm_eps):
super().__init__()
self.conv = nn.Conv2d(
in_channels=in_channels,
out_channels=out_channels,
kernel_size=(3, 3),
stride=(2, 2),
padding=1,
bias=False,
)
self.norm = nn.LayerNorm(out_channels, eps=norm_eps, elementwise_affine=True, bias=False)
self.act = nn.ReLU()
def forward(self, hidden_states: torch.Tensor, mask: torch.Tensor | None = None):
if mask is not None:
mask = mask.to(device=hidden_states.device)
hidden_states = hidden_states * mask[:, None, :, None]
hidden_states = self.conv(hidden_states.to(self.conv.weight.dtype))
hidden_states = self.act(self.norm(hidden_states.permute(0, 2, 3, 1)).permute(0, 3, 1, 2).contiguous())
if mask is not None:
mask = mask[:, ::2]
return hidden_states, mask
class Gemma4AudioSubSampleConvProjection(nn.Module):
def __init__(self, config: Gemma4AudioConfig):
super().__init__()
self.layer0 = Gemma4AudioSubSampleConvProjectionLayer(
in_channels=1,
out_channels=config.subsampling_conv_channels[0],
norm_eps=config.rms_norm_eps,
)
self.layer1 = Gemma4AudioSubSampleConvProjectionLayer(
in_channels=config.subsampling_conv_channels[0],
out_channels=config.subsampling_conv_channels[1],
norm_eps=config.rms_norm_eps,
)
proj_input_dim = (config.subsampling_conv_channels[0] // 4) * config.subsampling_conv_channels[1]
self.input_proj_linear = nn.Linear(proj_input_dim, config.hidden_size, bias=False)
def forward(
self,
input_features: torch.Tensor,
input_features_mask: torch.Tensor | None = None,
) -> tuple[torch.Tensor, torch.Tensor]:
hidden_states = input_features.unsqueeze(1)
hidden_states, mask = self.layer0(hidden_states, input_features_mask)
hidden_states, mask = self.layer1(hidden_states, mask)
batch_size, _, seq_len, _ = hidden_states.shape
hidden_states = hidden_states.permute(0, 2, 3, 1).contiguous().reshape(batch_size, seq_len, -1)
return self.input_proj_linear(hidden_states), mask
class Gemma4AudioFeedForward(nn.Module):
def __init__(self, config: Gemma4AudioConfig):
super().__init__()
self.config = config
self.ffw_layer_1 = Gemma4ClippableLinear(config, config.hidden_size, config.hidden_size * 4)
self.ffw_layer_2 = Gemma4ClippableLinear(config, config.hidden_size * 4, config.hidden_size)
self.pre_layer_norm = Gemma4RMSNorm(config.hidden_size)
self.post_layer_norm = Gemma4RMSNorm(config.hidden_size)
self.act_fn = ACT2FN[config.hidden_act]
self.gradient_clipping = config.gradient_clipping
self.post_layer_scale = config.residual_weight
def forward(self, hidden_states: torch.Tensor) -> torch.Tensor:
# This is needed to avoid any underflow/overflow issues when clipping
gradient_clipping = min(self.gradient_clipping, torch.finfo(self.ffw_layer_1.linear.weight.dtype).max)
residual = hidden_states
hidden_states = torch.clamp(hidden_states, -gradient_clipping, gradient_clipping)
hidden_states = self.pre_layer_norm(hidden_states)
hidden_states = self.ffw_layer_1(hidden_states)
hidden_states = self.act_fn(hidden_states)
hidden_states = self.ffw_layer_2(hidden_states)
hidden_states = torch.clamp(hidden_states, -gradient_clipping, gradient_clipping)
hidden_states = self.post_layer_norm(hidden_states)
hidden_states *= self.post_layer_scale
hidden_states += residual
return hidden_states
# TODO: this could be imported from Voxtral realtime
class Gemma4AudioCausalConv1d(nn.Conv1d):
# def __init__(
# self,
# in_channels: int,
# out_channels: int,
# kernel_size: int,
# # cache_key: str,
# stride: int = 1,
# dilation: int = 1,
# bias: bool = True,
# ):
# super().__init__(in_channels, out_channels, kernel_size, stride=stride, dilation=dilation, bias=bias)
# self.cache_key = cache_key
@cached_property
def left_pad(self):
effective_kernel_size = (self.kernel_size[0] - 1) * self.dilation[0] + 1
return effective_kernel_size - self.stride[0]
def forward(
self,
x: torch.Tensor,
# padding_cache: VoxtralRealtimeConv1dPaddingCache | None = None, # TODO: we might want to add a cache?
) -> torch.Tensor:
# if padding_cache is not None:
# x = padding_cache.update(x, self.cache_key, self)
# else:
# x = nn.functional.pad(x, (self.left_pad, 0))
x = nn.functional.pad(x, (self.left_pad, 0))
return super().forward(x)
class Gemma4AudioLightConv1d(nn.Module):
def __init__(self, config: Gemma4AudioConfig):
super().__init__()
self.config = config
self.linear_start = Gemma4ClippableLinear(config, config.hidden_size, config.hidden_size * 2)
self.linear_end = Gemma4ClippableLinear(config, config.hidden_size, config.hidden_size)
self.depthwise_conv1d = Gemma4AudioCausalConv1d(
in_channels=config.hidden_size,
out_channels=config.hidden_size,
kernel_size=config.conv_kernel_size,
groups=config.hidden_size,
bias=False,
)
self.pre_layer_norm = Gemma4RMSNorm(config.hidden_size, eps=config.rms_norm_eps, with_scale=True)
self.conv_norm = Gemma4RMSNorm(config.hidden_size, eps=config.rms_norm_eps, with_scale=True)
self.act_fn = ACT2FN[config.hidden_act]
self.gradient_clipping = config.gradient_clipping
def forward(self, hidden_states: torch.Tensor) -> torch.Tensor:
residual = hidden_states
hidden_states = self.pre_layer_norm(hidden_states)
hidden_states = self.linear_start(hidden_states)
hidden_states = nn.functional.glu(hidden_states, dim=-1)
hidden_states = self.depthwise_conv1d(hidden_states.transpose(1, 2)).transpose(1, 2)
# This is needed to avoid any underflow/overflow issues when clipping
gradient_clipping = min(self.gradient_clipping, torch.finfo(self.linear_start.linear.weight.dtype).max)
hidden_states = torch.clamp(hidden_states, -gradient_clipping, gradient_clipping)
hidden_states = self.conv_norm(hidden_states)
hidden_states = self.act_fn(hidden_states)
hidden_states = self.linear_end(hidden_states)
hidden_states += residual
return hidden_states
class Gemma4AudioLayer(nn.Module):
def __init__(self, config: Gemma4AudioConfig, layer_idx: int):
super().__init__()
self.config = config
self.feed_forward1 = Gemma4AudioFeedForward(config)
self.feed_forward2 = Gemma4AudioFeedForward(config)
self.self_attn = Gemma4AudioAttention(config, layer_idx)
self.lconv1d = Gemma4AudioLightConv1d(config)
self.norm_pre_attn = Gemma4RMSNorm(config.hidden_size)
self.norm_post_attn = Gemma4RMSNorm(config.hidden_size)
self.norm_out = Gemma4RMSNorm(config.hidden_size)
self.gradient_clipping = config.gradient_clipping
def forward(
self,
hidden_states: torch.Tensor,
attention_mask: torch.BoolTensor | None,
position_embeddings: torch.Tensor,
**kwargs: Unpack[TransformersKwargs],
) -> torch.Tensor:
# This is needed to avoid any underflow/overflow issues when clipping
gradient_clipping = min(self.gradient_clipping, torch.finfo(self.norm_pre_attn.weight.dtype).max)
hidden_states = self.feed_forward1(hidden_states)
residual = hidden_states
hidden_states = torch.clamp(hidden_states, -gradient_clipping, gradient_clipping)
hidden_states = self.norm_pre_attn(hidden_states)
hidden_states, _ = self.self_attn(
hidden_states=hidden_states,
position_embeddings=position_embeddings,
attention_mask=attention_mask,
)
hidden_states = torch.clamp(hidden_states, -gradient_clipping, gradient_clipping)
hidden_states = self.norm_post_attn(hidden_states)
hidden_states += residual
hidden_states = self.lconv1d(hidden_states)
hidden_states = self.feed_forward2(hidden_states)
hidden_states = torch.clamp(hidden_states, -gradient_clipping, gradient_clipping)
hidden_states = self.norm_out(hidden_states)
return hidden_states
# ---- Vision Encoder Layers ----
class Gemma4VisionPatchEmbedder(nn.Module):
def __init__(self, config: Gemma4VisionConfig):
super().__init__()
self.config = config
self.hidden_size = config.hidden_size
self.patch_size = config.patch_size
self.position_embedding_size = config.position_embedding_size
self.input_proj = nn.Linear(3 * self.patch_size**2, self.hidden_size, bias=False)
self.position_embedding_table = nn.Parameter(torch.ones(2, self.position_embedding_size, self.hidden_size))
def _position_embeddings(self, pixel_position_ids: torch.Tensor, padding_positions: torch.Tensor) -> torch.Tensor:
"""Prepare patch positions map for matmul with positon embedding table."""
# Expanding and permute patch positions to (batch_size, num_patches, 2, position_embedding_size) for matmul.
clamped_positions = pixel_position_ids.clamp(min=0)
one_hot = F.one_hot(clamped_positions, num_classes=self.position_embedding_size)
one_hot = one_hot.permute(0, 2, 1, 3).to(self.position_embedding_table)
# Compute positional embeddings and sum across x and y.
position_embeddings = one_hot @ self.position_embedding_table
position_embeddings = position_embeddings.sum(dim=1)
# Zero out embeddings for any padding patches.
position_embeddings = torch.where(padding_positions.unsqueeze(-1), 0.0, position_embeddings)
return position_embeddings
def forward(
self, pixel_values: torch.Tensor, pixel_position_ids: torch.Tensor, padding_positions: torch.Tensor
) -> torch.Tensor:
# Gemma4 applies no normalization and instead scales in model code
pixel_values = 2 * (pixel_values - 0.5)
hidden_states = self.input_proj(pixel_values.to(self.input_proj.weight.dtype))
position_embeddings = self._position_embeddings(pixel_position_ids, padding_positions)
return hidden_states + position_embeddings
class Gemma4VisionPooler(nn.Module):
"""Scaling and optional spatial pooling for vision encodings"""
def __init__(self, config: Gemma4VisionConfig):
super().__init__()
self.hidden_size = config.hidden_size
self.root_hidden_size = self.hidden_size**0.5
def _avg_pool_by_positions(
self, hidden_states: torch.Tensor, pixel_position_ids: torch.Tensor, length: int
) -> tuple[torch.Tensor, torch.Tensor]:
"""
2D spatial pooling according to patch positions.
Pools the input tokens by averaging patches within a `k^2` grid, where `k` is determined by the ratio between
input and output lengths
"""
input_seq_len = hidden_states.shape[1]
k = int((input_seq_len // length) ** 0.5)
k_squared = k**2
if k_squared * length != input_seq_len:
raise ValueError(
f"Cannot pool {hidden_states.shape} to {length}: {k=}^2 times {length=} must be {input_seq_len}."
)
# Clamp padding positions (which are -1) to 0 so they don't break one_hot.
# Padding patches have zero hidden states so they contribute nothing to the average.
clamped_positions = pixel_position_ids.clamp(min=0)
max_x = clamped_positions[..., 0].max(dim=-1, keepdim=True)[0] + 1
kernel_idxs = torch.div(clamped_positions, k, rounding_mode="floor")
kernel_idxs = kernel_idxs[..., 0] + (max_x // k) * kernel_idxs[..., 1]
weights = F.one_hot(kernel_idxs.long(), length).float() / k_squared
output = weights.transpose(1, 2) @ hidden_states.float()
mask = torch.logical_not((weights == 0).all(dim=1))
return output.to(hidden_states.dtype), mask
def forward(
self,
hidden_states: torch.Tensor,
pixel_position_ids: torch.Tensor,
padding_positions: torch.Tensor,
output_length: int | None = None,
) -> tuple[torch.Tensor, torch.Tensor]:
if output_length > hidden_states.shape[1]:
raise ValueError(
f"Cannot output more soft tokens (requested {output_length}) than there are patches"
f" ({hidden_states.shape[1]}). Change the value of `num_soft_tokens` when processing."
)
hidden_states = hidden_states.masked_fill(padding_positions.unsqueeze(-1), 0.0)
if hidden_states.shape[1] != output_length:
hidden_states, padding_positions = self._avg_pool_by_positions(
hidden_states, pixel_position_ids, output_length
)
hidden_states *= self.root_hidden_size
return hidden_states, padding_positions
class Gemma4VisionMLP(nn.Module):
def __init__(self, config: Gemma4VisionConfig):
super().__init__()
self.config = config
self.hidden_size = config.hidden_size
self.intermediate_size = config.intermediate_size
self.gate_proj = Gemma4ClippableLinear(config, self.hidden_size, self.intermediate_size)
self.up_proj = Gemma4ClippableLinear(config, self.hidden_size, self.intermediate_size)
self.down_proj = Gemma4ClippableLinear(config, self.intermediate_size, self.hidden_size)
self.act_fn = ACT2FN[config.hidden_activation]
def forward(self, x):
down_proj = self.down_proj(self.act_fn(self.gate_proj(x)) * self.up_proj(x))
return down_proj
class Gemma4VisionRotaryEmbedding(nn.Module):
inv_freq: torch.Tensor # fix linting for `register_buffer`
def __init__(self, config: Gemma4VisionConfig, device=None):
super().__init__()
self.max_seq_len_cached = config.max_position_embeddings
self.original_max_seq_len = config.max_position_embeddings
self.config = config
self.rope_type = self.config.rope_parameters["rope_type"]
rope_init_fn: Callable = self.compute_default_rope_parameters
if self.rope_type != "default":
rope_init_fn = ROPE_INIT_FUNCTIONS[self.rope_type]
inv_freq, self.attention_scaling = rope_init_fn(self.config, device)
self.register_buffer("inv_freq", inv_freq, persistent=False)
self.register_buffer("original_inv_freq", inv_freq.clone(), persistent=False)
@staticmethod
def compute_default_rope_parameters(
config: Gemma4VisionConfig | None = None,
device: torch.device | None = None,
seq_len: int | None = None,
) -> tuple["torch.Tensor", float]:
"""
Computes the inverse frequencies according to the original RoPE implementation
Args:
config ([`~transformers.PreTrainedConfig`]):
The model configuration.
device (`torch.device`):
The device to use for initialization of the inverse frequencies.
seq_len (`int`, *optional*):
The current sequence length. Unused for this type of RoPE.
Returns:
Tuple of (`torch.Tensor`, `float`), containing the inverse frequencies for the RoPE embeddings and the
post-processing scaling factor applied to the computed cos/sin (unused in this type of RoPE).
"""
base = config.rope_parameters["rope_theta"]
dim = getattr(config, "head_dim", None) or config.hidden_size // config.num_attention_heads
# The reference implementation computes RoPE frequencies INDEPENDENTLY
# for each spatial dimension using the partitioned head_dim (head_dim // ndim),
# so both x and y dimensions get identical frequency ranges.
# This is different from splitting the global inv_freq between dimensions.
spatial_dim = dim // 2
attention_factor = 1.0 # Unused in this type of RoPE
inv_freq = 1.0 / (
base
** (torch.arange(0, spatial_dim, 2, dtype=torch.int64).to(device=device, dtype=torch.float) / spatial_dim)
)
return inv_freq, attention_factor
@torch.no_grad()
@dynamic_rope_update # power user: used with advanced RoPE types (e.g. dynamic rope)
def forward(self, x, position_ids):
inv_freq_expanded = self.inv_freq[None, :, None].float().expand(position_ids.shape[0], -1, 1).to(x.device)
device_type = x.device.type if isinstance(x.device.type, str) and x.device.type != "mps" else "cpu"
# Multidimensional positions: [batch, num_patches, ndim]. Apply rotations to each spatial dim separately
all_cos, all_sin = [], []
for i in range(2):
dim_position_ids = position_ids[:, :, i]
dim_position_ids_expanded = dim_position_ids[:, None, :].float()
with maybe_autocast(device_type=device_type, enabled=False): # Force float32
freqs = (inv_freq_expanded.float() @ dim_position_ids_expanded.float()).transpose(1, 2)
emb = torch.cat((freqs, freqs), dim=-1)
cos = emb.cos() * self.attention_scaling
sin = emb.sin() * self.attention_scaling
all_cos.append(cos)
all_sin.append(sin)
cos = torch.cat(all_cos, dim=-1).to(dtype=x.dtype)
sin = torch.cat(all_sin, dim=-1).to(dtype=x.dtype)
return cos, sin
def rotate_half(x):
"""Rotates half the hidden dims of the input."""
x1 = x[..., : x.shape[-1] // 2]
x2 = x[..., x.shape[-1] // 2 :]
return torch.cat((-x2, x1), dim=-1)
def apply_rotary_pos_emb(x: torch.Tensor, cos: torch.Tensor, sin: torch.Tensor, unsqueeze_dim: int = 1):
"""Applies Rotary Position Embedding to the query and key tensors.
Args:
x (`torch.Tensor`): The tensor to embed.
cos (`torch.Tensor`): The cosine part of the rotary embedding.
sin (`torch.Tensor`): The sine part of the rotary embedding.
unsqueeze_dim (`int`, *optional*, defaults to 1):
The 'unsqueeze_dim' argument specifies the dimension along which to unsqueeze cos[position_ids] and
sin[position_ids] so that they can be properly broadcasted to the dimensions of q and k. For example, note
that cos[position_ids] and sin[position_ids] have the shape [batch_size, seq_len, head_dim]. Then, if q and
k have the shape [batch_size, heads, seq_len, head_dim], then setting unsqueeze_dim=1 makes
cos[position_ids] and sin[position_ids] broadcastable to the shapes of q and k. Similarly, if q and k have
the shape [batch_size, seq_len, heads, head_dim], then set unsqueeze_dim=2.
Returns:
`tuple(torch.Tensor)` comprising of the query and key tensors rotated using the Rotary Position Embedding.
"""
cos = cos.unsqueeze(unsqueeze_dim)
sin = sin.unsqueeze(unsqueeze_dim)
return (x * cos) + (rotate_half(x) * sin)
def repeat_kv(hidden_states: torch.Tensor, n_rep: int) -> torch.Tensor:
"""
This is the equivalent of torch.repeat_interleave(x, dim=1, repeats=n_rep). The hidden states go from (batch,
num_key_value_heads, seqlen, head_dim) to (batch, num_attention_heads, seqlen, head_dim)
"""
batch, num_key_value_heads, slen, head_dim = hidden_states.shape
if n_rep == 1:
return hidden_states
hidden_states = hidden_states[:, :, None, :, :].expand(batch, num_key_value_heads, n_rep, slen, head_dim)
return hidden_states.reshape(batch, num_key_value_heads * n_rep, slen, head_dim)
def eager_attention_forward(
module: nn.Module,
query: torch.Tensor,
key: torch.Tensor,
value: torch.Tensor,
attention_mask: torch.Tensor | None,
dropout: float | int = 0.0,
scaling: float | None = None,
softcap: float | None = None,
**kwargs,
) -> tuple[torch.Tensor, torch.Tensor]:
if scaling is None:
scaling = module.head_dim**-0.5
key_states = repeat_kv(key, module.num_key_value_groups)
value_states = repeat_kv(value, module.num_key_value_groups)
attn_weights = torch.matmul(query, key_states.transpose(2, 3)) * scaling
if softcap is not None:
attn_weights = attn_weights / softcap
attn_weights = torch.tanh(attn_weights)
attn_weights = attn_weights * softcap
if attention_mask is not None:
attn_weights = attn_weights + attention_mask
# upcast attention to fp32
attn_weights = nn.functional.softmax(attn_weights, dim=-1, dtype=torch.float32).to(query.dtype)
attn_weights = nn.functional.dropout(attn_weights, p=dropout, training=module.training)
attn_output = torch.matmul(attn_weights, value_states)
attn_output = attn_output.transpose(1, 2).contiguous()
return attn_output, attn_weights
def apply_multidimensional_rope(
x: torch.Tensor,
cos: torch.Tensor,
sin: torch.Tensor,
position_ids: torch.Tensor,
unsqueeze_dim: int = 2,
) -> torch.Tensor:
"""Applies multidimensional RoPE to inputs.
Args:
x (`torch.Tensor`): The tensor to embed.
cos (`torch.Tensor`): The cosine part of the rotary embedding.
sin (`torch.Tensor`): The sine part of the rotary embedding.
position_ids (`torch.Tensor`, *optional*):
If position_ids.ndim + 2 == x.ndim, then this function passes through to `apply_rotary_pos_emb()`.
Otherwise, position_ids is used to split the inputs, x, into multiple pieces, where each piece is fed to
`apply_rotary_pos_emb()`, and then concatenated back together.
unsqueeze_dim (`int`, *optional*, defaults to 1):
The 'unsqueeze_dim' argument specifies the dimension along which to unsqueeze cos[position_ids] and
sin[position_ids] so that they can be properly broadcasted to the dimensions of q and k. For example, note
that cos[position_ids] and sin[position_ids] have the shape [batch_size, seq_len, head_dim]. Then, if q and
k have the shape [batch_size, heads, seq_len, head_dim], then setting unsqueeze_dim=1 makes
cos[position_ids] and sin[position_ids] broadcastable to the shapes of q and k. Similarly, if q and k have
the shape [batch_size, seq_len, heads, head_dim], then set unsqueeze_dim=2.
Returns:
Tensor of shape [B, L, N, H] with RoPE applied.
"""
ndim = position_ids.shape[-1]
num_input_channels = x.shape[-1]
num_rotated_channels_per_dim = 2 * (num_input_channels // (2 * ndim))
if num_rotated_channels_per_dim <= 0:
raise ValueError(
"Invalid configuration: num_rotated_channels_per_dim must be > 0, got"
f" {num_rotated_channels_per_dim} (num_input_channels={num_input_channels},"
f" ndim={ndim})"
)
# Correctly split the input tensor into ndim parts
split_sizes = [num_rotated_channels_per_dim] * ndim
x_parts = torch.split(x, split_sizes, dim=-1)
cos_parts = torch.split(cos, split_sizes, dim=-1)
sin_parts = torch.split(sin, split_sizes, dim=-1)
y_parts = [
apply_rotary_pos_emb(
x=x_parts[k],
cos=cos_parts[k],
sin=sin_parts[k],
unsqueeze_dim=unsqueeze_dim,
)
for k in range(ndim)
]
return torch.cat(y_parts, dim=-1)
@use_kernelized_func(apply_rotary_pos_emb)
class Gemma4VisionAttention(nn.Module):
"""Multi-headed attention from 'Attention Is All You Need' paper"""
def __init__(self, config: Gemma4VisionConfig, layer_idx: int):
super().__init__()
self.layer_type = config.layer_types[layer_idx] if hasattr(config, "layer_types") else None
self.config = config
self.layer_idx = layer_idx
self.head_dim = getattr(config, "head_dim", config.hidden_size // config.num_attention_heads)
self.num_key_value_groups = config.num_attention_heads // config.num_key_value_heads
self.scaling = 1.0
self.attention_dropout = self.config.attention_dropout
self.is_causal = False
self.q_proj = Gemma4ClippableLinear(config, config.hidden_size, config.num_attention_heads * self.head_dim)
self.k_proj = Gemma4ClippableLinear(config, config.hidden_size, config.num_key_value_heads * self.head_dim)
self.v_proj = Gemma4ClippableLinear(config, config.hidden_size, config.num_key_value_heads * self.head_dim)
self.o_proj = Gemma4ClippableLinear(config, config.num_attention_heads * self.head_dim, config.hidden_size)
self.q_norm = Gemma4RMSNorm(dim=config.head_dim, eps=config.rms_norm_eps)
self.k_norm = Gemma4RMSNorm(dim=config.head_dim, eps=config.rms_norm_eps)
self.v_norm = Gemma4RMSNorm(self.head_dim, eps=config.rms_norm_eps, with_scale=False)
def forward(
self,
hidden_states: torch.Tensor,
position_embeddings: torch.Tensor = None,
attention_mask: torch.Tensor | None = None,
position_ids: torch.LongTensor | None = None,
**kwargs: Unpack[TransformersKwargs],
) -> tuple[torch.Tensor, torch.Tensor | None, tuple[torch.Tensor] | None]:
input_shape = hidden_states.shape[:-1]
hidden_shape = (*input_shape, -1, self.head_dim)
cos, sin = position_embeddings
query_states = self.q_proj(hidden_states).view(hidden_shape)
query_states = self.q_norm(query_states)
query_states = apply_multidimensional_rope(query_states, cos, sin, position_ids)
query_states = query_states.transpose(1, 2)
key_states = self.k_proj(hidden_states).view(hidden_shape)
key_states = self.k_norm(key_states)
key_states = apply_multidimensional_rope(key_states, cos, sin, position_ids)
key_states = key_states.transpose(1, 2)
value_states = self.v_proj(hidden_states).view(hidden_shape)
value_states = self.v_norm(value_states)
value_states = value_states.transpose(1, 2)
attention_interface: Callable = ALL_ATTENTION_FUNCTIONS.get_interface(
self.config._attn_implementation, eager_attention_forward
)
attn_output, attn_weights = attention_interface(
self,
query_states,
key_states,
value_states,
attention_mask,
dropout=self.attention_dropout if self.training else 0.0,
scaling=self.scaling,
**kwargs,
)
attn_output = attn_output.reshape(*input_shape, -1).contiguous()
attn_output = self.o_proj(attn_output)
return attn_output, attn_weights
class Gemma4VisionEncoderLayer(GradientCheckpointingLayer):
def __init__(self, config: Gemma4VisionConfig, layer_idx: int):
super().__init__()
self.config = config
self.hidden_size = config.hidden_size
self.layer_idx = layer_idx
self.self_attn = Gemma4VisionAttention(config=config, layer_idx=layer_idx)
self.mlp = Gemma4VisionMLP(config)
self.input_layernorm = Gemma4RMSNorm(self.hidden_size, eps=config.rms_norm_eps)
self.post_attention_layernorm = Gemma4RMSNorm(self.hidden_size, eps=config.rms_norm_eps)
self.pre_feedforward_layernorm = Gemma4RMSNorm(self.hidden_size, eps=config.rms_norm_eps)
self.post_feedforward_layernorm = Gemma4RMSNorm(self.hidden_size, eps=config.rms_norm_eps)
def forward(
self,
hidden_states: torch.Tensor,
position_embeddings: torch.Tensor = None,
attention_mask: torch.Tensor | None = None,
position_ids: torch.LongTensor | None = None,
**kwargs: Unpack[TransformersKwargs],
) -> tuple[torch.FloatTensor, tuple[torch.FloatTensor, torch.FloatTensor] | None]:
residual = hidden_states
hidden_states = self.input_layernorm(hidden_states)
hidden_states, _ = self.self_attn(
hidden_states=hidden_states,
position_embeddings=position_embeddings,
attention_mask=attention_mask,
position_ids=position_ids,
**kwargs,
)
hidden_states = self.post_attention_layernorm(hidden_states)
hidden_states = residual + hidden_states
residual = hidden_states
hidden_states = self.pre_feedforward_layernorm(hidden_states)
hidden_states = self.mlp(hidden_states)
hidden_states = self.post_feedforward_layernorm(hidden_states)
hidden_states = residual + hidden_states
return hidden_states
class Gemma4VisionEncoder(nn.Module):
def __init__(self, config: Gemma4VisionConfig):
super().__init__()
self.config = config
self.num_layers = config.num_hidden_layers
self.rotary_emb = Gemma4VisionRotaryEmbedding(config)
self.layers = nn.ModuleList(
[Gemma4VisionEncoderLayer(config=config, layer_idx=i) for i in range(self.num_layers)]
)
def forward(
self,
inputs_embeds: torch.Tensor,
attention_mask: torch.Tensor,
pixel_position_ids: torch.LongTensor | None = None,
**kwargs: Unpack[TransformersKwargs],
) -> BaseModelOutputWithPast:
r"""
pixel_position_ids (torch.Tensor):
Patch positions as (x, y) coordinates in the image as [batch, num_patches, 2].
"""
attention_mask = create_bidirectional_mask(
config=self.config,
inputs_embeds=inputs_embeds,
attention_mask=attention_mask,
)
# embed positions
hidden_states = inputs_embeds
position_embeddings = self.rotary_emb(hidden_states, pixel_position_ids)
# decoder layers
for decoder_layer in self.layers[: self.config.num_hidden_layers]:
hidden_states = decoder_layer(
hidden_states,
attention_mask=attention_mask,
position_embeddings=position_embeddings,
position_ids=pixel_position_ids,
**kwargs,
)
return BaseModelOutputWithPast(last_hidden_state=hidden_states)
class Gemma4TextMLP(nn.Module):
def __init__(self, config: Gemma4TextConfig, layer_idx: int):
super().__init__()
first_kv_shared_layer_idx = config.num_hidden_layers - config.num_kv_shared_layers
is_kv_shared_layer = layer_idx >= first_kv_shared_layer_idx > 0
use_double_wide_mlp = config.use_double_wide_mlp and is_kv_shared_layer
self.config = config
self.hidden_size = config.hidden_size
self.intermediate_size = config.intermediate_size * (2 if use_double_wide_mlp else 1)
self.gate_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=False)
self.up_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=False)
self.down_proj = nn.Linear(self.intermediate_size, self.hidden_size, bias=False)
self.act_fn = ACT2FN[config.hidden_activation]
def forward(self, x):
down_proj = self.down_proj(self.act_fn(self.gate_proj(x)) * self.up_proj(x))
return down_proj
class Gemma4TextRotaryEmbedding(nn.Module):
inv_freq: torch.Tensor # fix linting for `register_buffer`
def __init__(self, config: Gemma4TextConfig, device=None, layer_type=None):
super().__init__()
self.max_seq_len_cached = config.max_position_embeddings
self.original_max_seq_len = config.max_position_embeddings
self.config = config
self.layer_types = set(config.layer_types)
self.rope_init_fns: dict[str, Callable[..., tuple[torch.Tensor, float]]] = {}
self.rope_type: dict[str, str] = {}
for layer_type in self.layer_types:
rope_params = self.config.rope_parameters[layer_type]
if rope_params is None:
continue
if (rope_type := rope_params["rope_type"]) != "default":
rope_init_fn = ROPE_INIT_FUNCTIONS[rope_type]
else:
rope_init_fn = self.compute_default_rope_parameters
self.rope_init_fns[layer_type] = rope_init_fn
self.rope_type[layer_type] = rope_type
rope_init_fn_kwargs = {"device": device, "layer_type": layer_type}
if layer_type == "full_attention" and rope_type == "proportional":
rope_init_fn_kwargs["head_dim_key"] = "global_head_dim"
curr_inv_freq, curr_attention_scaling = rope_init_fn(self.config, **rope_init_fn_kwargs)
self.register_buffer(f"{layer_type}_inv_freq", curr_inv_freq, persistent=False)
self.register_buffer(f"{layer_type}_original_inv_freq", curr_inv_freq.clone(), persistent=False)
setattr(self, f"{layer_type}_attention_scaling", curr_attention_scaling)
@staticmethod
def compute_default_rope_parameters(
config: Gemma4TextConfig | None = None,
device: Optional["torch.device"] = None,
seq_len: int | None = None,
layer_type: str | None = None,
) -> tuple["torch.Tensor", float]:
"""
Computes the inverse frequencies according to the original RoPE implementation
Args:
config ([`~transformers.PreTrainedConfig`]):
The model configuration.
device (`torch.device`):
The device to use for initialization of the inverse frequencies.
seq_len (`int`, *optional*):
The current sequence length. Unused for this type of RoPE.
layer_type (`str`, *optional*):
The current layer type if the model has different RoPE parameters per type.
Should not be used unless `config.layer_types is not None`
Returns:
Tuple of (`torch.Tensor`, `float`), containing the inverse frequencies for the RoPE embeddings and the
post-processing scaling factor applied to the computed cos/sin (unused in this type of RoPE).
"""
# For backward compatibility standardize the `rope_parameters_dict` if it uses old format
base = config.rope_parameters[layer_type]["rope_theta"]
dim = getattr(config, "head_dim", None) or config.hidden_size // config.num_attention_heads
attention_factor = 1.0 # Unused in this type of RoPE
# Compute the inverse frequencies
inv_freq = 1.0 / (
base ** (torch.arange(0, dim, 2, dtype=torch.int64).to(device=device, dtype=torch.float) / dim)
)
return inv_freq, attention_factor
@torch.no_grad()
@dynamic_rope_update # power user: used with advanced RoPE types (e.g. dynamic rope)
def forward(self, x, position_ids, layer_type=None):
inv_freq = getattr(self, f"{layer_type}_inv_freq")
attention_scaling = getattr(self, f"{layer_type}_attention_scaling")
inv_freq_expanded = inv_freq[None, :, None].float().expand(position_ids.shape[0], -1, 1).to(x.device)
position_ids_expanded = position_ids[:, None, :].float()
device_type = x.device.type if isinstance(x.device.type, str) and x.device.type != "mps" else "cpu"
with maybe_autocast(device_type=device_type, enabled=False): # Force float32
freqs = (inv_freq_expanded.float() @ position_ids_expanded.float()).transpose(1, 2)
emb = torch.cat((freqs, freqs), dim=-1)
cos = emb.cos() * attention_scaling
sin = emb.sin() * attention_scaling
return cos.to(dtype=x.dtype), sin.to(dtype=x.dtype)
@use_kernelized_func(apply_rotary_pos_emb)
class Gemma4TextAttention(nn.Module):
"""Multi-headed attention from 'Attention Is All You Need' paper"""
def __init__(self, config: Gemma4TextConfig, layer_idx: int):
super().__init__()
self.layer_type = config.layer_types[layer_idx] if hasattr(config, "layer_types") else None
self.config = config
self.layer_idx = layer_idx
self.is_sliding = self.layer_type == "sliding_attention"
self.sliding_window = config.sliding_window if self.is_sliding else None
self.head_dim = config.global_head_dim if not self.is_sliding and config.global_head_dim else config.head_dim
self.use_alternative_attention = config.attention_k_eq_v and not self.is_sliding
num_key_value_heads = (
config.num_global_key_value_heads if self.use_alternative_attention else config.num_key_value_heads
)
self.num_key_value_groups = config.num_attention_heads // num_key_value_heads
self.scaling = 1.0
self.attention_dropout = self.config.attention_dropout
self.is_causal = config.use_bidirectional_attention != "all"
# Shared kv cache
first_kv_shared_layer_idx = self.config.num_hidden_layers - getattr(self.config, "num_kv_shared_layers", 0)
self.is_kv_shared_layer = layer_idx >= first_kv_shared_layer_idx > 0
prev_layers = config.layer_types[:first_kv_shared_layer_idx]
if self.is_kv_shared_layer:
# For shared layers, find the last non-shared layer of the same type before sharing starts
self.kv_shared_layer_index = len(prev_layers) - 1 - prev_layers[::-1].index(config.layer_types[layer_idx])
self.store_full_length_kv = False
else:
self.kv_shared_layer_index = None
# For non-shared layers, store full-length kv if this is the last non-shared layer of its type
self.store_full_length_kv = layer_idx == len(prev_layers) - 1 - prev_layers[::-1].index(
config.layer_types[layer_idx]
)
self.q_norm = Gemma4RMSNorm(dim=self.head_dim, eps=config.rms_norm_eps)
self.k_norm = Gemma4RMSNorm(dim=self.head_dim, eps=config.rms_norm_eps)
self.v_norm = Gemma4RMSNorm(self.head_dim, eps=config.rms_norm_eps, with_scale=False)
self.k_proj = nn.Linear(config.hidden_size, num_key_value_heads * self.head_dim, bias=config.attention_bias)
self.q_proj = nn.Linear(
config.hidden_size, config.num_attention_heads * self.head_dim, bias=config.attention_bias
)
self.v_proj = (
nn.Linear(config.hidden_size, num_key_value_heads * self.head_dim, bias=config.attention_bias)
if not self.use_alternative_attention
else None
)
self.o_proj = nn.Linear(
config.num_attention_heads * self.head_dim, config.hidden_size, bias=config.attention_bias
)
def forward(
self,
hidden_states: torch.Tensor,
position_embeddings: torch.Tensor,
attention_mask: torch.Tensor | None,
past_key_values: Cache | None = None,
**kwargs: Unpack[FlashAttentionKwargs],
) -> tuple[torch.Tensor, torch.Tensor | None]:
input_shape = hidden_states.shape[:-1]
hidden_shape = (*input_shape, -1, self.head_dim)
cos, sin = position_embeddings
query_states = self.q_proj(hidden_states).view(hidden_shape)
query_states = self.q_norm(query_states)
query_states = apply_rotary_pos_emb(query_states, cos, sin, unsqueeze_dim=2)
query_states = query_states.transpose(1, 2)
# For layers with shared KV (from kv sharing point onwards), we reuse the same keys/values states as the last non-sharing layer
if self.is_kv_shared_layer and past_key_values is not None:
key_states, value_states = past_key_values.shared_layers[self.kv_shared_layer_index]
# Device of past layer may be different from current one
key_states = key_states.to(query_states.device)
value_states = value_states.to(query_states.device)
else:
key_states = self.k_proj(hidden_states).view(hidden_shape)
value_states = self.v_proj(hidden_states).view(hidden_shape) if self.v_proj is not None else key_states
key_states = self.k_norm(key_states)
key_states = apply_rotary_pos_emb(key_states, cos, sin, unsqueeze_dim=2)
key_states = key_states.transpose(1, 2)
value_states = self.v_norm(value_states)
value_states = value_states.transpose(1, 2)
if past_key_values is not None:
if not self.is_kv_shared_layer:
key_states, value_states = past_key_values.update(key_states, value_states, self.layer_idx)
if self.store_full_length_kv:
if not hasattr(past_key_values, "shared_layers"):
past_key_values.shared_layers = {}
past_key_values.shared_layers[self.layer_idx] = key_states, value_states
attention_interface: Callable = eager_attention_forward
if self.config._attn_implementation != "eager":
attention_interface = ALL_ATTENTION_FUNCTIONS[self.config._attn_implementation]
attn_output, attn_weights = attention_interface(
self,
query_states,
key_states,
value_states,
attention_mask,
dropout=self.attention_dropout if self.training else 0.0,
scaling=self.scaling,
sliding_window=self.sliding_window,
**kwargs,
)
attn_output = attn_output.reshape(*input_shape, -1).contiguous()
attn_output = self.o_proj(attn_output)
return attn_output, attn_weights
# TODO: add a zero-compute under this abstraction
@use_experts_implementation
class Gemma4TextExperts(nn.Module):
"""Collection of expert weights stored as 3D tensors."""
def __init__(self, config: Gemma4TextConfig):
super().__init__()
self.num_experts = config.num_experts
self.hidden_dim = config.hidden_size
self.intermediate_dim = config.moe_intermediate_size
self.gate_up_proj = nn.Parameter(torch.empty(self.num_experts, 2 * self.intermediate_dim, self.hidden_dim))
self.down_proj = nn.Parameter(torch.empty(self.num_experts, self.hidden_dim, self.intermediate_dim))
self.act_fn = ACT2FN[config.hidden_activation]
def forward(
self,
hidden_states: torch.Tensor,
top_k_index: torch.Tensor,
top_k_weights: torch.Tensor,
) -> torch.Tensor:
final_hidden_states = torch.zeros_like(hidden_states)
with torch.no_grad():
expert_mask = torch.nn.functional.one_hot(top_k_index, num_classes=self.num_experts)
expert_mask = expert_mask.permute(2, 1, 0)
expert_hit = torch.greater(expert_mask.sum(dim=(-1, -2)), 0).nonzero()
for expert_idx in expert_hit:
expert_idx = expert_idx[0]
if expert_idx == self.num_experts:
continue
top_k_pos, token_idx = torch.where(expert_mask[expert_idx])
current_state = hidden_states[token_idx]
gate, up = nn.functional.linear(current_state, self.gate_up_proj[expert_idx]).chunk(2, dim=-1)
current_hidden_states = self.act_fn(gate) * up
current_hidden_states = nn.functional.linear(current_hidden_states, self.down_proj[expert_idx])
current_hidden_states = current_hidden_states * top_k_weights[token_idx, top_k_pos, None]
final_hidden_states.index_add_(0, token_idx, current_hidden_states.to(final_hidden_states.dtype))
return final_hidden_states
# TODO: Remake and self-distill for zero-compute
class Gemma4TextRouter(nn.Module):
def __init__(self, config: Gemma4TextConfig):
super().__init__()
self.config = config
self.hidden_size = config.hidden_size
self.scalar_root_size = self.hidden_size**-0.5
self.eps = config.rms_norm_eps
self.norm = Gemma4RMSNorm(self.hidden_size, eps=self.eps, with_scale=False)
self.proj = nn.Linear(config.hidden_size, config.num_experts, bias=False)
self.scale = nn.Parameter(torch.ones(self.hidden_size))
self.per_expert_scale = nn.Parameter(torch.ones(config.num_experts))
def forward(self, hidden_states: torch.Tensor) -> tuple[torch.Tensor, torch.Tensor]:
hidden_states = self.norm(hidden_states)
hidden_states = hidden_states * self.scale * self.scalar_root_size
expert_scores = self.proj(hidden_states) # [B*S, E]
router_probabilities = nn.functional.softmax(expert_scores, dim=-1)
# topk returns both values (probabilities) and indices directly
top_k_weights, top_k_index = torch.topk(
router_probabilities,
k=self.config.top_k_experts,
dim=-1,
) # both [B*S, K]
# Normalize the top-k weights so they sum to 1 per token
top_k_weights /= top_k_weights.sum(dim=-1, keepdim=True)
# Apply per-expert scale directly to the weights
top_k_weights = top_k_weights * self.per_expert_scale[top_k_index]
return router_probabilities, top_k_weights, top_k_index
# TODO: add a zero-compute under this abstraction and consider remaking without the need for post-feedforward norm
class Gemma4TextDecoderLayer(GradientCheckpointingLayer):
router_class = Gemma4TextRouter
experts_class = Gemma4TextExperts
def __init__(self, config: Gemma4TextConfig | Gemma4VisionConfig, layer_idx: int):
super().__init__()
self.config = config
self.hidden_size = config.hidden_size
self.layer_idx = layer_idx
self.self_attn = Gemma4TextAttention(config=config, layer_idx=layer_idx)
self.mlp = Gemma4TextMLP(config, layer_idx)
self.input_layernorm = Gemma4RMSNorm(self.hidden_size, eps=config.rms_norm_eps)
self.post_attention_layernorm = Gemma4RMSNorm(self.hidden_size, eps=config.rms_norm_eps)
self.pre_feedforward_layernorm = Gemma4RMSNorm(self.hidden_size, eps=config.rms_norm_eps)
self.post_feedforward_layernorm = Gemma4RMSNorm(self.hidden_size, eps=config.rms_norm_eps)
self.register_buffer("layer_scalar", torch.ones(1))
self.hidden_size_per_layer_input = config.hidden_size_per_layer_input
if self.hidden_size_per_layer_input:
self.act_fn = ACT2FN[config.hidden_activation]
self.per_layer_input_gate = nn.Linear(self.hidden_size, self.hidden_size_per_layer_input, bias=False)
self.per_layer_projection = nn.Linear(self.hidden_size_per_layer_input, self.hidden_size, bias=False)
self.post_per_layer_input_norm = Gemma4RMSNorm(self.hidden_size, eps=config.rms_norm_eps)
self.enable_moe_block = config.enable_moe_block
if self.enable_moe_block:
self.router = self.router_class(config)
self.experts = self.experts_class(config)
self.post_feedforward_layernorm_1 = Gemma4RMSNorm(self.hidden_size, eps=config.rms_norm_eps)
self.post_feedforward_layernorm_2 = Gemma4RMSNorm(self.hidden_size, eps=config.rms_norm_eps)
self.pre_feedforward_layernorm_2 = Gemma4RMSNorm(self.hidden_size, eps=config.rms_norm_eps)
def forward(
self,
hidden_states: torch.Tensor,
per_layer_input: torch.Tensor = None,
position_embeddings: torch.Tensor = None,
attention_mask: torch.Tensor | None = None,
position_ids: torch.LongTensor | None = None,
past_key_values: Cache | None = None,
**kwargs,
) -> torch.Tensor:
residual = hidden_states
hidden_states = self.input_layernorm(hidden_states)
hidden_states, _ = self.self_attn(
hidden_states=hidden_states,
position_embeddings=position_embeddings,
attention_mask=attention_mask,
position_ids=position_ids,
past_key_values=past_key_values,
**kwargs,
)
hidden_states = self.post_attention_layernorm(hidden_states)
hidden_states = residual + hidden_states
residual = hidden_states
hidden_states = self.pre_feedforward_layernorm(hidden_states)
hidden_states = self.mlp(hidden_states)
if self.enable_moe_block:
hidden_states_1 = self.post_feedforward_layernorm_1(hidden_states)
# Take hidden states before MLP here
hidden_states_flat = residual.reshape(-1, residual.shape[-1])
_, top_k_weights, top_k_index = self.router(hidden_states_flat)
hidden_states_2 = self.pre_feedforward_layernorm_2(hidden_states_flat)
hidden_states_2 = self.experts(hidden_states_2, top_k_index, top_k_weights)
hidden_states_2 = hidden_states_2.reshape(residual.shape)
hidden_states_2 = self.post_feedforward_layernorm_2(hidden_states_2)
# Combine mlp and moe outputs
hidden_states = hidden_states_1 + hidden_states_2
hidden_states = self.post_feedforward_layernorm(hidden_states)
hidden_states = residual + hidden_states
if self.hidden_size_per_layer_input:
residual = hidden_states
hidden_states = self.per_layer_input_gate(hidden_states)
hidden_states = self.act_fn(hidden_states)
hidden_states = hidden_states * per_layer_input
hidden_states = self.per_layer_projection(hidden_states)
hidden_states = self.post_per_layer_input_norm(hidden_states)
hidden_states = residual + hidden_states
hidden_states *= self.layer_scalar
return hidden_states
class Gemma4TextScaledWordEmbedding(nn.Embedding):
"""
This module overrides nn.Embeddings' forward by multiplying with embeddings scale.
"""
def __init__(self, num_embeddings: int, embedding_dim: int, padding_idx: int, embed_scale: float = 1.0):
super().__init__(num_embeddings, embedding_dim, padding_idx)
self.scalar_embed_scale = embed_scale
self.register_buffer("embed_scale", torch.tensor(embed_scale), persistent=False)
def forward(self, input_ids: torch.Tensor):
return super().forward(input_ids) * self.embed_scale.to(self.weight.dtype)
# ---- Model Classes ----
class Gemma4PreTrainedModel(PreTrainedModel):
config: Gemma4Config
supports_gradient_checkpointing = True
_supports_flash_attn = True
_supports_sdpa = True
_supports_flex_attn = True
_can_compile_fullgraph = True
_supports_attention_backend = True
_no_split_modules = ["Gemma4TextDecoderLayer", "Gemma4VisionEncoderLayer", "Gemma4AudioLayer"]
_skip_keys_device_placement = ["past_key_values"]
input_modalities = ("image", "text", "video", "audio")
@torch.no_grad()
def _init_weights(self, module):
super()._init_weights(module)
if isinstance(module, Gemma4VisionPatchEmbedder):
init.ones_(module.position_embedding_table)
elif isinstance(module, Gemma4AudioRelPositionalEncoding):
min_timescale = 1.0
max_timescale = 10000.0
num_timescales = module.hidden_size // 2
log_timescale_increment = math.log(max_timescale / min_timescale) / max(num_timescales - 1, 1)
inv_timescales = min_timescale * torch.exp(torch.arange(num_timescales) * -log_timescale_increment)
init.copy_(module.inv_timescales, inv_timescales.unsqueeze(0).unsqueeze(0))
elif isinstance(module, Gemma4AudioAttention):
init.constant_(module.softcap, module.attention_logits_soft_cap)
init.zeros_(module.per_dim_scale)
elif isinstance(module, Gemma4TextRotaryEmbedding):
for layer_type, rope_init_fn in module.rope_init_fns.items():
rope_init_fn_kwargs = {"layer_type": layer_type}
if layer_type == "full_attention" and module.rope_type[layer_type] == "proportional":
rope_init_fn_kwargs["head_dim_key"] = "global_head_dim"
curr_inv_freq, _ = rope_init_fn(module.config, **rope_init_fn_kwargs)
init.copy_(getattr(module, f"{layer_type}_inv_freq"), curr_inv_freq)
init.copy_(getattr(module, f"{layer_type}_original_inv_freq"), curr_inv_freq)
elif isinstance(module, Gemma4VisionRotaryEmbedding):
rope_fn = (
ROPE_INIT_FUNCTIONS[module.rope_type]
if module.rope_type != "default"
else module.compute_default_rope_parameters
)
buffer_value, _ = rope_fn(module.config)
init.copy_(module.inv_freq, buffer_value)
init.copy_(module.original_inv_freq, buffer_value)
elif isinstance(module, Gemma4TextScaledWordEmbedding):
init.constant_(module.embed_scale, module.scalar_embed_scale)
elif isinstance(module, Gemma4TextRouter):
init.ones_(module.scale)
init.ones_(module.per_expert_scale)
elif isinstance(module, Gemma4TextExperts):
std = self.config.initializer_range
init.normal_(module.gate_up_proj, mean=0.0, std=std)
init.normal_(module.down_proj, mean=0.0, std=std)
elif isinstance(module, Gemma4TextDecoderLayer):
init.ones_(module.layer_scalar)
elif isinstance(module, Gemma4ClippableLinear) and module.use_clipped_linears:
init.constant_(module.input_min, -float("inf"))
init.constant_(module.input_max, float("inf"))
init.constant_(module.output_min, -float("inf"))
init.constant_(module.output_max, float("inf"))
elif isinstance(module, Gemma4VisionModel) and module.config.standardize:
init.zeros_(module.std_bias)
init.ones_(module.std_scale)
@auto_docstring(custom_intro="The base Gemma 4 language model without a language modeling head.")
class Gemma4TextModel(Gemma4PreTrainedModel):
config: Gemma4TextConfig
decoder_layer_class = Gemma4TextDecoderLayer
input_modalities = ("text",)
_can_record_outputs = {
"router_logits": OutputRecorder(Gemma4TextRouter, index=0),
"hidden_states": Gemma4TextDecoderLayer,
"attentions": Gemma4TextAttention,
}
def __init__(self, config: Gemma4TextConfig):
super().__init__(config)
self.padding_idx = config.pad_token_id
self.vocab_size = config.vocab_size
# Gemma4 downcasts the below to bfloat16, causing sqrt(3072)=55.4256 to become 55.5. See https://github.com/huggingface/transformers/pull/29402
self.embed_tokens = Gemma4TextScaledWordEmbedding(
config.vocab_size, config.hidden_size, self.padding_idx, embed_scale=self.config.hidden_size**0.5
)
self.layers = nn.ModuleList([self.decoder_layer_class(config, layer_idx) for layer_idx in range(config.num_hidden_layers)])
self.norm = Gemma4RMSNorm(config.hidden_size, eps=config.rms_norm_eps)
self.rotary_emb = Gemma4TextRotaryEmbedding(config)
self.gradient_checkpointing = False
self.unique_layer_types = set(self.config.layer_types)
self.hidden_size_per_layer_input = config.hidden_size_per_layer_input
if self.hidden_size_per_layer_input:
self.embed_tokens_per_layer = Gemma4TextScaledWordEmbedding(
config.vocab_size_per_layer_input,
config.num_hidden_layers * config.hidden_size_per_layer_input,
self.padding_idx,
embed_scale=config.hidden_size_per_layer_input**0.5,
)
self.per_layer_input_scale = 2.0**-0.5
self.per_layer_model_projection = nn.Linear(
config.hidden_size,
config.num_hidden_layers * config.hidden_size_per_layer_input,
bias=False,
)
self.per_layer_model_projection_scale = config.hidden_size**-0.5
self.per_layer_projection_norm = Gemma4RMSNorm(config.hidden_size_per_layer_input, eps=config.rms_norm_eps)
# Initialize weights and apply final processing
self.post_init()
@merge_with_config_defaults
@capture_outputs
@auto_docstring
def forward(
self,
input_ids: torch.LongTensor | None = None,
attention_mask: torch.Tensor | None = None,
position_ids: torch.LongTensor | None = None,
past_key_values: Cache | None = None,
inputs_embeds: torch.FloatTensor | None = None,
per_layer_inputs: torch.Tensor | None = None,
use_cache: bool | None = None,
**kwargs: Unpack[TransformersKwargs],
) -> BaseModelOutputWithPast:
r"""
per_layer_inputs (`torch.Tensor` of shape `(batch_size, sequence_length, num_hidden_layers, hidden_size_per_layer_input)`, *optional*):
Pre-computed per-layer input embeddings. When provided, these are used directly instead of being
computed from `input_ids` via `get_per_layer_inputs()`. This is primarily used by the multimodal
model (`Gemma4Model`) which pre-computes per-layer inputs from the original `input_ids` *before*
merging multimodal soft tokens into `inputs_embeds` — at which point the original token ids are
no longer recoverable.
"""
use_cache = use_cache if use_cache is not None else self.config.use_cache
if (input_ids is None) ^ (inputs_embeds is not None):
raise ValueError("You must specify exactly one of input_ids or inputs_embeds")
if input_ids is not None:
inputs_embeds = self.embed_tokens(input_ids)
if self.hidden_size_per_layer_input:
if per_layer_inputs is None:
per_layer_inputs = self.get_per_layer_inputs(input_ids, inputs_embeds)
per_layer_inputs = self.project_per_layer_inputs(inputs_embeds, per_layer_inputs)
if use_cache and past_key_values is None:
past_key_values = DynamicCache(config=self.config)
if position_ids is None:
past_seen_tokens = past_key_values.get_seq_length() if past_key_values is not None else 0
position_ids = torch.arange(inputs_embeds.shape[1], device=inputs_embeds.device) + past_seen_tokens
position_ids = position_ids.unsqueeze(0)
# It may already have been prepared by e.g. `generate`
if not isinstance(causal_mask_mapping := attention_mask, dict):
# Prepare mask arguments
mask_kwargs = {
"config": self.config,
"inputs_embeds": inputs_embeds,
"attention_mask": attention_mask,
"past_key_values": past_key_values,
"position_ids": position_ids,
}
# Create the masks
causal_mask_mapping = {
"full_attention": create_causal_mask(**mask_kwargs),
"sliding_attention": create_sliding_window_causal_mask(**mask_kwargs),
}
# embed positions
hidden_states = inputs_embeds
position_embeddings = {}
for layer_type in self.unique_layer_types:
position_embeddings[layer_type] = self.rotary_emb(hidden_states, position_ids, layer_type)
# decoder layers
for i, decoder_layer in enumerate(self.layers[: self.config.num_hidden_layers]):
per_layer_input = per_layer_inputs[:, :, i, :] if per_layer_inputs is not None else None
hidden_states = decoder_layer(
hidden_states,
per_layer_input,
position_embeddings=position_embeddings[self.config.layer_types[i]],
attention_mask=causal_mask_mapping[self.config.layer_types[i]],
position_ids=position_ids,
past_key_values=past_key_values,
**kwargs,
)
hidden_states = self.norm(hidden_states)
return BaseModelOutputWithPast(
last_hidden_state=hidden_states,
past_key_values=past_key_values,
)
def get_per_layer_inputs(self, input_ids: torch.Tensor | None, inputs_embeds: torch.Tensor | None) -> torch.Tensor:
if not self.hidden_size_per_layer_input:
raise RuntimeError(
"Attempting to call get_per_layer_inputs() from a model initialized with a config that does not support"
f" per-layer embeddings. {self.config}"
)
# If only inputs_embeds are provided, reverse main embedding to find the input_ids - this allows to `generate`
# from `inputs_embeds` only as other models (otherwise it would need the value from both embeddings)
if input_ids is None:
with torch.no_grad():
input_ids = (
(
inputs_embeds[:, :, None, :]
== self.embed_tokens.weight[None, None, :, :] * self.config.hidden_size**0.5
)
.all(dim=3)
.nonzero()[:, 2]
)
try:
input_ids = input_ids.view(inputs_embeds.shape[:2])
except RuntimeError:
raise RuntimeError(
"It seems like you tried to call `forward` from `inputs_embeds` without providing `input_ids`, and that "
"the `inputs_embeds` you provided do not exactly match the embedding weights. Since Gemma4 needs to reverse "
"the embedding to compute another embedding, make sure you provide exact `inputs_embeds`"
)
return self.embed_tokens_per_layer(input_ids).reshape(
*input_ids.shape,
self.config.num_hidden_layers,
self.hidden_size_per_layer_input,
)
def project_per_layer_inputs(
self,
inputs_embeds: torch.Tensor,
per_layer_inputs: torch.Tensor | None = None,
) -> torch.Tensor:
if not self.hidden_size_per_layer_input:
raise RuntimeError(
"Attempting to call project_per_layer_inputs() from a model initialized with a config that does not"
f" support per-layer embeddings. {self.config}"
)
per_layer_projection = self.per_layer_model_projection(inputs_embeds) * self.per_layer_model_projection_scale
per_layer_projection = per_layer_projection.reshape(
*inputs_embeds.shape[:-1],
self.config.num_hidden_layers,
self.hidden_size_per_layer_input,
)
per_layer_projection = self.per_layer_projection_norm(per_layer_projection)
if per_layer_inputs is None:
return per_layer_projection
return (per_layer_projection + per_layer_inputs) * self.per_layer_input_scale
@auto_docstring(custom_intro="The base Gemma 4 language model with a language modeling head.")
class Gemma4ForCausalLM(Gemma4PreTrainedModel, GenerationMixin):
_tied_weights_keys = {"lm_head.weight": "model.embed_tokens.weight"}
_tp_plan = {"lm_head": "colwise_gather_output"}
_pp_plan = {"lm_head": (["hidden_states"], ["logits"])}
config: Gemma4TextConfig
base_model_prefix = "model"
text_model_class = Gemma4TextModel
def __init__(self, config: Gemma4TextConfig):
super().__init__(config)
self.model = self.text_model_class(config)
self.vocab_size = config.vocab_size
self.lm_head = nn.Linear(config.hidden_size, config.vocab_size, bias=False)
# Initialize weights and apply final processing
self.post_init()
@can_return_tuple
@auto_docstring
def forward(
self,
input_ids: torch.LongTensor | None = None,
attention_mask: torch.Tensor | None = None,
position_ids: torch.LongTensor | None = None,
past_key_values: Cache | None = None,
inputs_embeds: torch.FloatTensor | None = None,
labels: torch.LongTensor | None = None,
use_cache: bool | None = None,
logits_to_keep: int | torch.Tensor = 0,
**kwargs: Unpack[TransformersKwargs],
) -> CausalLMOutputWithPast:
r"""
Example:
```python
>>> from transformers import AutoTokenizer, Gemma4ForCausalLM
>>> model = Gemma4ForCausalLM.from_pretrained("google/gemma-2-9b")
>>> tokenizer = AutoTokenizer.from_pretrained("google/gemma-2-9b")
>>> prompt = "What is your favorite condiment?"
>>> inputs = tokenizer(prompt, return_tensors="pt")
>>> # Generate
>>> generate_ids = model.generate(inputs.input_ids, max_length=30)
>>> tokenizer.batch_decode(generate_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False)[0]
"What is your favorite condiment?"
```"""
# decoder outputs consists of (dec_features, layer_state, dec_hidden, dec_attn)
outputs: BaseModelOutputWithPast = self.model(
input_ids=input_ids,
attention_mask=attention_mask,
position_ids=position_ids,
past_key_values=past_key_values,
inputs_embeds=inputs_embeds,
use_cache=use_cache,
**kwargs,
)
hidden_states = outputs.last_hidden_state
# Only compute necessary logits, and do not upcast them to float if we are not computing the loss
slice_indices = slice(-logits_to_keep, None) if isinstance(logits_to_keep, int) else logits_to_keep
logits = self.lm_head(hidden_states[:, slice_indices, :])
if self.config.final_logit_softcapping is not None:
logits = logits / self.config.final_logit_softcapping
logits = torch.tanh(logits)
logits = logits * self.config.final_logit_softcapping
loss = None
if labels is not None:
loss = self.loss_function(logits, labels, self.vocab_size, **kwargs)
return CausalLMOutputWithPast(
loss=loss,
logits=logits,
past_key_values=outputs.past_key_values,
hidden_states=outputs.hidden_states,
attentions=outputs.attentions,
)
def sliding_window_mask_function(sliding_window: tuple[int, int]) -> Callable:
"""
This creates uni/bidirectional attention mask with sliding window.
"""
def inner_mask(batch_idx: int, head_idx: int, q_idx: int, kv_idx: int) -> bool:
left_window_size, right_window_size = sliding_window
dist = q_idx - kv_idx
left_mask = (dist >= 0) & (dist < left_window_size)
right_mask = (dist < 0) & (-dist < right_window_size)
return left_mask | right_mask
return inner_mask
class Gemma4AudioModel(Gemma4PreTrainedModel):
"""An audio encoder based on the [Universal Speech Model](https://huggingface.co/papers/2303.01037) architecture."""
config: Gemma4AudioConfig
main_input_name = "input_features"
base_model_prefix = "model.audio_tower" # prefix for Gemma4ForConditionalGeneration saved checkpoints, required for Gemma4AudioModel.from_pretrained()
_can_record_outputs = {
"hidden_states": Gemma4AudioLayer,
"attentions": Gemma4AudioAttention,
}
def __init__(self, config: Gemma4AudioConfig):
super().__init__(config)
self.config = config
self.subsample_conv_projection = Gemma4AudioSubSampleConvProjection(config)
self.rel_pos_enc = Gemma4AudioRelPositionalEncoding(config)
self.layers = nn.ModuleList(
[Gemma4AudioLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)]
)
self.output_proj = nn.Linear(config.hidden_size, config.output_proj_dims, bias=True)
self.post_init()
def _convert_4d_mask_to_blocked_5d(self, mask_4d: torch.Tensor) -> torch.Tensor:
"""
Convert a standard 4D attention mask `[batch_size, 1, seq_len, seq_len]` to the 5D blocked format
`[batch_size, 1, num_blocks, chunk_size, context_size]` expected by the chunked local attention,
"""
batch_size, _, seq_len, _ = mask_4d.shape
device = mask_4d.device
chunk_size = self.config.attention_chunk_size
max_past_horizon = self.config.attention_context_left - 1
max_future_horizon = self.config.attention_context_right
num_blocks = (seq_len + chunk_size - 1) // chunk_size
padded_seq_len = num_blocks * chunk_size
pad_amount = padded_seq_len - seq_len
mask_4d = F.pad(mask_4d, (0, pad_amount, 0, pad_amount), value=False)
mask_5d = mask_4d.reshape(batch_size, 1, num_blocks, chunk_size, padded_seq_len)
mask_5d = F.pad(mask_5d, (max_past_horizon, max_future_horizon), value=False)
block_starts = torch.arange(num_blocks, device=device) * chunk_size
offsets = torch.arange(chunk_size + max_past_horizon + max_future_horizon, device=device)
kv_indices = block_starts[:, None] + offsets[None, :]
kv_indices = kv_indices[None, None, :, None, :].expand(batch_size, 1, -1, chunk_size, -1)
return mask_5d.gather(-1, kv_indices)
@merge_with_config_defaults
@capture_outputs
@auto_docstring(custom_intro="Encodes audio features to soft tokens.")
def forward(
self,
input_features: torch.Tensor,
attention_mask: torch.Tensor | None = None,
**kwargs: Unpack[TransformersKwargs],
) -> tuple[torch.Tensor, torch.BoolTensor]:
hidden_states, output_mask = self.subsample_conv_projection(input_features, attention_mask)
position_embeddings = self.rel_pos_enc(hidden_states)
attention_mask = create_bidirectional_mask(
config=self.config,
inputs_embeds=hidden_states,
attention_mask=output_mask,
and_mask_function=sliding_window_mask_function(
(self.config.attention_context_left - 1, self.config.attention_context_right)
),
)
attention_mask = self._convert_4d_mask_to_blocked_5d(attention_mask)
for encoder_layer in self.layers[: self.config.num_hidden_layers]:
hidden_states = encoder_layer(
hidden_states,
attention_mask=attention_mask,
position_embeddings=position_embeddings,
**kwargs,
)
hidden_states = self.output_proj(hidden_states)
return Gemma4AudioModelOutput(last_hidden_state=hidden_states, attention_mask=output_mask)
class Gemma4VisionModel(Gemma4PreTrainedModel):
"""The Gemma 4 Vision Encoder."""
config = Gemma4VisionConfig
_can_record_outputs = {
"hidden_states": Gemma4VisionEncoderLayer,
"attentions": Gemma4VisionAttention,
}
def __init__(self, config: Gemma4VisionConfig):
super().__init__(config)
self.patch_embedder = Gemma4VisionPatchEmbedder(config)
self.encoder = Gemma4VisionEncoder(config)
self.pooler = Gemma4VisionPooler(config)
if self.config.standardize:
self.register_buffer("std_bias", torch.empty(self.config.hidden_size))
self.register_buffer("std_scale", torch.empty(self.config.hidden_size))
self.post_init()
@merge_with_config_defaults
@capture_outputs
@auto_docstring(custom_intro="Encodes image pixels to soft tokens from patches.")
def forward(
self,
pixel_values: torch.FloatTensor,
pixel_position_ids: torch.LongTensor,
**kwargs: Unpack[TransformersKwargs],
) -> BaseModelOutputWithPast:
r"""
pixel_values (`torch.FloatTensor` or `list[torch.FloatTensor]`):
The images to encode. Either a single `[batch, channels, height, width]` tensor
(all images same size) or a list of `[1, channels, height, width]` tensors (different sizes).
pixel_position_ids (`torch.LongTensor` of shape `(batch_size, max_patches, 2)`):
The patch positions as (x, y) coordinates in the image. Padding patches are indicated by (-1, -1).
"""
pooling_kernel_size = self.config.pooling_kernel_size
output_length = pixel_values.shape[-2] // (pooling_kernel_size * pooling_kernel_size)
padding_positions = (pixel_position_ids == -1).all(dim=-1)
inputs_embeds = self.patch_embedder(pixel_values, pixel_position_ids, padding_positions)
output = self.encoder(
inputs_embeds=inputs_embeds,
attention_mask=~padding_positions, # encoder expects True=valid, padding_positions is True=padding
pixel_position_ids=pixel_position_ids,
**kwargs,
)
hidden_states, pooler_mask = self.pooler(
hidden_states=output.last_hidden_state,
pixel_position_ids=pixel_position_ids,
padding_positions=padding_positions,
output_length=output_length,
)
# Strip padding tokens. pooler_mask is True = valid, False = padding.
hidden_states = hidden_states[pooler_mask]
if self.config.standardize:
hidden_states = (hidden_states - self.std_bias) * self.std_scale
return BaseModelOutputWithPast(last_hidden_state=hidden_states)
class Gemma4MultimodalEmbedder(nn.Module):
"""Embeds token ids or soft tokens for multimodal content into language model space."""
def __init__(
self,
multimodal_config: Gemma4AudioConfig | Gemma4VisionConfig,
text_config: Gemma4TextConfig,
):
super().__init__()
self.multimodal_hidden_size = getattr(multimodal_config, "output_proj_dims", multimodal_config.hidden_size)
self.eps = multimodal_config.rms_norm_eps
self.text_hidden_size = text_config.hidden_size
self.embedding_projection = nn.Linear(self.multimodal_hidden_size, self.text_hidden_size, bias=False)
self.embedding_pre_projection_norm = Gemma4RMSNorm(self.multimodal_hidden_size, eps=self.eps, with_scale=False)
def forward(self, inputs_embeds: torch.Tensor) -> torch.Tensor:
"""Embeds token ids or soft tokens for multimodal content into language model space.
Args:
inputs_embeds: A torch.Tensor containing the soft tokens to embed.
Returns:
A torch.Tensor of embeddings with shape `[batch_size, seq_len, self.config.text_config.hidden_size]`.
"""
embs_normed = self.embedding_pre_projection_norm(inputs_embeds)
return self.embedding_projection(embs_normed)
# Identical as Gemma3 but modular can't resolve if we simply import. FIXME: @cyril
def token_type_ids_mask_function(
token_type_ids: torch.Tensor | None,
image_group_ids: torch.Tensor | None,
) -> Callable | None:
"""
This function adds the correct offsets to the `q_idx` and `kv_idx` as the torch API can only accept lengths,
not start and end indices.
"""
# Do not return an additional mask in this case
if token_type_ids is None:
return None
def inner_mask(batch_idx: int, head_idx: int, q_idx: int, kv_idx: int) -> bool:
seq_length = image_group_ids.shape[-1]
# clamp indices because with static cache they can go beyond `image_group_ids.shape[-1]`
q_idx_clamped = q_idx.clamp(max=seq_length - 1)
kv_idx_clamped = kv_idx.clamp(max=seq_length - 1)
# Unmask if the q and kv come from same group which is not -1 (i.e. non-text)
q_group = image_group_ids[batch_idx, q_idx_clamped]
kv_group = image_group_ids[batch_idx, kv_idx_clamped]
q_group = torch.where(q_idx < seq_length, q_group, -1)
kv_group = torch.where(kv_idx < seq_length, kv_group, -1)
return (q_group == kv_group) & (q_group >= 0)
return inner_mask
# Similar to Gemma3 but `sliding_mask_kwargs` and `mask_kwargs` are different and `token_type_ids->mm_token_type_ids`
def create_causal_mask_mapping(
config: PreTrainedConfig,
inputs_embeds: torch.Tensor,
attention_mask: torch.Tensor | None,
past_key_values: Cache | None,
position_ids: torch.Tensor | None,
mm_token_type_ids: torch.Tensor | None = None,
pixel_values: torch.FloatTensor | None = None,
is_training: bool = False,
is_first_iteration: bool | None = None,
**kwargs,
) -> dict:
"""
Overwrites the base `create_masks_for_generate` with `token_type_ids` masking to create the causal mask mapping
for all kinds of forward passes. Gemma4 uses a bidirectional mask for images.
Uses `pixel_values` as an optional input to disambiguate edge cases.
"""
if is_training and mm_token_type_ids is None:
raise ValueError("`mm_token_type_ids` is required as a model input when training")
mask_kwargs = {
"config": config.get_text_config(),
"inputs_embeds": inputs_embeds,
"attention_mask": attention_mask,
"past_key_values": past_key_values,
"position_ids": position_ids,
}
sliding_mask_kwargs = mask_kwargs.copy()
# NOTE: this `may_have_image_input` logic is not flawless, it fails when we're using a cache eagerly initialized
# (e.g. compiled prefill) AND `pixel_values` are not provided (i.e. the image data is provided through other
# means). Determining prefill in that case requires checking data values, which is not compile-compatible.
is_first_iteration = (
is_first_iteration
if is_first_iteration is not None
else (past_key_values is None or not past_key_values.is_initialized or pixel_values is not None)
)
if mm_token_type_ids is not None and is_first_iteration:
# We need to pass an additional mask function to account for token type ids, and it needs to be an `or` (to
# undo the causal masking)
# First find where a new vision block starts. Vision tokens cannot attend to
# future vision tokens, but can attend to all prev tokens and to itself bidirectionally
is_vision = (mm_token_type_ids == 1) | (mm_token_type_ids == 2)
is_prev_vision = torch.roll(is_vision, shifts=1, dims=-1)
is_prev_vision[..., 0] = False
new_vision_starts = is_vision & ~is_prev_vision
vision_group_ids = torch.cumsum(new_vision_starts.int(), dim=1) - 1
vision_group_ids = torch.where(is_vision, vision_group_ids, -1)
sliding_mask_kwargs["or_mask_function"] = token_type_ids_mask_function(
mm_token_type_ids.to(inputs_embeds.device), vision_group_ids
)
return {
"full_attention": create_causal_mask(**mask_kwargs),
"sliding_attention": create_sliding_window_causal_mask(**sliding_mask_kwargs),
}
@auto_docstring(
custom_intro="""
The base Gemma 4 model comprising a vision backbone, an audio backbone, and a language model without a
language modeling head.
"""
)
class Gemma4Model(Gemma4PreTrainedModel):
# we are filtering the logits/labels so we shouldn't divide the loss based on num_items_in_batch
accepts_loss_kwargs = False
def __init__(self, config: Gemma4Config):
super().__init__(config)
self.vocab_size = config.text_config.vocab_size
language_model = AutoModel.from_config(config=config.text_config)
self.language_model = language_model
self.vocab_size_per_layer_input = config.text_config.vocab_size_per_layer_input
self.vision_tower = AutoModel.from_config(config.vision_config) if config.vision_config is not None else None
self.embed_vision = (
Gemma4MultimodalEmbedder(config.vision_config, config.text_config)
if config.vision_config is not None
else None
)
self.audio_tower = AutoModel.from_config(config.audio_config) if config.audio_config is not None else None
self.embed_audio = (
Gemma4MultimodalEmbedder(config.audio_config, config.text_config)
if config.audio_config is not None
else None
)
self.post_init()
def get_input_embeddings(self):
return self.language_model.get_input_embeddings()
def set_input_embeddings(self, value):
self.language_model.set_input_embeddings(value)
@can_return_tuple
@auto_docstring(custom_intro="Projects the last hidden state from the vision model into language model space.")
def get_image_features(
self,
pixel_values: torch.FloatTensor,
image_position_ids: torch.LongTensor | None = None,
**kwargs: Unpack[TransformersKwargs],
) -> BaseModelOutputWithPooling:
r"""
image_position_ids (`torch.LongTensor` of shape `(batch_size, max_patches, 2)`, *optional*):
The patch positions as (x, y) coordinates in the image. Padding patches are indicated by (-1, -1).
"""
vision_outputs = self.vision_tower(
pixel_values=pixel_values,
pixel_position_ids=image_position_ids,
**kwargs,
)
last_hidden_state = vision_outputs.last_hidden_state
vision_outputs.pooler_output = self.embed_vision(inputs_embeds=last_hidden_state)
return vision_outputs
def get_placeholder_mask(
self,
input_ids: torch.LongTensor | None = None,
inputs_embeds: torch.FloatTensor | None = None,
) -> tuple[torch.BoolTensor, torch.BoolTensor, torch.BoolTensor]:
"""
Obtains mask for multimodal placeholders (replaced by soft tokens) and hard text tokens.
Masks will be obtained from `mm_token_type_ids`, `input_ids`, or `inputs_embeds` as available and in that
precedence order. If passing `input_ids` or `inputs_embeds`, the image mask will be derived using
`config.image_token_id`. Same goes for audio and video masks
Args:
input_ids: A tensor containing the hard token IDs from the text tokenizer.
inputs_embeds: A tensor containing the embeddings for all hard text tokens.
Returns:
image_mask, video_mask, audio_mask
"""
if input_ids is not None:
special_image_mask = input_ids == self.config.image_token_id
special_video_mask = input_ids == self.config.video_token_id
special_audio_mask = input_ids == self.config.audio_token_id
else:
special_image_mask = (
inputs_embeds
== self.get_input_embeddings()(
torch.tensor(self.config.image_token_id, dtype=torch.long, device=inputs_embeds.device)
)
).all(-1)
special_video_mask = (
inputs_embeds
== self.get_input_embeddings()(
torch.tensor(self.config.video_token_id, dtype=torch.long, device=inputs_embeds.device)
)
).all(-1)
special_audio_mask = (
inputs_embeds
== self.get_input_embeddings()(
torch.tensor(self.config.audio_token_id, dtype=torch.long, device=inputs_embeds.device)
)
).all(-1)
return special_image_mask, special_video_mask, special_audio_mask
@merge_with_config_defaults
@can_return_tuple
@auto_docstring
def forward(
self,
input_ids: torch.LongTensor | None = None,
pixel_values: torch.FloatTensor | None = None,
pixel_values_videos: torch.FloatTensor | None = None,
input_features: torch.FloatTensor | None = None,
attention_mask: torch.Tensor | None = None,
input_features_mask: torch.Tensor | None = None,
position_ids: torch.LongTensor | None = None,
past_key_values: Cache | None = None,
mm_token_type_ids: torch.LongTensor | None = None,
inputs_embeds: torch.FloatTensor | None = None,
use_cache: bool | None = None,
image_position_ids: torch.LongTensor | None = None,
video_position_ids: torch.LongTensor | None = None,
**kwargs: Unpack[TransformersKwargs],
) -> Gemma4ModelOutputWithPast:
r"""
input_features_mask (`torch.FloatTensor]` of shape `(num_images, seq_length)`):
The attention mask for the input audio.
image_position_ids (`torch.LongTensor` of shape `(batch_size, max_patches, 2)`, *optional*):
2D patch position coordinates from the image processor, with `(-1, -1)` indicating padding.
Passed through to the vision encoder for positional embedding computation.
video_position_ids (`torch.LongTensor` of shape `(num_videos, num_frames, max_patches, 2)`, *optional*):
2D patch position coordinates from the video processor, with `(-1, -1)` indicating padding.
Passed through to the vision encoder for positional embedding computation.
"""
if (input_ids is None) ^ (inputs_embeds is not None):
raise ValueError("You must specify exactly one of input_ids or inputs_embeds")
image_mask, video_mask, audio_mask = self.get_placeholder_mask(input_ids, inputs_embeds)
multimodal_mask = image_mask | video_mask | audio_mask
# Replace image id with PAD if the image token if OOV, to avoid index-errors
llm_input_ids = None
if inputs_embeds is None:
llm_input_ids = input_ids.clone()
llm_input_ids[multimodal_mask] = self.config.text_config.pad_token_id
inputs_embeds = self.get_input_embeddings()(llm_input_ids)
if self.config.get_text_config().hidden_size_per_layer_input:
pad_embedding = self.language_model.embed_tokens.weight[self.config.text_config.pad_token_id, :]
llm_inputs_embeds = torch.where(multimodal_mask[..., None], pad_embedding.view(1, 1, -1), inputs_embeds)
per_layer_inputs = self.language_model.get_per_layer_inputs(llm_input_ids, llm_inputs_embeds)
else:
per_layer_inputs = None
# Merge text and images
if pixel_values is not None:
image_features = self.get_image_features(pixel_values, image_position_ids, return_dict=True).pooler_output
image_features = image_features.to(inputs_embeds.device, inputs_embeds.dtype)
# Confirm the number of soft tokens from the vision tower matches the number of slots in the embeddings.
n_image_tokens = image_mask.sum()
image_mask = image_mask.unsqueeze(-1).expand_as(inputs_embeds).to(inputs_embeds.device)
torch_compilable_check(
inputs_embeds[image_mask].numel() == image_features.numel(),
f"Image features and image tokens do not match, tokens: {n_image_tokens}, features:"
f" {image_features.shape[0]}",
)
inputs_embeds = inputs_embeds.masked_scatter(
image_mask.to(inputs_embeds.device), image_features.to(inputs_embeds.device)
)
if pixel_values_videos is not None:
video_features = self.get_video_features(
pixel_values_videos, video_position_ids, return_dict=True
).pooler_output
video_features = video_features.to(inputs_embeds.device, inputs_embeds.dtype)
# Confirm the number of soft tokens from the vision tower matches the number of slots in the embeddings.
n_video_tokens = video_mask.sum()
video_mask = video_mask.unsqueeze(-1).expand_as(inputs_embeds).to(inputs_embeds.device)
torch_compilable_check(
inputs_embeds[video_mask].numel() == video_features.numel(),
f"Video features and video tokens do not match, tokens: {n_video_tokens}, features:"
f" {video_features.shape[0]}",
)
inputs_embeds = inputs_embeds.masked_scatter(
video_mask.to(inputs_embeds.device), video_features.to(inputs_embeds.device)
)
# Merge text and audio
if input_features is not None and input_features_mask is not None:
audio_output = self.get_audio_features(input_features, input_features_mask, return_dict=True)
audio_features = audio_output.pooler_output
audio_mask_from_encoder = audio_output.attention_mask # True = valid
# Strip padding tokens: only keep real (non-padding) audio soft tokens.
# audio_mask_from_encoder is True for valid positions, False for padding tokens.
# This mirrors the vision encoder's padding stripping (see Gemma4VisionEncoder.forward).
audio_features = audio_features[audio_mask_from_encoder]
n_audio_tokens = audio_mask.sum()
audio_mask = audio_mask.unsqueeze(-1).expand_as(inputs_embeds).to(inputs_embeds.device)
torch_compilable_check(
inputs_embeds[audio_mask].numel() == audio_features.numel(),
f"Audio features and audio tokens do not match, tokens: {n_audio_tokens}, features:"
f" {audio_features.shape[0] * audio_features.shape[1]}",
)
inputs_embeds = inputs_embeds.masked_scatter(
audio_mask.to(inputs_embeds.device), audio_features.to(inputs_embeds.device)
)
# It may already have been prepared by, e.g., `generate`
if position_ids is None:
past_seen_tokens = past_key_values.get_seq_length() if past_key_values is not None else 0
position_ids = torch.arange(inputs_embeds.shape[1], device=inputs_embeds.device) + past_seen_tokens
position_ids = position_ids.unsqueeze(0)
if not isinstance(causal_mask_mapping := attention_mask, dict):
if self.config.get_text_config().use_bidirectional_attention == "vision":
# Larger Gemma 4 models use Gemma 3's bidirectional attention mask for vision inputs
causal_mask_mapping = create_causal_mask_mapping(
self.config,
inputs_embeds,
attention_mask,
past_key_values,
position_ids,
mm_token_type_ids,
pixel_values,
is_training=self.training,
)
else:
# Smaller Gemma models use a conventional casual attention mask
causal_mask_mapping = create_masks_for_generate(
self.config,
inputs_embeds,
attention_mask,
past_key_values,
position_ids,
)
outputs = self.language_model(
per_layer_inputs=per_layer_inputs,
attention_mask=causal_mask_mapping,
position_ids=position_ids,
past_key_values=past_key_values,
inputs_embeds=inputs_embeds,
use_cache=use_cache,
return_dict=True,
**kwargs,
)
return Gemma4ModelOutputWithPast(
last_hidden_state=outputs.last_hidden_state,
past_key_values=outputs.past_key_values,
hidden_states=outputs.hidden_states,
attentions=outputs.attentions,
image_hidden_states=image_features if pixel_values is not None else None,
audio_hidden_states=audio_features if input_features is not None else None,
)
@can_return_tuple
@auto_docstring(custom_intro="Projects the last hidden state from the audio encoder into language model space.")
def get_audio_features(
self,
input_features: torch.Tensor,
input_features_mask: torch.Tensor,
**kwargs: Unpack[TransformersKwargs],
) -> tuple | Gemma4AudioModelOutput:
r"""
input_features (`torch.FloatTensor]` of shape `(num_images, seq_length, num_features)`):
The tensors corresponding to the input audio.
input_features_mask (`torch.FloatTensor]` of shape `(num_images, seq_length)`):
The attention mask for the input audio.
"""
if self.audio_tower is None:
raise ValueError(
"Audio features were requested, but the model was initialized without an audio_config. "
"Cannot process audio without an audio tower and audio embedder."
)
audio_outputs = self.audio_tower(input_features, input_features_mask, return_dict=True, **kwargs)
audio_outputs.pooler_output = self.embed_audio(inputs_embeds=audio_outputs.last_hidden_state)
return audio_outputs
@can_return_tuple
@auto_docstring(custom_intro="Projects the last hidden state from the vision encoder into language model space.")
def get_video_features(
self,
pixel_values_videos: torch.FloatTensor,
video_position_ids: torch.LongTensor | None = None,
**kwargs: Unpack[TransformersKwargs],
) -> BaseModelOutputWithPooling:
r"""
video_position_ids (`torch.LongTensor` of shape `(num_videos, num_frames, max_patches, 2)`, *optional*):
2D patch position coordinates from the video processor, with `(-1, -1)` indicating padding.
Passed through to the vision encoder for positional embedding computation.
"""
pixel_values_videos = pixel_values_videos.flatten(0, 1)
video_position_ids = video_position_ids.flatten(0, 1)
vision_outputs = self.vision_tower(
pixel_values=pixel_values_videos,
pixel_position_ids=video_position_ids,
**kwargs,
)
last_hidden_state = vision_outputs.last_hidden_state
vision_outputs.pooler_output = self.embed_vision(inputs_embeds=last_hidden_state)
return vision_outputs
@auto_docstring(
custom_intro="""
The base Gemma 4 model comprising a vision backbone, an audio backbone, a language model, and a language modeling
head.
"""
)
class Gemma4ForConditionalGeneration(Gemma4PreTrainedModel, GenerationMixin):
_tied_weights_keys = {"lm_head.weight": "model.language_model.embed_tokens.weight"}
base_model_prefix = "model"
def __init__(self, config: Gemma4Config):
super().__init__(config)
self.model = Gemma4Model(config)
self.lm_head = nn.Linear(config.text_config.hidden_size, config.text_config.vocab_size, bias=False)
self.post_init()
def get_input_embeddings(self):
return self.model.get_input_embeddings()
def set_input_embeddings(self, value):
self.model.set_input_embeddings(value)
@auto_docstring
def get_image_features(
self,
pixel_values: torch.FloatTensor,
image_position_ids: torch.LongTensor | None = None,
**kwargs: Unpack[TransformersKwargs],
):
r"""
image_position_ids (`torch.LongTensor` of shape `(batch_size, max_patches, 2)`, *optional*):
2D patch position coordinates from the image processor, with `(-1, -1)` indicating padding.
Passed through to the vision encoder for positional embedding computation.
"""
return self.model.get_image_features(pixel_values, image_position_ids, **kwargs)
@can_return_tuple
@auto_docstring
def forward(
self,
input_ids: torch.LongTensor | None = None,
pixel_values: torch.FloatTensor | None = None,
pixel_values_videos: torch.FloatTensor | None = None,
input_features: torch.FloatTensor | None = None,
attention_mask: torch.Tensor | None = None,
input_features_mask: torch.Tensor | None = None,
position_ids: torch.LongTensor | None = None,
image_position_ids: torch.LongTensor | None = None,
video_position_ids: torch.LongTensor | None = None,
past_key_values: Cache | None = None,
mm_token_type_ids: torch.LongTensor | None = None,
inputs_embeds: torch.FloatTensor | None = None,
labels: torch.LongTensor | None = None,
use_cache: bool | None = None,
logits_to_keep: int | torch.Tensor = 0,
**kwargs: Unpack[TransformersKwargs],
) -> Gemma4CausalLMOutputWithPast:
r"""
input_features_mask (`torch.FloatTensor]` of shape `(num_images, seq_length)`):
The attention mask for the input audio.
image_position_ids (`torch.LongTensor` of shape `(batch_size, max_patches, 2)`, *optional*):
2D patch position coordinates from the image processor, with `(-1, -1)` indicating padding.
Passed through to the vision encoder for positional embedding computation.
video_position_ids (`torch.LongTensor` of shape `(num_videos, num_frames, max_patches, 2)`, *optional*):
2D patch position coordinates from the video processor, with `(-1, -1)` indicating padding.
Passed through to the vision encoder for positional embedding computation.
"""
outputs = self.model(
input_ids=input_ids,
pixel_values=pixel_values,
pixel_values_videos=pixel_values_videos,
input_features=input_features,
attention_mask=attention_mask,
input_features_mask=input_features_mask,
position_ids=position_ids,
past_key_values=past_key_values,
mm_token_type_ids=mm_token_type_ids,
inputs_embeds=inputs_embeds,
labels=labels,
use_cache=use_cache,
image_position_ids=image_position_ids,
video_position_ids=video_position_ids,
return_dict=True,
**kwargs,
)
hidden_states = outputs.last_hidden_state
# Only compute necessary logits, and do not upcast them to float if we are not computing the loss
slice_indices = slice(-logits_to_keep, None) if isinstance(logits_to_keep, int) else logits_to_keep
logits = self.lm_head(hidden_states[:, slice_indices, :])
if (final_logit_softcapping := self.config.get_text_config().final_logit_softcapping) is not None:
logits = logits / final_logit_softcapping
logits = torch.tanh(logits)
logits = logits * final_logit_softcapping
loss = None
if labels is not None:
# Upcast to float if we need to compute the loss to avoid potential precision issues
logits = logits.float()
shift_logits = logits[..., :-1, :]
shift_labels = labels[..., 1:]
if attention_mask is not None:
# we use the input attention mask to shift the logits and labels, because it is 2D.
# we also crop attn mask in case it is longer, which happens in PrefixTuning with peft
shift_attention_mask = attention_mask[:, -shift_logits.shape[1] :].to(logits.device)
shift_logits = shift_logits[shift_attention_mask.to(logits.device) != 0].contiguous()
shift_labels = shift_labels[shift_attention_mask.to(shift_labels.device) != 0].contiguous()
else:
shift_logits = shift_logits.contiguous()
shift_labels = shift_labels.contiguous()
# Flatten the tokens
loss_fct = nn.CrossEntropyLoss()
flat_logits = shift_logits.view(-1, self.config.get_text_config().vocab_size)
flat_labels = shift_labels.view(-1).to(shift_logits.device)
loss = loss_fct(flat_logits, flat_labels)
return Gemma4CausalLMOutputWithPast(
loss=loss,
logits=logits,
past_key_values=outputs.past_key_values,
hidden_states=outputs.hidden_states,
attentions=outputs.attentions,
image_hidden_states=outputs.image_hidden_states,
audio_hidden_states=outputs.audio_hidden_states,
)
def prepare_inputs_for_generation(
self,
input_ids,
past_key_values=None,
inputs_embeds=None,
position_ids=None,
pixel_values=None,
pixel_values_videos=None,
input_features=None,
attention_mask=None,
input_features_mask=None,
token_type_ids=None,
use_cache=True,
logits_to_keep=None,
labels=None,
is_first_iteration=False,
**kwargs,
):
# Overwritten -- custom `position_ids` and `pixel_values` handling
model_inputs = super().prepare_inputs_for_generation(
input_ids,
past_key_values=past_key_values,
inputs_embeds=inputs_embeds,
attention_mask=attention_mask,
position_ids=position_ids,
use_cache=use_cache,
logits_to_keep=logits_to_keep,
token_type_ids=token_type_ids,
is_first_iteration=is_first_iteration,
**kwargs,
)
# If we're in cached decoding stage, multimodal inputs are already cached and can be dropped
if is_first_iteration or not use_cache:
model_inputs["pixel_values"] = pixel_values
model_inputs["pixel_values_videos"] = pixel_values_videos
model_inputs["input_features"] = input_features
model_inputs["input_features_mask"] = input_features_mask
return model_inputs
@staticmethod
def create_masks_for_generate(
config: PreTrainedConfig,
inputs_embeds: torch.Tensor,
attention_mask: torch.Tensor | None,
past_key_values: Cache | None,
position_ids: torch.Tensor | None,
mm_token_type_ids: torch.Tensor | None = None,
is_first_iteration: bool | None = False,
**kwargs,
) -> dict:
if getattr(config.get_text_config(), "use_bidirectional_attention", None) == "vision":
# Larger Gemma 4 models use Gemma 3's bidirectional attention mask for vision inputs
return create_causal_mask_mapping(
config,
inputs_embeds,
attention_mask,
past_key_values,
position_ids,
mm_token_type_ids,
is_first_iteration=is_first_iteration,
**{k: v for k, v in kwargs.items() if k != "pixel_values"},
)
else:
# Smaller Gemma models use a conventional casual attention mask
return create_masks_for_generate(
config, inputs_embeds, attention_mask, past_key_values, position_ids, **kwargs
)
def resolve_gemma4_text_causal_lm_class(config: Gemma4TextConfig):
if getattr(config, "use_zero_compute_optimization", False):
from .gemma4_optimization import OptimizedGemma4ForCausalLM
return OptimizedGemma4ForCausalLM
return Gemma4ForCausalLM
def build_gemma4_text_causal_lm(config: Gemma4TextConfig):
return resolve_gemma4_text_causal_lm_class(config)(config)
__all__ = [
"build_gemma4_text_causal_lm",
"Gemma4AudioModel",
"Gemma4ForCausalLM",
"Gemma4ForConditionalGeneration",
"Gemma4Model",
"Gemma4PreTrainedModel",
"Gemma4TextModel",
"Gemma4VisionModel",
"resolve_gemma4_text_causal_lm_class",
]