diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000000000000000000000000000000000000..a6344aac8c09253b3b630fb776ae94478aa0275b --- /dev/null +++ b/.gitattributes @@ -0,0 +1,35 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text diff --git a/1280ms/.DS_Store b/1280ms/.DS_Store new file mode 100644 index 0000000000000000000000000000000000000000..6c0472dd5f7cb37790e42abfe607b3ecf31df01e Binary files /dev/null and b/1280ms/.DS_Store differ diff --git a/1280ms/convert_parakeet_eou.py b/1280ms/convert_parakeet_eou.py new file mode 100644 index 0000000000000000000000000000000000000000..9c865f90c8132840b5755dcd62fcfac15e3f0352 --- /dev/null +++ b/1280ms/convert_parakeet_eou.py @@ -0,0 +1,740 @@ +#!/usr/bin/env python3 +"""CLI for exporting Parakeet Realtime EOU 120M components to CoreML. + +This model is a cache-aware streaming FastConformer-RNNT model optimized for +low-latency speech recognition with end-of-utterance detection. + +Key differences from Parakeet TDT v3: +- Smaller model (120M vs 600M params) +- No duration outputs (standard RNNT, not TDT) +- Cache-aware streaming encoder (17 layers, attention context [70,1]) +- Special token for end-of-utterance detection +- Optimized for 80-160ms latency + +Reference: https://huggingface.co/nvidia/parakeet_realtime_eou_120m-v1 +""" +from __future__ import annotations + +import json +from dataclasses import asdict +from pathlib import Path +from typing import Dict, Optional, Tuple + +import coremltools as ct +import numpy as np +import soundfile as sf +import torch +import typer + +import nemo.collections.asr as nemo_asr + +from individual_components import ( + DecoderWrapper, + EncoderWrapper, + ExportSettings, + JointWrapper, + JointDecisionWrapper, + JointDecisionSingleStep, + PreprocessorWrapper, + MelEncoderWrapper, + _coreml_convert, +) + +def apply_stft_patch(): + # Monkey patch coremltools.stft to handle extra arguments from newer torch versions + try: + import coremltools.converters.mil.frontend.torch.ops as torch_ops + _original_stft = torch_ops.stft + + def patched_stft(context, node): + if len(node.inputs) > 8: + node.inputs = node.inputs[:8] + return _original_stft(context, node) + + torch_ops.stft = patched_stft + if "stft" in torch_ops._TORCH_OPS_REGISTRY: + torch_ops._TORCH_OPS_REGISTRY["stft"] = patched_stft + print("Monkey patched coremltools.stft for compatibility.") + except Exception as e: + print(f"Warning: Could not monkey patch stft: {e}") + +DEFAULT_MODEL_ID = "nvidia/parakeet_realtime_eou_120m-v1" +AUTHOR = "Fluid Inference" + + +def _compute_length(seconds: float, sample_rate: int) -> int: + return int(round(seconds * sample_rate)) + + +def _prepare_audio( + validation_audio: Optional[Path], + sample_rate: int, + max_samples: int, + seed: Optional[int], +) -> torch.Tensor: + if validation_audio is None: + if seed is not None: + torch.manual_seed(seed) + audio = torch.randn(1, max_samples, dtype=torch.float32) + return audio + + data, sr = sf.read(str(validation_audio), dtype="float32") + if sr != sample_rate: + raise typer.BadParameter( + f"Validation audio sample rate {sr} does not match model rate {sample_rate}" + ) + + if data.ndim > 1: + data = data[:, 0] + + if data.size == 0: + raise typer.BadParameter("Validation audio is empty") + + if data.size < max_samples: + pad_width = max_samples - data.size + data = np.pad(data, (0, pad_width)) + elif data.size > max_samples: + data = data[:max_samples] + + audio = torch.from_numpy(data).unsqueeze(0).to(dtype=torch.float32) + return audio + + +def _save_mlpackage(model: ct.models.MLModel, path: Path, description: str) -> None: + try: + model.minimum_deployment_target = ct.target.iOS17 + except Exception: + pass + model.short_description = description + model.author = AUTHOR + path.parent.mkdir(parents=True, exist_ok=True) + model.save(str(path)) + + +def _tensor_shape(tensor: torch.Tensor) -> Tuple[int, ...]: + return tuple(int(dim) for dim in tensor.shape) + + +def _parse_compute_units(name: str) -> ct.ComputeUnit: + """Parse a human-friendly compute units string into ct.ComputeUnit.""" + normalized = str(name).strip().upper() + mapping = { + "ALL": ct.ComputeUnit.ALL, + "CPU_ONLY": ct.ComputeUnit.CPU_ONLY, + "CPU_AND_GPU": ct.ComputeUnit.CPU_AND_GPU, + "CPU_AND_NE": ct.ComputeUnit.CPU_AND_NE, + "CPU_AND_NEURALENGINE": ct.ComputeUnit.CPU_AND_NE, + } + if normalized not in mapping: + raise typer.BadParameter( + f"Unknown compute units '{name}'. Choose from: " + ", ".join(mapping.keys()) + ) + return mapping[normalized] + + +def _parse_compute_precision(name: Optional[str]) -> Optional[ct.precision]: + """Parse compute precision string into ct.precision or None.""" + if name is None: + return None + normalized = str(name).strip().upper() + if normalized == "": + return None + mapping = { + "FLOAT32": ct.precision.FLOAT32, + "FLOAT16": ct.precision.FLOAT16, + } + if normalized not in mapping: + raise typer.BadParameter( + f"Unknown compute precision '{name}'. Choose from: " + + ", ".join(mapping.keys()) + ) + return mapping[normalized] + + +app = typer.Typer(add_completion=False, pretty_exceptions_show_locals=False) + + +@app.command() +def convert( + nemo_path: Optional[Path] = typer.Option( + None, + "--nemo-path", + exists=True, + resolve_path=True, + help="Path to parakeet_realtime_eou_120m-v1.nemo checkpoint (skip to auto-download)", + ), + model_id: str = typer.Option( + DEFAULT_MODEL_ID, + "--model-id", + help="Model identifier to download when --nemo-path is omitted", + ), + output_dir: Path = typer.Option( + Path("parakeet_eou_coreml"), + help="Directory where mlpackages and metadata will be written", + ), + preprocessor_cu: str = typer.Option( + "CPU_ONLY", + "--preprocessor-cu", + help="Compute units for preprocessor (default CPU_ONLY)", + ), + mel_encoder_cu: str = typer.Option( + "CPU_ONLY", + "--mel-encoder-cu", + help="Compute units for fused mel+encoder (default CPU_ONLY)", + ), + compute_precision: Optional[str] = typer.Option( + None, + "--compute-precision", + help="Export precision: FLOAT32 (default) or FLOAT16 to shrink non-quantized weights.", + ), + max_audio_seconds: float = typer.Option( + 15.0, + "--max-audio-seconds", + help="Maximum audio duration in seconds for the fixed window export", + ), + validation_audio: Optional[Path] = typer.Option( + None, + "--validation-audio", + exists=True, + resolve_path=True, + help="Path to a 16kHz WAV file for tracing (uses random if not provided)", + ), +) -> None: + """Export all Parakeet Realtime EOU sub-modules to CoreML. + + This exports the cache-aware streaming FastConformer-RNNT model for + low-latency speech recognition with end-of-utterance detection. + """ + export_settings = ExportSettings( + output_dir=output_dir, + compute_units=ct.ComputeUnit.CPU_ONLY, + deployment_target=ct.target.iOS17, + compute_precision=_parse_compute_precision(compute_precision), + max_audio_seconds=max_audio_seconds, + max_symbol_steps=1, + ) + + typer.echo("Export configuration:") + typer.echo(asdict(export_settings)) + + output_dir.mkdir(parents=True, exist_ok=True) + pre_cu = _parse_compute_units(preprocessor_cu) + melenc_cu = _parse_compute_units(mel_encoder_cu) + + if nemo_path is not None: + typer.echo(f"Loading NeMo model from {nemo_path}…") + # Try loading as generic ASRModel first, then specific class + try: + asr_model = nemo_asr.models.ASRModel.restore_from( + str(nemo_path), map_location="cpu" + ) + except Exception: + # Fallback to EncDecRNNTBPEModel + asr_model = nemo_asr.models.EncDecRNNTBPEModel.restore_from( + str(nemo_path), map_location="cpu" + ) + checkpoint_meta = { + "type": "file", + "path": str(nemo_path), + } + else: + typer.echo(f"Downloading NeMo model via {model_id}…") + # Use ASRModel.from_pretrained as recommended for this model + try: + asr_model = nemo_asr.models.ASRModel.from_pretrained( + model_id, map_location="cpu" + ) + except Exception: + # Fallback to EncDecRNNTBPEModel + asr_model = nemo_asr.models.EncDecRNNTBPEModel.from_pretrained( + model_id, map_location="cpu" + ) + checkpoint_meta = { + "type": "pretrained", + "model_id": model_id, + } + asr_model.eval() + + # Print model info + typer.echo(f"Model class: {type(asr_model).__name__}") + typer.echo(f"Encoder class: {type(asr_model.encoder).__name__}") + + sample_rate = int(asr_model.cfg.preprocessor.sample_rate) + max_samples = _compute_length(export_settings.max_audio_seconds, sample_rate) + + # Prepare audio for tracing + if validation_audio is not None: + typer.echo(f"Using validation audio: {validation_audio}") + audio_tensor = _prepare_audio(validation_audio, sample_rate, max_samples, seed=None) + else: + typer.echo("Using random audio for tracing (seed=42)") + audio_tensor = _prepare_audio(None, sample_rate, max_samples, seed=42) + + audio_length = torch.tensor([max_samples], dtype=torch.int32) + + preprocessor = PreprocessorWrapper(asr_model.preprocessor.eval()) + encoder = EncoderWrapper(asr_model.encoder.eval()) + decoder = DecoderWrapper(asr_model.decoder.eval()) + joint = JointWrapper(asr_model.joint.eval()) + + decoder_export_flag = getattr(asr_model.decoder, "_rnnt_export", False) + asr_model.decoder._rnnt_export = True + + try: + with torch.no_grad(): + mel_ref, mel_length_ref = preprocessor(audio_tensor, audio_length) + mel_length_ref = mel_length_ref.to(dtype=torch.int32) + encoder_ref, encoder_length_ref, frame_times_ref = encoder( + mel_ref, mel_length_ref + ) + encoder_length_ref = encoder_length_ref.to(dtype=torch.int32) + + # Clone tensors to drop inference flags + mel_ref = mel_ref.clone().detach() + mel_length_ref = mel_length_ref.clone().detach() + encoder_ref = encoder_ref.clone().detach() + encoder_length_ref = encoder_length_ref.clone().detach() + frame_times_ref = frame_times_ref.clone().detach() + + vocab_size = int(asr_model.tokenizer.vocab_size) + decoder_hidden = int(asr_model.decoder.pred_hidden) + decoder_layers = int(asr_model.decoder.pred_rnn_layers) + + # Check if model has extra outputs (TDT-style duration) + num_extra = getattr(asr_model.joint, "num_extra_outputs", 0) + typer.echo(f"Vocab size: {vocab_size}, num_extra_outputs: {num_extra}") + + targets = torch.full( + (1, export_settings.max_symbol_steps), + fill_value=asr_model.decoder.blank_idx, + dtype=torch.int32, + ) + target_lengths = torch.tensor( + [export_settings.max_symbol_steps], dtype=torch.int32 + ) + zero_state = torch.zeros( + decoder_layers, + 1, + decoder_hidden, + dtype=torch.float32, + ) + + with torch.no_grad(): + decoder_ref, h_ref, c_ref = decoder( + targets, target_lengths, zero_state, zero_state + ) + joint_ref = joint(encoder_ref, decoder_ref) + + decoder_ref = decoder_ref.clone() + h_ref = h_ref.clone() + c_ref = c_ref.clone() + joint_ref = joint_ref.clone() + + typer.echo(f"Encoder output shape: {encoder_ref.shape}") + typer.echo(f"Decoder output shape: {decoder_ref.shape}") + typer.echo(f"Joint output shape: {joint_ref.shape}") + + # === Export Preprocessor === + typer.echo("Tracing and converting preprocessor…") + preprocessor = preprocessor.cpu() + audio_tensor = audio_tensor.cpu() + audio_length = audio_length.cpu() + traced_preprocessor = torch.jit.trace( + preprocessor, (audio_tensor, audio_length), strict=False + ) + traced_preprocessor.eval() + preprocessor_inputs = [ + ct.TensorType( + name="audio_signal", + shape=(1, ct.RangeDim(1, max_samples)), + dtype=np.float32, + ), + ct.TensorType(name="audio_length", shape=(1,), dtype=np.int32), + ] + preprocessor_outputs = [ + ct.TensorType(name="mel", dtype=np.float32), + ct.TensorType(name="mel_length", dtype=np.int32), + ] + preprocessor_model = _coreml_convert( + traced_preprocessor, + preprocessor_inputs, + preprocessor_outputs, + export_settings, + compute_units_override=pre_cu, + ) + preprocessor_path = output_dir / "parakeet_eou_preprocessor.mlpackage" + _save_mlpackage( + preprocessor_model, + preprocessor_path, + f"Parakeet EOU preprocessor ({max_audio_seconds}s window)", + ) + + # === Export Encoder === + typer.echo("Tracing and converting encoder…") + traced_encoder = torch.jit.trace( + encoder, (mel_ref, mel_length_ref), strict=False + ) + traced_encoder.eval() + encoder_inputs = [ + ct.TensorType( + name="mel", shape=_tensor_shape(mel_ref), dtype=np.float32 + ), + ct.TensorType(name="mel_length", shape=(1,), dtype=np.int32), + ] + encoder_outputs = [ + ct.TensorType(name="encoder", dtype=np.float32), + ct.TensorType(name="encoder_length", dtype=np.int32), + ct.TensorType(name="frame_times", dtype=np.float32), + ] + encoder_model = _coreml_convert( + traced_encoder, + encoder_inputs, + encoder_outputs, + export_settings, + compute_units_override=ct.ComputeUnit.CPU_ONLY, + ) + encoder_path = output_dir / "parakeet_eou_encoder.mlpackage" + _save_mlpackage( + encoder_model, + encoder_path, + f"Parakeet EOU encoder ({max_audio_seconds}s window)", + ) + + # === Export Fused Mel+Encoder === + typer.echo("Tracing and converting fused mel+encoder…") + mel_encoder = MelEncoderWrapper(preprocessor, encoder) + traced_mel_encoder = torch.jit.trace( + mel_encoder, (audio_tensor, audio_length), strict=False + ) + traced_mel_encoder.eval() + mel_encoder_inputs = [ + ct.TensorType( + name="audio_signal", shape=(1, max_samples), dtype=np.float32 + ), + ct.TensorType(name="audio_length", shape=(1,), dtype=np.int32), + ] + mel_encoder_outputs = [ + ct.TensorType(name="encoder", dtype=np.float32), + ct.TensorType(name="encoder_length", dtype=np.int32), + ct.TensorType(name="frame_times", dtype=np.float32), + ] + mel_encoder_model = _coreml_convert( + traced_mel_encoder, + mel_encoder_inputs, + mel_encoder_outputs, + export_settings, + compute_units_override=melenc_cu, + ) + mel_encoder_path = output_dir / "parakeet_eou_mel_encoder.mlpackage" + _save_mlpackage( + mel_encoder_model, + mel_encoder_path, + f"Parakeet EOU fused Mel+Encoder ({max_audio_seconds}s window)", + ) + + # === Export Decoder === + typer.echo("Tracing and converting decoder…") + traced_decoder = torch.jit.trace( + decoder, + (targets, target_lengths, zero_state, zero_state), + strict=False, + ) + traced_decoder.eval() + decoder_inputs = [ + ct.TensorType( + name="targets", shape=_tensor_shape(targets), dtype=np.int32 + ), + ct.TensorType(name="target_length", shape=(1,), dtype=np.int32), + ct.TensorType( + name="h_in", shape=_tensor_shape(zero_state), dtype=np.float32 + ), + ct.TensorType( + name="c_in", shape=_tensor_shape(zero_state), dtype=np.float32 + ), + ] + decoder_outputs = [ + ct.TensorType(name="decoder", dtype=np.float32), + ct.TensorType(name="h_out", dtype=np.float32), + ct.TensorType(name="c_out", dtype=np.float32), + ] + decoder_model = _coreml_convert( + traced_decoder, + decoder_inputs, + decoder_outputs, + export_settings, + compute_units_override=ct.ComputeUnit.CPU_ONLY, + ) + decoder_path = output_dir / "parakeet_eou_decoder.mlpackage" + _save_mlpackage( + decoder_model, + decoder_path, + "Parakeet EOU decoder (RNNT prediction network)", + ) + + # === Export Joint === + typer.echo("Tracing and converting joint…") + traced_joint = torch.jit.trace( + joint, + (encoder_ref, decoder_ref), + strict=False, + ) + traced_joint.eval() + joint_inputs = [ + ct.TensorType( + name="encoder", shape=_tensor_shape(encoder_ref), dtype=np.float32 + ), + ct.TensorType( + name="decoder", shape=_tensor_shape(decoder_ref), dtype=np.float32 + ), + ] + joint_outputs = [ + ct.TensorType(name="logits", dtype=np.float32), + ] + joint_model = _coreml_convert( + traced_joint, + joint_inputs, + joint_outputs, + export_settings, + compute_units_override=ct.ComputeUnit.CPU_ONLY, + ) + joint_path = output_dir / "parakeet_eou_joint.mlpackage" + _save_mlpackage( + joint_model, + joint_path, + "Parakeet EOU joint network (RNNT)", + ) + + # === Export Joint Decision Head === + typer.echo("Tracing and converting joint decision head…") + joint_decision = JointDecisionWrapper(joint, vocab_size=vocab_size) + traced_joint_decision = torch.jit.trace( + joint_decision, + (encoder_ref, decoder_ref), + strict=False, + ) + traced_joint_decision.eval() + joint_decision_inputs = [ + ct.TensorType( + name="encoder", shape=_tensor_shape(encoder_ref), dtype=np.float32 + ), + ct.TensorType( + name="decoder", shape=_tensor_shape(decoder_ref), dtype=np.float32 + ), + ] + joint_decision_outputs = [ + ct.TensorType(name="token_id", dtype=np.int32), + ct.TensorType(name="token_prob", dtype=np.float32), + ] + joint_decision_model = _coreml_convert( + traced_joint_decision, + joint_decision_inputs, + joint_decision_outputs, + export_settings, + compute_units_override=ct.ComputeUnit.CPU_ONLY, + ) + joint_decision_path = output_dir / "parakeet_eou_joint_decision.mlpackage" + _save_mlpackage( + joint_decision_model, + joint_decision_path, + "Parakeet EOU joint + decision head (softmax, argmax)", + ) + + # === Export Single-Step Joint Decision === + typer.echo("Tracing and converting single-step joint decision…") + jd_single = JointDecisionSingleStep(joint, vocab_size=vocab_size) + # Create single-step slices from refs + enc_step = encoder_ref[:, :, :1].contiguous() + dec_step = decoder_ref[:, :, :1].contiguous() + traced_jd_single = torch.jit.trace( + jd_single, + (enc_step, dec_step), + strict=False, + ) + traced_jd_single.eval() + jd_single_inputs = [ + ct.TensorType( + name="encoder_step", + shape=(1, enc_step.shape[1], 1), + dtype=np.float32, + ), + ct.TensorType( + name="decoder_step", + shape=(1, dec_step.shape[1], 1), + dtype=np.float32, + ), + ] + jd_single_outputs = [ + ct.TensorType(name="token_id", dtype=np.int32), + ct.TensorType(name="token_prob", dtype=np.float32), + ct.TensorType(name="top_k_ids", dtype=np.int32), + ct.TensorType(name="top_k_logits", dtype=np.float32), + ] + jd_single_model = _coreml_convert( + traced_jd_single, + jd_single_inputs, + jd_single_outputs, + export_settings, + compute_units_override=ct.ComputeUnit.CPU_ONLY, + ) + jd_single_path = output_dir / "parakeet_eou_joint_decision_single_step.mlpackage" + _save_mlpackage( + jd_single_model, + jd_single_path, + "Parakeet EOU single-step joint decision (current frame)", + ) + + # === Save Metadata === + metadata: Dict[str, object] = { + "model_id": model_id, + "model_name": "parakeet_realtime_eou_120m-v1", + "model_class": type(asr_model).__name__, + "encoder_class": type(asr_model.encoder).__name__, + "sample_rate": sample_rate, + "max_audio_seconds": export_settings.max_audio_seconds, + "max_audio_samples": max_samples, + "max_symbol_steps": export_settings.max_symbol_steps, + "vocab_size": vocab_size, + "vocab_with_blank": vocab_size + 1, + "decoder_hidden": decoder_hidden, + "decoder_layers": decoder_layers, + "num_extra_outputs": num_extra, + "has_eou_token": True, + "checkpoint": checkpoint_meta, + "coreml": { + "compute_units": export_settings.compute_units.name, + "compute_precision": ( + export_settings.compute_precision.name + if export_settings.compute_precision is not None + else "FLOAT32" + ), + }, + "components": { + "preprocessor": { + "inputs": { + "audio_signal": [1, max_samples], + "audio_length": [1], + }, + "outputs": { + "mel": list(_tensor_shape(mel_ref)), + "mel_length": [1], + }, + "path": preprocessor_path.name, + }, + "encoder": { + "inputs": { + "mel": list(_tensor_shape(mel_ref)), + "mel_length": [1], + }, + "outputs": { + "encoder": list(_tensor_shape(encoder_ref)), + "encoder_length": [1], + "frame_times": [1, _tensor_shape(encoder_ref)[2]], + }, + "path": encoder_path.name, + }, + "mel_encoder": { + "inputs": { + "audio_signal": [1, max_samples], + "audio_length": [1], + }, + "outputs": { + "encoder": list(_tensor_shape(encoder_ref)), + "encoder_length": [1], + "frame_times": [1, _tensor_shape(encoder_ref)[2]], + }, + "path": mel_encoder_path.name, + }, + "decoder": { + "inputs": { + "targets": list(_tensor_shape(targets)), + "target_length": [1], + "h_in": list(_tensor_shape(zero_state)), + "c_in": list(_tensor_shape(zero_state)), + }, + "outputs": { + "decoder": list(_tensor_shape(decoder_ref)), + "h_out": list(_tensor_shape(h_ref)), + "c_out": list(_tensor_shape(c_ref)), + }, + "path": decoder_path.name, + }, + "joint": { + "inputs": { + "encoder": list(_tensor_shape(encoder_ref)), + "decoder": list(_tensor_shape(decoder_ref)), + }, + "outputs": { + "logits": list(_tensor_shape(joint_ref)), + }, + "path": joint_path.name, + }, + "joint_decision": { + "inputs": { + "encoder": list(_tensor_shape(encoder_ref)), + "decoder": list(_tensor_shape(decoder_ref)), + }, + "outputs": { + "token_id": [ + _tensor_shape(encoder_ref)[0], + _tensor_shape(encoder_ref)[2], + _tensor_shape(decoder_ref)[2], + ], + "token_prob": [ + _tensor_shape(encoder_ref)[0], + _tensor_shape(encoder_ref)[2], + _tensor_shape(decoder_ref)[2], + ], + }, + "path": joint_decision_path.name, + }, + "joint_decision_single_step": { + "inputs": { + "encoder_step": [1, _tensor_shape(encoder_ref)[1], 1], + "decoder_step": [1, _tensor_shape(decoder_ref)[1], 1], + }, + "outputs": { + "token_id": [1, 1, 1], + "token_prob": [1, 1, 1], + "top_k_ids": [1, 1, 1, 64], + "top_k_logits": [1, 1, 1, 64], + }, + "path": jd_single_path.name, + }, + }, + } + + # Export tokenizer vocab if available + try: + tokenizer = asr_model.tokenizer + vocab = { + "blank_id": int(asr_model.decoder.blank_idx), + "vocab_size": vocab_size, + } + # Try to get special tokens + if hasattr(tokenizer, "tokenizer"): + inner_tokenizer = tokenizer.tokenizer + if hasattr(inner_tokenizer, "get_vocab"): + full_vocab = inner_tokenizer.get_vocab() + # Find EOU token + eou_token = None + for token, idx in full_vocab.items(): + if "" in token.upper() or "eou" in token.lower(): + eou_token = {"token": token, "id": idx} + break + if eou_token: + vocab["eou_token"] = eou_token + metadata["tokenizer"] = vocab + except Exception as e: + typer.echo(f"Warning: Could not export tokenizer info: {e}") + + metadata_path = output_dir / "metadata.json" + metadata_path.write_text(json.dumps(metadata, indent=2)) + typer.echo(f"\nExport complete. Metadata written to {metadata_path}") + typer.echo(f"Output directory: {output_dir}") + + finally: + asr_model.decoder._rnnt_export = decoder_export_flag + + +if __name__ == "__main__": + app() diff --git a/1280ms/convert_streaming_encoder.py b/1280ms/convert_streaming_encoder.py new file mode 100644 index 0000000000000000000000000000000000000000..a1c199c8a6c540dab48b973ab4232db2901b8e72 --- /dev/null +++ b/1280ms/convert_streaming_encoder.py @@ -0,0 +1,193 @@ + +import torch +import torch.nn as nn +import coremltools as ct +import numpy as np +import typer +from pathlib import Path +from typing import Tuple, List, Optional +import json +import shutil + +# Iimport torch +import coremltools as ct +import numpy as np +import argparse +from nemo.collections.asr.models import EncDecRNNTBPEModel + +app = typer.Typer() + +class LoopbackEncoderWrapper(nn.Module): + """ + Wraps the entire Parakeet Encoder (PreEncode + Conformer) for CoreML Loopback Streaming. + + Inputs: + - audio_signal: [B, D, T] (Mel spectrogram chunk) + - audio_length: [B] + - pre_cache: [B, D, pre_cache_size] (Previous audio context) + - cache_last_channel: [layers, B, cache_size, hidden] + - cache_last_time: [layers, B, hidden, time_cache] + - cache_last_channel_len: [B] + + Outputs: + - encoded_output: [B, D_out, T_out] + - encoded_length: [B] + - new_pre_cache: [B, D, pre_cache_size] + - new_cache_last_channel + - new_cache_last_time + - new_cache_last_channel_len + """ + def __init__(self, encoder, pre_cache_size=16): + super().__init__() + self.encoder = encoder + self.pre_cache_size = pre_cache_size + + def forward( + self, + audio_signal: torch.Tensor, + audio_length: torch.Tensor, + pre_cache: torch.Tensor, + cache_last_channel: torch.Tensor, + cache_last_time: torch.Tensor, + cache_last_channel_len: torch.Tensor + ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor]: + + # 1. Prepend pre_cache to audio_signal + # audio_signal: [B, D, T] + # pre_cache: [B, D, T_cache] + full_input = torch.cat([pre_cache, audio_signal], dim=2) + full_length = audio_length + self.pre_cache_size + + # 2. Extract NEW pre_cache (last N frames of full_input) + # Note: We do this BEFORE processing because we want the raw audio context + new_pre_cache = full_input[:, :, -self.pre_cache_size:] + + # 3. Process with Encoder + # Reconstruct NeMo cache object + current_cache = [cache_last_channel, cache_last_time, cache_last_channel_len] + + encoded, encoded_len, new_cache_channel, new_cache_time, new_cache_len = self.encoder.cache_aware_stream_step( + processed_signal=full_input, + processed_signal_length=full_length, + cache_last_channel=cache_last_channel, + cache_last_time=cache_last_time, + cache_last_channel_len=cache_last_channel_len + ) + + # 4. Drop the first few frames corresponding to pre_cache? + # NeMo's cache_aware_stream_step usually handles the "valid" output frames. + # But since we manually prepended, we might get extra output frames. + # However, for streaming, we usually want the model to see the context but only output the new tokens. + # Let's trust NeMo's streaming logic for now, or check if we need to slice. + # Given we are using 'cache_aware_stream_step', it expects the full context window? + # Actually, standard usage is: input IS the new chunk, but internal convolution looks at past. + # But since we are stateless, we MUST provide the past. + # So passing (pre_cache + chunk) is correct. + + # Cast lengths to Int32 for CoreML + encoded_len_32 = encoded_len.to(dtype=torch.int32) + new_channel_len_32 = new_cache_len.to(dtype=torch.int32) + + return encoded, encoded_len_32, new_pre_cache, new_cache_channel, new_cache_time, new_channel_len_32 + +def _coreml_convert( + traced_model, + inputs, + outputs, + compute_units=ct.ComputeUnit.CPU_ONLY +): + return ct.convert( + traced_model, + inputs=inputs, + outputs=outputs, + compute_units=compute_units, + minimum_deployment_target=ct.target.macOS14, + ) + +def main(): + model_id: str = "nvidia/parakeet_realtime_eou_120m-v1" + output_dir: str = "temp_swift_models/StreamingLoopback" + output_path = Path(output_dir) + output_path.mkdir(parents=True, exist_ok=True) + + print(f"Loading model: {model_id}...") + asr_model = EncDecRNNTBPEModel.from_pretrained(model_name=model_id) + asr_model.eval() + + parser = argparse.ArgumentParser() + parser.add_argument("--chunk-frames", type=int, default=17, help="Number of frames in the input chunk (e.g. 17 for 160ms, 129 for 1.28s)") + args = parser.parse_args() + + encoder = asr_model.encoder + + # --- Configuration --- + # 160ms chunk = 16 frames (but preprocessor produces 17 with padding/centering) + # 1.28s chunk = 128 frames (preprocessor produces 129) + chunk_size_in = args.chunk_frames + mel_dim = 128 + hidden_dim = encoder.d_model # 512 + num_layers = len(encoder.layers) # 17 + + # Cache sizes + cache_channel_size = 70 + cache_time_size = 8 + pre_cache_size = 16 + + print(f"Config: Chunk={chunk_size_in}, Mel={mel_dim}, Hidden={hidden_dim}, Layers={num_layers}") + print(f"Cache: Channel={cache_channel_size}, Time={cache_time_size}, Pre={pre_cache_size}") + + # --- Wrapper --- + wrapper = LoopbackEncoderWrapper(encoder, pre_cache_size=pre_cache_size) + wrapper.eval() + + # --- Test Inputs (for Tracing) --- + batch_size = 1 + test_mel = torch.randn(batch_size, mel_dim, chunk_size_in) + test_mel_len = torch.tensor([chunk_size_in], dtype=torch.int32) + test_pre_cache = torch.zeros(batch_size, mel_dim, pre_cache_size) + + # Initial Cache (Zeros) + test_cache_channel = torch.zeros(num_layers, batch_size, cache_channel_size, hidden_dim) + test_cache_time = torch.zeros(num_layers, batch_size, hidden_dim, cache_time_size) + test_cache_len = torch.zeros(batch_size, dtype=torch.int32) + + print("Tracing model...") + traced_model = torch.jit.trace( + wrapper, + (test_mel, test_mel_len, test_pre_cache, test_cache_channel, test_cache_time, test_cache_len), + strict=False + ) + + # --- CoreML Conversion --- + print("Converting to CoreML...") + + inputs = [ + ct.TensorType(name="audio_signal", shape=(1, 128, chunk_size_in), dtype=np.float32), + ct.TensorType(name="audio_length", shape=(1,), dtype=np.int32), + ct.TensorType(name="pre_cache", shape=(1, 128, pre_cache_size), dtype=np.float32), + ct.TensorType(name="cache_last_channel", shape=(num_layers, 1, cache_channel_size, hidden_dim), dtype=np.float32), + ct.TensorType(name="cache_last_time", shape=(num_layers, 1, hidden_dim, cache_time_size), dtype=np.float32), + ct.TensorType(name="cache_last_channel_len", shape=(1,), dtype=np.int32), + ] + + outputs = [ + ct.TensorType(name="encoded_output", dtype=np.float32), + ct.TensorType(name="encoded_length", dtype=np.int32), + ct.TensorType(name="new_pre_cache", dtype=np.float32), + ct.TensorType(name="new_cache_last_channel", dtype=np.float32), + ct.TensorType(name="new_cache_last_time", dtype=np.float32), + ct.TensorType(name="new_cache_last_channel_len", dtype=np.int32), + ] + + mlmodel = _coreml_convert(traced_model, inputs, outputs) + + save_path = output_path / "streaming_encoder.mlpackage" + mlmodel.save(str(save_path)) + print(f"Saved: {save_path}") + + # Also export Preprocessor, Decoder, Joint for completeness? + # For now, let's assume we reuse the existing ones or export them separately if needed. + # But the user asked specifically for the Encoder loopback. + +if __name__ == "__main__": + main() diff --git a/1280ms/decoder.mlmodelc/analytics/coremldata.bin b/1280ms/decoder.mlmodelc/analytics/coremldata.bin new file mode 100644 index 0000000000000000000000000000000000000000..c6f51dedd2638184c5a2d71512339ceea086abc0 --- /dev/null +++ b/1280ms/decoder.mlmodelc/analytics/coremldata.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3996975a8cbc1949159c55605b3132b39b2484f51acbd55d796d93c70de02b49 +size 243 diff --git a/1280ms/decoder.mlmodelc/coremldata.bin b/1280ms/decoder.mlmodelc/coremldata.bin new file mode 100644 index 0000000000000000000000000000000000000000..71936e61ffe7697e4af55bae8870d95c1fbb66c5 --- /dev/null +++ b/1280ms/decoder.mlmodelc/coremldata.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3ccbff963d8cf07e2be2bd56ea3384a89ea49628922c6bd95ff62e2ae57dc34 +size 497 diff --git a/1280ms/decoder.mlmodelc/metadata.json b/1280ms/decoder.mlmodelc/metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..3581673bcb062f7e0e1e0efa9f59edc5f218d50e --- /dev/null +++ b/1280ms/decoder.mlmodelc/metadata.json @@ -0,0 +1,118 @@ +[ + { + "metadataOutputVersion" : "3.0", + "shortDescription" : "Parakeet EOU decoder (RNNT prediction network)", + "outputSchema" : [ + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 640 × 1)", + "shortDescription" : "", + "shape" : "[1, 640, 1]", + "name" : "decoder", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 1 × 640)", + "shortDescription" : "", + "shape" : "[1, 1, 640]", + "name" : "h_out", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 1 × 640)", + "shortDescription" : "", + "shape" : "[1, 1, 640]", + "name" : "c_out", + "type" : "MultiArray" + } + ], + "storagePrecision" : "Float16", + "modelParameters" : [ + + ], + "author" : "Fluid Inference", + "specificationVersion" : 8, + "mlProgramOperationTypeHistogram" : { + "Ios17.squeeze" : 2, + "Ios17.gather" : 1, + "Ios17.cast" : 6, + "Ios17.lstm" : 1, + "Ios17.transpose" : 2, + "Identity" : 1, + "Ios17.expandDims" : 2 + }, + "computePrecision" : "Mixed (Float16, Float32, Int16, Int32)", + "isUpdatable" : "0", + "stateSchema" : [ + + ], + "availability" : { + "macOS" : "14.0", + "tvOS" : "17.0", + "visionOS" : "1.0", + "watchOS" : "10.0", + "iOS" : "17.0", + "macCatalyst" : "17.0" + }, + "modelType" : { + "name" : "MLModelType_mlProgram" + }, + "inputSchema" : [ + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1 × 1)", + "shortDescription" : "", + "shape" : "[1, 1]", + "name" : "targets", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1)", + "shortDescription" : "", + "shape" : "[1]", + "name" : "target_length", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 1 × 640)", + "shortDescription" : "", + "shape" : "[1, 1, 640]", + "name" : "h_in", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 1 × 640)", + "shortDescription" : "", + "shape" : "[1, 1, 640]", + "name" : "c_in", + "type" : "MultiArray" + } + ], + "userDefinedMetadata" : { + "com.github.apple.coremltools.version" : "8.3.0", + "com.github.apple.coremltools.source" : "torch==2.4.0", + "com.github.apple.coremltools.source_dialect" : "TorchScript" + }, + "generatedClassName" : "parakeet_eou_decoder", + "method" : "predict" + } +] \ No newline at end of file diff --git a/1280ms/decoder.mlmodelc/model.mil b/1280ms/decoder.mlmodelc/model.mil new file mode 100644 index 0000000000000000000000000000000000000000..5ead417fbdc13f8bf0c05a26039b2828503d4eca --- /dev/null +++ b/1280ms/decoder.mlmodelc/model.mil @@ -0,0 +1,45 @@ +program(1.0) +[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.4.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] +{ + func main(tensor c_in, tensor h_in, tensor target_length, tensor targets) { + tensor y_axis_0 = const()[name = tensor("y_axis_0"), val = tensor(0)]; + tensor y_batch_dims_0 = const()[name = tensor("y_batch_dims_0"), val = tensor(0)]; + tensor y_validate_indices_0 = const()[name = tensor("y_validate_indices_0"), val = tensor(false)]; + tensor module_prediction_embed_weight_to_fp16 = const()[name = tensor("module_prediction_embed_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; + tensor targets_to_int16_dtype_0 = const()[name = tensor("targets_to_int16_dtype_0"), val = tensor("int16")]; + tensor targets_to_int16 = cast(dtype = targets_to_int16_dtype_0, x = targets)[name = tensor("cast_8")]; + tensor y_cast_fp16_cast_uint16 = gather(axis = y_axis_0, batch_dims = y_batch_dims_0, indices = targets_to_int16, validate_indices = y_validate_indices_0, x = module_prediction_embed_weight_to_fp16)[name = tensor("y_cast_fp16_cast_uint16")]; + tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([1, 0, 2])]; + tensor input_lstm_h0_squeeze_axes_0 = const()[name = tensor("input_lstm_h0_squeeze_axes_0"), val = tensor([0])]; + tensor h_in_to_fp16_dtype_0 = const()[name = tensor("h_in_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor h_in_to_fp16 = cast(dtype = h_in_to_fp16_dtype_0, x = h_in)[name = tensor("cast_7")]; + tensor input_lstm_h0_squeeze_cast_fp16 = squeeze(axes = input_lstm_h0_squeeze_axes_0, x = h_in_to_fp16)[name = tensor("input_lstm_h0_squeeze_cast_fp16")]; + tensor input_lstm_c0_squeeze_axes_0 = const()[name = tensor("input_lstm_c0_squeeze_axes_0"), val = tensor([0])]; + tensor c_in_to_fp16_dtype_0 = const()[name = tensor("c_in_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor c_in_to_fp16 = cast(dtype = c_in_to_fp16_dtype_0, x = c_in)[name = tensor("cast_6")]; + tensor input_lstm_c0_squeeze_cast_fp16 = squeeze(axes = input_lstm_c0_squeeze_axes_0, x = c_in_to_fp16)[name = tensor("input_lstm_c0_squeeze_cast_fp16")]; + tensor input_direction_0 = const()[name = tensor("input_direction_0"), val = tensor("forward")]; + tensor input_output_sequence_0 = const()[name = tensor("input_output_sequence_0"), val = tensor(true)]; + tensor input_recurrent_activation_0 = const()[name = tensor("input_recurrent_activation_0"), val = tensor("sigmoid")]; + tensor input_cell_activation_0 = const()[name = tensor("input_cell_activation_0"), val = tensor("tanh")]; + tensor input_activation_0 = const()[name = tensor("input_activation_0"), val = tensor("tanh")]; + tensor concat_1_to_fp16 = const()[name = tensor("concat_1_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1314688)))]; + tensor concat_2_to_fp16 = const()[name = tensor("concat_2_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4591552)))]; + tensor concat_0_to_fp16 = const()[name = tensor("concat_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7868416)))]; + tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = y_cast_fp16_cast_uint16)[name = tensor("transpose_2")]; + tensor input_cast_fp16_0, tensor input_cast_fp16_1, tensor input_cast_fp16_2 = lstm(activation = input_activation_0, bias = concat_0_to_fp16, cell_activation = input_cell_activation_0, direction = input_direction_0, initial_c = input_lstm_c0_squeeze_cast_fp16, initial_h = input_lstm_h0_squeeze_cast_fp16, output_sequence = input_output_sequence_0, recurrent_activation = input_recurrent_activation_0, weight_hh = concat_2_to_fp16, weight_ih = concat_1_to_fp16, x = input_3_cast_fp16)[name = tensor("input_cast_fp16")]; + tensor obj_3_axes_0 = const()[name = tensor("obj_3_axes_0"), val = tensor([0])]; + tensor obj_3_cast_fp16 = expand_dims(axes = obj_3_axes_0, x = input_cast_fp16_1)[name = tensor("obj_3_cast_fp16")]; + tensor obj_3_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_3_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor obj_axes_0 = const()[name = tensor("obj_axes_0"), val = tensor([0])]; + tensor obj_cast_fp16 = expand_dims(axes = obj_axes_0, x = input_cast_fp16_2)[name = tensor("obj_cast_fp16")]; + tensor obj_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor transpose_0_perm_0 = const()[name = tensor("transpose_0_perm_0"), val = tensor([1, 2, 0])]; + tensor transpose_0_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("transpose_0_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor transpose_0_cast_fp16 = transpose(perm = transpose_0_perm_0, x = input_cast_fp16_0)[name = tensor("transpose_1")]; + tensor decoder = cast(dtype = transpose_0_cast_fp16_to_fp32_dtype_0, x = transpose_0_cast_fp16)[name = tensor("cast_3")]; + tensor c_out = cast(dtype = obj_cast_fp16_to_fp32_dtype_0, x = obj_cast_fp16)[name = tensor("cast_4")]; + tensor h_out = cast(dtype = obj_3_cast_fp16_to_fp32_dtype_0, x = obj_3_cast_fp16)[name = tensor("cast_5")]; + tensor target_length_tmp = identity(x = target_length)[name = tensor("target_length_tmp")]; + } -> (decoder, h_out, c_out); +} \ No newline at end of file diff --git a/1280ms/decoder.mlmodelc/weights/weight.bin b/1280ms/decoder.mlmodelc/weights/weight.bin new file mode 100644 index 0000000000000000000000000000000000000000..cde1618cbed235b421984a302289e3bdd7e3df02 --- /dev/null +++ b/1280ms/decoder.mlmodelc/weights/weight.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b4cacecdcd9df79ab1e56de67230baf5a8664d2afe0bb8f3408eefa972cb2f4 +size 7873600 diff --git a/1280ms/individual_components.py b/1280ms/individual_components.py new file mode 100644 index 0000000000000000000000000000000000000000..47271397bc8d9d17cc0fabcf4bb63be7e7c2109c --- /dev/null +++ b/1280ms/individual_components.py @@ -0,0 +1,250 @@ +#!/usr/bin/env python3 +"""Export Parakeet Realtime EOU 120M RNNT components into CoreML.""" +from __future__ import annotations + +from dataclasses import dataclass +from pathlib import Path +from typing import Optional, Tuple + +import coremltools as ct +import torch + + +@dataclass +class ExportSettings: + output_dir: Path + compute_units: ct.ComputeUnit + deployment_target: Optional[ct.target] + compute_precision: Optional[ct.precision] + max_audio_seconds: float + max_symbol_steps: int + + +class PreprocessorWrapper(torch.nn.Module): + """Wrapper for the audio preprocessor (mel spectrogram extraction).""" + + def __init__(self, module: torch.nn.Module) -> None: + super().__init__() + self.module = module + + def forward( + self, audio_signal: torch.Tensor, length: torch.Tensor + ) -> Tuple[torch.Tensor, torch.Tensor]: + mel, mel_length = self.module( + input_signal=audio_signal, length=length.to(dtype=torch.long) + ) + return mel, mel_length + + +class EncoderWrapper(torch.nn.Module): + """Wrapper for the cache-aware FastConformer encoder. + + Note: For the realtime EOU model, the encoder is cache-aware which means + it can operate in a streaming fashion. For CoreML export, we export + without cache state for simplicity (full-context mode). + """ + + def __init__(self, module: torch.nn.Module) -> None: + super().__init__() + self.module = module + + def forward( + self, features: torch.Tensor, length: torch.Tensor + ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: + encoded, encoded_lengths = self.module( + audio_signal=features, length=length.to(dtype=torch.long) + ) + # Synthesize per-frame timestamps (seconds) using the 80 ms encoder stride. + # Shape: [B, T_enc] + frame_times = ( + torch.arange(encoded.shape[-1], device=encoded.device, dtype=torch.float32) + * 0.08 + ) + return encoded, encoded_lengths, frame_times + + +class DecoderWrapper(torch.nn.Module): + """Wrapper for the RNNT prediction network (decoder).""" + + def __init__(self, module: torch.nn.Module) -> None: + super().__init__() + self.module = module + + def forward( + self, + targets: torch.Tensor, + target_lengths: torch.Tensor, + h_in: torch.Tensor, + c_in: torch.Tensor, + ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: + state = [h_in, c_in] + decoder_output, _, new_state = self.module( + targets=targets.to(dtype=torch.long), + target_length=target_lengths.to(dtype=torch.long), + states=state, + ) + return decoder_output, new_state[0], new_state[1] + + +class JointWrapper(torch.nn.Module): + """Wrapper for the RNNT joint network. + + Note: Unlike Parakeet TDT v3, the realtime EOU model does NOT have + duration outputs (num_extra_outputs). The joint network outputs only + token logits over the vocabulary + blank. + """ + + def __init__(self, module: torch.nn.Module) -> None: + super().__init__() + self.module = module + + def forward( + self, encoder_outputs: torch.Tensor, decoder_outputs: torch.Tensor + ) -> torch.Tensor: + # Input: encoder_outputs [B, D, T], decoder_outputs [B, D, U] + # Transpose to match what projection layers expect + encoder_outputs = encoder_outputs.transpose(1, 2) # [B, T, D] + decoder_outputs = decoder_outputs.transpose(1, 2) # [B, U, D] + + # Apply projections + enc_proj = self.module.enc(encoder_outputs) # [B, T, joint_hidden] + dec_proj = self.module.pred(decoder_outputs) # [B, U, joint_hidden] + + # Explicit broadcasting along T and U to avoid converter ambiguity + x = enc_proj.unsqueeze(2) + dec_proj.unsqueeze(1) # [B, T, U, joint_hidden] + x = self.module.joint_net[0](x) # ReLU + x = self.module.joint_net[1](x) # Dropout (no-op in eval) + out = self.module.joint_net[2](x) # Linear -> logits [B, T, U, vocab+blank] + return out + + +class MelEncoderWrapper(torch.nn.Module): + """Fused wrapper: waveform -> mel -> encoder. + + Inputs: + - audio_signal: [B, S] + - audio_length: [B] + + Outputs: + - encoder: [B, D, T_enc] + - encoder_length: [B] + - frame_times: [T_enc] + """ + + def __init__( + self, preprocessor: PreprocessorWrapper, encoder: EncoderWrapper + ) -> None: + super().__init__() + self.preprocessor = preprocessor + self.encoder = encoder + + def forward( + self, audio_signal: torch.Tensor, audio_length: torch.Tensor + ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: + mel, mel_length = self.preprocessor(audio_signal, audio_length) + encoded, enc_len, frame_times = self.encoder(mel, mel_length.to(dtype=torch.int32)) + return encoded, enc_len, frame_times + + +class JointDecisionWrapper(torch.nn.Module): + """Joint + decision head: outputs label id and label prob. + + Unlike Parakeet TDT v3, this model does NOT have duration outputs. + + Inputs: + - encoder_outputs: [B, D, T] + - decoder_outputs: [B, D, U] + + Returns: + - token_id: [B, T, U] int32 + - token_prob: [B, T, U] float32 + """ + + def __init__(self, joint: JointWrapper, vocab_size: int) -> None: + super().__init__() + self.joint = joint + self.vocab_with_blank = int(vocab_size) + 1 + + def forward(self, encoder_outputs: torch.Tensor, decoder_outputs: torch.Tensor): + logits = self.joint(encoder_outputs, decoder_outputs) + + # Token selection + token_ids = torch.argmax(logits, dim=-1).to(dtype=torch.int32) + token_probs_all = torch.softmax(logits, dim=-1) + # gather expects int64 (long) indices; cast only for gather + token_prob = torch.gather( + token_probs_all, dim=-1, index=token_ids.long().unsqueeze(-1) + ).squeeze(-1) + + return token_ids, token_prob + + +class JointDecisionSingleStep(torch.nn.Module): + """Single-step variant for streaming: encoder_step -> token decision. + + Inputs: + - encoder_step: [B=1, D, T=1] + - decoder_step: [B=1, D, U=1] + + Returns: + - token_id: [1, 1, 1] int32 + - token_prob: [1, 1, 1] float32 + - top_k_ids: [1, 1, 1, K] int32 + - top_k_logits: [1, 1, 1, K] float32 + """ + + def __init__(self, joint: JointWrapper, vocab_size: int, top_k: int = 64) -> None: + super().__init__() + self.joint = joint + self.vocab_with_blank = int(vocab_size) + 1 + self.top_k = int(top_k) + + def forward(self, encoder_step: torch.Tensor, decoder_step: torch.Tensor): + # Reuse JointWrapper which expects [B, D, T] and [B, D, U] + logits = self.joint(encoder_step, decoder_step) # [1, 1, 1, V+blank] + + token_ids = torch.argmax(logits, dim=-1, keepdim=False).to(dtype=torch.int32) + token_probs_all = torch.softmax(logits, dim=-1) + token_prob = torch.gather( + token_probs_all, dim=-1, index=token_ids.long().unsqueeze(-1) + ).squeeze(-1) + + # Also expose top-K candidates for host-side processing + topk_logits, topk_ids_long = torch.topk( + logits, k=min(self.top_k, logits.shape[-1]), dim=-1 + ) + topk_ids = topk_ids_long.to(dtype=torch.int32) + return token_ids, token_prob, topk_ids, topk_logits + + +def _coreml_convert( + traced: torch.jit.ScriptModule, + inputs, + outputs, + settings: ExportSettings, + compute_units_override: Optional[ct.ComputeUnit] = None, + compute_precision: Optional[ct.precision] = None, +) -> ct.models.MLModel: + cu = ( + compute_units_override + if compute_units_override is not None + else settings.compute_units + ) + kwargs = { + "convert_to": "mlprogram", + "inputs": inputs, + "outputs": outputs, + "compute_units": cu, + } + print("Converting:", traced.__class__.__name__) + print("Conversion kwargs:", kwargs) + if settings.deployment_target is not None: + kwargs["minimum_deployment_target"] = settings.deployment_target + + # Priority: explicit argument > settings + if compute_precision is not None: + kwargs["compute_precision"] = compute_precision + elif settings.compute_precision is not None: + kwargs["compute_precision"] = settings.compute_precision + + return ct.convert(traced, **kwargs) diff --git a/1280ms/joint_decision.mlmodelc/analytics/coremldata.bin b/1280ms/joint_decision.mlmodelc/analytics/coremldata.bin new file mode 100644 index 0000000000000000000000000000000000000000..ff9929e12bae25c509655bda0ebda6b9fe13fade --- /dev/null +++ b/1280ms/joint_decision.mlmodelc/analytics/coremldata.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5bca32ad130dcad6605cc00044c752aa5b45ef57d14c17f2d1a2fa49d6cf55b5 +size 243 diff --git a/1280ms/joint_decision.mlmodelc/coremldata.bin b/1280ms/joint_decision.mlmodelc/coremldata.bin new file mode 100644 index 0000000000000000000000000000000000000000..1aa5afe7c02e93f9baba2757cad5c21f563157bf --- /dev/null +++ b/1280ms/joint_decision.mlmodelc/coremldata.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22d4abc4625b935ee035b5f8ce7cb28d1041b9b01c12173e287bf4b5f5d99625 +size 493 diff --git a/1280ms/joint_decision.mlmodelc/metadata.json b/1280ms/joint_decision.mlmodelc/metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..a0a319857f429f5ec4b106397cd52ec354ef2e96 --- /dev/null +++ b/1280ms/joint_decision.mlmodelc/metadata.json @@ -0,0 +1,112 @@ +[ + { + "metadataOutputVersion" : "3.0", + "shortDescription" : "Parakeet EOU single-step joint decision", + "outputSchema" : [ + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1 × 1 × 1)", + "shortDescription" : "", + "shape" : "[1, 1, 1]", + "name" : "token_id", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 1 × 1)", + "shortDescription" : "", + "shape" : "[1, 1, 1]", + "name" : "token_prob", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1 × 1 × 1 × 64)", + "shortDescription" : "", + "shape" : "[1, 1, 1, 64]", + "name" : "top_k_ids", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 1 × 1 × 64)", + "shortDescription" : "", + "shape" : "[1, 1, 1, 64]", + "name" : "top_k_logits", + "type" : "MultiArray" + } + ], + "storagePrecision" : "Float16", + "modelParameters" : [ + + ], + "author" : "Fluid Inference", + "specificationVersion" : 8, + "mlProgramOperationTypeHistogram" : { + "Ios17.reduceArgmax" : 1, + "Ios17.squeeze" : 1, + "Ios17.cast" : 6, + "Ios17.linear" : 3, + "Ios17.transpose" : 2, + "Ios17.add" : 1, + "Ios16.relu" : 1, + "Ios16.softmax" : 1, + "Ios17.gatherAlongAxis" : 1, + "Ios17.topk" : 1, + "Ios17.expandDims" : 3 + }, + "computePrecision" : "Mixed (Float16, Float32, Int16, Int32, UInt16)", + "isUpdatable" : "0", + "stateSchema" : [ + + ], + "availability" : { + "macOS" : "14.0", + "tvOS" : "17.0", + "visionOS" : "1.0", + "watchOS" : "10.0", + "iOS" : "17.0", + "macCatalyst" : "17.0" + }, + "modelType" : { + "name" : "MLModelType_mlProgram" + }, + "inputSchema" : [ + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 512 × 1)", + "shortDescription" : "", + "shape" : "[1, 512, 1]", + "name" : "encoder_step", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 640 × 1)", + "shortDescription" : "", + "shape" : "[1, 640, 1]", + "name" : "decoder_step", + "type" : "MultiArray" + } + ], + "userDefinedMetadata" : { + "com.github.apple.coremltools.source_dialect" : "TorchScript", + "com.github.apple.coremltools.version" : "8.3.0", + "com.github.apple.coremltools.source" : "torch==2.4.0" + }, + "generatedClassName" : "parakeet_eou_joint_decision_single_step", + "method" : "predict" + } +] \ No newline at end of file diff --git a/1280ms/joint_decision.mlmodelc/model.mil b/1280ms/joint_decision.mlmodelc/model.mil new file mode 100644 index 0000000000000000000000000000000000000000..172a7579866c84407b32b5f746b7f1ae132599d8 --- /dev/null +++ b/1280ms/joint_decision.mlmodelc/model.mil @@ -0,0 +1,57 @@ +program(1.0) +[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.4.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] +{ + func main(tensor decoder_step, tensor encoder_step) { + tensor input_1_perm_0 = const()[name = tensor("input_1_perm_0"), val = tensor([0, 2, 1])]; + tensor encoder_step_to_fp16_dtype_0 = const()[name = tensor("encoder_step_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([0, 2, 1])]; + tensor decoder_step_to_fp16_dtype_0 = const()[name = tensor("decoder_step_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor joint_module_enc_weight_to_fp16 = const()[name = tensor("joint_module_enc_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; + tensor joint_module_enc_bias_to_fp16 = const()[name = tensor("joint_module_enc_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(655488)))]; + tensor encoder_step_to_fp16 = cast(dtype = encoder_step_to_fp16_dtype_0, x = encoder_step)[name = tensor("cast_8")]; + tensor input_1_cast_fp16 = transpose(perm = input_1_perm_0, x = encoder_step_to_fp16)[name = tensor("transpose_1")]; + tensor linear_0_cast_fp16 = linear(bias = joint_module_enc_bias_to_fp16, weight = joint_module_enc_weight_to_fp16, x = input_1_cast_fp16)[name = tensor("linear_0_cast_fp16")]; + tensor joint_module_pred_weight_to_fp16 = const()[name = tensor("joint_module_pred_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(656832)))]; + tensor joint_module_pred_bias_to_fp16 = const()[name = tensor("joint_module_pred_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1476096)))]; + tensor decoder_step_to_fp16 = cast(dtype = decoder_step_to_fp16_dtype_0, x = decoder_step)[name = tensor("cast_7")]; + tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = decoder_step_to_fp16)[name = tensor("transpose_0")]; + tensor linear_1_cast_fp16 = linear(bias = joint_module_pred_bias_to_fp16, weight = joint_module_pred_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("linear_1_cast_fp16")]; + tensor var_23_axes_0 = const()[name = tensor("op_23_axes_0"), val = tensor([2])]; + tensor var_23_cast_fp16 = expand_dims(axes = var_23_axes_0, x = linear_0_cast_fp16)[name = tensor("op_23_cast_fp16")]; + tensor var_24_axes_0 = const()[name = tensor("op_24_axes_0"), val = tensor([1])]; + tensor var_24_cast_fp16 = expand_dims(axes = var_24_axes_0, x = linear_1_cast_fp16)[name = tensor("op_24_cast_fp16")]; + tensor input_5_cast_fp16 = add(x = var_23_cast_fp16, y = var_24_cast_fp16)[name = tensor("input_5_cast_fp16")]; + tensor input_7_cast_fp16 = relu(x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; + tensor joint_module_joint_net_2_weight_to_fp16 = const()[name = tensor("joint_module_joint_net_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1477440)))]; + tensor joint_module_joint_net_2_bias_to_fp16 = const()[name = tensor("joint_module_joint_net_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2792064)))]; + tensor linear_2_cast_fp16 = linear(bias = joint_module_joint_net_2_bias_to_fp16, weight = joint_module_joint_net_2_weight_to_fp16, x = input_7_cast_fp16)[name = tensor("linear_2_cast_fp16")]; + tensor var_38_axis_0 = const()[name = tensor("op_38_axis_0"), val = tensor(-1)]; + tensor var_38_keep_dims_0 = const()[name = tensor("op_38_keep_dims_0"), val = tensor(false)]; + tensor var_38_output_dtype_0 = const()[name = tensor("op_38_output_dtype_0"), val = tensor("int32")]; + tensor token_id = reduce_argmax(axis = var_38_axis_0, keep_dims = var_38_keep_dims_0, output_dtype = var_38_output_dtype_0, x = linear_2_cast_fp16)[name = tensor("op_38_cast_fp16")]; + tensor var_44 = const()[name = tensor("op_44"), val = tensor(-1)]; + tensor token_probs_all_cast_fp16 = softmax(axis = var_44, x = linear_2_cast_fp16)[name = tensor("token_probs_all_cast_fp16")]; + tensor var_53_axes_0 = const()[name = tensor("op_53_axes_0"), val = tensor([-1])]; + tensor var_53 = expand_dims(axes = var_53_axes_0, x = token_id)[name = tensor("op_53")]; + tensor var_54 = const()[name = tensor("op_54"), val = tensor(-1)]; + tensor var_56_validate_indices_0 = const()[name = tensor("op_56_validate_indices_0"), val = tensor(false)]; + tensor var_53_to_int16_dtype_0 = const()[name = tensor("op_53_to_int16_dtype_0"), val = tensor("int16")]; + tensor var_53_to_int16 = cast(dtype = var_53_to_int16_dtype_0, x = var_53)[name = tensor("cast_6")]; + tensor var_56_cast_fp16_cast_int16 = gather_along_axis(axis = var_54, indices = var_53_to_int16, validate_indices = var_56_validate_indices_0, x = token_probs_all_cast_fp16)[name = tensor("op_56_cast_fp16_cast_int16")]; + tensor var_58_axes_0 = const()[name = tensor("op_58_axes_0"), val = tensor([-1])]; + tensor var_58_cast_fp16 = squeeze(axes = var_58_axes_0, x = var_56_cast_fp16_cast_int16)[name = tensor("op_58_cast_fp16")]; + tensor var_58_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_58_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor var_59 = const()[name = tensor("op_59"), val = tensor(64)]; + tensor var_63_axis_0 = const()[name = tensor("op_63_axis_0"), val = tensor(-1)]; + tensor var_63_ascending_0 = const()[name = tensor("op_63_ascending_0"), val = tensor(false)]; + tensor var_63_sort_0 = const()[name = tensor("op_63_sort_0"), val = tensor(true)]; + tensor var_63_return_indices_0 = const()[name = tensor("op_63_return_indices_0"), val = tensor(true)]; + tensor var_63_cast_fp16_cast_int16_output_indices_dtype_0 = const()[name = tensor("op_63_cast_fp16_cast_int16_output_indices_dtype_0"), val = tensor("uint16")]; + tensor var_63_cast_fp16_cast_int16_0, tensor var_63_cast_fp16_cast_int16_1 = topk(ascending = var_63_ascending_0, axis = var_63_axis_0, k = var_59, output_indices_dtype = var_63_cast_fp16_cast_int16_output_indices_dtype_0, return_indices = var_63_return_indices_0, sort = var_63_sort_0, x = linear_2_cast_fp16)[name = tensor("op_63_cast_fp16_cast_int16")]; + tensor var_63_cast_fp16_cast_int16_1_to_int32_dtype_0 = const()[name = tensor("op_63_cast_fp16_cast_int16_1_to_int32_dtype_0"), val = tensor("int32")]; + tensor var_63_cast_fp16_0_to_fp32_dtype_0 = const()[name = tensor("op_63_cast_fp16_0_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor top_k_logits = cast(dtype = var_63_cast_fp16_0_to_fp32_dtype_0, x = var_63_cast_fp16_cast_int16_0)[name = tensor("cast_3")]; + tensor top_k_ids = cast(dtype = var_63_cast_fp16_cast_int16_1_to_int32_dtype_0, x = var_63_cast_fp16_cast_int16_1)[name = tensor("cast_4")]; + tensor token_prob = cast(dtype = var_58_cast_fp16_to_fp32_dtype_0, x = var_58_cast_fp16)[name = tensor("cast_5")]; + } -> (token_id, token_prob, top_k_ids, top_k_logits); +} \ No newline at end of file diff --git a/1280ms/joint_decision.mlmodelc/weights/weight.bin b/1280ms/joint_decision.mlmodelc/weights/weight.bin new file mode 100644 index 0000000000000000000000000000000000000000..ec3d2c1dea18b332fa1a3cd8d31981bae4e3f649 --- /dev/null +++ b/1280ms/joint_decision.mlmodelc/weights/weight.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7039b2010a269153f5a96edf28637f921a86ef8822f248f2d6712f7a6bce84b4 +size 2794182 diff --git a/1280ms/parakeet_eou_preprocessor.mlmodelc/analytics/coremldata.bin b/1280ms/parakeet_eou_preprocessor.mlmodelc/analytics/coremldata.bin new file mode 100644 index 0000000000000000000000000000000000000000..ca93efecf9fbf12e37943f7e192cd324a39e9f94 --- /dev/null +++ b/1280ms/parakeet_eou_preprocessor.mlmodelc/analytics/coremldata.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4ada8b0b99ac1d2ba7acbffacfbbf1a06cb69d30e9410d237ee0aa4c2b0ad63 +size 243 diff --git a/1280ms/parakeet_eou_preprocessor.mlmodelc/coremldata.bin b/1280ms/parakeet_eou_preprocessor.mlmodelc/coremldata.bin new file mode 100644 index 0000000000000000000000000000000000000000..ca69399b73b039b35bb1b4967b25b7a7a12e8178 --- /dev/null +++ b/1280ms/parakeet_eou_preprocessor.mlmodelc/coremldata.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc7252fa47622fe39577361233627062019a3bb740fdbb5366a7bae09df0ec5e +size 422 diff --git a/1280ms/parakeet_eou_preprocessor.mlmodelc/metadata.json b/1280ms/parakeet_eou_preprocessor.mlmodelc/metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..98b70ffcae507184b68f87d7adaad706e5c68c9a --- /dev/null +++ b/1280ms/parakeet_eou_preprocessor.mlmodelc/metadata.json @@ -0,0 +1,105 @@ +[ + { + "metadataOutputVersion" : "3.0", + "shortDescription" : "Parakeet EOU preprocessor", + "outputSchema" : [ + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32)", + "shortDescription" : "", + "shape" : "[]", + "name" : "mel", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1)", + "shortDescription" : "", + "shape" : "[1]", + "name" : "mel_length", + "type" : "MultiArray" + } + ], + "storagePrecision" : "Float32", + "modelParameters" : [ + + ], + "author" : "Fluid Inference", + "specificationVersion" : 8, + "mlProgramOperationTypeHistogram" : { + "Range1d" : 1, + "Ios17.reshape" : 2, + "Identity" : 1, + "Ios17.matmul" : 1, + "Ios17.expandDims" : 5, + "Select" : 1, + "Ios17.add" : 3, + "Ios17.sliceByIndex" : 3, + "Ios16.reduceSum" : 1, + "Shape" : 1, + "Ios17.gather" : 1, + "Pad" : 1, + "Ios17.log" : 1, + "Ios17.conv" : 2, + "Ios17.sub" : 2, + "Ios17.pow" : 1, + "Ios17.cast" : 2, + "Stack" : 1, + "Ios17.concat" : 1, + "Ios17.floorDiv" : 1, + "Ios17.greaterEqual" : 1, + "Ios17.mul" : 1 + }, + "computePrecision" : "Mixed (Float32, Int32)", + "isUpdatable" : "0", + "stateSchema" : [ + + ], + "availability" : { + "macOS" : "14.0", + "tvOS" : "17.0", + "visionOS" : "1.0", + "watchOS" : "10.0", + "iOS" : "17.0", + "macCatalyst" : "17.0" + }, + "modelType" : { + "name" : "MLModelType_mlProgram" + }, + "inputSchema" : [ + { + "dataType" : "Float32", + "hasShapeFlexibility" : "1", + "isOptional" : "0", + "shapeFlexibility" : "1 × 1...32000", + "shapeRange" : "[[1, 1], [1, 32000]]", + "formattedType" : "MultiArray (Float32 1 × 1)", + "type" : "MultiArray", + "shape" : "[1, 1]", + "name" : "audio_signal", + "shortDescription" : "" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1)", + "shortDescription" : "", + "shape" : "[1]", + "name" : "audio_length", + "type" : "MultiArray" + } + ], + "userDefinedMetadata" : { + "com.github.apple.coremltools.source_dialect" : "TorchScript", + "com.github.apple.coremltools.source" : "torch==2.4.0", + "com.github.apple.coremltools.version" : "8.3.0" + }, + "generatedClassName" : "parakeet_eou_preprocessor", + "method" : "predict" + } +] \ No newline at end of file diff --git a/1280ms/parakeet_eou_preprocessor.mlmodelc/model.mil b/1280ms/parakeet_eou_preprocessor.mlmodelc/model.mil new file mode 100644 index 0000000000000000000000000000000000000000..25db6a1374c1718340459f374949bfba2bff85bb --- /dev/null +++ b/1280ms/parakeet_eou_preprocessor.mlmodelc/model.mil @@ -0,0 +1,96 @@ +program(1.0) +[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.4.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] +{ + func main(tensor audio_length, tensor audio_signal) [FlexibleShapeInformation = tuple, dict, tensor>>, tuple, dict, list, ?>>>>((("DefaultShapes", {{"audio_signal", [1, 1]}}), ("RangeDims", {{"audio_signal", [[1, 1], [1, 32000]]}})))] { + tensor var_9 = const()[name = tensor("op_9"), val = tensor(1)]; + tensor var_10 = const()[name = tensor("op_10"), val = tensor(160)]; + tensor var_32 = const()[name = tensor("op_32"), val = tensor(512)]; + tensor var_33 = add(x = audio_length, y = var_32)[name = tensor("op_33")]; + tensor var_34 = const()[name = tensor("op_34"), val = tensor(512)]; + tensor var_35 = sub(x = var_33, y = var_34)[name = tensor("op_35")]; + tensor floor_div_0 = floor_div(x = var_35, y = var_10)[name = tensor("floor_div_0")]; + tensor var_36_dtype_0 = const()[name = tensor("op_36_dtype_0"), val = tensor("fp32")]; + tensor var_37_promoted = const()[name = tensor("op_37_promoted"), val = tensor(0x1p+0)]; + tensor var_36 = cast(dtype = var_36_dtype_0, x = floor_div_0)[name = tensor("cast_11")]; + tensor seq_len_1 = add(x = var_36, y = var_37_promoted)[name = tensor("seq_len_1")]; + tensor cast_2_dtype_0 = const()[name = tensor("cast_2_dtype_0"), val = tensor("int32")]; + tensor var_41_begin_0 = const()[name = tensor("op_41_begin_0"), val = tensor([0, 0])]; + tensor var_41_end_0 = const()[name = tensor("op_41_end_0"), val = tensor([1, 1])]; + tensor var_41_end_mask_0 = const()[name = tensor("op_41_end_mask_0"), val = tensor([true, false])]; + tensor var_41_squeeze_mask_0 = const()[name = tensor("op_41_squeeze_mask_0"), val = tensor([false, true])]; + tensor var_41 = slice_by_index(begin = var_41_begin_0, end = var_41_end_0, end_mask = var_41_end_mask_0, squeeze_mask = var_41_squeeze_mask_0, x = audio_signal)[name = tensor("op_41")]; + tensor var_42_axes_0 = const()[name = tensor("op_42_axes_0"), val = tensor([1])]; + tensor var_42 = expand_dims(axes = var_42_axes_0, x = var_41)[name = tensor("op_42")]; + tensor var_44_begin_0 = const()[name = tensor("op_44_begin_0"), val = tensor([0, 1])]; + tensor var_44_end_0 = const()[name = tensor("op_44_end_0"), val = tensor([1, 0])]; + tensor var_44_end_mask_0 = const()[name = tensor("op_44_end_mask_0"), val = tensor([true, true])]; + tensor var_44 = slice_by_index(begin = var_44_begin_0, end = var_44_end_0, end_mask = var_44_end_mask_0, x = audio_signal)[name = tensor("op_44")]; + tensor var_46_begin_0 = const()[name = tensor("op_46_begin_0"), val = tensor([0, 0])]; + tensor var_46_end_0 = const()[name = tensor("op_46_end_0"), val = tensor([1, -1])]; + tensor var_46_end_mask_0 = const()[name = tensor("op_46_end_mask_0"), val = tensor([true, false])]; + tensor var_46 = slice_by_index(begin = var_46_begin_0, end = var_46_end_0, end_mask = var_46_end_mask_0, x = audio_signal)[name = tensor("op_46")]; + tensor var_47 = const()[name = tensor("op_47"), val = tensor(0x1.f0a3d8p-1)]; + tensor var_48 = mul(x = var_46, y = var_47)[name = tensor("op_48")]; + tensor var_49 = sub(x = var_44, y = var_48)[name = tensor("op_49")]; + tensor input_1_interleave_0 = const()[name = tensor("input_1_interleave_0"), val = tensor(false)]; + tensor input_1 = concat(axis = var_9, interleave = input_1_interleave_0, values = (var_42, var_49))[name = tensor("input_1")]; + tensor concat_0x = const()[name = tensor("concat_0x"), val = tensor([1, 1, -1])]; + tensor input_3 = reshape(shape = concat_0x, x = input_1)[name = tensor("input_3")]; + tensor const_1 = const()[name = tensor("const_1"), val = tensor(0x0p+0)]; + tensor input_5_pad_0 = const()[name = tensor("input_5_pad_0"), val = tensor([0, 0, 0, 0, 256, 256])]; + tensor input_5_mode_0 = const()[name = tensor("input_5_mode_0"), val = tensor("reflect")]; + tensor input_5 = pad(constant_val = const_1, mode = input_5_mode_0, pad = input_5_pad_0, x = input_3)[name = tensor("input_5")]; + tensor concat_1x = const()[name = tensor("concat_1x"), val = tensor([1, -1])]; + tensor input = reshape(shape = concat_1x, x = input_5)[name = tensor("input")]; + tensor expand_dims_1 = const()[name = tensor("expand_dims_1"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; + tensor expand_dims_2 = const()[name = tensor("expand_dims_2"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526464)))]; + tensor expand_dims_3 = const()[name = tensor("expand_dims_3"), val = tensor([160])]; + tensor expand_dims_4_axes_0 = const()[name = tensor("expand_dims_4_axes_0"), val = tensor([1])]; + tensor expand_dims_4 = expand_dims(axes = expand_dims_4_axes_0, x = input)[name = tensor("expand_dims_4")]; + tensor conv_0_pad_type_0 = const()[name = tensor("conv_0_pad_type_0"), val = tensor("valid")]; + tensor conv_0_pad_0 = const()[name = tensor("conv_0_pad_0"), val = tensor([0, 0])]; + tensor conv_0_dilations_0 = const()[name = tensor("conv_0_dilations_0"), val = tensor([1])]; + tensor conv_0_groups_0 = const()[name = tensor("conv_0_groups_0"), val = tensor(1)]; + tensor conv_0 = conv(dilations = conv_0_dilations_0, groups = conv_0_groups_0, pad = conv_0_pad_0, pad_type = conv_0_pad_type_0, strides = expand_dims_3, weight = expand_dims_1, x = expand_dims_4)[name = tensor("conv_0")]; + tensor conv_1_pad_type_0 = const()[name = tensor("conv_1_pad_type_0"), val = tensor("valid")]; + tensor conv_1_pad_0 = const()[name = tensor("conv_1_pad_0"), val = tensor([0, 0])]; + tensor conv_1_dilations_0 = const()[name = tensor("conv_1_dilations_0"), val = tensor([1])]; + tensor conv_1_groups_0 = const()[name = tensor("conv_1_groups_0"), val = tensor(1)]; + tensor conv_1 = conv(dilations = conv_1_dilations_0, groups = conv_1_groups_0, pad = conv_1_pad_0, pad_type = conv_1_pad_type_0, strides = expand_dims_3, weight = expand_dims_2, x = expand_dims_4)[name = tensor("conv_1")]; + tensor stack_0_axis_0 = const()[name = tensor("stack_0_axis_0"), val = tensor(-1)]; + tensor stack_0 = stack(axis = stack_0_axis_0, values = (conv_0, conv_1))[name = tensor("stack_0")]; + tensor var_17_promoted = const()[name = tensor("op_17_promoted"), val = tensor(0x1p+1)]; + tensor var_65 = pow(x = stack_0, y = var_17_promoted)[name = tensor("op_65")]; + tensor var_67_axes_0 = const()[name = tensor("op_67_axes_0"), val = tensor([-1])]; + tensor var_67_keep_dims_0 = const()[name = tensor("op_67_keep_dims_0"), val = tensor(false)]; + tensor var_67 = reduce_sum(axes = var_67_axes_0, keep_dims = var_67_keep_dims_0, x = var_65)[name = tensor("op_67")]; + tensor x_9 = identity(x = var_67)[name = tensor("x_9")]; + tensor const_2 = const()[name = tensor("const_2"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1052864)))]; + tensor x_11_transpose_x_0 = const()[name = tensor("x_11_transpose_x_0"), val = tensor(false)]; + tensor x_11_transpose_y_0 = const()[name = tensor("x_11_transpose_y_0"), val = tensor(false)]; + tensor x_11 = matmul(transpose_x = x_11_transpose_x_0, transpose_y = x_11_transpose_y_0, x = const_2, y = x_9)[name = tensor("x_11")]; + tensor var_74 = const()[name = tensor("op_74"), val = tensor(0x1p-24)]; + tensor var_75 = add(x = x_11, y = var_74)[name = tensor("op_75")]; + tensor x_epsilon_0 = const()[name = tensor("x_epsilon_0"), val = tensor(0x1p-149)]; + tensor x = log(epsilon = x_epsilon_0, x = var_75)[name = tensor("x")]; + tensor var_77_shape = shape(x = x)[name = tensor("op_77_shape")]; + tensor select_4 = const()[name = tensor("select_4"), val = tensor(2)]; + tensor gather_4_axis_0 = const()[name = tensor("gather_4_axis_0"), val = tensor(0)]; + tensor gather_4_batch_dims_0 = const()[name = tensor("gather_4_batch_dims_0"), val = tensor(0)]; + tensor gather_4_validate_indices_0 = const()[name = tensor("gather_4_validate_indices_0"), val = tensor(false)]; + tensor gather_4 = gather(axis = gather_4_axis_0, batch_dims = gather_4_batch_dims_0, indices = select_4, validate_indices = gather_4_validate_indices_0, x = var_77_shape)[name = tensor("gather_4")]; + tensor const_3 = const()[name = tensor("const_3"), val = tensor(0)]; + tensor const_4 = const()[name = tensor("const_4"), val = tensor(1)]; + tensor mask_1 = range_1d(end = gather_4, start = const_3, step = const_4)[name = tensor("mask_1")]; + tensor expand_dims_0_axes_0 = const()[name = tensor("expand_dims_0_axes_0"), val = tensor([0])]; + tensor expand_dims_0 = expand_dims(axes = expand_dims_0_axes_0, x = mask_1)[name = tensor("expand_dims_0")]; + tensor var_82_axes_0 = const()[name = tensor("op_82_axes_0"), val = tensor([1])]; + tensor mel_length = cast(dtype = cast_2_dtype_0, x = seq_len_1)[name = tensor("cast_10")]; + tensor var_82 = expand_dims(axes = var_82_axes_0, x = mel_length)[name = tensor("op_82")]; + tensor mask = greater_equal(x = expand_dims_0, y = var_82)[name = tensor("mask")]; + tensor var_84_axes_0 = const()[name = tensor("op_84_axes_0"), val = tensor([1])]; + tensor var_84 = expand_dims(axes = var_84_axes_0, x = mask)[name = tensor("op_84")]; + tensor cast_7 = const()[name = tensor("cast_7"), val = tensor(0x0p+0)]; + tensor mel = select(a = cast_7, b = x, cond = var_84)[name = tensor("processed_signal")]; + } -> (mel, mel_length); +} \ No newline at end of file diff --git a/1280ms/parakeet_eou_preprocessor.mlmodelc/weights/weight.bin b/1280ms/parakeet_eou_preprocessor.mlmodelc/weights/weight.bin new file mode 100644 index 0000000000000000000000000000000000000000..99f6476377d3e7474aa570db365b7429f9d7b5b8 --- /dev/null +++ b/1280ms/parakeet_eou_preprocessor.mlmodelc/weights/weight.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:009bba4fde82dc55db9b55d77cf3ba5f791ce366c49f079285fe25a3b6e2291d +size 1184512 diff --git a/1280ms/streaming_encoder.mlmodelc/analytics/coremldata.bin b/1280ms/streaming_encoder.mlmodelc/analytics/coremldata.bin new file mode 100644 index 0000000000000000000000000000000000000000..e22bce589c25d1e6413bee12465c6eb2d68c6262 --- /dev/null +++ b/1280ms/streaming_encoder.mlmodelc/analytics/coremldata.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d0a3c84022a9d2dc769d38cf8f45e93423e20734d092e3c16db11fbf6dca4004 +size 243 diff --git a/1280ms/streaming_encoder.mlmodelc/coremldata.bin b/1280ms/streaming_encoder.mlmodelc/coremldata.bin new file mode 100644 index 0000000000000000000000000000000000000000..7e43604cbe391d85b29d5826bb5cee4f987575d2 --- /dev/null +++ b/1280ms/streaming_encoder.mlmodelc/coremldata.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:41ce3f96c3d6b3333796fc4ed82cb0c9b4ea99396b88f8eec3ba24394ba2bb78 +size 671 diff --git a/1280ms/streaming_encoder.mlmodelc/metadata.json b/1280ms/streaming_encoder.mlmodelc/metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..bd9ca5c8491ca0588ddec260eb9c532ffb9ef560 --- /dev/null +++ b/1280ms/streaming_encoder.mlmodelc/metadata.json @@ -0,0 +1,187 @@ +[ + { + "metadataOutputVersion" : "3.0", + "storagePrecision" : "Float16", + "outputSchema" : [ + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 512 × 17)", + "shortDescription" : "", + "shape" : "[1, 512, 17]", + "name" : "encoded_output", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1)", + "shortDescription" : "", + "shape" : "[1]", + "name" : "encoded_length", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 128 × 16)", + "shortDescription" : "", + "shape" : "[1, 128, 16]", + "name" : "new_pre_cache", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 17 × 1 × 70 × 512)", + "shortDescription" : "", + "shape" : "[17, 1, 70, 512]", + "name" : "new_cache_last_channel", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 17 × 1 × 512 × 8)", + "shortDescription" : "", + "shape" : "[17, 1, 512, 8]", + "name" : "new_cache_last_time", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1)", + "shortDescription" : "", + "shape" : "[1]", + "name" : "new_cache_last_channel_len", + "type" : "MultiArray" + } + ], + "modelParameters" : [ + + ], + "specificationVersion" : 8, + "mlProgramOperationTypeHistogram" : { + "Ios17.floor" : 3, + "Ios17.logicalAnd" : 3, + "Ios17.reshape" : 103, + "Ios16.softmax" : 17, + "Ios17.matmul" : 51, + "Ios17.transpose" : 157, + "Split" : 17, + "Ios17.expandDims" : 6, + "Select" : 51, + "Ios17.add" : 126, + "Tile" : 1, + "Ios17.sliceByIndex" : 106, + "Ios16.sigmoid" : 17, + "Pad" : 20, + "Ios17.logicalNot" : 2, + "Ios17.layerNorm" : 102, + "Ios17.less" : 1, + "Ios17.sub" : 1, + "Ios17.conv" : 56, + "Ios17.clip" : 2, + "Ios16.relu" : 3, + "Ios17.linear" : 137, + "Ios17.concat" : 52, + "Ios17.greaterEqual" : 1, + "Ios17.cast" : 14, + "Ios16.silu" : 51, + "Stack" : 2, + "Ios17.mul" : 72 + }, + "computePrecision" : "Mixed (Float16, Float32, Int32)", + "isUpdatable" : "0", + "stateSchema" : [ + + ], + "availability" : { + "macOS" : "14.0", + "tvOS" : "17.0", + "visionOS" : "1.0", + "watchOS" : "10.0", + "iOS" : "17.0", + "macCatalyst" : "17.0" + }, + "modelType" : { + "name" : "MLModelType_mlProgram" + }, + "userDefinedMetadata" : { + "com.github.apple.coremltools.source_dialect" : "TorchScript", + "com.github.apple.coremltools.source" : "torch==2.4.0", + "com.github.apple.coremltools.version" : "8.3.0" + }, + "inputSchema" : [ + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 128 × 129)", + "shortDescription" : "", + "shape" : "[1, 128, 129]", + "name" : "audio_signal", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1)", + "shortDescription" : "", + "shape" : "[1]", + "name" : "audio_length", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 128 × 16)", + "shortDescription" : "", + "shape" : "[1, 128, 16]", + "name" : "pre_cache", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 17 × 1 × 70 × 512)", + "shortDescription" : "", + "shape" : "[17, 1, 70, 512]", + "name" : "cache_last_channel", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 17 × 1 × 512 × 8)", + "shortDescription" : "", + "shape" : "[17, 1, 512, 8]", + "name" : "cache_last_time", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1)", + "shortDescription" : "", + "shape" : "[1]", + "name" : "cache_last_channel_len", + "type" : "MultiArray" + } + ], + "generatedClassName" : "streaming_encoder", + "method" : "predict" + } +] \ No newline at end of file diff --git a/1280ms/streaming_encoder.mlmodelc/model.mil b/1280ms/streaming_encoder.mlmodelc/model.mil new file mode 100644 index 0000000000000000000000000000000000000000..e87da5042d394a857c4bd11850c4ec70d6577d86 --- /dev/null +++ b/1280ms/streaming_encoder.mlmodelc/model.mil @@ -0,0 +1,2977 @@ +program(1.0) +[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.4.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] +{ + func main(tensor audio_length, tensor audio_signal, tensor cache_last_channel, tensor cache_last_channel_len, tensor cache_last_time, tensor pre_cache) { + tensor var_9 = const()[name = tensor("op_9"), val = tensor(2)]; + tensor full_input_interleave_0 = const()[name = tensor("full_input_interleave_0"), val = tensor(false)]; + tensor pre_cache_to_fp16_dtype_0 = const()[name = tensor("pre_cache_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor audio_signal_to_fp16_dtype_0 = const()[name = tensor("audio_signal_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor audio_signal_to_fp16 = cast(dtype = audio_signal_to_fp16_dtype_0, x = audio_signal)[name = tensor("cast_193")]; + tensor pre_cache_to_fp16 = cast(dtype = pre_cache_to_fp16_dtype_0, x = pre_cache)[name = tensor("cast_194")]; + tensor full_input_cast_fp16 = concat(axis = var_9, interleave = full_input_interleave_0, values = (pre_cache_to_fp16, audio_signal_to_fp16))[name = tensor("full_input_cast_fp16")]; + tensor var_12 = const()[name = tensor("op_12"), val = tensor(16)]; + tensor value_1 = add(x = audio_length, y = var_12)[name = tensor("value_1")]; + tensor var_28_begin_0 = const()[name = tensor("op_28_begin_0"), val = tensor([0, 0, 129])]; + tensor var_28_end_0 = const()[name = tensor("op_28_end_0"), val = tensor([1, 128, 145])]; + tensor var_28_end_mask_0 = const()[name = tensor("op_28_end_mask_0"), val = tensor([true, true, true])]; + tensor var_28_cast_fp16 = slice_by_index(begin = var_28_begin_0, end = var_28_end_0, end_mask = var_28_end_mask_0, x = full_input_cast_fp16)[name = tensor("op_28_cast_fp16")]; + tensor var_28_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_28_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor var_62 = const()[name = tensor("op_62"), val = tensor(-1)]; + tensor var_64 = const()[name = tensor("op_64"), val = tensor(1)]; + tensor x_1_perm_0 = const()[name = tensor("x_1_perm_0"), val = tensor([0, 2, 1])]; + tensor cast_0_to_fp16_dtype_0 = const()[name = tensor("cast_0_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor _inversed_108_y_0_to_fp16 = const()[name = tensor("_inversed_108_y_0_to_fp16"), val = tensor(0x1p-1)]; + tensor value_1_to_fp16 = cast(dtype = cast_0_to_fp16_dtype_0, x = value_1)[name = tensor("cast_191")]; + tensor _inversed_108_cast_fp16 = mul(x = value_1_to_fp16, y = _inversed_108_y_0_to_fp16)[name = tensor("_inversed_108_cast_fp16")]; + tensor var_109_to_fp16 = const()[name = tensor("op_109_to_fp16"), val = tensor(0x1p+0)]; + tensor lengths_1_cast_fp16 = add(x = _inversed_108_cast_fp16, y = var_109_to_fp16)[name = tensor("lengths_1_cast_fp16")]; + tensor lengths_3_cast_fp16 = floor(x = lengths_1_cast_fp16)[name = tensor("lengths_3_cast_fp16")]; + tensor _inversed_116_y_0_to_fp16 = const()[name = tensor("_inversed_116_y_0_to_fp16"), val = tensor(0x1p-1)]; + tensor _inversed_116_cast_fp16 = mul(x = lengths_3_cast_fp16, y = _inversed_116_y_0_to_fp16)[name = tensor("_inversed_116_cast_fp16")]; + tensor var_117_to_fp16 = const()[name = tensor("op_117_to_fp16"), val = tensor(0x1p+0)]; + tensor lengths_7_cast_fp16 = add(x = _inversed_116_cast_fp16, y = var_117_to_fp16)[name = tensor("lengths_7_cast_fp16")]; + tensor lengths_9_cast_fp16 = floor(x = lengths_7_cast_fp16)[name = tensor("lengths_9_cast_fp16")]; + tensor _inversed_124_y_0_to_fp16 = const()[name = tensor("_inversed_124_y_0_to_fp16"), val = tensor(0x1p-1)]; + tensor _inversed_124_cast_fp16 = mul(x = lengths_9_cast_fp16, y = _inversed_124_y_0_to_fp16)[name = tensor("_inversed_124_cast_fp16")]; + tensor var_125_to_fp16 = const()[name = tensor("op_125_to_fp16"), val = tensor(0x1p+0)]; + tensor lengths_13_cast_fp16 = add(x = _inversed_124_cast_fp16, y = var_125_to_fp16)[name = tensor("lengths_13_cast_fp16")]; + tensor lengths_cast_fp16 = floor(x = lengths_13_cast_fp16)[name = tensor("lengths_cast_fp16")]; + tensor cast_9_dtype_0 = const()[name = tensor("cast_9_dtype_0"), val = tensor("int32")]; + tensor input_1_axes_0 = const()[name = tensor("input_1_axes_0"), val = tensor([1])]; + tensor x_1_cast_fp16 = transpose(perm = x_1_perm_0, x = full_input_cast_fp16)[name = tensor("transpose_241")]; + tensor input_1_cast_fp16 = expand_dims(axes = input_1_axes_0, x = x_1_cast_fp16)[name = tensor("input_1_cast_fp16")]; + tensor input_3_pad_0 = const()[name = tensor("input_3_pad_0"), val = tensor([0, 0, 0, 0, 2, 1, 2, 1])]; + tensor input_3_mode_0 = const()[name = tensor("input_3_mode_0"), val = tensor("constant")]; + tensor const_0_to_fp16 = const()[name = tensor("const_0_to_fp16"), val = tensor(0x0p+0)]; + tensor input_3_cast_fp16 = pad(constant_val = const_0_to_fp16, mode = input_3_mode_0, pad = input_3_pad_0, x = input_1_cast_fp16)[name = tensor("input_3_cast_fp16")]; + tensor input_5_pad_type_0 = const()[name = tensor("input_5_pad_type_0"), val = tensor("valid")]; + tensor input_5_strides_0 = const()[name = tensor("input_5_strides_0"), val = tensor([2, 2])]; + tensor input_5_pad_0 = const()[name = tensor("input_5_pad_0"), val = tensor([0, 0, 0, 0])]; + tensor input_5_dilations_0 = const()[name = tensor("input_5_dilations_0"), val = tensor([1, 1])]; + tensor input_5_groups_0 = const()[name = tensor("input_5_groups_0"), val = tensor(1)]; + tensor encoder_pre_encode_conv_0_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; + tensor encoder_pre_encode_conv_0_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4736)))]; + tensor input_5_cast_fp16 = conv(bias = encoder_pre_encode_conv_0_bias_to_fp16, dilations = input_5_dilations_0, groups = input_5_groups_0, pad = input_5_pad_0, pad_type = input_5_pad_type_0, strides = input_5_strides_0, weight = encoder_pre_encode_conv_0_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; + tensor input_7_cast_fp16 = relu(x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; + tensor input_9_pad_0 = const()[name = tensor("input_9_pad_0"), val = tensor([0, 0, 0, 0, 2, 1, 2, 1])]; + tensor input_9_mode_0 = const()[name = tensor("input_9_mode_0"), val = tensor("constant")]; + tensor const_1_to_fp16 = const()[name = tensor("const_1_to_fp16"), val = tensor(0x0p+0)]; + tensor input_9_cast_fp16 = pad(constant_val = const_1_to_fp16, mode = input_9_mode_0, pad = input_9_pad_0, x = input_7_cast_fp16)[name = tensor("input_9_cast_fp16")]; + tensor input_11_pad_type_0 = const()[name = tensor("input_11_pad_type_0"), val = tensor("valid")]; + tensor input_11_strides_0 = const()[name = tensor("input_11_strides_0"), val = tensor([2, 2])]; + tensor input_11_groups_0 = const()[name = tensor("input_11_groups_0"), val = tensor(256)]; + tensor input_11_pad_0 = const()[name = tensor("input_11_pad_0"), val = tensor([0, 0, 0, 0])]; + tensor input_11_dilations_0 = const()[name = tensor("input_11_dilations_0"), val = tensor([1, 1])]; + tensor encoder_pre_encode_conv_2_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5312)))]; + tensor encoder_pre_encode_conv_2_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9984)))]; + tensor input_11_cast_fp16 = conv(bias = encoder_pre_encode_conv_2_bias_to_fp16, dilations = input_11_dilations_0, groups = input_11_groups_0, pad = input_11_pad_0, pad_type = input_11_pad_type_0, strides = input_11_strides_0, weight = encoder_pre_encode_conv_2_weight_to_fp16, x = input_9_cast_fp16)[name = tensor("input_11_cast_fp16")]; + tensor input_13_pad_type_0 = const()[name = tensor("input_13_pad_type_0"), val = tensor("valid")]; + tensor input_13_strides_0 = const()[name = tensor("input_13_strides_0"), val = tensor([1, 1])]; + tensor input_13_pad_0 = const()[name = tensor("input_13_pad_0"), val = tensor([0, 0, 0, 0])]; + tensor input_13_dilations_0 = const()[name = tensor("input_13_dilations_0"), val = tensor([1, 1])]; + tensor input_13_groups_0 = const()[name = tensor("input_13_groups_0"), val = tensor(1)]; + tensor encoder_pre_encode_conv_3_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_3_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10560)))]; + tensor encoder_pre_encode_conv_3_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_3_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141696)))]; + tensor input_13_cast_fp16 = conv(bias = encoder_pre_encode_conv_3_bias_to_fp16, dilations = input_13_dilations_0, groups = input_13_groups_0, pad = input_13_pad_0, pad_type = input_13_pad_type_0, strides = input_13_strides_0, weight = encoder_pre_encode_conv_3_weight_to_fp16, x = input_11_cast_fp16)[name = tensor("input_13_cast_fp16")]; + tensor input_15_cast_fp16 = relu(x = input_13_cast_fp16)[name = tensor("input_15_cast_fp16")]; + tensor input_17_pad_0 = const()[name = tensor("input_17_pad_0"), val = tensor([0, 0, 0, 0, 2, 1, 2, 1])]; + tensor input_17_mode_0 = const()[name = tensor("input_17_mode_0"), val = tensor("constant")]; + tensor const_2_to_fp16 = const()[name = tensor("const_2_to_fp16"), val = tensor(0x0p+0)]; + tensor input_17_cast_fp16 = pad(constant_val = const_2_to_fp16, mode = input_17_mode_0, pad = input_17_pad_0, x = input_15_cast_fp16)[name = tensor("input_17_cast_fp16")]; + tensor input_19_pad_type_0 = const()[name = tensor("input_19_pad_type_0"), val = tensor("valid")]; + tensor input_19_strides_0 = const()[name = tensor("input_19_strides_0"), val = tensor([2, 2])]; + tensor input_19_groups_0 = const()[name = tensor("input_19_groups_0"), val = tensor(256)]; + tensor input_19_pad_0 = const()[name = tensor("input_19_pad_0"), val = tensor([0, 0, 0, 0])]; + tensor input_19_dilations_0 = const()[name = tensor("input_19_dilations_0"), val = tensor([1, 1])]; + tensor encoder_pre_encode_conv_5_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_5_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142272)))]; + tensor encoder_pre_encode_conv_5_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_5_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146944)))]; + tensor input_19_cast_fp16 = conv(bias = encoder_pre_encode_conv_5_bias_to_fp16, dilations = input_19_dilations_0, groups = input_19_groups_0, pad = input_19_pad_0, pad_type = input_19_pad_type_0, strides = input_19_strides_0, weight = encoder_pre_encode_conv_5_weight_to_fp16, x = input_17_cast_fp16)[name = tensor("input_19_cast_fp16")]; + tensor input_21_pad_type_0 = const()[name = tensor("input_21_pad_type_0"), val = tensor("valid")]; + tensor input_21_strides_0 = const()[name = tensor("input_21_strides_0"), val = tensor([1, 1])]; + tensor input_21_pad_0 = const()[name = tensor("input_21_pad_0"), val = tensor([0, 0, 0, 0])]; + tensor input_21_dilations_0 = const()[name = tensor("input_21_dilations_0"), val = tensor([1, 1])]; + tensor input_21_groups_0 = const()[name = tensor("input_21_groups_0"), val = tensor(1)]; + tensor encoder_pre_encode_conv_6_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_6_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147520)))]; + tensor encoder_pre_encode_conv_6_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_6_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(278656)))]; + tensor input_21_cast_fp16 = conv(bias = encoder_pre_encode_conv_6_bias_to_fp16, dilations = input_21_dilations_0, groups = input_21_groups_0, pad = input_21_pad_0, pad_type = input_21_pad_type_0, strides = input_21_strides_0, weight = encoder_pre_encode_conv_6_weight_to_fp16, x = input_19_cast_fp16)[name = tensor("input_21_cast_fp16")]; + tensor x_3_cast_fp16 = relu(x = input_21_cast_fp16)[name = tensor("x_3_cast_fp16")]; + tensor var_181_perm_0 = const()[name = tensor("op_181_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_182 = const()[name = tensor("op_182"), val = tensor([1, 19, -1])]; + tensor var_181_cast_fp16 = transpose(perm = var_181_perm_0, x = x_3_cast_fp16)[name = tensor("transpose_240")]; + tensor input_23_cast_fp16 = reshape(shape = var_182, x = var_181_cast_fp16)[name = tensor("input_23_cast_fp16")]; + tensor encoder_pre_encode_out_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279232)))]; + tensor encoder_pre_encode_out_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4735744)))]; + tensor linear_0_cast_fp16 = linear(bias = encoder_pre_encode_out_bias_to_fp16, weight = encoder_pre_encode_out_weight_to_fp16, x = input_23_cast_fp16)[name = tensor("linear_0_cast_fp16")]; + tensor var_192_begin_0 = const()[name = tensor("op_192_begin_0"), val = tensor([0, 2, 0])]; + tensor var_192_end_0 = const()[name = tensor("op_192_end_0"), val = tensor([1, 19, 512])]; + tensor var_192_end_mask_0 = const()[name = tensor("op_192_end_mask_0"), val = tensor([true, true, true])]; + tensor var_192_cast_fp16 = slice_by_index(begin = var_192_begin_0, end = var_192_end_0, end_mask = var_192_end_mask_0, x = linear_0_cast_fp16)[name = tensor("op_192_cast_fp16")]; + tensor var_194 = const()[name = tensor("op_194"), val = tensor(2)]; + tensor lengths_cast_fp16_to_int32 = cast(dtype = cast_9_dtype_0, x = lengths_cast_fp16)[name = tensor("cast_190")]; + tensor var_195 = sub(x = lengths_cast_fp16_to_int32, y = var_194)[name = tensor("op_195")]; + tensor var_195_promoted_to_fp16_dtype_0 = const()[name = tensor("op_195_promoted_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor var_60_promoted_to_fp16 = const()[name = tensor("op_60_promoted_to_fp16"), val = tensor(0x0p+0)]; + tensor const_5_to_fp16 = const()[name = tensor("const_5_to_fp16"), val = tensor(inf)]; + tensor var_195_to_fp16 = cast(dtype = var_195_promoted_to_fp16_dtype_0, x = var_195)[name = tensor("cast_189")]; + tensor clip_0_cast_fp16 = clip(alpha = var_60_promoted_to_fp16, beta = const_5_to_fp16, x = var_195_to_fp16)[name = tensor("clip_0_cast_fp16")]; + tensor max_audio_length_1 = const()[name = tensor("max_audio_length_1"), val = tensor([17])]; + tensor var_211_promoted_to_fp16 = const()[name = tensor("op_211_promoted_to_fp16"), val = tensor(0x1.18p+6)]; + tensor padding_length_cast_fp16 = add(x = clip_0_cast_fp16, y = var_211_promoted_to_fp16)[name = tensor("padding_length_cast_fp16")]; + tensor const_7 = const()[name = tensor("const_7"), val = tensor(-1)]; + tensor var_213 = mul(x = cache_last_channel_len, y = const_7)[name = tensor("op_213")]; + tensor var_214 = const()[name = tensor("op_214"), val = tensor(70)]; + tensor offset = add(x = var_213, y = var_214)[name = tensor("offset")]; + tensor var_254_axes_0 = const()[name = tensor("op_254_axes_0"), val = tensor([-1])]; + tensor var_254_cast_fp16 = expand_dims(axes = var_254_axes_0, x = padding_length_cast_fp16)[name = tensor("op_254_cast_fp16")]; + tensor var_253_promoted_to_fp16 = const()[name = tensor("op_253_promoted_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4736832)))]; + tensor pad_mask_1_cast_fp16 = less(x = var_253_promoted_to_fp16, y = var_254_cast_fp16)[name = tensor("pad_mask_1_cast_fp16")]; + tensor expand_dims_1 = const()[name = tensor("expand_dims_1"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86]])]; + tensor var_260_axes_0 = const()[name = tensor("op_260_axes_0"), val = tensor([-1])]; + tensor var_260 = expand_dims(axes = var_260_axes_0, x = offset)[name = tensor("op_260")]; + tensor pad_mask_off = greater_equal(x = expand_dims_1, y = var_260)[name = tensor("pad_mask_off")]; + tensor pad_mask_3 = logical_and(x = pad_mask_off, y = pad_mask_1_cast_fp16)[name = tensor("pad_mask_3")]; + tensor var_263_axes_0 = const()[name = tensor("op_263_axes_0"), val = tensor([1])]; + tensor var_263 = expand_dims(axes = var_263_axes_0, x = pad_mask_3)[name = tensor("op_263")]; + tensor var_264 = const()[name = tensor("op_264"), val = tensor([1, 87, 1])]; + tensor pad_mask_for_att_mask_1 = tile(reps = var_264, x = var_263)[name = tensor("pad_mask_for_att_mask_1")]; + tensor var_266_perm_0 = const()[name = tensor("op_266_perm_0"), val = tensor([0, 2, 1])]; + tensor var_266 = transpose(perm = var_266_perm_0, x = pad_mask_for_att_mask_1)[name = tensor("transpose_239")]; + tensor pad_mask_for_att_mask = logical_and(x = pad_mask_for_att_mask_1, y = var_266)[name = tensor("pad_mask_for_att_mask")]; + tensor const_15 = const()[name = tensor("const_15"), val = tensor([[[true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false], [false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false], [false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false], [false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false], [false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false], [false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false], [false, false, false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false], [false, false, false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false], [false, false, false, false, false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false], [false, false, false, false, false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false], [false, false, false, false, false, false, false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false], [false, false, false, false, false, false, false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false], [false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false], [false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false], [false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true]]])]; + tensor att_mask_9 = logical_and(x = pad_mask_for_att_mask, y = const_15)[name = tensor("att_mask_9")]; + tensor att_mask = logical_not(x = att_mask_9)[name = tensor("att_mask")]; + tensor pad_mask_5 = logical_not(x = pad_mask_3)[name = tensor("pad_mask_5")]; + tensor pad_mask_begin_0 = const()[name = tensor("pad_mask_begin_0"), val = tensor([0, 70])]; + tensor pad_mask_end_0 = const()[name = tensor("pad_mask_end_0"), val = tensor([1, 87])]; + tensor pad_mask_end_mask_0 = const()[name = tensor("pad_mask_end_mask_0"), val = tensor([true, true])]; + tensor pad_mask = slice_by_index(begin = pad_mask_begin_0, end = pad_mask_end_0, end_mask = pad_mask_end_mask_0, x = pad_mask_5)[name = tensor("pad_mask")]; + tensor mask_1_begin_0 = const()[name = tensor("mask_1_begin_0"), val = tensor([0, 70, 0])]; + tensor mask_1_end_0 = const()[name = tensor("mask_1_end_0"), val = tensor([1, 87, 87])]; + tensor mask_1_end_mask_0 = const()[name = tensor("mask_1_end_mask_0"), val = tensor([true, true, true])]; + tensor mask_1 = slice_by_index(begin = mask_1_begin_0, end = mask_1_end_0, end_mask = mask_1_end_mask_0, x = att_mask)[name = tensor("mask_1")]; + tensor cache_1_begin_0 = const()[name = tensor("cache_1_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor cache_1_end_0 = const()[name = tensor("cache_1_end_0"), val = tensor([1, 1, 70, 512])]; + tensor cache_1_end_mask_0 = const()[name = tensor("cache_1_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_1_squeeze_mask_0 = const()[name = tensor("cache_1_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_last_channel_to_fp16_dtype_0 = const()[name = tensor("cache_last_channel_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor cache_last_channel_to_fp16 = cast(dtype = cache_last_channel_to_fp16_dtype_0, x = cache_last_channel)[name = tensor("cast_188")]; + tensor cache_1_cast_fp16 = slice_by_index(begin = cache_1_begin_0, end = cache_1_end_0, end_mask = cache_1_end_mask_0, squeeze_mask = cache_1_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_1_cast_fp16")]; + tensor cache_3_begin_0 = const()[name = tensor("cache_3_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor cache_3_end_0 = const()[name = tensor("cache_3_end_0"), val = tensor([1, 1, 512, 8])]; + tensor cache_3_end_mask_0 = const()[name = tensor("cache_3_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_3_squeeze_mask_0 = const()[name = tensor("cache_3_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_last_time_to_fp16_dtype_0 = const()[name = tensor("cache_last_time_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor cache_last_time_to_fp16 = cast(dtype = cache_last_time_to_fp16_dtype_0, x = cache_last_time)[name = tensor("cast_187")]; + tensor cache_3_cast_fp16 = slice_by_index(begin = cache_3_begin_0, end = cache_3_end_0, end_mask = cache_3_end_mask_0, squeeze_mask = cache_3_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_3_cast_fp16")]; + tensor input_27_axes_0 = const()[name = tensor("input_27_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4737088)))]; + tensor encoder_layers_0_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4738176)))]; + tensor var_38_to_fp16 = const()[name = tensor("op_38_to_fp16"), val = tensor(0x1.5p-17)]; + tensor input_27_cast_fp16 = layer_norm(axes = input_27_axes_0, beta = encoder_layers_0_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_norm_feed_forward1_weight_to_fp16, x = var_192_cast_fp16)[name = tensor("input_27_cast_fp16")]; + tensor encoder_layers_0_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_0_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4739264)))]; + tensor linear_1_bias_0_to_fp16 = const()[name = tensor("linear_1_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6836480)))]; + tensor linear_1_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_0_feed_forward1_linear1_weight_to_fp16, x = input_27_cast_fp16)[name = tensor("linear_1_cast_fp16")]; + tensor input_31_cast_fp16 = silu(x = linear_1_cast_fp16)[name = tensor("input_31_cast_fp16")]; + tensor encoder_layers_0_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_0_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6840640)))]; + tensor linear_2_bias_0_to_fp16 = const()[name = tensor("linear_2_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8937856)))]; + tensor linear_2_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_feed_forward1_linear2_weight_to_fp16, x = input_31_cast_fp16)[name = tensor("linear_2_cast_fp16")]; + tensor var_303_to_fp16 = const()[name = tensor("op_303_to_fp16"), val = tensor(0x1p-1)]; + tensor var_304_cast_fp16 = mul(x = linear_2_cast_fp16, y = var_303_to_fp16)[name = tensor("op_304_cast_fp16")]; + tensor input_37_cast_fp16 = add(x = var_192_cast_fp16, y = var_304_cast_fp16)[name = tensor("input_37_cast_fp16")]; + tensor key_1_axes_0 = const()[name = tensor("key_1_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8938944)))]; + tensor encoder_layers_0_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8940032)))]; + tensor key_1_cast_fp16 = layer_norm(axes = key_1_axes_0, beta = encoder_layers_0_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_norm_self_att_weight_to_fp16, x = input_37_cast_fp16)[name = tensor("key_1_cast_fp16")]; + tensor input_39_interleave_0 = const()[name = tensor("input_39_interleave_0"), val = tensor(false)]; + tensor input_39_cast_fp16 = concat(axis = var_64, interleave = input_39_interleave_0, values = (cache_1_cast_fp16, key_1_cast_fp16))[name = tensor("input_39_cast_fp16")]; + tensor var_326_begin_0 = const()[name = tensor("op_326_begin_0"), val = tensor([0, 17, 0])]; + tensor var_326_end_0 = const()[name = tensor("op_326_end_0"), val = tensor([1, 70, 512])]; + tensor var_326_end_mask_0 = const()[name = tensor("op_326_end_mask_0"), val = tensor([true, true, true])]; + tensor var_326_cast_fp16 = slice_by_index(begin = var_326_begin_0, end = var_326_end_0, end_mask = var_326_end_mask_0, x = cache_1_cast_fp16)[name = tensor("op_326_cast_fp16")]; + tensor var_332_interleave_0 = const()[name = tensor("op_332_interleave_0"), val = tensor(false)]; + tensor var_332_cast_fp16 = concat(axis = var_64, interleave = var_332_interleave_0, values = (var_326_cast_fp16, key_1_cast_fp16))[name = tensor("op_332_cast_fp16")]; + tensor encoder_layers_0_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8941120)))]; + tensor linear_3_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_self_attn_linear_q_weight_to_fp16, x = key_1_cast_fp16)[name = tensor("linear_3_cast_fp16")]; + tensor var_336 = const()[name = tensor("op_336"), val = tensor([1, -1, 8, 64])]; + tensor q_1_cast_fp16 = reshape(shape = var_336, x = linear_3_cast_fp16)[name = tensor("q_1_cast_fp16")]; + tensor encoder_layers_0_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9465472)))]; + tensor linear_4_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_self_attn_linear_k_weight_to_fp16, x = input_39_cast_fp16)[name = tensor("linear_4_cast_fp16")]; + tensor var_340 = const()[name = tensor("op_340"), val = tensor([1, -1, 8, 64])]; + tensor k_1_cast_fp16 = reshape(shape = var_340, x = linear_4_cast_fp16)[name = tensor("k_1_cast_fp16")]; + tensor encoder_layers_0_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9989824)))]; + tensor linear_5_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_self_attn_linear_v_weight_to_fp16, x = input_39_cast_fp16)[name = tensor("linear_5_cast_fp16")]; + tensor var_344 = const()[name = tensor("op_344"), val = tensor([1, -1, 8, 64])]; + tensor v_1_cast_fp16 = reshape(shape = var_344, x = linear_5_cast_fp16)[name = tensor("v_1_cast_fp16")]; + tensor value_3_perm_0 = const()[name = tensor("value_3_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_0_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10514176)))]; + tensor var_356_cast_fp16 = add(x = q_1_cast_fp16, y = encoder_layers_0_self_attn_pos_bias_u_to_fp16)[name = tensor("op_356_cast_fp16")]; + tensor encoder_layers_0_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10515264)))]; + tensor var_358_cast_fp16 = add(x = q_1_cast_fp16, y = encoder_layers_0_self_attn_pos_bias_v_to_fp16)[name = tensor("op_358_cast_fp16")]; + tensor q_with_bias_v_1_perm_0 = const()[name = tensor("q_with_bias_v_1_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_7_transpose_x_0 = const()[name = tensor("x_7_transpose_x_0"), val = tensor(false)]; + tensor x_7_transpose_y_0 = const()[name = tensor("x_7_transpose_y_0"), val = tensor(false)]; + tensor var_360_to_fp16 = const()[name = tensor("op_360_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10516352)))]; + tensor q_with_bias_v_1_cast_fp16 = transpose(perm = q_with_bias_v_1_perm_0, x = var_358_cast_fp16)[name = tensor("transpose_237")]; + tensor x_7_cast_fp16 = matmul(transpose_x = x_7_transpose_x_0, transpose_y = x_7_transpose_y_0, x = q_with_bias_v_1_cast_fp16, y = var_360_to_fp16)[name = tensor("x_7_cast_fp16")]; + tensor x_9_pad_0 = const()[name = tensor("x_9_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_9_mode_0 = const()[name = tensor("x_9_mode_0"), val = tensor("constant")]; + tensor const_23_to_fp16 = const()[name = tensor("const_23_to_fp16"), val = tensor(0x0p+0)]; + tensor x_9_cast_fp16 = pad(constant_val = const_23_to_fp16, mode = x_9_mode_0, pad = x_9_pad_0, x = x_7_cast_fp16)[name = tensor("x_9_cast_fp16")]; + tensor var_368 = const()[name = tensor("op_368"), val = tensor([1, 8, -1, 17])]; + tensor x_11_cast_fp16 = reshape(shape = var_368, x = x_9_cast_fp16)[name = tensor("x_11_cast_fp16")]; + tensor var_372_begin_0 = const()[name = tensor("op_372_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_372_end_0 = const()[name = tensor("op_372_end_0"), val = tensor([1, 8, 174, 17])]; + tensor var_372_end_mask_0 = const()[name = tensor("op_372_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_372_cast_fp16 = slice_by_index(begin = var_372_begin_0, end = var_372_end_0, end_mask = var_372_end_mask_0, x = x_11_cast_fp16)[name = tensor("op_372_cast_fp16")]; + tensor var_373 = const()[name = tensor("op_373"), val = tensor([1, 8, 17, 173])]; + tensor matrix_bd_1_cast_fp16 = reshape(shape = var_373, x = var_372_cast_fp16)[name = tensor("matrix_bd_1_cast_fp16")]; + tensor matrix_ac_1_transpose_x_0 = const()[name = tensor("matrix_ac_1_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_1_transpose_y_0 = const()[name = tensor("matrix_ac_1_transpose_y_0"), val = tensor(false)]; + tensor transpose_51_perm_0 = const()[name = tensor("transpose_51_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_52_perm_0 = const()[name = tensor("transpose_52_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_52 = transpose(perm = transpose_52_perm_0, x = k_1_cast_fp16)[name = tensor("transpose_235")]; + tensor transpose_51 = transpose(perm = transpose_51_perm_0, x = var_356_cast_fp16)[name = tensor("transpose_236")]; + tensor matrix_ac_1_cast_fp16 = matmul(transpose_x = matrix_ac_1_transpose_x_0, transpose_y = matrix_ac_1_transpose_y_0, x = transpose_51, y = transpose_52)[name = tensor("matrix_ac_1_cast_fp16")]; + tensor matrix_bd_3_begin_0 = const()[name = tensor("matrix_bd_3_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_3_end_0 = const()[name = tensor("matrix_bd_3_end_0"), val = tensor([1, 8, 17, 87])]; + tensor matrix_bd_3_end_mask_0 = const()[name = tensor("matrix_bd_3_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_3_cast_fp16 = slice_by_index(begin = matrix_bd_3_begin_0, end = matrix_bd_3_end_0, end_mask = matrix_bd_3_end_mask_0, x = matrix_bd_1_cast_fp16)[name = tensor("matrix_bd_3_cast_fp16")]; + tensor var_382_cast_fp16 = add(x = matrix_ac_1_cast_fp16, y = matrix_bd_3_cast_fp16)[name = tensor("op_382_cast_fp16")]; + tensor _inversed_scores_1_y_0_to_fp16 = const()[name = tensor("_inversed_scores_1_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_1_cast_fp16 = mul(x = var_382_cast_fp16, y = _inversed_scores_1_y_0_to_fp16)[name = tensor("_inversed_scores_1_cast_fp16")]; + tensor mask_3_axes_0 = const()[name = tensor("mask_3_axes_0"), val = tensor([1])]; + tensor mask_3 = expand_dims(axes = mask_3_axes_0, x = mask_1)[name = tensor("mask_3")]; + tensor var_41_to_fp16 = const()[name = tensor("op_41_to_fp16"), val = tensor(-0x1.388p+13)]; + tensor scores_3_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_1_cast_fp16, cond = mask_3)[name = tensor("scores_3_cast_fp16")]; + tensor var_388_cast_fp16 = softmax(axis = var_62, x = scores_3_cast_fp16)[name = tensor("op_388_cast_fp16")]; + tensor var_40_to_fp16 = const()[name = tensor("op_40_to_fp16"), val = tensor(0x0p+0)]; + tensor input_41_cast_fp16 = select(a = var_40_to_fp16, b = var_388_cast_fp16, cond = mask_3)[name = tensor("input_41_cast_fp16")]; + tensor x_13_transpose_x_0 = const()[name = tensor("x_13_transpose_x_0"), val = tensor(false)]; + tensor x_13_transpose_y_0 = const()[name = tensor("x_13_transpose_y_0"), val = tensor(false)]; + tensor value_3_cast_fp16 = transpose(perm = value_3_perm_0, x = v_1_cast_fp16)[name = tensor("transpose_238")]; + tensor x_13_cast_fp16 = matmul(transpose_x = x_13_transpose_x_0, transpose_y = x_13_transpose_y_0, x = input_41_cast_fp16, y = value_3_cast_fp16)[name = tensor("x_13_cast_fp16")]; + tensor var_392_perm_0 = const()[name = tensor("op_392_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_393 = const()[name = tensor("op_393"), val = tensor([1, -1, 512])]; + tensor var_392_cast_fp16 = transpose(perm = var_392_perm_0, x = x_13_cast_fp16)[name = tensor("transpose_234")]; + tensor input_43_cast_fp16 = reshape(shape = var_393, x = var_392_cast_fp16)[name = tensor("input_43_cast_fp16")]; + tensor encoder_layers_0_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10693568)))]; + tensor linear_7_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_self_attn_linear_out_weight_to_fp16, x = input_43_cast_fp16)[name = tensor("linear_7_cast_fp16")]; + tensor input_47_cast_fp16 = add(x = input_37_cast_fp16, y = linear_7_cast_fp16)[name = tensor("input_47_cast_fp16")]; + tensor x_17_axes_0 = const()[name = tensor("x_17_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11217920)))]; + tensor encoder_layers_0_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11219008)))]; + tensor x_17_cast_fp16 = layer_norm(axes = x_17_axes_0, beta = encoder_layers_0_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_norm_conv_weight_to_fp16, x = input_47_cast_fp16)[name = tensor("x_17_cast_fp16")]; + tensor input_49_perm_0 = const()[name = tensor("input_49_perm_0"), val = tensor([0, 2, 1])]; + tensor input_51_pad_type_0 = const()[name = tensor("input_51_pad_type_0"), val = tensor("valid")]; + tensor input_51_strides_0 = const()[name = tensor("input_51_strides_0"), val = tensor([1])]; + tensor input_51_pad_0 = const()[name = tensor("input_51_pad_0"), val = tensor([0, 0])]; + tensor input_51_dilations_0 = const()[name = tensor("input_51_dilations_0"), val = tensor([1])]; + tensor input_51_groups_0 = const()[name = tensor("input_51_groups_0"), val = tensor(1)]; + tensor encoder_layers_0_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_0_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11220096)))]; + tensor input_49_cast_fp16 = transpose(perm = input_49_perm_0, x = x_17_cast_fp16)[name = tensor("transpose_233")]; + tensor input_51_cast_fp16 = conv(dilations = input_51_dilations_0, groups = input_51_groups_0, pad = input_51_pad_0, pad_type = input_51_pad_type_0, strides = input_51_strides_0, weight = encoder_layers_0_conv_pointwise_conv1_weight_to_fp16, x = input_49_cast_fp16)[name = tensor("input_51_cast_fp16")]; + tensor x_19_split_num_splits_0 = const()[name = tensor("x_19_split_num_splits_0"), val = tensor(2)]; + tensor x_19_split_axis_0 = const()[name = tensor("x_19_split_axis_0"), val = tensor(1)]; + tensor x_19_split_cast_fp16_0, tensor x_19_split_cast_fp16_1 = split(axis = x_19_split_axis_0, num_splits = x_19_split_num_splits_0, x = input_51_cast_fp16)[name = tensor("x_19_split_cast_fp16")]; + tensor x_19_split_1_sigmoid_cast_fp16 = sigmoid(x = x_19_split_cast_fp16_1)[name = tensor("x_19_split_1_sigmoid_cast_fp16")]; + tensor x_19_cast_fp16 = mul(x = x_19_split_cast_fp16_0, y = x_19_split_1_sigmoid_cast_fp16)[name = tensor("x_19_cast_fp16")]; + tensor var_418_axes_0 = const()[name = tensor("op_418_axes_0"), val = tensor([1])]; + tensor var_418 = expand_dims(axes = var_418_axes_0, x = pad_mask)[name = tensor("op_418")]; + tensor input_53_cast_fp16 = select(a = var_40_to_fp16, b = x_19_cast_fp16, cond = var_418)[name = tensor("input_53_cast_fp16")]; + tensor new_x_3_interleave_0 = const()[name = tensor("new_x_3_interleave_0"), val = tensor(false)]; + tensor new_x_3_cast_fp16 = concat(axis = var_62, interleave = new_x_3_interleave_0, values = (cache_3_cast_fp16, input_53_cast_fp16))[name = tensor("new_x_3_cast_fp16")]; + tensor var_431_begin_0 = const()[name = tensor("op_431_begin_0"), val = tensor([0, 0, 17])]; + tensor var_431_end_0 = const()[name = tensor("op_431_end_0"), val = tensor([1, 512, 25])]; + tensor var_431_end_mask_0 = const()[name = tensor("op_431_end_mask_0"), val = tensor([true, true, true])]; + tensor var_431_cast_fp16 = slice_by_index(begin = var_431_begin_0, end = var_431_end_0, end_mask = var_431_end_mask_0, x = new_x_3_cast_fp16)[name = tensor("op_431_cast_fp16")]; + tensor x_21_pad_type_0 = const()[name = tensor("x_21_pad_type_0"), val = tensor("valid")]; + tensor x_21_groups_0 = const()[name = tensor("x_21_groups_0"), val = tensor(512)]; + tensor x_21_strides_0 = const()[name = tensor("x_21_strides_0"), val = tensor([1])]; + tensor x_21_pad_0 = const()[name = tensor("x_21_pad_0"), val = tensor([0, 0])]; + tensor x_21_dilations_0 = const()[name = tensor("x_21_dilations_0"), val = tensor([1])]; + tensor encoder_layers_0_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_0_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12268736)))]; + tensor x_21_cast_fp16 = conv(dilations = x_21_dilations_0, groups = x_21_groups_0, pad = x_21_pad_0, pad_type = x_21_pad_type_0, strides = x_21_strides_0, weight = encoder_layers_0_conv_depthwise_conv_weight_to_fp16, x = new_x_3_cast_fp16)[name = tensor("x_21_cast_fp16")]; + tensor input_55_perm_0 = const()[name = tensor("input_55_perm_0"), val = tensor([0, 2, 1])]; + tensor x_23_axes_0 = const()[name = tensor("x_23_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_0_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12278016)))]; + tensor encoder_layers_0_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_0_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12279104)))]; + tensor input_55_cast_fp16 = transpose(perm = input_55_perm_0, x = x_21_cast_fp16)[name = tensor("transpose_232")]; + tensor x_23_cast_fp16 = layer_norm(axes = x_23_axes_0, beta = encoder_layers_0_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_conv_batch_norm_weight_to_fp16, x = input_55_cast_fp16)[name = tensor("x_23_cast_fp16")]; + tensor input_57_perm_0 = const()[name = tensor("input_57_perm_0"), val = tensor([0, 2, 1])]; + tensor input_57_cast_fp16 = transpose(perm = input_57_perm_0, x = x_23_cast_fp16)[name = tensor("transpose_231")]; + tensor input_59_cast_fp16 = silu(x = input_57_cast_fp16)[name = tensor("input_59_cast_fp16")]; + tensor x_25_pad_type_0 = const()[name = tensor("x_25_pad_type_0"), val = tensor("valid")]; + tensor x_25_strides_0 = const()[name = tensor("x_25_strides_0"), val = tensor([1])]; + tensor x_25_pad_0 = const()[name = tensor("x_25_pad_0"), val = tensor([0, 0])]; + tensor x_25_dilations_0 = const()[name = tensor("x_25_dilations_0"), val = tensor([1])]; + tensor x_25_groups_0 = const()[name = tensor("x_25_groups_0"), val = tensor(1)]; + tensor encoder_layers_0_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_0_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12280192)))]; + tensor x_25_cast_fp16 = conv(dilations = x_25_dilations_0, groups = x_25_groups_0, pad = x_25_pad_0, pad_type = x_25_pad_type_0, strides = x_25_strides_0, weight = encoder_layers_0_conv_pointwise_conv2_weight_to_fp16, x = input_59_cast_fp16)[name = tensor("x_25_cast_fp16")]; + tensor input_61_perm_0 = const()[name = tensor("input_61_perm_0"), val = tensor([0, 2, 1])]; + tensor input_61_cast_fp16 = transpose(perm = input_61_perm_0, x = x_25_cast_fp16)[name = tensor("transpose_230")]; + tensor input_63_cast_fp16 = add(x = input_47_cast_fp16, y = input_61_cast_fp16)[name = tensor("input_63_cast_fp16")]; + tensor input_65_axes_0 = const()[name = tensor("input_65_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12804544)))]; + tensor encoder_layers_0_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12805632)))]; + tensor input_65_cast_fp16 = layer_norm(axes = input_65_axes_0, beta = encoder_layers_0_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_norm_feed_forward2_weight_to_fp16, x = input_63_cast_fp16)[name = tensor("input_65_cast_fp16")]; + tensor encoder_layers_0_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_0_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12806720)))]; + tensor linear_8_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_0_feed_forward2_linear1_weight_to_fp16, x = input_65_cast_fp16)[name = tensor("linear_8_cast_fp16")]; + tensor input_69_cast_fp16 = silu(x = linear_8_cast_fp16)[name = tensor("input_69_cast_fp16")]; + tensor encoder_layers_0_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_0_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14903936)))]; + tensor linear_9_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_feed_forward2_linear2_weight_to_fp16, x = input_69_cast_fp16)[name = tensor("linear_9_cast_fp16")]; + tensor var_472_to_fp16 = const()[name = tensor("op_472_to_fp16"), val = tensor(0x1p-1)]; + tensor var_473_cast_fp16 = mul(x = linear_9_cast_fp16, y = var_472_to_fp16)[name = tensor("op_473_cast_fp16")]; + tensor input_75_cast_fp16 = add(x = input_63_cast_fp16, y = var_473_cast_fp16)[name = tensor("input_75_cast_fp16")]; + tensor input_77_axes_0 = const()[name = tensor("input_77_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17001152)))]; + tensor encoder_layers_0_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17002240)))]; + tensor input_77_cast_fp16 = layer_norm(axes = input_77_axes_0, beta = encoder_layers_0_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_norm_out_weight_to_fp16, x = input_75_cast_fp16)[name = tensor("input_77_cast_fp16")]; + tensor cache_5_begin_0 = const()[name = tensor("cache_5_begin_0"), val = tensor([1, 0, 0, 0])]; + tensor cache_5_end_0 = const()[name = tensor("cache_5_end_0"), val = tensor([2, 1, 70, 512])]; + tensor cache_5_end_mask_0 = const()[name = tensor("cache_5_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_5_squeeze_mask_0 = const()[name = tensor("cache_5_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_5_cast_fp16 = slice_by_index(begin = cache_5_begin_0, end = cache_5_end_0, end_mask = cache_5_end_mask_0, squeeze_mask = cache_5_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_5_cast_fp16")]; + tensor cache_7_begin_0 = const()[name = tensor("cache_7_begin_0"), val = tensor([1, 0, 0, 0])]; + tensor cache_7_end_0 = const()[name = tensor("cache_7_end_0"), val = tensor([2, 1, 512, 8])]; + tensor cache_7_end_mask_0 = const()[name = tensor("cache_7_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_7_squeeze_mask_0 = const()[name = tensor("cache_7_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_7_cast_fp16 = slice_by_index(begin = cache_7_begin_0, end = cache_7_end_0, end_mask = cache_7_end_mask_0, squeeze_mask = cache_7_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_7_cast_fp16")]; + tensor input_79_axes_0 = const()[name = tensor("input_79_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17003328)))]; + tensor encoder_layers_1_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17004416)))]; + tensor input_79_cast_fp16 = layer_norm(axes = input_79_axes_0, beta = encoder_layers_1_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_norm_feed_forward1_weight_to_fp16, x = input_77_cast_fp16)[name = tensor("input_79_cast_fp16")]; + tensor encoder_layers_1_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_1_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17005504)))]; + tensor linear_10_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_1_feed_forward1_linear1_weight_to_fp16, x = input_79_cast_fp16)[name = tensor("linear_10_cast_fp16")]; + tensor input_83_cast_fp16 = silu(x = linear_10_cast_fp16)[name = tensor("input_83_cast_fp16")]; + tensor encoder_layers_1_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_1_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19102720)))]; + tensor linear_11_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_feed_forward1_linear2_weight_to_fp16, x = input_83_cast_fp16)[name = tensor("linear_11_cast_fp16")]; + tensor var_507_to_fp16 = const()[name = tensor("op_507_to_fp16"), val = tensor(0x1p-1)]; + tensor var_508_cast_fp16 = mul(x = linear_11_cast_fp16, y = var_507_to_fp16)[name = tensor("op_508_cast_fp16")]; + tensor input_89_cast_fp16 = add(x = input_77_cast_fp16, y = var_508_cast_fp16)[name = tensor("input_89_cast_fp16")]; + tensor key_3_axes_0 = const()[name = tensor("key_3_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21199936)))]; + tensor encoder_layers_1_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21201024)))]; + tensor key_3_cast_fp16 = layer_norm(axes = key_3_axes_0, beta = encoder_layers_1_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_norm_self_att_weight_to_fp16, x = input_89_cast_fp16)[name = tensor("key_3_cast_fp16")]; + tensor input_91_interleave_0 = const()[name = tensor("input_91_interleave_0"), val = tensor(false)]; + tensor input_91_cast_fp16 = concat(axis = var_64, interleave = input_91_interleave_0, values = (cache_5_cast_fp16, key_3_cast_fp16))[name = tensor("input_91_cast_fp16")]; + tensor var_530_begin_0 = const()[name = tensor("op_530_begin_0"), val = tensor([0, 17, 0])]; + tensor var_530_end_0 = const()[name = tensor("op_530_end_0"), val = tensor([1, 70, 512])]; + tensor var_530_end_mask_0 = const()[name = tensor("op_530_end_mask_0"), val = tensor([true, true, true])]; + tensor var_530_cast_fp16 = slice_by_index(begin = var_530_begin_0, end = var_530_end_0, end_mask = var_530_end_mask_0, x = cache_5_cast_fp16)[name = tensor("op_530_cast_fp16")]; + tensor var_536_interleave_0 = const()[name = tensor("op_536_interleave_0"), val = tensor(false)]; + tensor var_536_cast_fp16 = concat(axis = var_64, interleave = var_536_interleave_0, values = (var_530_cast_fp16, key_3_cast_fp16))[name = tensor("op_536_cast_fp16")]; + tensor encoder_layers_1_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21202112)))]; + tensor linear_12_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_self_attn_linear_q_weight_to_fp16, x = key_3_cast_fp16)[name = tensor("linear_12_cast_fp16")]; + tensor var_540 = const()[name = tensor("op_540"), val = tensor([1, -1, 8, 64])]; + tensor q_7_cast_fp16 = reshape(shape = var_540, x = linear_12_cast_fp16)[name = tensor("q_7_cast_fp16")]; + tensor encoder_layers_1_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21726464)))]; + tensor linear_13_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_self_attn_linear_k_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("linear_13_cast_fp16")]; + tensor var_544 = const()[name = tensor("op_544"), val = tensor([1, -1, 8, 64])]; + tensor k_5_cast_fp16 = reshape(shape = var_544, x = linear_13_cast_fp16)[name = tensor("k_5_cast_fp16")]; + tensor encoder_layers_1_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22250816)))]; + tensor linear_14_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_self_attn_linear_v_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("linear_14_cast_fp16")]; + tensor var_548 = const()[name = tensor("op_548"), val = tensor([1, -1, 8, 64])]; + tensor v_3_cast_fp16 = reshape(shape = var_548, x = linear_14_cast_fp16)[name = tensor("v_3_cast_fp16")]; + tensor value_5_perm_0 = const()[name = tensor("value_5_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_1_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22775168)))]; + tensor var_560_cast_fp16 = add(x = q_7_cast_fp16, y = encoder_layers_1_self_attn_pos_bias_u_to_fp16)[name = tensor("op_560_cast_fp16")]; + tensor encoder_layers_1_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22776256)))]; + tensor var_562_cast_fp16 = add(x = q_7_cast_fp16, y = encoder_layers_1_self_attn_pos_bias_v_to_fp16)[name = tensor("op_562_cast_fp16")]; + tensor q_with_bias_v_3_perm_0 = const()[name = tensor("q_with_bias_v_3_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_33_transpose_x_0 = const()[name = tensor("x_33_transpose_x_0"), val = tensor(false)]; + tensor x_33_transpose_y_0 = const()[name = tensor("x_33_transpose_y_0"), val = tensor(false)]; + tensor var_564_to_fp16 = const()[name = tensor("op_564_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22777344)))]; + tensor q_with_bias_v_3_cast_fp16 = transpose(perm = q_with_bias_v_3_perm_0, x = var_562_cast_fp16)[name = tensor("transpose_228")]; + tensor x_33_cast_fp16 = matmul(transpose_x = x_33_transpose_x_0, transpose_y = x_33_transpose_y_0, x = q_with_bias_v_3_cast_fp16, y = var_564_to_fp16)[name = tensor("x_33_cast_fp16")]; + tensor x_35_pad_0 = const()[name = tensor("x_35_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_35_mode_0 = const()[name = tensor("x_35_mode_0"), val = tensor("constant")]; + tensor const_36_to_fp16 = const()[name = tensor("const_36_to_fp16"), val = tensor(0x0p+0)]; + tensor x_35_cast_fp16 = pad(constant_val = const_36_to_fp16, mode = x_35_mode_0, pad = x_35_pad_0, x = x_33_cast_fp16)[name = tensor("x_35_cast_fp16")]; + tensor var_572 = const()[name = tensor("op_572"), val = tensor([1, 8, -1, 17])]; + tensor x_37_cast_fp16 = reshape(shape = var_572, x = x_35_cast_fp16)[name = tensor("x_37_cast_fp16")]; + tensor var_576_begin_0 = const()[name = tensor("op_576_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_576_end_0 = const()[name = tensor("op_576_end_0"), val = tensor([1, 8, 174, 17])]; + tensor var_576_end_mask_0 = const()[name = tensor("op_576_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_576_cast_fp16 = slice_by_index(begin = var_576_begin_0, end = var_576_end_0, end_mask = var_576_end_mask_0, x = x_37_cast_fp16)[name = tensor("op_576_cast_fp16")]; + tensor var_577 = const()[name = tensor("op_577"), val = tensor([1, 8, 17, 173])]; + tensor matrix_bd_5_cast_fp16 = reshape(shape = var_577, x = var_576_cast_fp16)[name = tensor("matrix_bd_5_cast_fp16")]; + tensor matrix_ac_3_transpose_x_0 = const()[name = tensor("matrix_ac_3_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_3_transpose_y_0 = const()[name = tensor("matrix_ac_3_transpose_y_0"), val = tensor(false)]; + tensor transpose_53_perm_0 = const()[name = tensor("transpose_53_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_54_perm_0 = const()[name = tensor("transpose_54_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_54 = transpose(perm = transpose_54_perm_0, x = k_5_cast_fp16)[name = tensor("transpose_226")]; + tensor transpose_53 = transpose(perm = transpose_53_perm_0, x = var_560_cast_fp16)[name = tensor("transpose_227")]; + tensor matrix_ac_3_cast_fp16 = matmul(transpose_x = matrix_ac_3_transpose_x_0, transpose_y = matrix_ac_3_transpose_y_0, x = transpose_53, y = transpose_54)[name = tensor("matrix_ac_3_cast_fp16")]; + tensor matrix_bd_7_begin_0 = const()[name = tensor("matrix_bd_7_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_7_end_0 = const()[name = tensor("matrix_bd_7_end_0"), val = tensor([1, 8, 17, 87])]; + tensor matrix_bd_7_end_mask_0 = const()[name = tensor("matrix_bd_7_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_7_cast_fp16 = slice_by_index(begin = matrix_bd_7_begin_0, end = matrix_bd_7_end_0, end_mask = matrix_bd_7_end_mask_0, x = matrix_bd_5_cast_fp16)[name = tensor("matrix_bd_7_cast_fp16")]; + tensor var_586_cast_fp16 = add(x = matrix_ac_3_cast_fp16, y = matrix_bd_7_cast_fp16)[name = tensor("op_586_cast_fp16")]; + tensor _inversed_scores_5_y_0_to_fp16 = const()[name = tensor("_inversed_scores_5_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_5_cast_fp16 = mul(x = var_586_cast_fp16, y = _inversed_scores_5_y_0_to_fp16)[name = tensor("_inversed_scores_5_cast_fp16")]; + tensor scores_7_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_5_cast_fp16, cond = mask_3)[name = tensor("scores_7_cast_fp16")]; + tensor var_592_cast_fp16 = softmax(axis = var_62, x = scores_7_cast_fp16)[name = tensor("op_592_cast_fp16")]; + tensor input_93_cast_fp16 = select(a = var_40_to_fp16, b = var_592_cast_fp16, cond = mask_3)[name = tensor("input_93_cast_fp16")]; + tensor x_39_transpose_x_0 = const()[name = tensor("x_39_transpose_x_0"), val = tensor(false)]; + tensor x_39_transpose_y_0 = const()[name = tensor("x_39_transpose_y_0"), val = tensor(false)]; + tensor value_5_cast_fp16 = transpose(perm = value_5_perm_0, x = v_3_cast_fp16)[name = tensor("transpose_229")]; + tensor x_39_cast_fp16 = matmul(transpose_x = x_39_transpose_x_0, transpose_y = x_39_transpose_y_0, x = input_93_cast_fp16, y = value_5_cast_fp16)[name = tensor("x_39_cast_fp16")]; + tensor var_596_perm_0 = const()[name = tensor("op_596_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_597 = const()[name = tensor("op_597"), val = tensor([1, -1, 512])]; + tensor var_596_cast_fp16 = transpose(perm = var_596_perm_0, x = x_39_cast_fp16)[name = tensor("transpose_225")]; + tensor input_95_cast_fp16 = reshape(shape = var_597, x = var_596_cast_fp16)[name = tensor("input_95_cast_fp16")]; + tensor encoder_layers_1_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22954560)))]; + tensor linear_16_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_self_attn_linear_out_weight_to_fp16, x = input_95_cast_fp16)[name = tensor("linear_16_cast_fp16")]; + tensor input_99_cast_fp16 = add(x = input_89_cast_fp16, y = linear_16_cast_fp16)[name = tensor("input_99_cast_fp16")]; + tensor x_43_axes_0 = const()[name = tensor("x_43_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23478912)))]; + tensor encoder_layers_1_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23480000)))]; + tensor x_43_cast_fp16 = layer_norm(axes = x_43_axes_0, beta = encoder_layers_1_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_norm_conv_weight_to_fp16, x = input_99_cast_fp16)[name = tensor("x_43_cast_fp16")]; + tensor input_101_perm_0 = const()[name = tensor("input_101_perm_0"), val = tensor([0, 2, 1])]; + tensor input_103_pad_type_0 = const()[name = tensor("input_103_pad_type_0"), val = tensor("valid")]; + tensor input_103_strides_0 = const()[name = tensor("input_103_strides_0"), val = tensor([1])]; + tensor input_103_pad_0 = const()[name = tensor("input_103_pad_0"), val = tensor([0, 0])]; + tensor input_103_dilations_0 = const()[name = tensor("input_103_dilations_0"), val = tensor([1])]; + tensor input_103_groups_0 = const()[name = tensor("input_103_groups_0"), val = tensor(1)]; + tensor encoder_layers_1_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_1_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23481088)))]; + tensor input_101_cast_fp16 = transpose(perm = input_101_perm_0, x = x_43_cast_fp16)[name = tensor("transpose_224")]; + tensor input_103_cast_fp16 = conv(dilations = input_103_dilations_0, groups = input_103_groups_0, pad = input_103_pad_0, pad_type = input_103_pad_type_0, strides = input_103_strides_0, weight = encoder_layers_1_conv_pointwise_conv1_weight_to_fp16, x = input_101_cast_fp16)[name = tensor("input_103_cast_fp16")]; + tensor x_45_split_num_splits_0 = const()[name = tensor("x_45_split_num_splits_0"), val = tensor(2)]; + tensor x_45_split_axis_0 = const()[name = tensor("x_45_split_axis_0"), val = tensor(1)]; + tensor x_45_split_cast_fp16_0, tensor x_45_split_cast_fp16_1 = split(axis = x_45_split_axis_0, num_splits = x_45_split_num_splits_0, x = input_103_cast_fp16)[name = tensor("x_45_split_cast_fp16")]; + tensor x_45_split_1_sigmoid_cast_fp16 = sigmoid(x = x_45_split_cast_fp16_1)[name = tensor("x_45_split_1_sigmoid_cast_fp16")]; + tensor x_45_cast_fp16 = mul(x = x_45_split_cast_fp16_0, y = x_45_split_1_sigmoid_cast_fp16)[name = tensor("x_45_cast_fp16")]; + tensor input_105_cast_fp16 = select(a = var_40_to_fp16, b = x_45_cast_fp16, cond = var_418)[name = tensor("input_105_cast_fp16")]; + tensor new_x_7_interleave_0 = const()[name = tensor("new_x_7_interleave_0"), val = tensor(false)]; + tensor new_x_7_cast_fp16 = concat(axis = var_62, interleave = new_x_7_interleave_0, values = (cache_7_cast_fp16, input_105_cast_fp16))[name = tensor("new_x_7_cast_fp16")]; + tensor var_635_begin_0 = const()[name = tensor("op_635_begin_0"), val = tensor([0, 0, 17])]; + tensor var_635_end_0 = const()[name = tensor("op_635_end_0"), val = tensor([1, 512, 25])]; + tensor var_635_end_mask_0 = const()[name = tensor("op_635_end_mask_0"), val = tensor([true, true, true])]; + tensor var_635_cast_fp16 = slice_by_index(begin = var_635_begin_0, end = var_635_end_0, end_mask = var_635_end_mask_0, x = new_x_7_cast_fp16)[name = tensor("op_635_cast_fp16")]; + tensor x_47_pad_type_0 = const()[name = tensor("x_47_pad_type_0"), val = tensor("valid")]; + tensor x_47_groups_0 = const()[name = tensor("x_47_groups_0"), val = tensor(512)]; + tensor x_47_strides_0 = const()[name = tensor("x_47_strides_0"), val = tensor([1])]; + tensor x_47_pad_0 = const()[name = tensor("x_47_pad_0"), val = tensor([0, 0])]; + tensor x_47_dilations_0 = const()[name = tensor("x_47_dilations_0"), val = tensor([1])]; + tensor encoder_layers_1_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_1_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24529728)))]; + tensor x_47_cast_fp16 = conv(dilations = x_47_dilations_0, groups = x_47_groups_0, pad = x_47_pad_0, pad_type = x_47_pad_type_0, strides = x_47_strides_0, weight = encoder_layers_1_conv_depthwise_conv_weight_to_fp16, x = new_x_7_cast_fp16)[name = tensor("x_47_cast_fp16")]; + tensor input_107_perm_0 = const()[name = tensor("input_107_perm_0"), val = tensor([0, 2, 1])]; + tensor x_49_axes_0 = const()[name = tensor("x_49_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_1_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24539008)))]; + tensor encoder_layers_1_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_1_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24540096)))]; + tensor input_107_cast_fp16 = transpose(perm = input_107_perm_0, x = x_47_cast_fp16)[name = tensor("transpose_223")]; + tensor x_49_cast_fp16 = layer_norm(axes = x_49_axes_0, beta = encoder_layers_1_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_conv_batch_norm_weight_to_fp16, x = input_107_cast_fp16)[name = tensor("x_49_cast_fp16")]; + tensor input_109_perm_0 = const()[name = tensor("input_109_perm_0"), val = tensor([0, 2, 1])]; + tensor input_109_cast_fp16 = transpose(perm = input_109_perm_0, x = x_49_cast_fp16)[name = tensor("transpose_222")]; + tensor input_111_cast_fp16 = silu(x = input_109_cast_fp16)[name = tensor("input_111_cast_fp16")]; + tensor x_51_pad_type_0 = const()[name = tensor("x_51_pad_type_0"), val = tensor("valid")]; + tensor x_51_strides_0 = const()[name = tensor("x_51_strides_0"), val = tensor([1])]; + tensor x_51_pad_0 = const()[name = tensor("x_51_pad_0"), val = tensor([0, 0])]; + tensor x_51_dilations_0 = const()[name = tensor("x_51_dilations_0"), val = tensor([1])]; + tensor x_51_groups_0 = const()[name = tensor("x_51_groups_0"), val = tensor(1)]; + tensor encoder_layers_1_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_1_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24541184)))]; + tensor x_51_cast_fp16 = conv(dilations = x_51_dilations_0, groups = x_51_groups_0, pad = x_51_pad_0, pad_type = x_51_pad_type_0, strides = x_51_strides_0, weight = encoder_layers_1_conv_pointwise_conv2_weight_to_fp16, x = input_111_cast_fp16)[name = tensor("x_51_cast_fp16")]; + tensor input_113_perm_0 = const()[name = tensor("input_113_perm_0"), val = tensor([0, 2, 1])]; + tensor input_113_cast_fp16 = transpose(perm = input_113_perm_0, x = x_51_cast_fp16)[name = tensor("transpose_221")]; + tensor input_115_cast_fp16 = add(x = input_99_cast_fp16, y = input_113_cast_fp16)[name = tensor("input_115_cast_fp16")]; + tensor input_117_axes_0 = const()[name = tensor("input_117_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25065536)))]; + tensor encoder_layers_1_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25066624)))]; + tensor input_117_cast_fp16 = layer_norm(axes = input_117_axes_0, beta = encoder_layers_1_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_norm_feed_forward2_weight_to_fp16, x = input_115_cast_fp16)[name = tensor("input_117_cast_fp16")]; + tensor encoder_layers_1_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_1_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25067712)))]; + tensor linear_17_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_1_feed_forward2_linear1_weight_to_fp16, x = input_117_cast_fp16)[name = tensor("linear_17_cast_fp16")]; + tensor input_121_cast_fp16 = silu(x = linear_17_cast_fp16)[name = tensor("input_121_cast_fp16")]; + tensor encoder_layers_1_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_1_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27164928)))]; + tensor linear_18_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_feed_forward2_linear2_weight_to_fp16, x = input_121_cast_fp16)[name = tensor("linear_18_cast_fp16")]; + tensor var_676_to_fp16 = const()[name = tensor("op_676_to_fp16"), val = tensor(0x1p-1)]; + tensor var_677_cast_fp16 = mul(x = linear_18_cast_fp16, y = var_676_to_fp16)[name = tensor("op_677_cast_fp16")]; + tensor input_127_cast_fp16 = add(x = input_115_cast_fp16, y = var_677_cast_fp16)[name = tensor("input_127_cast_fp16")]; + tensor input_129_axes_0 = const()[name = tensor("input_129_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29262144)))]; + tensor encoder_layers_1_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29263232)))]; + tensor input_129_cast_fp16 = layer_norm(axes = input_129_axes_0, beta = encoder_layers_1_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_norm_out_weight_to_fp16, x = input_127_cast_fp16)[name = tensor("input_129_cast_fp16")]; + tensor cache_9_begin_0 = const()[name = tensor("cache_9_begin_0"), val = tensor([2, 0, 0, 0])]; + tensor cache_9_end_0 = const()[name = tensor("cache_9_end_0"), val = tensor([3, 1, 70, 512])]; + tensor cache_9_end_mask_0 = const()[name = tensor("cache_9_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_9_squeeze_mask_0 = const()[name = tensor("cache_9_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_9_cast_fp16 = slice_by_index(begin = cache_9_begin_0, end = cache_9_end_0, end_mask = cache_9_end_mask_0, squeeze_mask = cache_9_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_9_cast_fp16")]; + tensor cache_11_begin_0 = const()[name = tensor("cache_11_begin_0"), val = tensor([2, 0, 0, 0])]; + tensor cache_11_end_0 = const()[name = tensor("cache_11_end_0"), val = tensor([3, 1, 512, 8])]; + tensor cache_11_end_mask_0 = const()[name = tensor("cache_11_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_11_squeeze_mask_0 = const()[name = tensor("cache_11_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_11_cast_fp16 = slice_by_index(begin = cache_11_begin_0, end = cache_11_end_0, end_mask = cache_11_end_mask_0, squeeze_mask = cache_11_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_11_cast_fp16")]; + tensor input_131_axes_0 = const()[name = tensor("input_131_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29264320)))]; + tensor encoder_layers_2_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29265408)))]; + tensor input_131_cast_fp16 = layer_norm(axes = input_131_axes_0, beta = encoder_layers_2_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_norm_feed_forward1_weight_to_fp16, x = input_129_cast_fp16)[name = tensor("input_131_cast_fp16")]; + tensor encoder_layers_2_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_2_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29266496)))]; + tensor linear_19_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_2_feed_forward1_linear1_weight_to_fp16, x = input_131_cast_fp16)[name = tensor("linear_19_cast_fp16")]; + tensor input_135_cast_fp16 = silu(x = linear_19_cast_fp16)[name = tensor("input_135_cast_fp16")]; + tensor encoder_layers_2_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_2_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31363712)))]; + tensor linear_20_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_feed_forward1_linear2_weight_to_fp16, x = input_135_cast_fp16)[name = tensor("linear_20_cast_fp16")]; + tensor var_711_to_fp16 = const()[name = tensor("op_711_to_fp16"), val = tensor(0x1p-1)]; + tensor var_712_cast_fp16 = mul(x = linear_20_cast_fp16, y = var_711_to_fp16)[name = tensor("op_712_cast_fp16")]; + tensor input_141_cast_fp16 = add(x = input_129_cast_fp16, y = var_712_cast_fp16)[name = tensor("input_141_cast_fp16")]; + tensor key_5_axes_0 = const()[name = tensor("key_5_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33460928)))]; + tensor encoder_layers_2_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33462016)))]; + tensor key_5_cast_fp16 = layer_norm(axes = key_5_axes_0, beta = encoder_layers_2_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_norm_self_att_weight_to_fp16, x = input_141_cast_fp16)[name = tensor("key_5_cast_fp16")]; + tensor input_143_interleave_0 = const()[name = tensor("input_143_interleave_0"), val = tensor(false)]; + tensor input_143_cast_fp16 = concat(axis = var_64, interleave = input_143_interleave_0, values = (cache_9_cast_fp16, key_5_cast_fp16))[name = tensor("input_143_cast_fp16")]; + tensor var_734_begin_0 = const()[name = tensor("op_734_begin_0"), val = tensor([0, 17, 0])]; + tensor var_734_end_0 = const()[name = tensor("op_734_end_0"), val = tensor([1, 70, 512])]; + tensor var_734_end_mask_0 = const()[name = tensor("op_734_end_mask_0"), val = tensor([true, true, true])]; + tensor var_734_cast_fp16 = slice_by_index(begin = var_734_begin_0, end = var_734_end_0, end_mask = var_734_end_mask_0, x = cache_9_cast_fp16)[name = tensor("op_734_cast_fp16")]; + tensor var_740_interleave_0 = const()[name = tensor("op_740_interleave_0"), val = tensor(false)]; + tensor var_740_cast_fp16 = concat(axis = var_64, interleave = var_740_interleave_0, values = (var_734_cast_fp16, key_5_cast_fp16))[name = tensor("op_740_cast_fp16")]; + tensor encoder_layers_2_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33463104)))]; + tensor linear_21_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_self_attn_linear_q_weight_to_fp16, x = key_5_cast_fp16)[name = tensor("linear_21_cast_fp16")]; + tensor var_744 = const()[name = tensor("op_744"), val = tensor([1, -1, 8, 64])]; + tensor q_13_cast_fp16 = reshape(shape = var_744, x = linear_21_cast_fp16)[name = tensor("q_13_cast_fp16")]; + tensor encoder_layers_2_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33987456)))]; + tensor linear_22_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_self_attn_linear_k_weight_to_fp16, x = input_143_cast_fp16)[name = tensor("linear_22_cast_fp16")]; + tensor var_748 = const()[name = tensor("op_748"), val = tensor([1, -1, 8, 64])]; + tensor k_9_cast_fp16 = reshape(shape = var_748, x = linear_22_cast_fp16)[name = tensor("k_9_cast_fp16")]; + tensor encoder_layers_2_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34511808)))]; + tensor linear_23_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_self_attn_linear_v_weight_to_fp16, x = input_143_cast_fp16)[name = tensor("linear_23_cast_fp16")]; + tensor var_752 = const()[name = tensor("op_752"), val = tensor([1, -1, 8, 64])]; + tensor v_5_cast_fp16 = reshape(shape = var_752, x = linear_23_cast_fp16)[name = tensor("v_5_cast_fp16")]; + tensor value_7_perm_0 = const()[name = tensor("value_7_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_2_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35036160)))]; + tensor var_764_cast_fp16 = add(x = q_13_cast_fp16, y = encoder_layers_2_self_attn_pos_bias_u_to_fp16)[name = tensor("op_764_cast_fp16")]; + tensor encoder_layers_2_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35037248)))]; + tensor var_766_cast_fp16 = add(x = q_13_cast_fp16, y = encoder_layers_2_self_attn_pos_bias_v_to_fp16)[name = tensor("op_766_cast_fp16")]; + tensor q_with_bias_v_5_perm_0 = const()[name = tensor("q_with_bias_v_5_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_59_transpose_x_0 = const()[name = tensor("x_59_transpose_x_0"), val = tensor(false)]; + tensor x_59_transpose_y_0 = const()[name = tensor("x_59_transpose_y_0"), val = tensor(false)]; + tensor var_768_to_fp16 = const()[name = tensor("op_768_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35038336)))]; + tensor q_with_bias_v_5_cast_fp16 = transpose(perm = q_with_bias_v_5_perm_0, x = var_766_cast_fp16)[name = tensor("transpose_219")]; + tensor x_59_cast_fp16 = matmul(transpose_x = x_59_transpose_x_0, transpose_y = x_59_transpose_y_0, x = q_with_bias_v_5_cast_fp16, y = var_768_to_fp16)[name = tensor("x_59_cast_fp16")]; + tensor x_61_pad_0 = const()[name = tensor("x_61_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_61_mode_0 = const()[name = tensor("x_61_mode_0"), val = tensor("constant")]; + tensor const_49_to_fp16 = const()[name = tensor("const_49_to_fp16"), val = tensor(0x0p+0)]; + tensor x_61_cast_fp16 = pad(constant_val = const_49_to_fp16, mode = x_61_mode_0, pad = x_61_pad_0, x = x_59_cast_fp16)[name = tensor("x_61_cast_fp16")]; + tensor var_776 = const()[name = tensor("op_776"), val = tensor([1, 8, -1, 17])]; + tensor x_63_cast_fp16 = reshape(shape = var_776, x = x_61_cast_fp16)[name = tensor("x_63_cast_fp16")]; + tensor var_780_begin_0 = const()[name = tensor("op_780_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_780_end_0 = const()[name = tensor("op_780_end_0"), val = tensor([1, 8, 174, 17])]; + tensor var_780_end_mask_0 = const()[name = tensor("op_780_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_780_cast_fp16 = slice_by_index(begin = var_780_begin_0, end = var_780_end_0, end_mask = var_780_end_mask_0, x = x_63_cast_fp16)[name = tensor("op_780_cast_fp16")]; + tensor var_781 = const()[name = tensor("op_781"), val = tensor([1, 8, 17, 173])]; + tensor matrix_bd_9_cast_fp16 = reshape(shape = var_781, x = var_780_cast_fp16)[name = tensor("matrix_bd_9_cast_fp16")]; + tensor matrix_ac_5_transpose_x_0 = const()[name = tensor("matrix_ac_5_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_5_transpose_y_0 = const()[name = tensor("matrix_ac_5_transpose_y_0"), val = tensor(false)]; + tensor transpose_55_perm_0 = const()[name = tensor("transpose_55_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_56_perm_0 = const()[name = tensor("transpose_56_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_56 = transpose(perm = transpose_56_perm_0, x = k_9_cast_fp16)[name = tensor("transpose_217")]; + tensor transpose_55 = transpose(perm = transpose_55_perm_0, x = var_764_cast_fp16)[name = tensor("transpose_218")]; + tensor matrix_ac_5_cast_fp16 = matmul(transpose_x = matrix_ac_5_transpose_x_0, transpose_y = matrix_ac_5_transpose_y_0, x = transpose_55, y = transpose_56)[name = tensor("matrix_ac_5_cast_fp16")]; + tensor matrix_bd_11_begin_0 = const()[name = tensor("matrix_bd_11_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_11_end_0 = const()[name = tensor("matrix_bd_11_end_0"), val = tensor([1, 8, 17, 87])]; + tensor matrix_bd_11_end_mask_0 = const()[name = tensor("matrix_bd_11_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_11_cast_fp16 = slice_by_index(begin = matrix_bd_11_begin_0, end = matrix_bd_11_end_0, end_mask = matrix_bd_11_end_mask_0, x = matrix_bd_9_cast_fp16)[name = tensor("matrix_bd_11_cast_fp16")]; + tensor var_790_cast_fp16 = add(x = matrix_ac_5_cast_fp16, y = matrix_bd_11_cast_fp16)[name = tensor("op_790_cast_fp16")]; + tensor _inversed_scores_9_y_0_to_fp16 = const()[name = tensor("_inversed_scores_9_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_9_cast_fp16 = mul(x = var_790_cast_fp16, y = _inversed_scores_9_y_0_to_fp16)[name = tensor("_inversed_scores_9_cast_fp16")]; + tensor scores_11_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_9_cast_fp16, cond = mask_3)[name = tensor("scores_11_cast_fp16")]; + tensor var_796_cast_fp16 = softmax(axis = var_62, x = scores_11_cast_fp16)[name = tensor("op_796_cast_fp16")]; + tensor input_145_cast_fp16 = select(a = var_40_to_fp16, b = var_796_cast_fp16, cond = mask_3)[name = tensor("input_145_cast_fp16")]; + tensor x_65_transpose_x_0 = const()[name = tensor("x_65_transpose_x_0"), val = tensor(false)]; + tensor x_65_transpose_y_0 = const()[name = tensor("x_65_transpose_y_0"), val = tensor(false)]; + tensor value_7_cast_fp16 = transpose(perm = value_7_perm_0, x = v_5_cast_fp16)[name = tensor("transpose_220")]; + tensor x_65_cast_fp16 = matmul(transpose_x = x_65_transpose_x_0, transpose_y = x_65_transpose_y_0, x = input_145_cast_fp16, y = value_7_cast_fp16)[name = tensor("x_65_cast_fp16")]; + tensor var_800_perm_0 = const()[name = tensor("op_800_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_801 = const()[name = tensor("op_801"), val = tensor([1, -1, 512])]; + tensor var_800_cast_fp16 = transpose(perm = var_800_perm_0, x = x_65_cast_fp16)[name = tensor("transpose_216")]; + tensor input_147_cast_fp16 = reshape(shape = var_801, x = var_800_cast_fp16)[name = tensor("input_147_cast_fp16")]; + tensor encoder_layers_2_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35215552)))]; + tensor linear_25_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_self_attn_linear_out_weight_to_fp16, x = input_147_cast_fp16)[name = tensor("linear_25_cast_fp16")]; + tensor input_151_cast_fp16 = add(x = input_141_cast_fp16, y = linear_25_cast_fp16)[name = tensor("input_151_cast_fp16")]; + tensor x_69_axes_0 = const()[name = tensor("x_69_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35739904)))]; + tensor encoder_layers_2_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35740992)))]; + tensor x_69_cast_fp16 = layer_norm(axes = x_69_axes_0, beta = encoder_layers_2_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_norm_conv_weight_to_fp16, x = input_151_cast_fp16)[name = tensor("x_69_cast_fp16")]; + tensor input_153_perm_0 = const()[name = tensor("input_153_perm_0"), val = tensor([0, 2, 1])]; + tensor input_155_pad_type_0 = const()[name = tensor("input_155_pad_type_0"), val = tensor("valid")]; + tensor input_155_strides_0 = const()[name = tensor("input_155_strides_0"), val = tensor([1])]; + tensor input_155_pad_0 = const()[name = tensor("input_155_pad_0"), val = tensor([0, 0])]; + tensor input_155_dilations_0 = const()[name = tensor("input_155_dilations_0"), val = tensor([1])]; + tensor input_155_groups_0 = const()[name = tensor("input_155_groups_0"), val = tensor(1)]; + tensor encoder_layers_2_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_2_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35742080)))]; + tensor input_153_cast_fp16 = transpose(perm = input_153_perm_0, x = x_69_cast_fp16)[name = tensor("transpose_215")]; + tensor input_155_cast_fp16 = conv(dilations = input_155_dilations_0, groups = input_155_groups_0, pad = input_155_pad_0, pad_type = input_155_pad_type_0, strides = input_155_strides_0, weight = encoder_layers_2_conv_pointwise_conv1_weight_to_fp16, x = input_153_cast_fp16)[name = tensor("input_155_cast_fp16")]; + tensor x_71_split_num_splits_0 = const()[name = tensor("x_71_split_num_splits_0"), val = tensor(2)]; + tensor x_71_split_axis_0 = const()[name = tensor("x_71_split_axis_0"), val = tensor(1)]; + tensor x_71_split_cast_fp16_0, tensor x_71_split_cast_fp16_1 = split(axis = x_71_split_axis_0, num_splits = x_71_split_num_splits_0, x = input_155_cast_fp16)[name = tensor("x_71_split_cast_fp16")]; + tensor x_71_split_1_sigmoid_cast_fp16 = sigmoid(x = x_71_split_cast_fp16_1)[name = tensor("x_71_split_1_sigmoid_cast_fp16")]; + tensor x_71_cast_fp16 = mul(x = x_71_split_cast_fp16_0, y = x_71_split_1_sigmoid_cast_fp16)[name = tensor("x_71_cast_fp16")]; + tensor input_157_cast_fp16 = select(a = var_40_to_fp16, b = x_71_cast_fp16, cond = var_418)[name = tensor("input_157_cast_fp16")]; + tensor new_x_11_interleave_0 = const()[name = tensor("new_x_11_interleave_0"), val = tensor(false)]; + tensor new_x_11_cast_fp16 = concat(axis = var_62, interleave = new_x_11_interleave_0, values = (cache_11_cast_fp16, input_157_cast_fp16))[name = tensor("new_x_11_cast_fp16")]; + tensor var_839_begin_0 = const()[name = tensor("op_839_begin_0"), val = tensor([0, 0, 17])]; + tensor var_839_end_0 = const()[name = tensor("op_839_end_0"), val = tensor([1, 512, 25])]; + tensor var_839_end_mask_0 = const()[name = tensor("op_839_end_mask_0"), val = tensor([true, true, true])]; + tensor var_839_cast_fp16 = slice_by_index(begin = var_839_begin_0, end = var_839_end_0, end_mask = var_839_end_mask_0, x = new_x_11_cast_fp16)[name = tensor("op_839_cast_fp16")]; + tensor x_73_pad_type_0 = const()[name = tensor("x_73_pad_type_0"), val = tensor("valid")]; + tensor x_73_groups_0 = const()[name = tensor("x_73_groups_0"), val = tensor(512)]; + tensor x_73_strides_0 = const()[name = tensor("x_73_strides_0"), val = tensor([1])]; + tensor x_73_pad_0 = const()[name = tensor("x_73_pad_0"), val = tensor([0, 0])]; + tensor x_73_dilations_0 = const()[name = tensor("x_73_dilations_0"), val = tensor([1])]; + tensor encoder_layers_2_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_2_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36790720)))]; + tensor x_73_cast_fp16 = conv(dilations = x_73_dilations_0, groups = x_73_groups_0, pad = x_73_pad_0, pad_type = x_73_pad_type_0, strides = x_73_strides_0, weight = encoder_layers_2_conv_depthwise_conv_weight_to_fp16, x = new_x_11_cast_fp16)[name = tensor("x_73_cast_fp16")]; + tensor input_159_perm_0 = const()[name = tensor("input_159_perm_0"), val = tensor([0, 2, 1])]; + tensor x_75_axes_0 = const()[name = tensor("x_75_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_2_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36800000)))]; + tensor encoder_layers_2_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_2_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36801088)))]; + tensor input_159_cast_fp16 = transpose(perm = input_159_perm_0, x = x_73_cast_fp16)[name = tensor("transpose_214")]; + tensor x_75_cast_fp16 = layer_norm(axes = x_75_axes_0, beta = encoder_layers_2_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_conv_batch_norm_weight_to_fp16, x = input_159_cast_fp16)[name = tensor("x_75_cast_fp16")]; + tensor input_161_perm_0 = const()[name = tensor("input_161_perm_0"), val = tensor([0, 2, 1])]; + tensor input_161_cast_fp16 = transpose(perm = input_161_perm_0, x = x_75_cast_fp16)[name = tensor("transpose_213")]; + tensor input_163_cast_fp16 = silu(x = input_161_cast_fp16)[name = tensor("input_163_cast_fp16")]; + tensor x_77_pad_type_0 = const()[name = tensor("x_77_pad_type_0"), val = tensor("valid")]; + tensor x_77_strides_0 = const()[name = tensor("x_77_strides_0"), val = tensor([1])]; + tensor x_77_pad_0 = const()[name = tensor("x_77_pad_0"), val = tensor([0, 0])]; + tensor x_77_dilations_0 = const()[name = tensor("x_77_dilations_0"), val = tensor([1])]; + tensor x_77_groups_0 = const()[name = tensor("x_77_groups_0"), val = tensor(1)]; + tensor encoder_layers_2_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_2_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36802176)))]; + tensor x_77_cast_fp16 = conv(dilations = x_77_dilations_0, groups = x_77_groups_0, pad = x_77_pad_0, pad_type = x_77_pad_type_0, strides = x_77_strides_0, weight = encoder_layers_2_conv_pointwise_conv2_weight_to_fp16, x = input_163_cast_fp16)[name = tensor("x_77_cast_fp16")]; + tensor input_165_perm_0 = const()[name = tensor("input_165_perm_0"), val = tensor([0, 2, 1])]; + tensor input_165_cast_fp16 = transpose(perm = input_165_perm_0, x = x_77_cast_fp16)[name = tensor("transpose_212")]; + tensor input_167_cast_fp16 = add(x = input_151_cast_fp16, y = input_165_cast_fp16)[name = tensor("input_167_cast_fp16")]; + tensor input_169_axes_0 = const()[name = tensor("input_169_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37326528)))]; + tensor encoder_layers_2_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37327616)))]; + tensor input_169_cast_fp16 = layer_norm(axes = input_169_axes_0, beta = encoder_layers_2_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_norm_feed_forward2_weight_to_fp16, x = input_167_cast_fp16)[name = tensor("input_169_cast_fp16")]; + tensor encoder_layers_2_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_2_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37328704)))]; + tensor linear_26_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_2_feed_forward2_linear1_weight_to_fp16, x = input_169_cast_fp16)[name = tensor("linear_26_cast_fp16")]; + tensor input_173_cast_fp16 = silu(x = linear_26_cast_fp16)[name = tensor("input_173_cast_fp16")]; + tensor encoder_layers_2_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_2_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39425920)))]; + tensor linear_27_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_feed_forward2_linear2_weight_to_fp16, x = input_173_cast_fp16)[name = tensor("linear_27_cast_fp16")]; + tensor var_880_to_fp16 = const()[name = tensor("op_880_to_fp16"), val = tensor(0x1p-1)]; + tensor var_881_cast_fp16 = mul(x = linear_27_cast_fp16, y = var_880_to_fp16)[name = tensor("op_881_cast_fp16")]; + tensor input_179_cast_fp16 = add(x = input_167_cast_fp16, y = var_881_cast_fp16)[name = tensor("input_179_cast_fp16")]; + tensor input_181_axes_0 = const()[name = tensor("input_181_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41523136)))]; + tensor encoder_layers_2_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41524224)))]; + tensor input_181_cast_fp16 = layer_norm(axes = input_181_axes_0, beta = encoder_layers_2_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_norm_out_weight_to_fp16, x = input_179_cast_fp16)[name = tensor("input_181_cast_fp16")]; + tensor cache_13_begin_0 = const()[name = tensor("cache_13_begin_0"), val = tensor([3, 0, 0, 0])]; + tensor cache_13_end_0 = const()[name = tensor("cache_13_end_0"), val = tensor([4, 1, 70, 512])]; + tensor cache_13_end_mask_0 = const()[name = tensor("cache_13_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_13_squeeze_mask_0 = const()[name = tensor("cache_13_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_13_cast_fp16 = slice_by_index(begin = cache_13_begin_0, end = cache_13_end_0, end_mask = cache_13_end_mask_0, squeeze_mask = cache_13_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_13_cast_fp16")]; + tensor cache_15_begin_0 = const()[name = tensor("cache_15_begin_0"), val = tensor([3, 0, 0, 0])]; + tensor cache_15_end_0 = const()[name = tensor("cache_15_end_0"), val = tensor([4, 1, 512, 8])]; + tensor cache_15_end_mask_0 = const()[name = tensor("cache_15_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_15_squeeze_mask_0 = const()[name = tensor("cache_15_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_15_cast_fp16 = slice_by_index(begin = cache_15_begin_0, end = cache_15_end_0, end_mask = cache_15_end_mask_0, squeeze_mask = cache_15_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_15_cast_fp16")]; + tensor input_183_axes_0 = const()[name = tensor("input_183_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41525312)))]; + tensor encoder_layers_3_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41526400)))]; + tensor input_183_cast_fp16 = layer_norm(axes = input_183_axes_0, beta = encoder_layers_3_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_norm_feed_forward1_weight_to_fp16, x = input_181_cast_fp16)[name = tensor("input_183_cast_fp16")]; + tensor encoder_layers_3_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_3_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41527488)))]; + tensor linear_28_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_3_feed_forward1_linear1_weight_to_fp16, x = input_183_cast_fp16)[name = tensor("linear_28_cast_fp16")]; + tensor input_187_cast_fp16 = silu(x = linear_28_cast_fp16)[name = tensor("input_187_cast_fp16")]; + tensor encoder_layers_3_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_3_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43624704)))]; + tensor linear_29_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_feed_forward1_linear2_weight_to_fp16, x = input_187_cast_fp16)[name = tensor("linear_29_cast_fp16")]; + tensor var_915_to_fp16 = const()[name = tensor("op_915_to_fp16"), val = tensor(0x1p-1)]; + tensor var_916_cast_fp16 = mul(x = linear_29_cast_fp16, y = var_915_to_fp16)[name = tensor("op_916_cast_fp16")]; + tensor input_193_cast_fp16 = add(x = input_181_cast_fp16, y = var_916_cast_fp16)[name = tensor("input_193_cast_fp16")]; + tensor key_7_axes_0 = const()[name = tensor("key_7_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45721920)))]; + tensor encoder_layers_3_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45723008)))]; + tensor key_7_cast_fp16 = layer_norm(axes = key_7_axes_0, beta = encoder_layers_3_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_norm_self_att_weight_to_fp16, x = input_193_cast_fp16)[name = tensor("key_7_cast_fp16")]; + tensor input_195_interleave_0 = const()[name = tensor("input_195_interleave_0"), val = tensor(false)]; + tensor input_195_cast_fp16 = concat(axis = var_64, interleave = input_195_interleave_0, values = (cache_13_cast_fp16, key_7_cast_fp16))[name = tensor("input_195_cast_fp16")]; + tensor var_938_begin_0 = const()[name = tensor("op_938_begin_0"), val = tensor([0, 17, 0])]; + tensor var_938_end_0 = const()[name = tensor("op_938_end_0"), val = tensor([1, 70, 512])]; + tensor var_938_end_mask_0 = const()[name = tensor("op_938_end_mask_0"), val = tensor([true, true, true])]; + tensor var_938_cast_fp16 = slice_by_index(begin = var_938_begin_0, end = var_938_end_0, end_mask = var_938_end_mask_0, x = cache_13_cast_fp16)[name = tensor("op_938_cast_fp16")]; + tensor var_944_interleave_0 = const()[name = tensor("op_944_interleave_0"), val = tensor(false)]; + tensor var_944_cast_fp16 = concat(axis = var_64, interleave = var_944_interleave_0, values = (var_938_cast_fp16, key_7_cast_fp16))[name = tensor("op_944_cast_fp16")]; + tensor encoder_layers_3_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45724096)))]; + tensor linear_30_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_self_attn_linear_q_weight_to_fp16, x = key_7_cast_fp16)[name = tensor("linear_30_cast_fp16")]; + tensor var_948 = const()[name = tensor("op_948"), val = tensor([1, -1, 8, 64])]; + tensor q_19_cast_fp16 = reshape(shape = var_948, x = linear_30_cast_fp16)[name = tensor("q_19_cast_fp16")]; + tensor encoder_layers_3_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46248448)))]; + tensor linear_31_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_self_attn_linear_k_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("linear_31_cast_fp16")]; + tensor var_952 = const()[name = tensor("op_952"), val = tensor([1, -1, 8, 64])]; + tensor k_13_cast_fp16 = reshape(shape = var_952, x = linear_31_cast_fp16)[name = tensor("k_13_cast_fp16")]; + tensor encoder_layers_3_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46772800)))]; + tensor linear_32_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_self_attn_linear_v_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("linear_32_cast_fp16")]; + tensor var_956 = const()[name = tensor("op_956"), val = tensor([1, -1, 8, 64])]; + tensor v_7_cast_fp16 = reshape(shape = var_956, x = linear_32_cast_fp16)[name = tensor("v_7_cast_fp16")]; + tensor value_9_perm_0 = const()[name = tensor("value_9_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_3_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47297152)))]; + tensor var_968_cast_fp16 = add(x = q_19_cast_fp16, y = encoder_layers_3_self_attn_pos_bias_u_to_fp16)[name = tensor("op_968_cast_fp16")]; + tensor encoder_layers_3_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47298240)))]; + tensor var_970_cast_fp16 = add(x = q_19_cast_fp16, y = encoder_layers_3_self_attn_pos_bias_v_to_fp16)[name = tensor("op_970_cast_fp16")]; + tensor q_with_bias_v_7_perm_0 = const()[name = tensor("q_with_bias_v_7_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_85_transpose_x_0 = const()[name = tensor("x_85_transpose_x_0"), val = tensor(false)]; + tensor x_85_transpose_y_0 = const()[name = tensor("x_85_transpose_y_0"), val = tensor(false)]; + tensor var_972_to_fp16 = const()[name = tensor("op_972_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47299328)))]; + tensor q_with_bias_v_7_cast_fp16 = transpose(perm = q_with_bias_v_7_perm_0, x = var_970_cast_fp16)[name = tensor("transpose_210")]; + tensor x_85_cast_fp16 = matmul(transpose_x = x_85_transpose_x_0, transpose_y = x_85_transpose_y_0, x = q_with_bias_v_7_cast_fp16, y = var_972_to_fp16)[name = tensor("x_85_cast_fp16")]; + tensor x_87_pad_0 = const()[name = tensor("x_87_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_87_mode_0 = const()[name = tensor("x_87_mode_0"), val = tensor("constant")]; + tensor const_62_to_fp16 = const()[name = tensor("const_62_to_fp16"), val = tensor(0x0p+0)]; + tensor x_87_cast_fp16 = pad(constant_val = const_62_to_fp16, mode = x_87_mode_0, pad = x_87_pad_0, x = x_85_cast_fp16)[name = tensor("x_87_cast_fp16")]; + tensor var_980 = const()[name = tensor("op_980"), val = tensor([1, 8, -1, 17])]; + tensor x_89_cast_fp16 = reshape(shape = var_980, x = x_87_cast_fp16)[name = tensor("x_89_cast_fp16")]; + tensor var_984_begin_0 = const()[name = tensor("op_984_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_984_end_0 = const()[name = tensor("op_984_end_0"), val = tensor([1, 8, 174, 17])]; + tensor var_984_end_mask_0 = const()[name = tensor("op_984_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_984_cast_fp16 = slice_by_index(begin = var_984_begin_0, end = var_984_end_0, end_mask = var_984_end_mask_0, x = x_89_cast_fp16)[name = tensor("op_984_cast_fp16")]; + tensor var_985 = const()[name = tensor("op_985"), val = tensor([1, 8, 17, 173])]; + tensor matrix_bd_13_cast_fp16 = reshape(shape = var_985, x = var_984_cast_fp16)[name = tensor("matrix_bd_13_cast_fp16")]; + tensor matrix_ac_7_transpose_x_0 = const()[name = tensor("matrix_ac_7_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_7_transpose_y_0 = const()[name = tensor("matrix_ac_7_transpose_y_0"), val = tensor(false)]; + tensor transpose_57_perm_0 = const()[name = tensor("transpose_57_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_58_perm_0 = const()[name = tensor("transpose_58_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_58 = transpose(perm = transpose_58_perm_0, x = k_13_cast_fp16)[name = tensor("transpose_208")]; + tensor transpose_57 = transpose(perm = transpose_57_perm_0, x = var_968_cast_fp16)[name = tensor("transpose_209")]; + tensor matrix_ac_7_cast_fp16 = matmul(transpose_x = matrix_ac_7_transpose_x_0, transpose_y = matrix_ac_7_transpose_y_0, x = transpose_57, y = transpose_58)[name = tensor("matrix_ac_7_cast_fp16")]; + tensor matrix_bd_15_begin_0 = const()[name = tensor("matrix_bd_15_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_15_end_0 = const()[name = tensor("matrix_bd_15_end_0"), val = tensor([1, 8, 17, 87])]; + tensor matrix_bd_15_end_mask_0 = const()[name = tensor("matrix_bd_15_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_15_cast_fp16 = slice_by_index(begin = matrix_bd_15_begin_0, end = matrix_bd_15_end_0, end_mask = matrix_bd_15_end_mask_0, x = matrix_bd_13_cast_fp16)[name = tensor("matrix_bd_15_cast_fp16")]; + tensor var_994_cast_fp16 = add(x = matrix_ac_7_cast_fp16, y = matrix_bd_15_cast_fp16)[name = tensor("op_994_cast_fp16")]; + tensor _inversed_scores_13_y_0_to_fp16 = const()[name = tensor("_inversed_scores_13_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_13_cast_fp16 = mul(x = var_994_cast_fp16, y = _inversed_scores_13_y_0_to_fp16)[name = tensor("_inversed_scores_13_cast_fp16")]; + tensor scores_15_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_13_cast_fp16, cond = mask_3)[name = tensor("scores_15_cast_fp16")]; + tensor var_1000_cast_fp16 = softmax(axis = var_62, x = scores_15_cast_fp16)[name = tensor("op_1000_cast_fp16")]; + tensor input_197_cast_fp16 = select(a = var_40_to_fp16, b = var_1000_cast_fp16, cond = mask_3)[name = tensor("input_197_cast_fp16")]; + tensor x_91_transpose_x_0 = const()[name = tensor("x_91_transpose_x_0"), val = tensor(false)]; + tensor x_91_transpose_y_0 = const()[name = tensor("x_91_transpose_y_0"), val = tensor(false)]; + tensor value_9_cast_fp16 = transpose(perm = value_9_perm_0, x = v_7_cast_fp16)[name = tensor("transpose_211")]; + tensor x_91_cast_fp16 = matmul(transpose_x = x_91_transpose_x_0, transpose_y = x_91_transpose_y_0, x = input_197_cast_fp16, y = value_9_cast_fp16)[name = tensor("x_91_cast_fp16")]; + tensor var_1004_perm_0 = const()[name = tensor("op_1004_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1005 = const()[name = tensor("op_1005"), val = tensor([1, -1, 512])]; + tensor var_1004_cast_fp16 = transpose(perm = var_1004_perm_0, x = x_91_cast_fp16)[name = tensor("transpose_207")]; + tensor input_199_cast_fp16 = reshape(shape = var_1005, x = var_1004_cast_fp16)[name = tensor("input_199_cast_fp16")]; + tensor encoder_layers_3_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47476544)))]; + tensor linear_34_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_self_attn_linear_out_weight_to_fp16, x = input_199_cast_fp16)[name = tensor("linear_34_cast_fp16")]; + tensor input_203_cast_fp16 = add(x = input_193_cast_fp16, y = linear_34_cast_fp16)[name = tensor("input_203_cast_fp16")]; + tensor x_95_axes_0 = const()[name = tensor("x_95_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48000896)))]; + tensor encoder_layers_3_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48001984)))]; + tensor x_95_cast_fp16 = layer_norm(axes = x_95_axes_0, beta = encoder_layers_3_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_norm_conv_weight_to_fp16, x = input_203_cast_fp16)[name = tensor("x_95_cast_fp16")]; + tensor input_205_perm_0 = const()[name = tensor("input_205_perm_0"), val = tensor([0, 2, 1])]; + tensor input_207_pad_type_0 = const()[name = tensor("input_207_pad_type_0"), val = tensor("valid")]; + tensor input_207_strides_0 = const()[name = tensor("input_207_strides_0"), val = tensor([1])]; + tensor input_207_pad_0 = const()[name = tensor("input_207_pad_0"), val = tensor([0, 0])]; + tensor input_207_dilations_0 = const()[name = tensor("input_207_dilations_0"), val = tensor([1])]; + tensor input_207_groups_0 = const()[name = tensor("input_207_groups_0"), val = tensor(1)]; + tensor encoder_layers_3_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_3_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48003072)))]; + tensor input_205_cast_fp16 = transpose(perm = input_205_perm_0, x = x_95_cast_fp16)[name = tensor("transpose_206")]; + tensor input_207_cast_fp16 = conv(dilations = input_207_dilations_0, groups = input_207_groups_0, pad = input_207_pad_0, pad_type = input_207_pad_type_0, strides = input_207_strides_0, weight = encoder_layers_3_conv_pointwise_conv1_weight_to_fp16, x = input_205_cast_fp16)[name = tensor("input_207_cast_fp16")]; + tensor x_97_split_num_splits_0 = const()[name = tensor("x_97_split_num_splits_0"), val = tensor(2)]; + tensor x_97_split_axis_0 = const()[name = tensor("x_97_split_axis_0"), val = tensor(1)]; + tensor x_97_split_cast_fp16_0, tensor x_97_split_cast_fp16_1 = split(axis = x_97_split_axis_0, num_splits = x_97_split_num_splits_0, x = input_207_cast_fp16)[name = tensor("x_97_split_cast_fp16")]; + tensor x_97_split_1_sigmoid_cast_fp16 = sigmoid(x = x_97_split_cast_fp16_1)[name = tensor("x_97_split_1_sigmoid_cast_fp16")]; + tensor x_97_cast_fp16 = mul(x = x_97_split_cast_fp16_0, y = x_97_split_1_sigmoid_cast_fp16)[name = tensor("x_97_cast_fp16")]; + tensor input_209_cast_fp16 = select(a = var_40_to_fp16, b = x_97_cast_fp16, cond = var_418)[name = tensor("input_209_cast_fp16")]; + tensor new_x_15_interleave_0 = const()[name = tensor("new_x_15_interleave_0"), val = tensor(false)]; + tensor new_x_15_cast_fp16 = concat(axis = var_62, interleave = new_x_15_interleave_0, values = (cache_15_cast_fp16, input_209_cast_fp16))[name = tensor("new_x_15_cast_fp16")]; + tensor var_1043_begin_0 = const()[name = tensor("op_1043_begin_0"), val = tensor([0, 0, 17])]; + tensor var_1043_end_0 = const()[name = tensor("op_1043_end_0"), val = tensor([1, 512, 25])]; + tensor var_1043_end_mask_0 = const()[name = tensor("op_1043_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1043_cast_fp16 = slice_by_index(begin = var_1043_begin_0, end = var_1043_end_0, end_mask = var_1043_end_mask_0, x = new_x_15_cast_fp16)[name = tensor("op_1043_cast_fp16")]; + tensor x_99_pad_type_0 = const()[name = tensor("x_99_pad_type_0"), val = tensor("valid")]; + tensor x_99_groups_0 = const()[name = tensor("x_99_groups_0"), val = tensor(512)]; + tensor x_99_strides_0 = const()[name = tensor("x_99_strides_0"), val = tensor([1])]; + tensor x_99_pad_0 = const()[name = tensor("x_99_pad_0"), val = tensor([0, 0])]; + tensor x_99_dilations_0 = const()[name = tensor("x_99_dilations_0"), val = tensor([1])]; + tensor encoder_layers_3_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_3_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49051712)))]; + tensor x_99_cast_fp16 = conv(dilations = x_99_dilations_0, groups = x_99_groups_0, pad = x_99_pad_0, pad_type = x_99_pad_type_0, strides = x_99_strides_0, weight = encoder_layers_3_conv_depthwise_conv_weight_to_fp16, x = new_x_15_cast_fp16)[name = tensor("x_99_cast_fp16")]; + tensor input_211_perm_0 = const()[name = tensor("input_211_perm_0"), val = tensor([0, 2, 1])]; + tensor x_101_axes_0 = const()[name = tensor("x_101_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_3_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49060992)))]; + tensor encoder_layers_3_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_3_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49062080)))]; + tensor input_211_cast_fp16 = transpose(perm = input_211_perm_0, x = x_99_cast_fp16)[name = tensor("transpose_205")]; + tensor x_101_cast_fp16 = layer_norm(axes = x_101_axes_0, beta = encoder_layers_3_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_conv_batch_norm_weight_to_fp16, x = input_211_cast_fp16)[name = tensor("x_101_cast_fp16")]; + tensor input_213_perm_0 = const()[name = tensor("input_213_perm_0"), val = tensor([0, 2, 1])]; + tensor input_213_cast_fp16 = transpose(perm = input_213_perm_0, x = x_101_cast_fp16)[name = tensor("transpose_204")]; + tensor input_215_cast_fp16 = silu(x = input_213_cast_fp16)[name = tensor("input_215_cast_fp16")]; + tensor x_103_pad_type_0 = const()[name = tensor("x_103_pad_type_0"), val = tensor("valid")]; + tensor x_103_strides_0 = const()[name = tensor("x_103_strides_0"), val = tensor([1])]; + tensor x_103_pad_0 = const()[name = tensor("x_103_pad_0"), val = tensor([0, 0])]; + tensor x_103_dilations_0 = const()[name = tensor("x_103_dilations_0"), val = tensor([1])]; + tensor x_103_groups_0 = const()[name = tensor("x_103_groups_0"), val = tensor(1)]; + tensor encoder_layers_3_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_3_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49063168)))]; + tensor x_103_cast_fp16 = conv(dilations = x_103_dilations_0, groups = x_103_groups_0, pad = x_103_pad_0, pad_type = x_103_pad_type_0, strides = x_103_strides_0, weight = encoder_layers_3_conv_pointwise_conv2_weight_to_fp16, x = input_215_cast_fp16)[name = tensor("x_103_cast_fp16")]; + tensor input_217_perm_0 = const()[name = tensor("input_217_perm_0"), val = tensor([0, 2, 1])]; + tensor input_217_cast_fp16 = transpose(perm = input_217_perm_0, x = x_103_cast_fp16)[name = tensor("transpose_203")]; + tensor input_219_cast_fp16 = add(x = input_203_cast_fp16, y = input_217_cast_fp16)[name = tensor("input_219_cast_fp16")]; + tensor input_221_axes_0 = const()[name = tensor("input_221_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49587520)))]; + tensor encoder_layers_3_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49588608)))]; + tensor input_221_cast_fp16 = layer_norm(axes = input_221_axes_0, beta = encoder_layers_3_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_norm_feed_forward2_weight_to_fp16, x = input_219_cast_fp16)[name = tensor("input_221_cast_fp16")]; + tensor encoder_layers_3_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_3_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49589696)))]; + tensor linear_35_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_3_feed_forward2_linear1_weight_to_fp16, x = input_221_cast_fp16)[name = tensor("linear_35_cast_fp16")]; + tensor input_225_cast_fp16 = silu(x = linear_35_cast_fp16)[name = tensor("input_225_cast_fp16")]; + tensor encoder_layers_3_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_3_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51686912)))]; + tensor linear_36_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_feed_forward2_linear2_weight_to_fp16, x = input_225_cast_fp16)[name = tensor("linear_36_cast_fp16")]; + tensor var_1084_to_fp16 = const()[name = tensor("op_1084_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1085_cast_fp16 = mul(x = linear_36_cast_fp16, y = var_1084_to_fp16)[name = tensor("op_1085_cast_fp16")]; + tensor input_231_cast_fp16 = add(x = input_219_cast_fp16, y = var_1085_cast_fp16)[name = tensor("input_231_cast_fp16")]; + tensor input_233_axes_0 = const()[name = tensor("input_233_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53784128)))]; + tensor encoder_layers_3_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53785216)))]; + tensor input_233_cast_fp16 = layer_norm(axes = input_233_axes_0, beta = encoder_layers_3_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_norm_out_weight_to_fp16, x = input_231_cast_fp16)[name = tensor("input_233_cast_fp16")]; + tensor cache_17_begin_0 = const()[name = tensor("cache_17_begin_0"), val = tensor([4, 0, 0, 0])]; + tensor cache_17_end_0 = const()[name = tensor("cache_17_end_0"), val = tensor([5, 1, 70, 512])]; + tensor cache_17_end_mask_0 = const()[name = tensor("cache_17_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_17_squeeze_mask_0 = const()[name = tensor("cache_17_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_17_cast_fp16 = slice_by_index(begin = cache_17_begin_0, end = cache_17_end_0, end_mask = cache_17_end_mask_0, squeeze_mask = cache_17_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_17_cast_fp16")]; + tensor cache_19_begin_0 = const()[name = tensor("cache_19_begin_0"), val = tensor([4, 0, 0, 0])]; + tensor cache_19_end_0 = const()[name = tensor("cache_19_end_0"), val = tensor([5, 1, 512, 8])]; + tensor cache_19_end_mask_0 = const()[name = tensor("cache_19_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_19_squeeze_mask_0 = const()[name = tensor("cache_19_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_19_cast_fp16 = slice_by_index(begin = cache_19_begin_0, end = cache_19_end_0, end_mask = cache_19_end_mask_0, squeeze_mask = cache_19_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_19_cast_fp16")]; + tensor input_235_axes_0 = const()[name = tensor("input_235_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53786304)))]; + tensor encoder_layers_4_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53787392)))]; + tensor input_235_cast_fp16 = layer_norm(axes = input_235_axes_0, beta = encoder_layers_4_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_norm_feed_forward1_weight_to_fp16, x = input_233_cast_fp16)[name = tensor("input_235_cast_fp16")]; + tensor encoder_layers_4_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_4_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53788480)))]; + tensor linear_37_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_4_feed_forward1_linear1_weight_to_fp16, x = input_235_cast_fp16)[name = tensor("linear_37_cast_fp16")]; + tensor input_239_cast_fp16 = silu(x = linear_37_cast_fp16)[name = tensor("input_239_cast_fp16")]; + tensor encoder_layers_4_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_4_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55885696)))]; + tensor linear_38_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_feed_forward1_linear2_weight_to_fp16, x = input_239_cast_fp16)[name = tensor("linear_38_cast_fp16")]; + tensor var_1119_to_fp16 = const()[name = tensor("op_1119_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1120_cast_fp16 = mul(x = linear_38_cast_fp16, y = var_1119_to_fp16)[name = tensor("op_1120_cast_fp16")]; + tensor input_245_cast_fp16 = add(x = input_233_cast_fp16, y = var_1120_cast_fp16)[name = tensor("input_245_cast_fp16")]; + tensor key_9_axes_0 = const()[name = tensor("key_9_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57982912)))]; + tensor encoder_layers_4_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57984000)))]; + tensor key_9_cast_fp16 = layer_norm(axes = key_9_axes_0, beta = encoder_layers_4_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_norm_self_att_weight_to_fp16, x = input_245_cast_fp16)[name = tensor("key_9_cast_fp16")]; + tensor input_247_interleave_0 = const()[name = tensor("input_247_interleave_0"), val = tensor(false)]; + tensor input_247_cast_fp16 = concat(axis = var_64, interleave = input_247_interleave_0, values = (cache_17_cast_fp16, key_9_cast_fp16))[name = tensor("input_247_cast_fp16")]; + tensor var_1142_begin_0 = const()[name = tensor("op_1142_begin_0"), val = tensor([0, 17, 0])]; + tensor var_1142_end_0 = const()[name = tensor("op_1142_end_0"), val = tensor([1, 70, 512])]; + tensor var_1142_end_mask_0 = const()[name = tensor("op_1142_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1142_cast_fp16 = slice_by_index(begin = var_1142_begin_0, end = var_1142_end_0, end_mask = var_1142_end_mask_0, x = cache_17_cast_fp16)[name = tensor("op_1142_cast_fp16")]; + tensor var_1148_interleave_0 = const()[name = tensor("op_1148_interleave_0"), val = tensor(false)]; + tensor var_1148_cast_fp16 = concat(axis = var_64, interleave = var_1148_interleave_0, values = (var_1142_cast_fp16, key_9_cast_fp16))[name = tensor("op_1148_cast_fp16")]; + tensor encoder_layers_4_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57985088)))]; + tensor linear_39_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_self_attn_linear_q_weight_to_fp16, x = key_9_cast_fp16)[name = tensor("linear_39_cast_fp16")]; + tensor var_1152 = const()[name = tensor("op_1152"), val = tensor([1, -1, 8, 64])]; + tensor q_25_cast_fp16 = reshape(shape = var_1152, x = linear_39_cast_fp16)[name = tensor("q_25_cast_fp16")]; + tensor encoder_layers_4_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58509440)))]; + tensor linear_40_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_self_attn_linear_k_weight_to_fp16, x = input_247_cast_fp16)[name = tensor("linear_40_cast_fp16")]; + tensor var_1156 = const()[name = tensor("op_1156"), val = tensor([1, -1, 8, 64])]; + tensor k_17_cast_fp16 = reshape(shape = var_1156, x = linear_40_cast_fp16)[name = tensor("k_17_cast_fp16")]; + tensor encoder_layers_4_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59033792)))]; + tensor linear_41_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_self_attn_linear_v_weight_to_fp16, x = input_247_cast_fp16)[name = tensor("linear_41_cast_fp16")]; + tensor var_1160 = const()[name = tensor("op_1160"), val = tensor([1, -1, 8, 64])]; + tensor v_9_cast_fp16 = reshape(shape = var_1160, x = linear_41_cast_fp16)[name = tensor("v_9_cast_fp16")]; + tensor value_11_perm_0 = const()[name = tensor("value_11_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_4_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59558144)))]; + tensor var_1172_cast_fp16 = add(x = q_25_cast_fp16, y = encoder_layers_4_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1172_cast_fp16")]; + tensor encoder_layers_4_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59559232)))]; + tensor var_1174_cast_fp16 = add(x = q_25_cast_fp16, y = encoder_layers_4_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1174_cast_fp16")]; + tensor q_with_bias_v_9_perm_0 = const()[name = tensor("q_with_bias_v_9_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_111_transpose_x_0 = const()[name = tensor("x_111_transpose_x_0"), val = tensor(false)]; + tensor x_111_transpose_y_0 = const()[name = tensor("x_111_transpose_y_0"), val = tensor(false)]; + tensor var_1176_to_fp16 = const()[name = tensor("op_1176_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59560320)))]; + tensor q_with_bias_v_9_cast_fp16 = transpose(perm = q_with_bias_v_9_perm_0, x = var_1174_cast_fp16)[name = tensor("transpose_201")]; + tensor x_111_cast_fp16 = matmul(transpose_x = x_111_transpose_x_0, transpose_y = x_111_transpose_y_0, x = q_with_bias_v_9_cast_fp16, y = var_1176_to_fp16)[name = tensor("x_111_cast_fp16")]; + tensor x_113_pad_0 = const()[name = tensor("x_113_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_113_mode_0 = const()[name = tensor("x_113_mode_0"), val = tensor("constant")]; + tensor const_75_to_fp16 = const()[name = tensor("const_75_to_fp16"), val = tensor(0x0p+0)]; + tensor x_113_cast_fp16 = pad(constant_val = const_75_to_fp16, mode = x_113_mode_0, pad = x_113_pad_0, x = x_111_cast_fp16)[name = tensor("x_113_cast_fp16")]; + tensor var_1184 = const()[name = tensor("op_1184"), val = tensor([1, 8, -1, 17])]; + tensor x_115_cast_fp16 = reshape(shape = var_1184, x = x_113_cast_fp16)[name = tensor("x_115_cast_fp16")]; + tensor var_1188_begin_0 = const()[name = tensor("op_1188_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1188_end_0 = const()[name = tensor("op_1188_end_0"), val = tensor([1, 8, 174, 17])]; + tensor var_1188_end_mask_0 = const()[name = tensor("op_1188_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1188_cast_fp16 = slice_by_index(begin = var_1188_begin_0, end = var_1188_end_0, end_mask = var_1188_end_mask_0, x = x_115_cast_fp16)[name = tensor("op_1188_cast_fp16")]; + tensor var_1189 = const()[name = tensor("op_1189"), val = tensor([1, 8, 17, 173])]; + tensor matrix_bd_17_cast_fp16 = reshape(shape = var_1189, x = var_1188_cast_fp16)[name = tensor("matrix_bd_17_cast_fp16")]; + tensor matrix_ac_9_transpose_x_0 = const()[name = tensor("matrix_ac_9_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_9_transpose_y_0 = const()[name = tensor("matrix_ac_9_transpose_y_0"), val = tensor(false)]; + tensor transpose_59_perm_0 = const()[name = tensor("transpose_59_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_60_perm_0 = const()[name = tensor("transpose_60_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_60 = transpose(perm = transpose_60_perm_0, x = k_17_cast_fp16)[name = tensor("transpose_199")]; + tensor transpose_59 = transpose(perm = transpose_59_perm_0, x = var_1172_cast_fp16)[name = tensor("transpose_200")]; + tensor matrix_ac_9_cast_fp16 = matmul(transpose_x = matrix_ac_9_transpose_x_0, transpose_y = matrix_ac_9_transpose_y_0, x = transpose_59, y = transpose_60)[name = tensor("matrix_ac_9_cast_fp16")]; + tensor matrix_bd_19_begin_0 = const()[name = tensor("matrix_bd_19_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_19_end_0 = const()[name = tensor("matrix_bd_19_end_0"), val = tensor([1, 8, 17, 87])]; + tensor matrix_bd_19_end_mask_0 = const()[name = tensor("matrix_bd_19_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_19_cast_fp16 = slice_by_index(begin = matrix_bd_19_begin_0, end = matrix_bd_19_end_0, end_mask = matrix_bd_19_end_mask_0, x = matrix_bd_17_cast_fp16)[name = tensor("matrix_bd_19_cast_fp16")]; + tensor var_1198_cast_fp16 = add(x = matrix_ac_9_cast_fp16, y = matrix_bd_19_cast_fp16)[name = tensor("op_1198_cast_fp16")]; + tensor _inversed_scores_17_y_0_to_fp16 = const()[name = tensor("_inversed_scores_17_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_17_cast_fp16 = mul(x = var_1198_cast_fp16, y = _inversed_scores_17_y_0_to_fp16)[name = tensor("_inversed_scores_17_cast_fp16")]; + tensor scores_19_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_17_cast_fp16, cond = mask_3)[name = tensor("scores_19_cast_fp16")]; + tensor var_1204_cast_fp16 = softmax(axis = var_62, x = scores_19_cast_fp16)[name = tensor("op_1204_cast_fp16")]; + tensor input_249_cast_fp16 = select(a = var_40_to_fp16, b = var_1204_cast_fp16, cond = mask_3)[name = tensor("input_249_cast_fp16")]; + tensor x_117_transpose_x_0 = const()[name = tensor("x_117_transpose_x_0"), val = tensor(false)]; + tensor x_117_transpose_y_0 = const()[name = tensor("x_117_transpose_y_0"), val = tensor(false)]; + tensor value_11_cast_fp16 = transpose(perm = value_11_perm_0, x = v_9_cast_fp16)[name = tensor("transpose_202")]; + tensor x_117_cast_fp16 = matmul(transpose_x = x_117_transpose_x_0, transpose_y = x_117_transpose_y_0, x = input_249_cast_fp16, y = value_11_cast_fp16)[name = tensor("x_117_cast_fp16")]; + tensor var_1208_perm_0 = const()[name = tensor("op_1208_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1209 = const()[name = tensor("op_1209"), val = tensor([1, -1, 512])]; + tensor var_1208_cast_fp16 = transpose(perm = var_1208_perm_0, x = x_117_cast_fp16)[name = tensor("transpose_198")]; + tensor input_251_cast_fp16 = reshape(shape = var_1209, x = var_1208_cast_fp16)[name = tensor("input_251_cast_fp16")]; + tensor encoder_layers_4_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59737536)))]; + tensor linear_43_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_self_attn_linear_out_weight_to_fp16, x = input_251_cast_fp16)[name = tensor("linear_43_cast_fp16")]; + tensor input_255_cast_fp16 = add(x = input_245_cast_fp16, y = linear_43_cast_fp16)[name = tensor("input_255_cast_fp16")]; + tensor x_121_axes_0 = const()[name = tensor("x_121_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60261888)))]; + tensor encoder_layers_4_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60262976)))]; + tensor x_121_cast_fp16 = layer_norm(axes = x_121_axes_0, beta = encoder_layers_4_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_norm_conv_weight_to_fp16, x = input_255_cast_fp16)[name = tensor("x_121_cast_fp16")]; + tensor input_257_perm_0 = const()[name = tensor("input_257_perm_0"), val = tensor([0, 2, 1])]; + tensor input_259_pad_type_0 = const()[name = tensor("input_259_pad_type_0"), val = tensor("valid")]; + tensor input_259_strides_0 = const()[name = tensor("input_259_strides_0"), val = tensor([1])]; + tensor input_259_pad_0 = const()[name = tensor("input_259_pad_0"), val = tensor([0, 0])]; + tensor input_259_dilations_0 = const()[name = tensor("input_259_dilations_0"), val = tensor([1])]; + tensor input_259_groups_0 = const()[name = tensor("input_259_groups_0"), val = tensor(1)]; + tensor encoder_layers_4_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_4_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60264064)))]; + tensor input_257_cast_fp16 = transpose(perm = input_257_perm_0, x = x_121_cast_fp16)[name = tensor("transpose_197")]; + tensor input_259_cast_fp16 = conv(dilations = input_259_dilations_0, groups = input_259_groups_0, pad = input_259_pad_0, pad_type = input_259_pad_type_0, strides = input_259_strides_0, weight = encoder_layers_4_conv_pointwise_conv1_weight_to_fp16, x = input_257_cast_fp16)[name = tensor("input_259_cast_fp16")]; + tensor x_123_split_num_splits_0 = const()[name = tensor("x_123_split_num_splits_0"), val = tensor(2)]; + tensor x_123_split_axis_0 = const()[name = tensor("x_123_split_axis_0"), val = tensor(1)]; + tensor x_123_split_cast_fp16_0, tensor x_123_split_cast_fp16_1 = split(axis = x_123_split_axis_0, num_splits = x_123_split_num_splits_0, x = input_259_cast_fp16)[name = tensor("x_123_split_cast_fp16")]; + tensor x_123_split_1_sigmoid_cast_fp16 = sigmoid(x = x_123_split_cast_fp16_1)[name = tensor("x_123_split_1_sigmoid_cast_fp16")]; + tensor x_123_cast_fp16 = mul(x = x_123_split_cast_fp16_0, y = x_123_split_1_sigmoid_cast_fp16)[name = tensor("x_123_cast_fp16")]; + tensor input_261_cast_fp16 = select(a = var_40_to_fp16, b = x_123_cast_fp16, cond = var_418)[name = tensor("input_261_cast_fp16")]; + tensor new_x_19_interleave_0 = const()[name = tensor("new_x_19_interleave_0"), val = tensor(false)]; + tensor new_x_19_cast_fp16 = concat(axis = var_62, interleave = new_x_19_interleave_0, values = (cache_19_cast_fp16, input_261_cast_fp16))[name = tensor("new_x_19_cast_fp16")]; + tensor var_1247_begin_0 = const()[name = tensor("op_1247_begin_0"), val = tensor([0, 0, 17])]; + tensor var_1247_end_0 = const()[name = tensor("op_1247_end_0"), val = tensor([1, 512, 25])]; + tensor var_1247_end_mask_0 = const()[name = tensor("op_1247_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1247_cast_fp16 = slice_by_index(begin = var_1247_begin_0, end = var_1247_end_0, end_mask = var_1247_end_mask_0, x = new_x_19_cast_fp16)[name = tensor("op_1247_cast_fp16")]; + tensor x_125_pad_type_0 = const()[name = tensor("x_125_pad_type_0"), val = tensor("valid")]; + tensor x_125_groups_0 = const()[name = tensor("x_125_groups_0"), val = tensor(512)]; + tensor x_125_strides_0 = const()[name = tensor("x_125_strides_0"), val = tensor([1])]; + tensor x_125_pad_0 = const()[name = tensor("x_125_pad_0"), val = tensor([0, 0])]; + tensor x_125_dilations_0 = const()[name = tensor("x_125_dilations_0"), val = tensor([1])]; + tensor encoder_layers_4_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_4_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61312704)))]; + tensor x_125_cast_fp16 = conv(dilations = x_125_dilations_0, groups = x_125_groups_0, pad = x_125_pad_0, pad_type = x_125_pad_type_0, strides = x_125_strides_0, weight = encoder_layers_4_conv_depthwise_conv_weight_to_fp16, x = new_x_19_cast_fp16)[name = tensor("x_125_cast_fp16")]; + tensor input_263_perm_0 = const()[name = tensor("input_263_perm_0"), val = tensor([0, 2, 1])]; + tensor x_127_axes_0 = const()[name = tensor("x_127_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_4_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61321984)))]; + tensor encoder_layers_4_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_4_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61323072)))]; + tensor input_263_cast_fp16 = transpose(perm = input_263_perm_0, x = x_125_cast_fp16)[name = tensor("transpose_196")]; + tensor x_127_cast_fp16 = layer_norm(axes = x_127_axes_0, beta = encoder_layers_4_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_conv_batch_norm_weight_to_fp16, x = input_263_cast_fp16)[name = tensor("x_127_cast_fp16")]; + tensor input_265_perm_0 = const()[name = tensor("input_265_perm_0"), val = tensor([0, 2, 1])]; + tensor input_265_cast_fp16 = transpose(perm = input_265_perm_0, x = x_127_cast_fp16)[name = tensor("transpose_195")]; + tensor input_267_cast_fp16 = silu(x = input_265_cast_fp16)[name = tensor("input_267_cast_fp16")]; + tensor x_129_pad_type_0 = const()[name = tensor("x_129_pad_type_0"), val = tensor("valid")]; + tensor x_129_strides_0 = const()[name = tensor("x_129_strides_0"), val = tensor([1])]; + tensor x_129_pad_0 = const()[name = tensor("x_129_pad_0"), val = tensor([0, 0])]; + tensor x_129_dilations_0 = const()[name = tensor("x_129_dilations_0"), val = tensor([1])]; + tensor x_129_groups_0 = const()[name = tensor("x_129_groups_0"), val = tensor(1)]; + tensor encoder_layers_4_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_4_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61324160)))]; + tensor x_129_cast_fp16 = conv(dilations = x_129_dilations_0, groups = x_129_groups_0, pad = x_129_pad_0, pad_type = x_129_pad_type_0, strides = x_129_strides_0, weight = encoder_layers_4_conv_pointwise_conv2_weight_to_fp16, x = input_267_cast_fp16)[name = tensor("x_129_cast_fp16")]; + tensor input_269_perm_0 = const()[name = tensor("input_269_perm_0"), val = tensor([0, 2, 1])]; + tensor input_269_cast_fp16 = transpose(perm = input_269_perm_0, x = x_129_cast_fp16)[name = tensor("transpose_194")]; + tensor input_271_cast_fp16 = add(x = input_255_cast_fp16, y = input_269_cast_fp16)[name = tensor("input_271_cast_fp16")]; + tensor input_273_axes_0 = const()[name = tensor("input_273_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61848512)))]; + tensor encoder_layers_4_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61849600)))]; + tensor input_273_cast_fp16 = layer_norm(axes = input_273_axes_0, beta = encoder_layers_4_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_norm_feed_forward2_weight_to_fp16, x = input_271_cast_fp16)[name = tensor("input_273_cast_fp16")]; + tensor encoder_layers_4_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_4_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61850688)))]; + tensor linear_44_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_4_feed_forward2_linear1_weight_to_fp16, x = input_273_cast_fp16)[name = tensor("linear_44_cast_fp16")]; + tensor input_277_cast_fp16 = silu(x = linear_44_cast_fp16)[name = tensor("input_277_cast_fp16")]; + tensor encoder_layers_4_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_4_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63947904)))]; + tensor linear_45_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_feed_forward2_linear2_weight_to_fp16, x = input_277_cast_fp16)[name = tensor("linear_45_cast_fp16")]; + tensor var_1288_to_fp16 = const()[name = tensor("op_1288_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1289_cast_fp16 = mul(x = linear_45_cast_fp16, y = var_1288_to_fp16)[name = tensor("op_1289_cast_fp16")]; + tensor input_283_cast_fp16 = add(x = input_271_cast_fp16, y = var_1289_cast_fp16)[name = tensor("input_283_cast_fp16")]; + tensor input_285_axes_0 = const()[name = tensor("input_285_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66045120)))]; + tensor encoder_layers_4_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66046208)))]; + tensor input_285_cast_fp16 = layer_norm(axes = input_285_axes_0, beta = encoder_layers_4_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_norm_out_weight_to_fp16, x = input_283_cast_fp16)[name = tensor("input_285_cast_fp16")]; + tensor cache_21_begin_0 = const()[name = tensor("cache_21_begin_0"), val = tensor([5, 0, 0, 0])]; + tensor cache_21_end_0 = const()[name = tensor("cache_21_end_0"), val = tensor([6, 1, 70, 512])]; + tensor cache_21_end_mask_0 = const()[name = tensor("cache_21_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_21_squeeze_mask_0 = const()[name = tensor("cache_21_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_21_cast_fp16 = slice_by_index(begin = cache_21_begin_0, end = cache_21_end_0, end_mask = cache_21_end_mask_0, squeeze_mask = cache_21_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_21_cast_fp16")]; + tensor cache_23_begin_0 = const()[name = tensor("cache_23_begin_0"), val = tensor([5, 0, 0, 0])]; + tensor cache_23_end_0 = const()[name = tensor("cache_23_end_0"), val = tensor([6, 1, 512, 8])]; + tensor cache_23_end_mask_0 = const()[name = tensor("cache_23_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_23_squeeze_mask_0 = const()[name = tensor("cache_23_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_23_cast_fp16 = slice_by_index(begin = cache_23_begin_0, end = cache_23_end_0, end_mask = cache_23_end_mask_0, squeeze_mask = cache_23_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_23_cast_fp16")]; + tensor input_287_axes_0 = const()[name = tensor("input_287_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66047296)))]; + tensor encoder_layers_5_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66048384)))]; + tensor input_287_cast_fp16 = layer_norm(axes = input_287_axes_0, beta = encoder_layers_5_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_norm_feed_forward1_weight_to_fp16, x = input_285_cast_fp16)[name = tensor("input_287_cast_fp16")]; + tensor encoder_layers_5_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_5_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66049472)))]; + tensor linear_46_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_5_feed_forward1_linear1_weight_to_fp16, x = input_287_cast_fp16)[name = tensor("linear_46_cast_fp16")]; + tensor input_291_cast_fp16 = silu(x = linear_46_cast_fp16)[name = tensor("input_291_cast_fp16")]; + tensor encoder_layers_5_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_5_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68146688)))]; + tensor linear_47_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_feed_forward1_linear2_weight_to_fp16, x = input_291_cast_fp16)[name = tensor("linear_47_cast_fp16")]; + tensor var_1323_to_fp16 = const()[name = tensor("op_1323_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1324_cast_fp16 = mul(x = linear_47_cast_fp16, y = var_1323_to_fp16)[name = tensor("op_1324_cast_fp16")]; + tensor input_297_cast_fp16 = add(x = input_285_cast_fp16, y = var_1324_cast_fp16)[name = tensor("input_297_cast_fp16")]; + tensor key_11_axes_0 = const()[name = tensor("key_11_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70243904)))]; + tensor encoder_layers_5_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70244992)))]; + tensor key_11_cast_fp16 = layer_norm(axes = key_11_axes_0, beta = encoder_layers_5_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_norm_self_att_weight_to_fp16, x = input_297_cast_fp16)[name = tensor("key_11_cast_fp16")]; + tensor input_299_interleave_0 = const()[name = tensor("input_299_interleave_0"), val = tensor(false)]; + tensor input_299_cast_fp16 = concat(axis = var_64, interleave = input_299_interleave_0, values = (cache_21_cast_fp16, key_11_cast_fp16))[name = tensor("input_299_cast_fp16")]; + tensor var_1346_begin_0 = const()[name = tensor("op_1346_begin_0"), val = tensor([0, 17, 0])]; + tensor var_1346_end_0 = const()[name = tensor("op_1346_end_0"), val = tensor([1, 70, 512])]; + tensor var_1346_end_mask_0 = const()[name = tensor("op_1346_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1346_cast_fp16 = slice_by_index(begin = var_1346_begin_0, end = var_1346_end_0, end_mask = var_1346_end_mask_0, x = cache_21_cast_fp16)[name = tensor("op_1346_cast_fp16")]; + tensor var_1352_interleave_0 = const()[name = tensor("op_1352_interleave_0"), val = tensor(false)]; + tensor var_1352_cast_fp16 = concat(axis = var_64, interleave = var_1352_interleave_0, values = (var_1346_cast_fp16, key_11_cast_fp16))[name = tensor("op_1352_cast_fp16")]; + tensor encoder_layers_5_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70246080)))]; + tensor linear_48_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_self_attn_linear_q_weight_to_fp16, x = key_11_cast_fp16)[name = tensor("linear_48_cast_fp16")]; + tensor var_1356 = const()[name = tensor("op_1356"), val = tensor([1, -1, 8, 64])]; + tensor q_31_cast_fp16 = reshape(shape = var_1356, x = linear_48_cast_fp16)[name = tensor("q_31_cast_fp16")]; + tensor encoder_layers_5_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70770432)))]; + tensor linear_49_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_self_attn_linear_k_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("linear_49_cast_fp16")]; + tensor var_1360 = const()[name = tensor("op_1360"), val = tensor([1, -1, 8, 64])]; + tensor k_21_cast_fp16 = reshape(shape = var_1360, x = linear_49_cast_fp16)[name = tensor("k_21_cast_fp16")]; + tensor encoder_layers_5_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71294784)))]; + tensor linear_50_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_self_attn_linear_v_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("linear_50_cast_fp16")]; + tensor var_1364 = const()[name = tensor("op_1364"), val = tensor([1, -1, 8, 64])]; + tensor v_11_cast_fp16 = reshape(shape = var_1364, x = linear_50_cast_fp16)[name = tensor("v_11_cast_fp16")]; + tensor value_13_perm_0 = const()[name = tensor("value_13_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_5_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71819136)))]; + tensor var_1376_cast_fp16 = add(x = q_31_cast_fp16, y = encoder_layers_5_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1376_cast_fp16")]; + tensor encoder_layers_5_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71820224)))]; + tensor var_1378_cast_fp16 = add(x = q_31_cast_fp16, y = encoder_layers_5_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1378_cast_fp16")]; + tensor q_with_bias_v_11_perm_0 = const()[name = tensor("q_with_bias_v_11_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_137_transpose_x_0 = const()[name = tensor("x_137_transpose_x_0"), val = tensor(false)]; + tensor x_137_transpose_y_0 = const()[name = tensor("x_137_transpose_y_0"), val = tensor(false)]; + tensor var_1380_to_fp16 = const()[name = tensor("op_1380_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71821312)))]; + tensor q_with_bias_v_11_cast_fp16 = transpose(perm = q_with_bias_v_11_perm_0, x = var_1378_cast_fp16)[name = tensor("transpose_192")]; + tensor x_137_cast_fp16 = matmul(transpose_x = x_137_transpose_x_0, transpose_y = x_137_transpose_y_0, x = q_with_bias_v_11_cast_fp16, y = var_1380_to_fp16)[name = tensor("x_137_cast_fp16")]; + tensor x_139_pad_0 = const()[name = tensor("x_139_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_139_mode_0 = const()[name = tensor("x_139_mode_0"), val = tensor("constant")]; + tensor const_88_to_fp16 = const()[name = tensor("const_88_to_fp16"), val = tensor(0x0p+0)]; + tensor x_139_cast_fp16 = pad(constant_val = const_88_to_fp16, mode = x_139_mode_0, pad = x_139_pad_0, x = x_137_cast_fp16)[name = tensor("x_139_cast_fp16")]; + tensor var_1388 = const()[name = tensor("op_1388"), val = tensor([1, 8, -1, 17])]; + tensor x_141_cast_fp16 = reshape(shape = var_1388, x = x_139_cast_fp16)[name = tensor("x_141_cast_fp16")]; + tensor var_1392_begin_0 = const()[name = tensor("op_1392_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1392_end_0 = const()[name = tensor("op_1392_end_0"), val = tensor([1, 8, 174, 17])]; + tensor var_1392_end_mask_0 = const()[name = tensor("op_1392_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1392_cast_fp16 = slice_by_index(begin = var_1392_begin_0, end = var_1392_end_0, end_mask = var_1392_end_mask_0, x = x_141_cast_fp16)[name = tensor("op_1392_cast_fp16")]; + tensor var_1393 = const()[name = tensor("op_1393"), val = tensor([1, 8, 17, 173])]; + tensor matrix_bd_21_cast_fp16 = reshape(shape = var_1393, x = var_1392_cast_fp16)[name = tensor("matrix_bd_21_cast_fp16")]; + tensor matrix_ac_11_transpose_x_0 = const()[name = tensor("matrix_ac_11_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_11_transpose_y_0 = const()[name = tensor("matrix_ac_11_transpose_y_0"), val = tensor(false)]; + tensor transpose_61_perm_0 = const()[name = tensor("transpose_61_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_62_perm_0 = const()[name = tensor("transpose_62_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_62 = transpose(perm = transpose_62_perm_0, x = k_21_cast_fp16)[name = tensor("transpose_190")]; + tensor transpose_61 = transpose(perm = transpose_61_perm_0, x = var_1376_cast_fp16)[name = tensor("transpose_191")]; + tensor matrix_ac_11_cast_fp16 = matmul(transpose_x = matrix_ac_11_transpose_x_0, transpose_y = matrix_ac_11_transpose_y_0, x = transpose_61, y = transpose_62)[name = tensor("matrix_ac_11_cast_fp16")]; + tensor matrix_bd_23_begin_0 = const()[name = tensor("matrix_bd_23_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_23_end_0 = const()[name = tensor("matrix_bd_23_end_0"), val = tensor([1, 8, 17, 87])]; + tensor matrix_bd_23_end_mask_0 = const()[name = tensor("matrix_bd_23_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_23_cast_fp16 = slice_by_index(begin = matrix_bd_23_begin_0, end = matrix_bd_23_end_0, end_mask = matrix_bd_23_end_mask_0, x = matrix_bd_21_cast_fp16)[name = tensor("matrix_bd_23_cast_fp16")]; + tensor var_1402_cast_fp16 = add(x = matrix_ac_11_cast_fp16, y = matrix_bd_23_cast_fp16)[name = tensor("op_1402_cast_fp16")]; + tensor _inversed_scores_21_y_0_to_fp16 = const()[name = tensor("_inversed_scores_21_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_21_cast_fp16 = mul(x = var_1402_cast_fp16, y = _inversed_scores_21_y_0_to_fp16)[name = tensor("_inversed_scores_21_cast_fp16")]; + tensor scores_23_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_21_cast_fp16, cond = mask_3)[name = tensor("scores_23_cast_fp16")]; + tensor var_1408_cast_fp16 = softmax(axis = var_62, x = scores_23_cast_fp16)[name = tensor("op_1408_cast_fp16")]; + tensor input_301_cast_fp16 = select(a = var_40_to_fp16, b = var_1408_cast_fp16, cond = mask_3)[name = tensor("input_301_cast_fp16")]; + tensor x_143_transpose_x_0 = const()[name = tensor("x_143_transpose_x_0"), val = tensor(false)]; + tensor x_143_transpose_y_0 = const()[name = tensor("x_143_transpose_y_0"), val = tensor(false)]; + tensor value_13_cast_fp16 = transpose(perm = value_13_perm_0, x = v_11_cast_fp16)[name = tensor("transpose_193")]; + tensor x_143_cast_fp16 = matmul(transpose_x = x_143_transpose_x_0, transpose_y = x_143_transpose_y_0, x = input_301_cast_fp16, y = value_13_cast_fp16)[name = tensor("x_143_cast_fp16")]; + tensor var_1412_perm_0 = const()[name = tensor("op_1412_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1413 = const()[name = tensor("op_1413"), val = tensor([1, -1, 512])]; + tensor var_1412_cast_fp16 = transpose(perm = var_1412_perm_0, x = x_143_cast_fp16)[name = tensor("transpose_189")]; + tensor input_303_cast_fp16 = reshape(shape = var_1413, x = var_1412_cast_fp16)[name = tensor("input_303_cast_fp16")]; + tensor encoder_layers_5_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71998528)))]; + tensor linear_52_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_self_attn_linear_out_weight_to_fp16, x = input_303_cast_fp16)[name = tensor("linear_52_cast_fp16")]; + tensor input_307_cast_fp16 = add(x = input_297_cast_fp16, y = linear_52_cast_fp16)[name = tensor("input_307_cast_fp16")]; + tensor x_147_axes_0 = const()[name = tensor("x_147_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72522880)))]; + tensor encoder_layers_5_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72523968)))]; + tensor x_147_cast_fp16 = layer_norm(axes = x_147_axes_0, beta = encoder_layers_5_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_norm_conv_weight_to_fp16, x = input_307_cast_fp16)[name = tensor("x_147_cast_fp16")]; + tensor input_309_perm_0 = const()[name = tensor("input_309_perm_0"), val = tensor([0, 2, 1])]; + tensor input_311_pad_type_0 = const()[name = tensor("input_311_pad_type_0"), val = tensor("valid")]; + tensor input_311_strides_0 = const()[name = tensor("input_311_strides_0"), val = tensor([1])]; + tensor input_311_pad_0 = const()[name = tensor("input_311_pad_0"), val = tensor([0, 0])]; + tensor input_311_dilations_0 = const()[name = tensor("input_311_dilations_0"), val = tensor([1])]; + tensor input_311_groups_0 = const()[name = tensor("input_311_groups_0"), val = tensor(1)]; + tensor encoder_layers_5_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_5_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72525056)))]; + tensor input_309_cast_fp16 = transpose(perm = input_309_perm_0, x = x_147_cast_fp16)[name = tensor("transpose_188")]; + tensor input_311_cast_fp16 = conv(dilations = input_311_dilations_0, groups = input_311_groups_0, pad = input_311_pad_0, pad_type = input_311_pad_type_0, strides = input_311_strides_0, weight = encoder_layers_5_conv_pointwise_conv1_weight_to_fp16, x = input_309_cast_fp16)[name = tensor("input_311_cast_fp16")]; + tensor x_149_split_num_splits_0 = const()[name = tensor("x_149_split_num_splits_0"), val = tensor(2)]; + tensor x_149_split_axis_0 = const()[name = tensor("x_149_split_axis_0"), val = tensor(1)]; + tensor x_149_split_cast_fp16_0, tensor x_149_split_cast_fp16_1 = split(axis = x_149_split_axis_0, num_splits = x_149_split_num_splits_0, x = input_311_cast_fp16)[name = tensor("x_149_split_cast_fp16")]; + tensor x_149_split_1_sigmoid_cast_fp16 = sigmoid(x = x_149_split_cast_fp16_1)[name = tensor("x_149_split_1_sigmoid_cast_fp16")]; + tensor x_149_cast_fp16 = mul(x = x_149_split_cast_fp16_0, y = x_149_split_1_sigmoid_cast_fp16)[name = tensor("x_149_cast_fp16")]; + tensor input_313_cast_fp16 = select(a = var_40_to_fp16, b = x_149_cast_fp16, cond = var_418)[name = tensor("input_313_cast_fp16")]; + tensor new_x_23_interleave_0 = const()[name = tensor("new_x_23_interleave_0"), val = tensor(false)]; + tensor new_x_23_cast_fp16 = concat(axis = var_62, interleave = new_x_23_interleave_0, values = (cache_23_cast_fp16, input_313_cast_fp16))[name = tensor("new_x_23_cast_fp16")]; + tensor var_1451_begin_0 = const()[name = tensor("op_1451_begin_0"), val = tensor([0, 0, 17])]; + tensor var_1451_end_0 = const()[name = tensor("op_1451_end_0"), val = tensor([1, 512, 25])]; + tensor var_1451_end_mask_0 = const()[name = tensor("op_1451_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1451_cast_fp16 = slice_by_index(begin = var_1451_begin_0, end = var_1451_end_0, end_mask = var_1451_end_mask_0, x = new_x_23_cast_fp16)[name = tensor("op_1451_cast_fp16")]; + tensor x_151_pad_type_0 = const()[name = tensor("x_151_pad_type_0"), val = tensor("valid")]; + tensor x_151_groups_0 = const()[name = tensor("x_151_groups_0"), val = tensor(512)]; + tensor x_151_strides_0 = const()[name = tensor("x_151_strides_0"), val = tensor([1])]; + tensor x_151_pad_0 = const()[name = tensor("x_151_pad_0"), val = tensor([0, 0])]; + tensor x_151_dilations_0 = const()[name = tensor("x_151_dilations_0"), val = tensor([1])]; + tensor encoder_layers_5_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_5_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73573696)))]; + tensor x_151_cast_fp16 = conv(dilations = x_151_dilations_0, groups = x_151_groups_0, pad = x_151_pad_0, pad_type = x_151_pad_type_0, strides = x_151_strides_0, weight = encoder_layers_5_conv_depthwise_conv_weight_to_fp16, x = new_x_23_cast_fp16)[name = tensor("x_151_cast_fp16")]; + tensor input_315_perm_0 = const()[name = tensor("input_315_perm_0"), val = tensor([0, 2, 1])]; + tensor x_153_axes_0 = const()[name = tensor("x_153_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_5_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73582976)))]; + tensor encoder_layers_5_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_5_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73584064)))]; + tensor input_315_cast_fp16 = transpose(perm = input_315_perm_0, x = x_151_cast_fp16)[name = tensor("transpose_187")]; + tensor x_153_cast_fp16 = layer_norm(axes = x_153_axes_0, beta = encoder_layers_5_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_conv_batch_norm_weight_to_fp16, x = input_315_cast_fp16)[name = tensor("x_153_cast_fp16")]; + tensor input_317_perm_0 = const()[name = tensor("input_317_perm_0"), val = tensor([0, 2, 1])]; + tensor input_317_cast_fp16 = transpose(perm = input_317_perm_0, x = x_153_cast_fp16)[name = tensor("transpose_186")]; + tensor input_319_cast_fp16 = silu(x = input_317_cast_fp16)[name = tensor("input_319_cast_fp16")]; + tensor x_155_pad_type_0 = const()[name = tensor("x_155_pad_type_0"), val = tensor("valid")]; + tensor x_155_strides_0 = const()[name = tensor("x_155_strides_0"), val = tensor([1])]; + tensor x_155_pad_0 = const()[name = tensor("x_155_pad_0"), val = tensor([0, 0])]; + tensor x_155_dilations_0 = const()[name = tensor("x_155_dilations_0"), val = tensor([1])]; + tensor x_155_groups_0 = const()[name = tensor("x_155_groups_0"), val = tensor(1)]; + tensor encoder_layers_5_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_5_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73585152)))]; + tensor x_155_cast_fp16 = conv(dilations = x_155_dilations_0, groups = x_155_groups_0, pad = x_155_pad_0, pad_type = x_155_pad_type_0, strides = x_155_strides_0, weight = encoder_layers_5_conv_pointwise_conv2_weight_to_fp16, x = input_319_cast_fp16)[name = tensor("x_155_cast_fp16")]; + tensor input_321_perm_0 = const()[name = tensor("input_321_perm_0"), val = tensor([0, 2, 1])]; + tensor input_321_cast_fp16 = transpose(perm = input_321_perm_0, x = x_155_cast_fp16)[name = tensor("transpose_185")]; + tensor input_323_cast_fp16 = add(x = input_307_cast_fp16, y = input_321_cast_fp16)[name = tensor("input_323_cast_fp16")]; + tensor input_325_axes_0 = const()[name = tensor("input_325_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74109504)))]; + tensor encoder_layers_5_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74110592)))]; + tensor input_325_cast_fp16 = layer_norm(axes = input_325_axes_0, beta = encoder_layers_5_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_norm_feed_forward2_weight_to_fp16, x = input_323_cast_fp16)[name = tensor("input_325_cast_fp16")]; + tensor encoder_layers_5_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_5_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74111680)))]; + tensor linear_53_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_5_feed_forward2_linear1_weight_to_fp16, x = input_325_cast_fp16)[name = tensor("linear_53_cast_fp16")]; + tensor input_329_cast_fp16 = silu(x = linear_53_cast_fp16)[name = tensor("input_329_cast_fp16")]; + tensor encoder_layers_5_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_5_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76208896)))]; + tensor linear_54_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_feed_forward2_linear2_weight_to_fp16, x = input_329_cast_fp16)[name = tensor("linear_54_cast_fp16")]; + tensor var_1492_to_fp16 = const()[name = tensor("op_1492_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1493_cast_fp16 = mul(x = linear_54_cast_fp16, y = var_1492_to_fp16)[name = tensor("op_1493_cast_fp16")]; + tensor input_335_cast_fp16 = add(x = input_323_cast_fp16, y = var_1493_cast_fp16)[name = tensor("input_335_cast_fp16")]; + tensor input_337_axes_0 = const()[name = tensor("input_337_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78306112)))]; + tensor encoder_layers_5_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78307200)))]; + tensor input_337_cast_fp16 = layer_norm(axes = input_337_axes_0, beta = encoder_layers_5_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_norm_out_weight_to_fp16, x = input_335_cast_fp16)[name = tensor("input_337_cast_fp16")]; + tensor cache_25_begin_0 = const()[name = tensor("cache_25_begin_0"), val = tensor([6, 0, 0, 0])]; + tensor cache_25_end_0 = const()[name = tensor("cache_25_end_0"), val = tensor([7, 1, 70, 512])]; + tensor cache_25_end_mask_0 = const()[name = tensor("cache_25_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_25_squeeze_mask_0 = const()[name = tensor("cache_25_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_25_cast_fp16 = slice_by_index(begin = cache_25_begin_0, end = cache_25_end_0, end_mask = cache_25_end_mask_0, squeeze_mask = cache_25_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_25_cast_fp16")]; + tensor cache_27_begin_0 = const()[name = tensor("cache_27_begin_0"), val = tensor([6, 0, 0, 0])]; + tensor cache_27_end_0 = const()[name = tensor("cache_27_end_0"), val = tensor([7, 1, 512, 8])]; + tensor cache_27_end_mask_0 = const()[name = tensor("cache_27_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_27_squeeze_mask_0 = const()[name = tensor("cache_27_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_27_cast_fp16 = slice_by_index(begin = cache_27_begin_0, end = cache_27_end_0, end_mask = cache_27_end_mask_0, squeeze_mask = cache_27_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_27_cast_fp16")]; + tensor input_339_axes_0 = const()[name = tensor("input_339_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78308288)))]; + tensor encoder_layers_6_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78309376)))]; + tensor input_339_cast_fp16 = layer_norm(axes = input_339_axes_0, beta = encoder_layers_6_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_norm_feed_forward1_weight_to_fp16, x = input_337_cast_fp16)[name = tensor("input_339_cast_fp16")]; + tensor encoder_layers_6_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_6_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78310464)))]; + tensor linear_55_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_6_feed_forward1_linear1_weight_to_fp16, x = input_339_cast_fp16)[name = tensor("linear_55_cast_fp16")]; + tensor input_343_cast_fp16 = silu(x = linear_55_cast_fp16)[name = tensor("input_343_cast_fp16")]; + tensor encoder_layers_6_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_6_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80407680)))]; + tensor linear_56_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_feed_forward1_linear2_weight_to_fp16, x = input_343_cast_fp16)[name = tensor("linear_56_cast_fp16")]; + tensor var_1527_to_fp16 = const()[name = tensor("op_1527_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1528_cast_fp16 = mul(x = linear_56_cast_fp16, y = var_1527_to_fp16)[name = tensor("op_1528_cast_fp16")]; + tensor input_349_cast_fp16 = add(x = input_337_cast_fp16, y = var_1528_cast_fp16)[name = tensor("input_349_cast_fp16")]; + tensor key_13_axes_0 = const()[name = tensor("key_13_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82504896)))]; + tensor encoder_layers_6_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82505984)))]; + tensor key_13_cast_fp16 = layer_norm(axes = key_13_axes_0, beta = encoder_layers_6_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_norm_self_att_weight_to_fp16, x = input_349_cast_fp16)[name = tensor("key_13_cast_fp16")]; + tensor input_351_interleave_0 = const()[name = tensor("input_351_interleave_0"), val = tensor(false)]; + tensor input_351_cast_fp16 = concat(axis = var_64, interleave = input_351_interleave_0, values = (cache_25_cast_fp16, key_13_cast_fp16))[name = tensor("input_351_cast_fp16")]; + tensor var_1550_begin_0 = const()[name = tensor("op_1550_begin_0"), val = tensor([0, 17, 0])]; + tensor var_1550_end_0 = const()[name = tensor("op_1550_end_0"), val = tensor([1, 70, 512])]; + tensor var_1550_end_mask_0 = const()[name = tensor("op_1550_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1550_cast_fp16 = slice_by_index(begin = var_1550_begin_0, end = var_1550_end_0, end_mask = var_1550_end_mask_0, x = cache_25_cast_fp16)[name = tensor("op_1550_cast_fp16")]; + tensor var_1556_interleave_0 = const()[name = tensor("op_1556_interleave_0"), val = tensor(false)]; + tensor var_1556_cast_fp16 = concat(axis = var_64, interleave = var_1556_interleave_0, values = (var_1550_cast_fp16, key_13_cast_fp16))[name = tensor("op_1556_cast_fp16")]; + tensor encoder_layers_6_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82507072)))]; + tensor linear_57_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_self_attn_linear_q_weight_to_fp16, x = key_13_cast_fp16)[name = tensor("linear_57_cast_fp16")]; + tensor var_1560 = const()[name = tensor("op_1560"), val = tensor([1, -1, 8, 64])]; + tensor q_37_cast_fp16 = reshape(shape = var_1560, x = linear_57_cast_fp16)[name = tensor("q_37_cast_fp16")]; + tensor encoder_layers_6_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83031424)))]; + tensor linear_58_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_self_attn_linear_k_weight_to_fp16, x = input_351_cast_fp16)[name = tensor("linear_58_cast_fp16")]; + tensor var_1564 = const()[name = tensor("op_1564"), val = tensor([1, -1, 8, 64])]; + tensor k_25_cast_fp16 = reshape(shape = var_1564, x = linear_58_cast_fp16)[name = tensor("k_25_cast_fp16")]; + tensor encoder_layers_6_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83555776)))]; + tensor linear_59_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_self_attn_linear_v_weight_to_fp16, x = input_351_cast_fp16)[name = tensor("linear_59_cast_fp16")]; + tensor var_1568 = const()[name = tensor("op_1568"), val = tensor([1, -1, 8, 64])]; + tensor v_13_cast_fp16 = reshape(shape = var_1568, x = linear_59_cast_fp16)[name = tensor("v_13_cast_fp16")]; + tensor value_15_perm_0 = const()[name = tensor("value_15_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_6_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84080128)))]; + tensor var_1580_cast_fp16 = add(x = q_37_cast_fp16, y = encoder_layers_6_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1580_cast_fp16")]; + tensor encoder_layers_6_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84081216)))]; + tensor var_1582_cast_fp16 = add(x = q_37_cast_fp16, y = encoder_layers_6_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1582_cast_fp16")]; + tensor q_with_bias_v_13_perm_0 = const()[name = tensor("q_with_bias_v_13_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_163_transpose_x_0 = const()[name = tensor("x_163_transpose_x_0"), val = tensor(false)]; + tensor x_163_transpose_y_0 = const()[name = tensor("x_163_transpose_y_0"), val = tensor(false)]; + tensor var_1584_to_fp16 = const()[name = tensor("op_1584_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84082304)))]; + tensor q_with_bias_v_13_cast_fp16 = transpose(perm = q_with_bias_v_13_perm_0, x = var_1582_cast_fp16)[name = tensor("transpose_183")]; + tensor x_163_cast_fp16 = matmul(transpose_x = x_163_transpose_x_0, transpose_y = x_163_transpose_y_0, x = q_with_bias_v_13_cast_fp16, y = var_1584_to_fp16)[name = tensor("x_163_cast_fp16")]; + tensor x_165_pad_0 = const()[name = tensor("x_165_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_165_mode_0 = const()[name = tensor("x_165_mode_0"), val = tensor("constant")]; + tensor const_101_to_fp16 = const()[name = tensor("const_101_to_fp16"), val = tensor(0x0p+0)]; + tensor x_165_cast_fp16 = pad(constant_val = const_101_to_fp16, mode = x_165_mode_0, pad = x_165_pad_0, x = x_163_cast_fp16)[name = tensor("x_165_cast_fp16")]; + tensor var_1592 = const()[name = tensor("op_1592"), val = tensor([1, 8, -1, 17])]; + tensor x_167_cast_fp16 = reshape(shape = var_1592, x = x_165_cast_fp16)[name = tensor("x_167_cast_fp16")]; + tensor var_1596_begin_0 = const()[name = tensor("op_1596_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1596_end_0 = const()[name = tensor("op_1596_end_0"), val = tensor([1, 8, 174, 17])]; + tensor var_1596_end_mask_0 = const()[name = tensor("op_1596_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1596_cast_fp16 = slice_by_index(begin = var_1596_begin_0, end = var_1596_end_0, end_mask = var_1596_end_mask_0, x = x_167_cast_fp16)[name = tensor("op_1596_cast_fp16")]; + tensor var_1597 = const()[name = tensor("op_1597"), val = tensor([1, 8, 17, 173])]; + tensor matrix_bd_25_cast_fp16 = reshape(shape = var_1597, x = var_1596_cast_fp16)[name = tensor("matrix_bd_25_cast_fp16")]; + tensor matrix_ac_13_transpose_x_0 = const()[name = tensor("matrix_ac_13_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_13_transpose_y_0 = const()[name = tensor("matrix_ac_13_transpose_y_0"), val = tensor(false)]; + tensor transpose_63_perm_0 = const()[name = tensor("transpose_63_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_64_perm_0 = const()[name = tensor("transpose_64_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_64 = transpose(perm = transpose_64_perm_0, x = k_25_cast_fp16)[name = tensor("transpose_181")]; + tensor transpose_63 = transpose(perm = transpose_63_perm_0, x = var_1580_cast_fp16)[name = tensor("transpose_182")]; + tensor matrix_ac_13_cast_fp16 = matmul(transpose_x = matrix_ac_13_transpose_x_0, transpose_y = matrix_ac_13_transpose_y_0, x = transpose_63, y = transpose_64)[name = tensor("matrix_ac_13_cast_fp16")]; + tensor matrix_bd_27_begin_0 = const()[name = tensor("matrix_bd_27_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_27_end_0 = const()[name = tensor("matrix_bd_27_end_0"), val = tensor([1, 8, 17, 87])]; + tensor matrix_bd_27_end_mask_0 = const()[name = tensor("matrix_bd_27_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_27_cast_fp16 = slice_by_index(begin = matrix_bd_27_begin_0, end = matrix_bd_27_end_0, end_mask = matrix_bd_27_end_mask_0, x = matrix_bd_25_cast_fp16)[name = tensor("matrix_bd_27_cast_fp16")]; + tensor var_1606_cast_fp16 = add(x = matrix_ac_13_cast_fp16, y = matrix_bd_27_cast_fp16)[name = tensor("op_1606_cast_fp16")]; + tensor _inversed_scores_25_y_0_to_fp16 = const()[name = tensor("_inversed_scores_25_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_25_cast_fp16 = mul(x = var_1606_cast_fp16, y = _inversed_scores_25_y_0_to_fp16)[name = tensor("_inversed_scores_25_cast_fp16")]; + tensor scores_27_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_25_cast_fp16, cond = mask_3)[name = tensor("scores_27_cast_fp16")]; + tensor var_1612_cast_fp16 = softmax(axis = var_62, x = scores_27_cast_fp16)[name = tensor("op_1612_cast_fp16")]; + tensor input_353_cast_fp16 = select(a = var_40_to_fp16, b = var_1612_cast_fp16, cond = mask_3)[name = tensor("input_353_cast_fp16")]; + tensor x_169_transpose_x_0 = const()[name = tensor("x_169_transpose_x_0"), val = tensor(false)]; + tensor x_169_transpose_y_0 = const()[name = tensor("x_169_transpose_y_0"), val = tensor(false)]; + tensor value_15_cast_fp16 = transpose(perm = value_15_perm_0, x = v_13_cast_fp16)[name = tensor("transpose_184")]; + tensor x_169_cast_fp16 = matmul(transpose_x = x_169_transpose_x_0, transpose_y = x_169_transpose_y_0, x = input_353_cast_fp16, y = value_15_cast_fp16)[name = tensor("x_169_cast_fp16")]; + tensor var_1616_perm_0 = const()[name = tensor("op_1616_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1617 = const()[name = tensor("op_1617"), val = tensor([1, -1, 512])]; + tensor var_1616_cast_fp16 = transpose(perm = var_1616_perm_0, x = x_169_cast_fp16)[name = tensor("transpose_180")]; + tensor input_355_cast_fp16 = reshape(shape = var_1617, x = var_1616_cast_fp16)[name = tensor("input_355_cast_fp16")]; + tensor encoder_layers_6_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84259520)))]; + tensor linear_61_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_self_attn_linear_out_weight_to_fp16, x = input_355_cast_fp16)[name = tensor("linear_61_cast_fp16")]; + tensor input_359_cast_fp16 = add(x = input_349_cast_fp16, y = linear_61_cast_fp16)[name = tensor("input_359_cast_fp16")]; + tensor x_173_axes_0 = const()[name = tensor("x_173_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84783872)))]; + tensor encoder_layers_6_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84784960)))]; + tensor x_173_cast_fp16 = layer_norm(axes = x_173_axes_0, beta = encoder_layers_6_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_norm_conv_weight_to_fp16, x = input_359_cast_fp16)[name = tensor("x_173_cast_fp16")]; + tensor input_361_perm_0 = const()[name = tensor("input_361_perm_0"), val = tensor([0, 2, 1])]; + tensor input_363_pad_type_0 = const()[name = tensor("input_363_pad_type_0"), val = tensor("valid")]; + tensor input_363_strides_0 = const()[name = tensor("input_363_strides_0"), val = tensor([1])]; + tensor input_363_pad_0 = const()[name = tensor("input_363_pad_0"), val = tensor([0, 0])]; + tensor input_363_dilations_0 = const()[name = tensor("input_363_dilations_0"), val = tensor([1])]; + tensor input_363_groups_0 = const()[name = tensor("input_363_groups_0"), val = tensor(1)]; + tensor encoder_layers_6_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_6_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84786048)))]; + tensor input_361_cast_fp16 = transpose(perm = input_361_perm_0, x = x_173_cast_fp16)[name = tensor("transpose_179")]; + tensor input_363_cast_fp16 = conv(dilations = input_363_dilations_0, groups = input_363_groups_0, pad = input_363_pad_0, pad_type = input_363_pad_type_0, strides = input_363_strides_0, weight = encoder_layers_6_conv_pointwise_conv1_weight_to_fp16, x = input_361_cast_fp16)[name = tensor("input_363_cast_fp16")]; + tensor x_175_split_num_splits_0 = const()[name = tensor("x_175_split_num_splits_0"), val = tensor(2)]; + tensor x_175_split_axis_0 = const()[name = tensor("x_175_split_axis_0"), val = tensor(1)]; + tensor x_175_split_cast_fp16_0, tensor x_175_split_cast_fp16_1 = split(axis = x_175_split_axis_0, num_splits = x_175_split_num_splits_0, x = input_363_cast_fp16)[name = tensor("x_175_split_cast_fp16")]; + tensor x_175_split_1_sigmoid_cast_fp16 = sigmoid(x = x_175_split_cast_fp16_1)[name = tensor("x_175_split_1_sigmoid_cast_fp16")]; + tensor x_175_cast_fp16 = mul(x = x_175_split_cast_fp16_0, y = x_175_split_1_sigmoid_cast_fp16)[name = tensor("x_175_cast_fp16")]; + tensor input_365_cast_fp16 = select(a = var_40_to_fp16, b = x_175_cast_fp16, cond = var_418)[name = tensor("input_365_cast_fp16")]; + tensor new_x_27_interleave_0 = const()[name = tensor("new_x_27_interleave_0"), val = tensor(false)]; + tensor new_x_27_cast_fp16 = concat(axis = var_62, interleave = new_x_27_interleave_0, values = (cache_27_cast_fp16, input_365_cast_fp16))[name = tensor("new_x_27_cast_fp16")]; + tensor var_1655_begin_0 = const()[name = tensor("op_1655_begin_0"), val = tensor([0, 0, 17])]; + tensor var_1655_end_0 = const()[name = tensor("op_1655_end_0"), val = tensor([1, 512, 25])]; + tensor var_1655_end_mask_0 = const()[name = tensor("op_1655_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1655_cast_fp16 = slice_by_index(begin = var_1655_begin_0, end = var_1655_end_0, end_mask = var_1655_end_mask_0, x = new_x_27_cast_fp16)[name = tensor("op_1655_cast_fp16")]; + tensor x_177_pad_type_0 = const()[name = tensor("x_177_pad_type_0"), val = tensor("valid")]; + tensor x_177_groups_0 = const()[name = tensor("x_177_groups_0"), val = tensor(512)]; + tensor x_177_strides_0 = const()[name = tensor("x_177_strides_0"), val = tensor([1])]; + tensor x_177_pad_0 = const()[name = tensor("x_177_pad_0"), val = tensor([0, 0])]; + tensor x_177_dilations_0 = const()[name = tensor("x_177_dilations_0"), val = tensor([1])]; + tensor encoder_layers_6_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_6_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85834688)))]; + tensor x_177_cast_fp16 = conv(dilations = x_177_dilations_0, groups = x_177_groups_0, pad = x_177_pad_0, pad_type = x_177_pad_type_0, strides = x_177_strides_0, weight = encoder_layers_6_conv_depthwise_conv_weight_to_fp16, x = new_x_27_cast_fp16)[name = tensor("x_177_cast_fp16")]; + tensor input_367_perm_0 = const()[name = tensor("input_367_perm_0"), val = tensor([0, 2, 1])]; + tensor x_179_axes_0 = const()[name = tensor("x_179_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_6_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85843968)))]; + tensor encoder_layers_6_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_6_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85845056)))]; + tensor input_367_cast_fp16 = transpose(perm = input_367_perm_0, x = x_177_cast_fp16)[name = tensor("transpose_178")]; + tensor x_179_cast_fp16 = layer_norm(axes = x_179_axes_0, beta = encoder_layers_6_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_conv_batch_norm_weight_to_fp16, x = input_367_cast_fp16)[name = tensor("x_179_cast_fp16")]; + tensor input_369_perm_0 = const()[name = tensor("input_369_perm_0"), val = tensor([0, 2, 1])]; + tensor input_369_cast_fp16 = transpose(perm = input_369_perm_0, x = x_179_cast_fp16)[name = tensor("transpose_177")]; + tensor input_371_cast_fp16 = silu(x = input_369_cast_fp16)[name = tensor("input_371_cast_fp16")]; + tensor x_181_pad_type_0 = const()[name = tensor("x_181_pad_type_0"), val = tensor("valid")]; + tensor x_181_strides_0 = const()[name = tensor("x_181_strides_0"), val = tensor([1])]; + tensor x_181_pad_0 = const()[name = tensor("x_181_pad_0"), val = tensor([0, 0])]; + tensor x_181_dilations_0 = const()[name = tensor("x_181_dilations_0"), val = tensor([1])]; + tensor x_181_groups_0 = const()[name = tensor("x_181_groups_0"), val = tensor(1)]; + tensor encoder_layers_6_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_6_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85846144)))]; + tensor x_181_cast_fp16 = conv(dilations = x_181_dilations_0, groups = x_181_groups_0, pad = x_181_pad_0, pad_type = x_181_pad_type_0, strides = x_181_strides_0, weight = encoder_layers_6_conv_pointwise_conv2_weight_to_fp16, x = input_371_cast_fp16)[name = tensor("x_181_cast_fp16")]; + tensor input_373_perm_0 = const()[name = tensor("input_373_perm_0"), val = tensor([0, 2, 1])]; + tensor input_373_cast_fp16 = transpose(perm = input_373_perm_0, x = x_181_cast_fp16)[name = tensor("transpose_176")]; + tensor input_375_cast_fp16 = add(x = input_359_cast_fp16, y = input_373_cast_fp16)[name = tensor("input_375_cast_fp16")]; + tensor input_377_axes_0 = const()[name = tensor("input_377_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86370496)))]; + tensor encoder_layers_6_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86371584)))]; + tensor input_377_cast_fp16 = layer_norm(axes = input_377_axes_0, beta = encoder_layers_6_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_norm_feed_forward2_weight_to_fp16, x = input_375_cast_fp16)[name = tensor("input_377_cast_fp16")]; + tensor encoder_layers_6_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_6_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86372672)))]; + tensor linear_62_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_6_feed_forward2_linear1_weight_to_fp16, x = input_377_cast_fp16)[name = tensor("linear_62_cast_fp16")]; + tensor input_381_cast_fp16 = silu(x = linear_62_cast_fp16)[name = tensor("input_381_cast_fp16")]; + tensor encoder_layers_6_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_6_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88469888)))]; + tensor linear_63_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_feed_forward2_linear2_weight_to_fp16, x = input_381_cast_fp16)[name = tensor("linear_63_cast_fp16")]; + tensor var_1696_to_fp16 = const()[name = tensor("op_1696_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1697_cast_fp16 = mul(x = linear_63_cast_fp16, y = var_1696_to_fp16)[name = tensor("op_1697_cast_fp16")]; + tensor input_387_cast_fp16 = add(x = input_375_cast_fp16, y = var_1697_cast_fp16)[name = tensor("input_387_cast_fp16")]; + tensor input_389_axes_0 = const()[name = tensor("input_389_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90567104)))]; + tensor encoder_layers_6_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90568192)))]; + tensor input_389_cast_fp16 = layer_norm(axes = input_389_axes_0, beta = encoder_layers_6_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_norm_out_weight_to_fp16, x = input_387_cast_fp16)[name = tensor("input_389_cast_fp16")]; + tensor cache_29_begin_0 = const()[name = tensor("cache_29_begin_0"), val = tensor([7, 0, 0, 0])]; + tensor cache_29_end_0 = const()[name = tensor("cache_29_end_0"), val = tensor([8, 1, 70, 512])]; + tensor cache_29_end_mask_0 = const()[name = tensor("cache_29_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_29_squeeze_mask_0 = const()[name = tensor("cache_29_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_29_cast_fp16 = slice_by_index(begin = cache_29_begin_0, end = cache_29_end_0, end_mask = cache_29_end_mask_0, squeeze_mask = cache_29_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_29_cast_fp16")]; + tensor cache_31_begin_0 = const()[name = tensor("cache_31_begin_0"), val = tensor([7, 0, 0, 0])]; + tensor cache_31_end_0 = const()[name = tensor("cache_31_end_0"), val = tensor([8, 1, 512, 8])]; + tensor cache_31_end_mask_0 = const()[name = tensor("cache_31_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_31_squeeze_mask_0 = const()[name = tensor("cache_31_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_31_cast_fp16 = slice_by_index(begin = cache_31_begin_0, end = cache_31_end_0, end_mask = cache_31_end_mask_0, squeeze_mask = cache_31_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_31_cast_fp16")]; + tensor input_391_axes_0 = const()[name = tensor("input_391_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90569280)))]; + tensor encoder_layers_7_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90570368)))]; + tensor input_391_cast_fp16 = layer_norm(axes = input_391_axes_0, beta = encoder_layers_7_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_norm_feed_forward1_weight_to_fp16, x = input_389_cast_fp16)[name = tensor("input_391_cast_fp16")]; + tensor encoder_layers_7_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_7_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90571456)))]; + tensor linear_64_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_7_feed_forward1_linear1_weight_to_fp16, x = input_391_cast_fp16)[name = tensor("linear_64_cast_fp16")]; + tensor input_395_cast_fp16 = silu(x = linear_64_cast_fp16)[name = tensor("input_395_cast_fp16")]; + tensor encoder_layers_7_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_7_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92668672)))]; + tensor linear_65_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_feed_forward1_linear2_weight_to_fp16, x = input_395_cast_fp16)[name = tensor("linear_65_cast_fp16")]; + tensor var_1731_to_fp16 = const()[name = tensor("op_1731_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1732_cast_fp16 = mul(x = linear_65_cast_fp16, y = var_1731_to_fp16)[name = tensor("op_1732_cast_fp16")]; + tensor input_401_cast_fp16 = add(x = input_389_cast_fp16, y = var_1732_cast_fp16)[name = tensor("input_401_cast_fp16")]; + tensor key_15_axes_0 = const()[name = tensor("key_15_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94765888)))]; + tensor encoder_layers_7_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94766976)))]; + tensor key_15_cast_fp16 = layer_norm(axes = key_15_axes_0, beta = encoder_layers_7_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_norm_self_att_weight_to_fp16, x = input_401_cast_fp16)[name = tensor("key_15_cast_fp16")]; + tensor input_403_interleave_0 = const()[name = tensor("input_403_interleave_0"), val = tensor(false)]; + tensor input_403_cast_fp16 = concat(axis = var_64, interleave = input_403_interleave_0, values = (cache_29_cast_fp16, key_15_cast_fp16))[name = tensor("input_403_cast_fp16")]; + tensor var_1754_begin_0 = const()[name = tensor("op_1754_begin_0"), val = tensor([0, 17, 0])]; + tensor var_1754_end_0 = const()[name = tensor("op_1754_end_0"), val = tensor([1, 70, 512])]; + tensor var_1754_end_mask_0 = const()[name = tensor("op_1754_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1754_cast_fp16 = slice_by_index(begin = var_1754_begin_0, end = var_1754_end_0, end_mask = var_1754_end_mask_0, x = cache_29_cast_fp16)[name = tensor("op_1754_cast_fp16")]; + tensor var_1760_interleave_0 = const()[name = tensor("op_1760_interleave_0"), val = tensor(false)]; + tensor var_1760_cast_fp16 = concat(axis = var_64, interleave = var_1760_interleave_0, values = (var_1754_cast_fp16, key_15_cast_fp16))[name = tensor("op_1760_cast_fp16")]; + tensor encoder_layers_7_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94768064)))]; + tensor linear_66_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_self_attn_linear_q_weight_to_fp16, x = key_15_cast_fp16)[name = tensor("linear_66_cast_fp16")]; + tensor var_1764 = const()[name = tensor("op_1764"), val = tensor([1, -1, 8, 64])]; + tensor q_43_cast_fp16 = reshape(shape = var_1764, x = linear_66_cast_fp16)[name = tensor("q_43_cast_fp16")]; + tensor encoder_layers_7_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95292416)))]; + tensor linear_67_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_self_attn_linear_k_weight_to_fp16, x = input_403_cast_fp16)[name = tensor("linear_67_cast_fp16")]; + tensor var_1768 = const()[name = tensor("op_1768"), val = tensor([1, -1, 8, 64])]; + tensor k_29_cast_fp16 = reshape(shape = var_1768, x = linear_67_cast_fp16)[name = tensor("k_29_cast_fp16")]; + tensor encoder_layers_7_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95816768)))]; + tensor linear_68_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_self_attn_linear_v_weight_to_fp16, x = input_403_cast_fp16)[name = tensor("linear_68_cast_fp16")]; + tensor var_1772 = const()[name = tensor("op_1772"), val = tensor([1, -1, 8, 64])]; + tensor v_15_cast_fp16 = reshape(shape = var_1772, x = linear_68_cast_fp16)[name = tensor("v_15_cast_fp16")]; + tensor value_17_perm_0 = const()[name = tensor("value_17_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_7_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96341120)))]; + tensor var_1784_cast_fp16 = add(x = q_43_cast_fp16, y = encoder_layers_7_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1784_cast_fp16")]; + tensor encoder_layers_7_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96342208)))]; + tensor var_1786_cast_fp16 = add(x = q_43_cast_fp16, y = encoder_layers_7_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1786_cast_fp16")]; + tensor q_with_bias_v_15_perm_0 = const()[name = tensor("q_with_bias_v_15_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_189_transpose_x_0 = const()[name = tensor("x_189_transpose_x_0"), val = tensor(false)]; + tensor x_189_transpose_y_0 = const()[name = tensor("x_189_transpose_y_0"), val = tensor(false)]; + tensor var_1788_to_fp16 = const()[name = tensor("op_1788_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96343296)))]; + tensor q_with_bias_v_15_cast_fp16 = transpose(perm = q_with_bias_v_15_perm_0, x = var_1786_cast_fp16)[name = tensor("transpose_174")]; + tensor x_189_cast_fp16 = matmul(transpose_x = x_189_transpose_x_0, transpose_y = x_189_transpose_y_0, x = q_with_bias_v_15_cast_fp16, y = var_1788_to_fp16)[name = tensor("x_189_cast_fp16")]; + tensor x_191_pad_0 = const()[name = tensor("x_191_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_191_mode_0 = const()[name = tensor("x_191_mode_0"), val = tensor("constant")]; + tensor const_114_to_fp16 = const()[name = tensor("const_114_to_fp16"), val = tensor(0x0p+0)]; + tensor x_191_cast_fp16 = pad(constant_val = const_114_to_fp16, mode = x_191_mode_0, pad = x_191_pad_0, x = x_189_cast_fp16)[name = tensor("x_191_cast_fp16")]; + tensor var_1796 = const()[name = tensor("op_1796"), val = tensor([1, 8, -1, 17])]; + tensor x_193_cast_fp16 = reshape(shape = var_1796, x = x_191_cast_fp16)[name = tensor("x_193_cast_fp16")]; + tensor var_1800_begin_0 = const()[name = tensor("op_1800_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1800_end_0 = const()[name = tensor("op_1800_end_0"), val = tensor([1, 8, 174, 17])]; + tensor var_1800_end_mask_0 = const()[name = tensor("op_1800_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1800_cast_fp16 = slice_by_index(begin = var_1800_begin_0, end = var_1800_end_0, end_mask = var_1800_end_mask_0, x = x_193_cast_fp16)[name = tensor("op_1800_cast_fp16")]; + tensor var_1801 = const()[name = tensor("op_1801"), val = tensor([1, 8, 17, 173])]; + tensor matrix_bd_29_cast_fp16 = reshape(shape = var_1801, x = var_1800_cast_fp16)[name = tensor("matrix_bd_29_cast_fp16")]; + tensor matrix_ac_15_transpose_x_0 = const()[name = tensor("matrix_ac_15_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_15_transpose_y_0 = const()[name = tensor("matrix_ac_15_transpose_y_0"), val = tensor(false)]; + tensor transpose_65_perm_0 = const()[name = tensor("transpose_65_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_66_perm_0 = const()[name = tensor("transpose_66_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_66 = transpose(perm = transpose_66_perm_0, x = k_29_cast_fp16)[name = tensor("transpose_172")]; + tensor transpose_65 = transpose(perm = transpose_65_perm_0, x = var_1784_cast_fp16)[name = tensor("transpose_173")]; + tensor matrix_ac_15_cast_fp16 = matmul(transpose_x = matrix_ac_15_transpose_x_0, transpose_y = matrix_ac_15_transpose_y_0, x = transpose_65, y = transpose_66)[name = tensor("matrix_ac_15_cast_fp16")]; + tensor matrix_bd_31_begin_0 = const()[name = tensor("matrix_bd_31_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_31_end_0 = const()[name = tensor("matrix_bd_31_end_0"), val = tensor([1, 8, 17, 87])]; + tensor matrix_bd_31_end_mask_0 = const()[name = tensor("matrix_bd_31_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_31_cast_fp16 = slice_by_index(begin = matrix_bd_31_begin_0, end = matrix_bd_31_end_0, end_mask = matrix_bd_31_end_mask_0, x = matrix_bd_29_cast_fp16)[name = tensor("matrix_bd_31_cast_fp16")]; + tensor var_1810_cast_fp16 = add(x = matrix_ac_15_cast_fp16, y = matrix_bd_31_cast_fp16)[name = tensor("op_1810_cast_fp16")]; + tensor _inversed_scores_29_y_0_to_fp16 = const()[name = tensor("_inversed_scores_29_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_29_cast_fp16 = mul(x = var_1810_cast_fp16, y = _inversed_scores_29_y_0_to_fp16)[name = tensor("_inversed_scores_29_cast_fp16")]; + tensor scores_31_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_29_cast_fp16, cond = mask_3)[name = tensor("scores_31_cast_fp16")]; + tensor var_1816_cast_fp16 = softmax(axis = var_62, x = scores_31_cast_fp16)[name = tensor("op_1816_cast_fp16")]; + tensor input_405_cast_fp16 = select(a = var_40_to_fp16, b = var_1816_cast_fp16, cond = mask_3)[name = tensor("input_405_cast_fp16")]; + tensor x_195_transpose_x_0 = const()[name = tensor("x_195_transpose_x_0"), val = tensor(false)]; + tensor x_195_transpose_y_0 = const()[name = tensor("x_195_transpose_y_0"), val = tensor(false)]; + tensor value_17_cast_fp16 = transpose(perm = value_17_perm_0, x = v_15_cast_fp16)[name = tensor("transpose_175")]; + tensor x_195_cast_fp16 = matmul(transpose_x = x_195_transpose_x_0, transpose_y = x_195_transpose_y_0, x = input_405_cast_fp16, y = value_17_cast_fp16)[name = tensor("x_195_cast_fp16")]; + tensor var_1820_perm_0 = const()[name = tensor("op_1820_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1821 = const()[name = tensor("op_1821"), val = tensor([1, -1, 512])]; + tensor var_1820_cast_fp16 = transpose(perm = var_1820_perm_0, x = x_195_cast_fp16)[name = tensor("transpose_171")]; + tensor input_407_cast_fp16 = reshape(shape = var_1821, x = var_1820_cast_fp16)[name = tensor("input_407_cast_fp16")]; + tensor encoder_layers_7_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96520512)))]; + tensor linear_70_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_self_attn_linear_out_weight_to_fp16, x = input_407_cast_fp16)[name = tensor("linear_70_cast_fp16")]; + tensor input_411_cast_fp16 = add(x = input_401_cast_fp16, y = linear_70_cast_fp16)[name = tensor("input_411_cast_fp16")]; + tensor x_199_axes_0 = const()[name = tensor("x_199_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97044864)))]; + tensor encoder_layers_7_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97045952)))]; + tensor x_199_cast_fp16 = layer_norm(axes = x_199_axes_0, beta = encoder_layers_7_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_norm_conv_weight_to_fp16, x = input_411_cast_fp16)[name = tensor("x_199_cast_fp16")]; + tensor input_413_perm_0 = const()[name = tensor("input_413_perm_0"), val = tensor([0, 2, 1])]; + tensor input_415_pad_type_0 = const()[name = tensor("input_415_pad_type_0"), val = tensor("valid")]; + tensor input_415_strides_0 = const()[name = tensor("input_415_strides_0"), val = tensor([1])]; + tensor input_415_pad_0 = const()[name = tensor("input_415_pad_0"), val = tensor([0, 0])]; + tensor input_415_dilations_0 = const()[name = tensor("input_415_dilations_0"), val = tensor([1])]; + tensor input_415_groups_0 = const()[name = tensor("input_415_groups_0"), val = tensor(1)]; + tensor encoder_layers_7_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_7_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97047040)))]; + tensor input_413_cast_fp16 = transpose(perm = input_413_perm_0, x = x_199_cast_fp16)[name = tensor("transpose_170")]; + tensor input_415_cast_fp16 = conv(dilations = input_415_dilations_0, groups = input_415_groups_0, pad = input_415_pad_0, pad_type = input_415_pad_type_0, strides = input_415_strides_0, weight = encoder_layers_7_conv_pointwise_conv1_weight_to_fp16, x = input_413_cast_fp16)[name = tensor("input_415_cast_fp16")]; + tensor x_201_split_num_splits_0 = const()[name = tensor("x_201_split_num_splits_0"), val = tensor(2)]; + tensor x_201_split_axis_0 = const()[name = tensor("x_201_split_axis_0"), val = tensor(1)]; + tensor x_201_split_cast_fp16_0, tensor x_201_split_cast_fp16_1 = split(axis = x_201_split_axis_0, num_splits = x_201_split_num_splits_0, x = input_415_cast_fp16)[name = tensor("x_201_split_cast_fp16")]; + tensor x_201_split_1_sigmoid_cast_fp16 = sigmoid(x = x_201_split_cast_fp16_1)[name = tensor("x_201_split_1_sigmoid_cast_fp16")]; + tensor x_201_cast_fp16 = mul(x = x_201_split_cast_fp16_0, y = x_201_split_1_sigmoid_cast_fp16)[name = tensor("x_201_cast_fp16")]; + tensor input_417_cast_fp16 = select(a = var_40_to_fp16, b = x_201_cast_fp16, cond = var_418)[name = tensor("input_417_cast_fp16")]; + tensor new_x_31_interleave_0 = const()[name = tensor("new_x_31_interleave_0"), val = tensor(false)]; + tensor new_x_31_cast_fp16 = concat(axis = var_62, interleave = new_x_31_interleave_0, values = (cache_31_cast_fp16, input_417_cast_fp16))[name = tensor("new_x_31_cast_fp16")]; + tensor var_1859_begin_0 = const()[name = tensor("op_1859_begin_0"), val = tensor([0, 0, 17])]; + tensor var_1859_end_0 = const()[name = tensor("op_1859_end_0"), val = tensor([1, 512, 25])]; + tensor var_1859_end_mask_0 = const()[name = tensor("op_1859_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1859_cast_fp16 = slice_by_index(begin = var_1859_begin_0, end = var_1859_end_0, end_mask = var_1859_end_mask_0, x = new_x_31_cast_fp16)[name = tensor("op_1859_cast_fp16")]; + tensor x_203_pad_type_0 = const()[name = tensor("x_203_pad_type_0"), val = tensor("valid")]; + tensor x_203_groups_0 = const()[name = tensor("x_203_groups_0"), val = tensor(512)]; + tensor x_203_strides_0 = const()[name = tensor("x_203_strides_0"), val = tensor([1])]; + tensor x_203_pad_0 = const()[name = tensor("x_203_pad_0"), val = tensor([0, 0])]; + tensor x_203_dilations_0 = const()[name = tensor("x_203_dilations_0"), val = tensor([1])]; + tensor encoder_layers_7_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_7_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98095680)))]; + tensor x_203_cast_fp16 = conv(dilations = x_203_dilations_0, groups = x_203_groups_0, pad = x_203_pad_0, pad_type = x_203_pad_type_0, strides = x_203_strides_0, weight = encoder_layers_7_conv_depthwise_conv_weight_to_fp16, x = new_x_31_cast_fp16)[name = tensor("x_203_cast_fp16")]; + tensor input_419_perm_0 = const()[name = tensor("input_419_perm_0"), val = tensor([0, 2, 1])]; + tensor x_205_axes_0 = const()[name = tensor("x_205_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_7_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98104960)))]; + tensor encoder_layers_7_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_7_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98106048)))]; + tensor input_419_cast_fp16 = transpose(perm = input_419_perm_0, x = x_203_cast_fp16)[name = tensor("transpose_169")]; + tensor x_205_cast_fp16 = layer_norm(axes = x_205_axes_0, beta = encoder_layers_7_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_conv_batch_norm_weight_to_fp16, x = input_419_cast_fp16)[name = tensor("x_205_cast_fp16")]; + tensor input_421_perm_0 = const()[name = tensor("input_421_perm_0"), val = tensor([0, 2, 1])]; + tensor input_421_cast_fp16 = transpose(perm = input_421_perm_0, x = x_205_cast_fp16)[name = tensor("transpose_168")]; + tensor input_423_cast_fp16 = silu(x = input_421_cast_fp16)[name = tensor("input_423_cast_fp16")]; + tensor x_207_pad_type_0 = const()[name = tensor("x_207_pad_type_0"), val = tensor("valid")]; + tensor x_207_strides_0 = const()[name = tensor("x_207_strides_0"), val = tensor([1])]; + tensor x_207_pad_0 = const()[name = tensor("x_207_pad_0"), val = tensor([0, 0])]; + tensor x_207_dilations_0 = const()[name = tensor("x_207_dilations_0"), val = tensor([1])]; + tensor x_207_groups_0 = const()[name = tensor("x_207_groups_0"), val = tensor(1)]; + tensor encoder_layers_7_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_7_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98107136)))]; + tensor x_207_cast_fp16 = conv(dilations = x_207_dilations_0, groups = x_207_groups_0, pad = x_207_pad_0, pad_type = x_207_pad_type_0, strides = x_207_strides_0, weight = encoder_layers_7_conv_pointwise_conv2_weight_to_fp16, x = input_423_cast_fp16)[name = tensor("x_207_cast_fp16")]; + tensor input_425_perm_0 = const()[name = tensor("input_425_perm_0"), val = tensor([0, 2, 1])]; + tensor input_425_cast_fp16 = transpose(perm = input_425_perm_0, x = x_207_cast_fp16)[name = tensor("transpose_167")]; + tensor input_427_cast_fp16 = add(x = input_411_cast_fp16, y = input_425_cast_fp16)[name = tensor("input_427_cast_fp16")]; + tensor input_429_axes_0 = const()[name = tensor("input_429_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98631488)))]; + tensor encoder_layers_7_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98632576)))]; + tensor input_429_cast_fp16 = layer_norm(axes = input_429_axes_0, beta = encoder_layers_7_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_norm_feed_forward2_weight_to_fp16, x = input_427_cast_fp16)[name = tensor("input_429_cast_fp16")]; + tensor encoder_layers_7_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_7_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98633664)))]; + tensor linear_71_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_7_feed_forward2_linear1_weight_to_fp16, x = input_429_cast_fp16)[name = tensor("linear_71_cast_fp16")]; + tensor input_433_cast_fp16 = silu(x = linear_71_cast_fp16)[name = tensor("input_433_cast_fp16")]; + tensor encoder_layers_7_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_7_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100730880)))]; + tensor linear_72_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_feed_forward2_linear2_weight_to_fp16, x = input_433_cast_fp16)[name = tensor("linear_72_cast_fp16")]; + tensor var_1900_to_fp16 = const()[name = tensor("op_1900_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1901_cast_fp16 = mul(x = linear_72_cast_fp16, y = var_1900_to_fp16)[name = tensor("op_1901_cast_fp16")]; + tensor input_439_cast_fp16 = add(x = input_427_cast_fp16, y = var_1901_cast_fp16)[name = tensor("input_439_cast_fp16")]; + tensor input_441_axes_0 = const()[name = tensor("input_441_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102828096)))]; + tensor encoder_layers_7_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102829184)))]; + tensor input_441_cast_fp16 = layer_norm(axes = input_441_axes_0, beta = encoder_layers_7_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_norm_out_weight_to_fp16, x = input_439_cast_fp16)[name = tensor("input_441_cast_fp16")]; + tensor cache_33_begin_0 = const()[name = tensor("cache_33_begin_0"), val = tensor([8, 0, 0, 0])]; + tensor cache_33_end_0 = const()[name = tensor("cache_33_end_0"), val = tensor([9, 1, 70, 512])]; + tensor cache_33_end_mask_0 = const()[name = tensor("cache_33_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_33_squeeze_mask_0 = const()[name = tensor("cache_33_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_33_cast_fp16 = slice_by_index(begin = cache_33_begin_0, end = cache_33_end_0, end_mask = cache_33_end_mask_0, squeeze_mask = cache_33_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_33_cast_fp16")]; + tensor cache_35_begin_0 = const()[name = tensor("cache_35_begin_0"), val = tensor([8, 0, 0, 0])]; + tensor cache_35_end_0 = const()[name = tensor("cache_35_end_0"), val = tensor([9, 1, 512, 8])]; + tensor cache_35_end_mask_0 = const()[name = tensor("cache_35_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_35_squeeze_mask_0 = const()[name = tensor("cache_35_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_35_cast_fp16 = slice_by_index(begin = cache_35_begin_0, end = cache_35_end_0, end_mask = cache_35_end_mask_0, squeeze_mask = cache_35_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_35_cast_fp16")]; + tensor input_443_axes_0 = const()[name = tensor("input_443_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102830272)))]; + tensor encoder_layers_8_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102831360)))]; + tensor input_443_cast_fp16 = layer_norm(axes = input_443_axes_0, beta = encoder_layers_8_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_norm_feed_forward1_weight_to_fp16, x = input_441_cast_fp16)[name = tensor("input_443_cast_fp16")]; + tensor encoder_layers_8_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_8_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102832448)))]; + tensor linear_73_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_8_feed_forward1_linear1_weight_to_fp16, x = input_443_cast_fp16)[name = tensor("linear_73_cast_fp16")]; + tensor input_447_cast_fp16 = silu(x = linear_73_cast_fp16)[name = tensor("input_447_cast_fp16")]; + tensor encoder_layers_8_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_8_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(104929664)))]; + tensor linear_74_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_feed_forward1_linear2_weight_to_fp16, x = input_447_cast_fp16)[name = tensor("linear_74_cast_fp16")]; + tensor var_1935_to_fp16 = const()[name = tensor("op_1935_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1936_cast_fp16 = mul(x = linear_74_cast_fp16, y = var_1935_to_fp16)[name = tensor("op_1936_cast_fp16")]; + tensor input_453_cast_fp16 = add(x = input_441_cast_fp16, y = var_1936_cast_fp16)[name = tensor("input_453_cast_fp16")]; + tensor key_17_axes_0 = const()[name = tensor("key_17_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107026880)))]; + tensor encoder_layers_8_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107027968)))]; + tensor key_17_cast_fp16 = layer_norm(axes = key_17_axes_0, beta = encoder_layers_8_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_norm_self_att_weight_to_fp16, x = input_453_cast_fp16)[name = tensor("key_17_cast_fp16")]; + tensor input_455_interleave_0 = const()[name = tensor("input_455_interleave_0"), val = tensor(false)]; + tensor input_455_cast_fp16 = concat(axis = var_64, interleave = input_455_interleave_0, values = (cache_33_cast_fp16, key_17_cast_fp16))[name = tensor("input_455_cast_fp16")]; + tensor var_1958_begin_0 = const()[name = tensor("op_1958_begin_0"), val = tensor([0, 17, 0])]; + tensor var_1958_end_0 = const()[name = tensor("op_1958_end_0"), val = tensor([1, 70, 512])]; + tensor var_1958_end_mask_0 = const()[name = tensor("op_1958_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1958_cast_fp16 = slice_by_index(begin = var_1958_begin_0, end = var_1958_end_0, end_mask = var_1958_end_mask_0, x = cache_33_cast_fp16)[name = tensor("op_1958_cast_fp16")]; + tensor var_1964_interleave_0 = const()[name = tensor("op_1964_interleave_0"), val = tensor(false)]; + tensor var_1964_cast_fp16 = concat(axis = var_64, interleave = var_1964_interleave_0, values = (var_1958_cast_fp16, key_17_cast_fp16))[name = tensor("op_1964_cast_fp16")]; + tensor encoder_layers_8_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107029056)))]; + tensor linear_75_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_self_attn_linear_q_weight_to_fp16, x = key_17_cast_fp16)[name = tensor("linear_75_cast_fp16")]; + tensor var_1968 = const()[name = tensor("op_1968"), val = tensor([1, -1, 8, 64])]; + tensor q_49_cast_fp16 = reshape(shape = var_1968, x = linear_75_cast_fp16)[name = tensor("q_49_cast_fp16")]; + tensor encoder_layers_8_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107553408)))]; + tensor linear_76_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_self_attn_linear_k_weight_to_fp16, x = input_455_cast_fp16)[name = tensor("linear_76_cast_fp16")]; + tensor var_1972 = const()[name = tensor("op_1972"), val = tensor([1, -1, 8, 64])]; + tensor k_33_cast_fp16 = reshape(shape = var_1972, x = linear_76_cast_fp16)[name = tensor("k_33_cast_fp16")]; + tensor encoder_layers_8_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108077760)))]; + tensor linear_77_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_self_attn_linear_v_weight_to_fp16, x = input_455_cast_fp16)[name = tensor("linear_77_cast_fp16")]; + tensor var_1976 = const()[name = tensor("op_1976"), val = tensor([1, -1, 8, 64])]; + tensor v_17_cast_fp16 = reshape(shape = var_1976, x = linear_77_cast_fp16)[name = tensor("v_17_cast_fp16")]; + tensor value_19_perm_0 = const()[name = tensor("value_19_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_8_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108602112)))]; + tensor var_1988_cast_fp16 = add(x = q_49_cast_fp16, y = encoder_layers_8_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1988_cast_fp16")]; + tensor encoder_layers_8_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108603200)))]; + tensor var_1990_cast_fp16 = add(x = q_49_cast_fp16, y = encoder_layers_8_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1990_cast_fp16")]; + tensor q_with_bias_v_17_perm_0 = const()[name = tensor("q_with_bias_v_17_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_215_transpose_x_0 = const()[name = tensor("x_215_transpose_x_0"), val = tensor(false)]; + tensor x_215_transpose_y_0 = const()[name = tensor("x_215_transpose_y_0"), val = tensor(false)]; + tensor var_1992_to_fp16 = const()[name = tensor("op_1992_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108604288)))]; + tensor q_with_bias_v_17_cast_fp16 = transpose(perm = q_with_bias_v_17_perm_0, x = var_1990_cast_fp16)[name = tensor("transpose_165")]; + tensor x_215_cast_fp16 = matmul(transpose_x = x_215_transpose_x_0, transpose_y = x_215_transpose_y_0, x = q_with_bias_v_17_cast_fp16, y = var_1992_to_fp16)[name = tensor("x_215_cast_fp16")]; + tensor x_217_pad_0 = const()[name = tensor("x_217_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_217_mode_0 = const()[name = tensor("x_217_mode_0"), val = tensor("constant")]; + tensor const_127_to_fp16 = const()[name = tensor("const_127_to_fp16"), val = tensor(0x0p+0)]; + tensor x_217_cast_fp16 = pad(constant_val = const_127_to_fp16, mode = x_217_mode_0, pad = x_217_pad_0, x = x_215_cast_fp16)[name = tensor("x_217_cast_fp16")]; + tensor var_2000 = const()[name = tensor("op_2000"), val = tensor([1, 8, -1, 17])]; + tensor x_219_cast_fp16 = reshape(shape = var_2000, x = x_217_cast_fp16)[name = tensor("x_219_cast_fp16")]; + tensor var_2004_begin_0 = const()[name = tensor("op_2004_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2004_end_0 = const()[name = tensor("op_2004_end_0"), val = tensor([1, 8, 174, 17])]; + tensor var_2004_end_mask_0 = const()[name = tensor("op_2004_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2004_cast_fp16 = slice_by_index(begin = var_2004_begin_0, end = var_2004_end_0, end_mask = var_2004_end_mask_0, x = x_219_cast_fp16)[name = tensor("op_2004_cast_fp16")]; + tensor var_2005 = const()[name = tensor("op_2005"), val = tensor([1, 8, 17, 173])]; + tensor matrix_bd_33_cast_fp16 = reshape(shape = var_2005, x = var_2004_cast_fp16)[name = tensor("matrix_bd_33_cast_fp16")]; + tensor matrix_ac_17_transpose_x_0 = const()[name = tensor("matrix_ac_17_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_17_transpose_y_0 = const()[name = tensor("matrix_ac_17_transpose_y_0"), val = tensor(false)]; + tensor transpose_67_perm_0 = const()[name = tensor("transpose_67_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_68_perm_0 = const()[name = tensor("transpose_68_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_68 = transpose(perm = transpose_68_perm_0, x = k_33_cast_fp16)[name = tensor("transpose_163")]; + tensor transpose_67 = transpose(perm = transpose_67_perm_0, x = var_1988_cast_fp16)[name = tensor("transpose_164")]; + tensor matrix_ac_17_cast_fp16 = matmul(transpose_x = matrix_ac_17_transpose_x_0, transpose_y = matrix_ac_17_transpose_y_0, x = transpose_67, y = transpose_68)[name = tensor("matrix_ac_17_cast_fp16")]; + tensor matrix_bd_35_begin_0 = const()[name = tensor("matrix_bd_35_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_35_end_0 = const()[name = tensor("matrix_bd_35_end_0"), val = tensor([1, 8, 17, 87])]; + tensor matrix_bd_35_end_mask_0 = const()[name = tensor("matrix_bd_35_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_35_cast_fp16 = slice_by_index(begin = matrix_bd_35_begin_0, end = matrix_bd_35_end_0, end_mask = matrix_bd_35_end_mask_0, x = matrix_bd_33_cast_fp16)[name = tensor("matrix_bd_35_cast_fp16")]; + tensor var_2014_cast_fp16 = add(x = matrix_ac_17_cast_fp16, y = matrix_bd_35_cast_fp16)[name = tensor("op_2014_cast_fp16")]; + tensor _inversed_scores_33_y_0_to_fp16 = const()[name = tensor("_inversed_scores_33_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_33_cast_fp16 = mul(x = var_2014_cast_fp16, y = _inversed_scores_33_y_0_to_fp16)[name = tensor("_inversed_scores_33_cast_fp16")]; + tensor scores_35_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_33_cast_fp16, cond = mask_3)[name = tensor("scores_35_cast_fp16")]; + tensor var_2020_cast_fp16 = softmax(axis = var_62, x = scores_35_cast_fp16)[name = tensor("op_2020_cast_fp16")]; + tensor input_457_cast_fp16 = select(a = var_40_to_fp16, b = var_2020_cast_fp16, cond = mask_3)[name = tensor("input_457_cast_fp16")]; + tensor x_221_transpose_x_0 = const()[name = tensor("x_221_transpose_x_0"), val = tensor(false)]; + tensor x_221_transpose_y_0 = const()[name = tensor("x_221_transpose_y_0"), val = tensor(false)]; + tensor value_19_cast_fp16 = transpose(perm = value_19_perm_0, x = v_17_cast_fp16)[name = tensor("transpose_166")]; + tensor x_221_cast_fp16 = matmul(transpose_x = x_221_transpose_x_0, transpose_y = x_221_transpose_y_0, x = input_457_cast_fp16, y = value_19_cast_fp16)[name = tensor("x_221_cast_fp16")]; + tensor var_2024_perm_0 = const()[name = tensor("op_2024_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2025 = const()[name = tensor("op_2025"), val = tensor([1, -1, 512])]; + tensor var_2024_cast_fp16 = transpose(perm = var_2024_perm_0, x = x_221_cast_fp16)[name = tensor("transpose_162")]; + tensor input_459_cast_fp16 = reshape(shape = var_2025, x = var_2024_cast_fp16)[name = tensor("input_459_cast_fp16")]; + tensor encoder_layers_8_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108781504)))]; + tensor linear_79_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_self_attn_linear_out_weight_to_fp16, x = input_459_cast_fp16)[name = tensor("linear_79_cast_fp16")]; + tensor input_463_cast_fp16 = add(x = input_453_cast_fp16, y = linear_79_cast_fp16)[name = tensor("input_463_cast_fp16")]; + tensor x_225_axes_0 = const()[name = tensor("x_225_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109305856)))]; + tensor encoder_layers_8_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109306944)))]; + tensor x_225_cast_fp16 = layer_norm(axes = x_225_axes_0, beta = encoder_layers_8_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_norm_conv_weight_to_fp16, x = input_463_cast_fp16)[name = tensor("x_225_cast_fp16")]; + tensor input_465_perm_0 = const()[name = tensor("input_465_perm_0"), val = tensor([0, 2, 1])]; + tensor input_467_pad_type_0 = const()[name = tensor("input_467_pad_type_0"), val = tensor("valid")]; + tensor input_467_strides_0 = const()[name = tensor("input_467_strides_0"), val = tensor([1])]; + tensor input_467_pad_0 = const()[name = tensor("input_467_pad_0"), val = tensor([0, 0])]; + tensor input_467_dilations_0 = const()[name = tensor("input_467_dilations_0"), val = tensor([1])]; + tensor input_467_groups_0 = const()[name = tensor("input_467_groups_0"), val = tensor(1)]; + tensor encoder_layers_8_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_8_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109308032)))]; + tensor input_465_cast_fp16 = transpose(perm = input_465_perm_0, x = x_225_cast_fp16)[name = tensor("transpose_161")]; + tensor input_467_cast_fp16 = conv(dilations = input_467_dilations_0, groups = input_467_groups_0, pad = input_467_pad_0, pad_type = input_467_pad_type_0, strides = input_467_strides_0, weight = encoder_layers_8_conv_pointwise_conv1_weight_to_fp16, x = input_465_cast_fp16)[name = tensor("input_467_cast_fp16")]; + tensor x_227_split_num_splits_0 = const()[name = tensor("x_227_split_num_splits_0"), val = tensor(2)]; + tensor x_227_split_axis_0 = const()[name = tensor("x_227_split_axis_0"), val = tensor(1)]; + tensor x_227_split_cast_fp16_0, tensor x_227_split_cast_fp16_1 = split(axis = x_227_split_axis_0, num_splits = x_227_split_num_splits_0, x = input_467_cast_fp16)[name = tensor("x_227_split_cast_fp16")]; + tensor x_227_split_1_sigmoid_cast_fp16 = sigmoid(x = x_227_split_cast_fp16_1)[name = tensor("x_227_split_1_sigmoid_cast_fp16")]; + tensor x_227_cast_fp16 = mul(x = x_227_split_cast_fp16_0, y = x_227_split_1_sigmoid_cast_fp16)[name = tensor("x_227_cast_fp16")]; + tensor input_469_cast_fp16 = select(a = var_40_to_fp16, b = x_227_cast_fp16, cond = var_418)[name = tensor("input_469_cast_fp16")]; + tensor new_x_35_interleave_0 = const()[name = tensor("new_x_35_interleave_0"), val = tensor(false)]; + tensor new_x_35_cast_fp16 = concat(axis = var_62, interleave = new_x_35_interleave_0, values = (cache_35_cast_fp16, input_469_cast_fp16))[name = tensor("new_x_35_cast_fp16")]; + tensor var_2063_begin_0 = const()[name = tensor("op_2063_begin_0"), val = tensor([0, 0, 17])]; + tensor var_2063_end_0 = const()[name = tensor("op_2063_end_0"), val = tensor([1, 512, 25])]; + tensor var_2063_end_mask_0 = const()[name = tensor("op_2063_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2063_cast_fp16 = slice_by_index(begin = var_2063_begin_0, end = var_2063_end_0, end_mask = var_2063_end_mask_0, x = new_x_35_cast_fp16)[name = tensor("op_2063_cast_fp16")]; + tensor x_229_pad_type_0 = const()[name = tensor("x_229_pad_type_0"), val = tensor("valid")]; + tensor x_229_groups_0 = const()[name = tensor("x_229_groups_0"), val = tensor(512)]; + tensor x_229_strides_0 = const()[name = tensor("x_229_strides_0"), val = tensor([1])]; + tensor x_229_pad_0 = const()[name = tensor("x_229_pad_0"), val = tensor([0, 0])]; + tensor x_229_dilations_0 = const()[name = tensor("x_229_dilations_0"), val = tensor([1])]; + tensor encoder_layers_8_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_8_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110356672)))]; + tensor x_229_cast_fp16 = conv(dilations = x_229_dilations_0, groups = x_229_groups_0, pad = x_229_pad_0, pad_type = x_229_pad_type_0, strides = x_229_strides_0, weight = encoder_layers_8_conv_depthwise_conv_weight_to_fp16, x = new_x_35_cast_fp16)[name = tensor("x_229_cast_fp16")]; + tensor input_471_perm_0 = const()[name = tensor("input_471_perm_0"), val = tensor([0, 2, 1])]; + tensor x_231_axes_0 = const()[name = tensor("x_231_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_8_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110365952)))]; + tensor encoder_layers_8_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_8_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110367040)))]; + tensor input_471_cast_fp16 = transpose(perm = input_471_perm_0, x = x_229_cast_fp16)[name = tensor("transpose_160")]; + tensor x_231_cast_fp16 = layer_norm(axes = x_231_axes_0, beta = encoder_layers_8_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_conv_batch_norm_weight_to_fp16, x = input_471_cast_fp16)[name = tensor("x_231_cast_fp16")]; + tensor input_473_perm_0 = const()[name = tensor("input_473_perm_0"), val = tensor([0, 2, 1])]; + tensor input_473_cast_fp16 = transpose(perm = input_473_perm_0, x = x_231_cast_fp16)[name = tensor("transpose_159")]; + tensor input_475_cast_fp16 = silu(x = input_473_cast_fp16)[name = tensor("input_475_cast_fp16")]; + tensor x_233_pad_type_0 = const()[name = tensor("x_233_pad_type_0"), val = tensor("valid")]; + tensor x_233_strides_0 = const()[name = tensor("x_233_strides_0"), val = tensor([1])]; + tensor x_233_pad_0 = const()[name = tensor("x_233_pad_0"), val = tensor([0, 0])]; + tensor x_233_dilations_0 = const()[name = tensor("x_233_dilations_0"), val = tensor([1])]; + tensor x_233_groups_0 = const()[name = tensor("x_233_groups_0"), val = tensor(1)]; + tensor encoder_layers_8_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_8_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110368128)))]; + tensor x_233_cast_fp16 = conv(dilations = x_233_dilations_0, groups = x_233_groups_0, pad = x_233_pad_0, pad_type = x_233_pad_type_0, strides = x_233_strides_0, weight = encoder_layers_8_conv_pointwise_conv2_weight_to_fp16, x = input_475_cast_fp16)[name = tensor("x_233_cast_fp16")]; + tensor input_477_perm_0 = const()[name = tensor("input_477_perm_0"), val = tensor([0, 2, 1])]; + tensor input_477_cast_fp16 = transpose(perm = input_477_perm_0, x = x_233_cast_fp16)[name = tensor("transpose_158")]; + tensor input_479_cast_fp16 = add(x = input_463_cast_fp16, y = input_477_cast_fp16)[name = tensor("input_479_cast_fp16")]; + tensor input_481_axes_0 = const()[name = tensor("input_481_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110892480)))]; + tensor encoder_layers_8_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110893568)))]; + tensor input_481_cast_fp16 = layer_norm(axes = input_481_axes_0, beta = encoder_layers_8_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_norm_feed_forward2_weight_to_fp16, x = input_479_cast_fp16)[name = tensor("input_481_cast_fp16")]; + tensor encoder_layers_8_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_8_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110894656)))]; + tensor linear_80_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_8_feed_forward2_linear1_weight_to_fp16, x = input_481_cast_fp16)[name = tensor("linear_80_cast_fp16")]; + tensor input_485_cast_fp16 = silu(x = linear_80_cast_fp16)[name = tensor("input_485_cast_fp16")]; + tensor encoder_layers_8_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_8_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(112991872)))]; + tensor linear_81_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_feed_forward2_linear2_weight_to_fp16, x = input_485_cast_fp16)[name = tensor("linear_81_cast_fp16")]; + tensor var_2104_to_fp16 = const()[name = tensor("op_2104_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2105_cast_fp16 = mul(x = linear_81_cast_fp16, y = var_2104_to_fp16)[name = tensor("op_2105_cast_fp16")]; + tensor input_491_cast_fp16 = add(x = input_479_cast_fp16, y = var_2105_cast_fp16)[name = tensor("input_491_cast_fp16")]; + tensor input_493_axes_0 = const()[name = tensor("input_493_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115089088)))]; + tensor encoder_layers_8_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115090176)))]; + tensor input_493_cast_fp16 = layer_norm(axes = input_493_axes_0, beta = encoder_layers_8_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_norm_out_weight_to_fp16, x = input_491_cast_fp16)[name = tensor("input_493_cast_fp16")]; + tensor cache_37_begin_0 = const()[name = tensor("cache_37_begin_0"), val = tensor([9, 0, 0, 0])]; + tensor cache_37_end_0 = const()[name = tensor("cache_37_end_0"), val = tensor([10, 1, 70, 512])]; + tensor cache_37_end_mask_0 = const()[name = tensor("cache_37_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_37_squeeze_mask_0 = const()[name = tensor("cache_37_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_37_cast_fp16 = slice_by_index(begin = cache_37_begin_0, end = cache_37_end_0, end_mask = cache_37_end_mask_0, squeeze_mask = cache_37_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_37_cast_fp16")]; + tensor cache_39_begin_0 = const()[name = tensor("cache_39_begin_0"), val = tensor([9, 0, 0, 0])]; + tensor cache_39_end_0 = const()[name = tensor("cache_39_end_0"), val = tensor([10, 1, 512, 8])]; + tensor cache_39_end_mask_0 = const()[name = tensor("cache_39_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_39_squeeze_mask_0 = const()[name = tensor("cache_39_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_39_cast_fp16 = slice_by_index(begin = cache_39_begin_0, end = cache_39_end_0, end_mask = cache_39_end_mask_0, squeeze_mask = cache_39_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_39_cast_fp16")]; + tensor input_495_axes_0 = const()[name = tensor("input_495_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115091264)))]; + tensor encoder_layers_9_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115092352)))]; + tensor input_495_cast_fp16 = layer_norm(axes = input_495_axes_0, beta = encoder_layers_9_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_norm_feed_forward1_weight_to_fp16, x = input_493_cast_fp16)[name = tensor("input_495_cast_fp16")]; + tensor encoder_layers_9_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_9_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115093440)))]; + tensor linear_82_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_9_feed_forward1_linear1_weight_to_fp16, x = input_495_cast_fp16)[name = tensor("linear_82_cast_fp16")]; + tensor input_499_cast_fp16 = silu(x = linear_82_cast_fp16)[name = tensor("input_499_cast_fp16")]; + tensor encoder_layers_9_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_9_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117190656)))]; + tensor linear_83_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_feed_forward1_linear2_weight_to_fp16, x = input_499_cast_fp16)[name = tensor("linear_83_cast_fp16")]; + tensor var_2139_to_fp16 = const()[name = tensor("op_2139_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2140_cast_fp16 = mul(x = linear_83_cast_fp16, y = var_2139_to_fp16)[name = tensor("op_2140_cast_fp16")]; + tensor input_505_cast_fp16 = add(x = input_493_cast_fp16, y = var_2140_cast_fp16)[name = tensor("input_505_cast_fp16")]; + tensor key_19_axes_0 = const()[name = tensor("key_19_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119287872)))]; + tensor encoder_layers_9_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119288960)))]; + tensor key_19_cast_fp16 = layer_norm(axes = key_19_axes_0, beta = encoder_layers_9_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_norm_self_att_weight_to_fp16, x = input_505_cast_fp16)[name = tensor("key_19_cast_fp16")]; + tensor input_507_interleave_0 = const()[name = tensor("input_507_interleave_0"), val = tensor(false)]; + tensor input_507_cast_fp16 = concat(axis = var_64, interleave = input_507_interleave_0, values = (cache_37_cast_fp16, key_19_cast_fp16))[name = tensor("input_507_cast_fp16")]; + tensor var_2162_begin_0 = const()[name = tensor("op_2162_begin_0"), val = tensor([0, 17, 0])]; + tensor var_2162_end_0 = const()[name = tensor("op_2162_end_0"), val = tensor([1, 70, 512])]; + tensor var_2162_end_mask_0 = const()[name = tensor("op_2162_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2162_cast_fp16 = slice_by_index(begin = var_2162_begin_0, end = var_2162_end_0, end_mask = var_2162_end_mask_0, x = cache_37_cast_fp16)[name = tensor("op_2162_cast_fp16")]; + tensor var_2168_interleave_0 = const()[name = tensor("op_2168_interleave_0"), val = tensor(false)]; + tensor var_2168_cast_fp16 = concat(axis = var_64, interleave = var_2168_interleave_0, values = (var_2162_cast_fp16, key_19_cast_fp16))[name = tensor("op_2168_cast_fp16")]; + tensor encoder_layers_9_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119290048)))]; + tensor linear_84_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_self_attn_linear_q_weight_to_fp16, x = key_19_cast_fp16)[name = tensor("linear_84_cast_fp16")]; + tensor var_2172 = const()[name = tensor("op_2172"), val = tensor([1, -1, 8, 64])]; + tensor q_55_cast_fp16 = reshape(shape = var_2172, x = linear_84_cast_fp16)[name = tensor("q_55_cast_fp16")]; + tensor encoder_layers_9_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119814400)))]; + tensor linear_85_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_self_attn_linear_k_weight_to_fp16, x = input_507_cast_fp16)[name = tensor("linear_85_cast_fp16")]; + tensor var_2176 = const()[name = tensor("op_2176"), val = tensor([1, -1, 8, 64])]; + tensor k_37_cast_fp16 = reshape(shape = var_2176, x = linear_85_cast_fp16)[name = tensor("k_37_cast_fp16")]; + tensor encoder_layers_9_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120338752)))]; + tensor linear_86_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_self_attn_linear_v_weight_to_fp16, x = input_507_cast_fp16)[name = tensor("linear_86_cast_fp16")]; + tensor var_2180 = const()[name = tensor("op_2180"), val = tensor([1, -1, 8, 64])]; + tensor v_19_cast_fp16 = reshape(shape = var_2180, x = linear_86_cast_fp16)[name = tensor("v_19_cast_fp16")]; + tensor value_21_perm_0 = const()[name = tensor("value_21_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_9_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120863104)))]; + tensor var_2192_cast_fp16 = add(x = q_55_cast_fp16, y = encoder_layers_9_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2192_cast_fp16")]; + tensor encoder_layers_9_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120864192)))]; + tensor var_2194_cast_fp16 = add(x = q_55_cast_fp16, y = encoder_layers_9_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2194_cast_fp16")]; + tensor q_with_bias_v_19_perm_0 = const()[name = tensor("q_with_bias_v_19_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_241_transpose_x_0 = const()[name = tensor("x_241_transpose_x_0"), val = tensor(false)]; + tensor x_241_transpose_y_0 = const()[name = tensor("x_241_transpose_y_0"), val = tensor(false)]; + tensor var_2196_to_fp16 = const()[name = tensor("op_2196_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120865280)))]; + tensor q_with_bias_v_19_cast_fp16 = transpose(perm = q_with_bias_v_19_perm_0, x = var_2194_cast_fp16)[name = tensor("transpose_156")]; + tensor x_241_cast_fp16 = matmul(transpose_x = x_241_transpose_x_0, transpose_y = x_241_transpose_y_0, x = q_with_bias_v_19_cast_fp16, y = var_2196_to_fp16)[name = tensor("x_241_cast_fp16")]; + tensor x_243_pad_0 = const()[name = tensor("x_243_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_243_mode_0 = const()[name = tensor("x_243_mode_0"), val = tensor("constant")]; + tensor const_140_to_fp16 = const()[name = tensor("const_140_to_fp16"), val = tensor(0x0p+0)]; + tensor x_243_cast_fp16 = pad(constant_val = const_140_to_fp16, mode = x_243_mode_0, pad = x_243_pad_0, x = x_241_cast_fp16)[name = tensor("x_243_cast_fp16")]; + tensor var_2204 = const()[name = tensor("op_2204"), val = tensor([1, 8, -1, 17])]; + tensor x_245_cast_fp16 = reshape(shape = var_2204, x = x_243_cast_fp16)[name = tensor("x_245_cast_fp16")]; + tensor var_2208_begin_0 = const()[name = tensor("op_2208_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2208_end_0 = const()[name = tensor("op_2208_end_0"), val = tensor([1, 8, 174, 17])]; + tensor var_2208_end_mask_0 = const()[name = tensor("op_2208_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2208_cast_fp16 = slice_by_index(begin = var_2208_begin_0, end = var_2208_end_0, end_mask = var_2208_end_mask_0, x = x_245_cast_fp16)[name = tensor("op_2208_cast_fp16")]; + tensor var_2209 = const()[name = tensor("op_2209"), val = tensor([1, 8, 17, 173])]; + tensor matrix_bd_37_cast_fp16 = reshape(shape = var_2209, x = var_2208_cast_fp16)[name = tensor("matrix_bd_37_cast_fp16")]; + tensor matrix_ac_19_transpose_x_0 = const()[name = tensor("matrix_ac_19_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_19_transpose_y_0 = const()[name = tensor("matrix_ac_19_transpose_y_0"), val = tensor(false)]; + tensor transpose_69_perm_0 = const()[name = tensor("transpose_69_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_70_perm_0 = const()[name = tensor("transpose_70_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_70 = transpose(perm = transpose_70_perm_0, x = k_37_cast_fp16)[name = tensor("transpose_154")]; + tensor transpose_69 = transpose(perm = transpose_69_perm_0, x = var_2192_cast_fp16)[name = tensor("transpose_155")]; + tensor matrix_ac_19_cast_fp16 = matmul(transpose_x = matrix_ac_19_transpose_x_0, transpose_y = matrix_ac_19_transpose_y_0, x = transpose_69, y = transpose_70)[name = tensor("matrix_ac_19_cast_fp16")]; + tensor matrix_bd_39_begin_0 = const()[name = tensor("matrix_bd_39_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_39_end_0 = const()[name = tensor("matrix_bd_39_end_0"), val = tensor([1, 8, 17, 87])]; + tensor matrix_bd_39_end_mask_0 = const()[name = tensor("matrix_bd_39_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_39_cast_fp16 = slice_by_index(begin = matrix_bd_39_begin_0, end = matrix_bd_39_end_0, end_mask = matrix_bd_39_end_mask_0, x = matrix_bd_37_cast_fp16)[name = tensor("matrix_bd_39_cast_fp16")]; + tensor var_2218_cast_fp16 = add(x = matrix_ac_19_cast_fp16, y = matrix_bd_39_cast_fp16)[name = tensor("op_2218_cast_fp16")]; + tensor _inversed_scores_37_y_0_to_fp16 = const()[name = tensor("_inversed_scores_37_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_37_cast_fp16 = mul(x = var_2218_cast_fp16, y = _inversed_scores_37_y_0_to_fp16)[name = tensor("_inversed_scores_37_cast_fp16")]; + tensor scores_39_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_37_cast_fp16, cond = mask_3)[name = tensor("scores_39_cast_fp16")]; + tensor var_2224_cast_fp16 = softmax(axis = var_62, x = scores_39_cast_fp16)[name = tensor("op_2224_cast_fp16")]; + tensor input_509_cast_fp16 = select(a = var_40_to_fp16, b = var_2224_cast_fp16, cond = mask_3)[name = tensor("input_509_cast_fp16")]; + tensor x_247_transpose_x_0 = const()[name = tensor("x_247_transpose_x_0"), val = tensor(false)]; + tensor x_247_transpose_y_0 = const()[name = tensor("x_247_transpose_y_0"), val = tensor(false)]; + tensor value_21_cast_fp16 = transpose(perm = value_21_perm_0, x = v_19_cast_fp16)[name = tensor("transpose_157")]; + tensor x_247_cast_fp16 = matmul(transpose_x = x_247_transpose_x_0, transpose_y = x_247_transpose_y_0, x = input_509_cast_fp16, y = value_21_cast_fp16)[name = tensor("x_247_cast_fp16")]; + tensor var_2228_perm_0 = const()[name = tensor("op_2228_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2229 = const()[name = tensor("op_2229"), val = tensor([1, -1, 512])]; + tensor var_2228_cast_fp16 = transpose(perm = var_2228_perm_0, x = x_247_cast_fp16)[name = tensor("transpose_153")]; + tensor input_511_cast_fp16 = reshape(shape = var_2229, x = var_2228_cast_fp16)[name = tensor("input_511_cast_fp16")]; + tensor encoder_layers_9_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121042496)))]; + tensor linear_88_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_self_attn_linear_out_weight_to_fp16, x = input_511_cast_fp16)[name = tensor("linear_88_cast_fp16")]; + tensor input_515_cast_fp16 = add(x = input_505_cast_fp16, y = linear_88_cast_fp16)[name = tensor("input_515_cast_fp16")]; + tensor x_251_axes_0 = const()[name = tensor("x_251_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121566848)))]; + tensor encoder_layers_9_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121567936)))]; + tensor x_251_cast_fp16 = layer_norm(axes = x_251_axes_0, beta = encoder_layers_9_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_norm_conv_weight_to_fp16, x = input_515_cast_fp16)[name = tensor("x_251_cast_fp16")]; + tensor input_517_perm_0 = const()[name = tensor("input_517_perm_0"), val = tensor([0, 2, 1])]; + tensor input_519_pad_type_0 = const()[name = tensor("input_519_pad_type_0"), val = tensor("valid")]; + tensor input_519_strides_0 = const()[name = tensor("input_519_strides_0"), val = tensor([1])]; + tensor input_519_pad_0 = const()[name = tensor("input_519_pad_0"), val = tensor([0, 0])]; + tensor input_519_dilations_0 = const()[name = tensor("input_519_dilations_0"), val = tensor([1])]; + tensor input_519_groups_0 = const()[name = tensor("input_519_groups_0"), val = tensor(1)]; + tensor encoder_layers_9_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_9_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121569024)))]; + tensor input_517_cast_fp16 = transpose(perm = input_517_perm_0, x = x_251_cast_fp16)[name = tensor("transpose_152")]; + tensor input_519_cast_fp16 = conv(dilations = input_519_dilations_0, groups = input_519_groups_0, pad = input_519_pad_0, pad_type = input_519_pad_type_0, strides = input_519_strides_0, weight = encoder_layers_9_conv_pointwise_conv1_weight_to_fp16, x = input_517_cast_fp16)[name = tensor("input_519_cast_fp16")]; + tensor x_253_split_num_splits_0 = const()[name = tensor("x_253_split_num_splits_0"), val = tensor(2)]; + tensor x_253_split_axis_0 = const()[name = tensor("x_253_split_axis_0"), val = tensor(1)]; + tensor x_253_split_cast_fp16_0, tensor x_253_split_cast_fp16_1 = split(axis = x_253_split_axis_0, num_splits = x_253_split_num_splits_0, x = input_519_cast_fp16)[name = tensor("x_253_split_cast_fp16")]; + tensor x_253_split_1_sigmoid_cast_fp16 = sigmoid(x = x_253_split_cast_fp16_1)[name = tensor("x_253_split_1_sigmoid_cast_fp16")]; + tensor x_253_cast_fp16 = mul(x = x_253_split_cast_fp16_0, y = x_253_split_1_sigmoid_cast_fp16)[name = tensor("x_253_cast_fp16")]; + tensor input_521_cast_fp16 = select(a = var_40_to_fp16, b = x_253_cast_fp16, cond = var_418)[name = tensor("input_521_cast_fp16")]; + tensor new_x_39_interleave_0 = const()[name = tensor("new_x_39_interleave_0"), val = tensor(false)]; + tensor new_x_39_cast_fp16 = concat(axis = var_62, interleave = new_x_39_interleave_0, values = (cache_39_cast_fp16, input_521_cast_fp16))[name = tensor("new_x_39_cast_fp16")]; + tensor var_2267_begin_0 = const()[name = tensor("op_2267_begin_0"), val = tensor([0, 0, 17])]; + tensor var_2267_end_0 = const()[name = tensor("op_2267_end_0"), val = tensor([1, 512, 25])]; + tensor var_2267_end_mask_0 = const()[name = tensor("op_2267_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2267_cast_fp16 = slice_by_index(begin = var_2267_begin_0, end = var_2267_end_0, end_mask = var_2267_end_mask_0, x = new_x_39_cast_fp16)[name = tensor("op_2267_cast_fp16")]; + tensor x_255_pad_type_0 = const()[name = tensor("x_255_pad_type_0"), val = tensor("valid")]; + tensor x_255_groups_0 = const()[name = tensor("x_255_groups_0"), val = tensor(512)]; + tensor x_255_strides_0 = const()[name = tensor("x_255_strides_0"), val = tensor([1])]; + tensor x_255_pad_0 = const()[name = tensor("x_255_pad_0"), val = tensor([0, 0])]; + tensor x_255_dilations_0 = const()[name = tensor("x_255_dilations_0"), val = tensor([1])]; + tensor encoder_layers_9_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_9_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122617664)))]; + tensor x_255_cast_fp16 = conv(dilations = x_255_dilations_0, groups = x_255_groups_0, pad = x_255_pad_0, pad_type = x_255_pad_type_0, strides = x_255_strides_0, weight = encoder_layers_9_conv_depthwise_conv_weight_to_fp16, x = new_x_39_cast_fp16)[name = tensor("x_255_cast_fp16")]; + tensor input_523_perm_0 = const()[name = tensor("input_523_perm_0"), val = tensor([0, 2, 1])]; + tensor x_257_axes_0 = const()[name = tensor("x_257_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_9_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122626944)))]; + tensor encoder_layers_9_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_9_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122628032)))]; + tensor input_523_cast_fp16 = transpose(perm = input_523_perm_0, x = x_255_cast_fp16)[name = tensor("transpose_151")]; + tensor x_257_cast_fp16 = layer_norm(axes = x_257_axes_0, beta = encoder_layers_9_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_conv_batch_norm_weight_to_fp16, x = input_523_cast_fp16)[name = tensor("x_257_cast_fp16")]; + tensor input_525_perm_0 = const()[name = tensor("input_525_perm_0"), val = tensor([0, 2, 1])]; + tensor input_525_cast_fp16 = transpose(perm = input_525_perm_0, x = x_257_cast_fp16)[name = tensor("transpose_150")]; + tensor input_527_cast_fp16 = silu(x = input_525_cast_fp16)[name = tensor("input_527_cast_fp16")]; + tensor x_259_pad_type_0 = const()[name = tensor("x_259_pad_type_0"), val = tensor("valid")]; + tensor x_259_strides_0 = const()[name = tensor("x_259_strides_0"), val = tensor([1])]; + tensor x_259_pad_0 = const()[name = tensor("x_259_pad_0"), val = tensor([0, 0])]; + tensor x_259_dilations_0 = const()[name = tensor("x_259_dilations_0"), val = tensor([1])]; + tensor x_259_groups_0 = const()[name = tensor("x_259_groups_0"), val = tensor(1)]; + tensor encoder_layers_9_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_9_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122629120)))]; + tensor x_259_cast_fp16 = conv(dilations = x_259_dilations_0, groups = x_259_groups_0, pad = x_259_pad_0, pad_type = x_259_pad_type_0, strides = x_259_strides_0, weight = encoder_layers_9_conv_pointwise_conv2_weight_to_fp16, x = input_527_cast_fp16)[name = tensor("x_259_cast_fp16")]; + tensor input_529_perm_0 = const()[name = tensor("input_529_perm_0"), val = tensor([0, 2, 1])]; + tensor input_529_cast_fp16 = transpose(perm = input_529_perm_0, x = x_259_cast_fp16)[name = tensor("transpose_149")]; + tensor input_531_cast_fp16 = add(x = input_515_cast_fp16, y = input_529_cast_fp16)[name = tensor("input_531_cast_fp16")]; + tensor input_533_axes_0 = const()[name = tensor("input_533_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123153472)))]; + tensor encoder_layers_9_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123154560)))]; + tensor input_533_cast_fp16 = layer_norm(axes = input_533_axes_0, beta = encoder_layers_9_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_norm_feed_forward2_weight_to_fp16, x = input_531_cast_fp16)[name = tensor("input_533_cast_fp16")]; + tensor encoder_layers_9_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_9_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123155648)))]; + tensor linear_89_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_9_feed_forward2_linear1_weight_to_fp16, x = input_533_cast_fp16)[name = tensor("linear_89_cast_fp16")]; + tensor input_537_cast_fp16 = silu(x = linear_89_cast_fp16)[name = tensor("input_537_cast_fp16")]; + tensor encoder_layers_9_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_9_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125252864)))]; + tensor linear_90_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_feed_forward2_linear2_weight_to_fp16, x = input_537_cast_fp16)[name = tensor("linear_90_cast_fp16")]; + tensor var_2308_to_fp16 = const()[name = tensor("op_2308_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2309_cast_fp16 = mul(x = linear_90_cast_fp16, y = var_2308_to_fp16)[name = tensor("op_2309_cast_fp16")]; + tensor input_543_cast_fp16 = add(x = input_531_cast_fp16, y = var_2309_cast_fp16)[name = tensor("input_543_cast_fp16")]; + tensor input_545_axes_0 = const()[name = tensor("input_545_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127350080)))]; + tensor encoder_layers_9_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127351168)))]; + tensor input_545_cast_fp16 = layer_norm(axes = input_545_axes_0, beta = encoder_layers_9_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_norm_out_weight_to_fp16, x = input_543_cast_fp16)[name = tensor("input_545_cast_fp16")]; + tensor cache_41_begin_0 = const()[name = tensor("cache_41_begin_0"), val = tensor([10, 0, 0, 0])]; + tensor cache_41_end_0 = const()[name = tensor("cache_41_end_0"), val = tensor([11, 1, 70, 512])]; + tensor cache_41_end_mask_0 = const()[name = tensor("cache_41_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_41_squeeze_mask_0 = const()[name = tensor("cache_41_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_41_cast_fp16 = slice_by_index(begin = cache_41_begin_0, end = cache_41_end_0, end_mask = cache_41_end_mask_0, squeeze_mask = cache_41_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_41_cast_fp16")]; + tensor cache_43_begin_0 = const()[name = tensor("cache_43_begin_0"), val = tensor([10, 0, 0, 0])]; + tensor cache_43_end_0 = const()[name = tensor("cache_43_end_0"), val = tensor([11, 1, 512, 8])]; + tensor cache_43_end_mask_0 = const()[name = tensor("cache_43_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_43_squeeze_mask_0 = const()[name = tensor("cache_43_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_43_cast_fp16 = slice_by_index(begin = cache_43_begin_0, end = cache_43_end_0, end_mask = cache_43_end_mask_0, squeeze_mask = cache_43_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_43_cast_fp16")]; + tensor input_547_axes_0 = const()[name = tensor("input_547_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127352256)))]; + tensor encoder_layers_10_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127353344)))]; + tensor input_547_cast_fp16 = layer_norm(axes = input_547_axes_0, beta = encoder_layers_10_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_norm_feed_forward1_weight_to_fp16, x = input_545_cast_fp16)[name = tensor("input_547_cast_fp16")]; + tensor encoder_layers_10_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_10_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127354432)))]; + tensor linear_91_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_10_feed_forward1_linear1_weight_to_fp16, x = input_547_cast_fp16)[name = tensor("linear_91_cast_fp16")]; + tensor input_551_cast_fp16 = silu(x = linear_91_cast_fp16)[name = tensor("input_551_cast_fp16")]; + tensor encoder_layers_10_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_10_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129451648)))]; + tensor linear_92_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_feed_forward1_linear2_weight_to_fp16, x = input_551_cast_fp16)[name = tensor("linear_92_cast_fp16")]; + tensor var_2343_to_fp16 = const()[name = tensor("op_2343_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2344_cast_fp16 = mul(x = linear_92_cast_fp16, y = var_2343_to_fp16)[name = tensor("op_2344_cast_fp16")]; + tensor input_557_cast_fp16 = add(x = input_545_cast_fp16, y = var_2344_cast_fp16)[name = tensor("input_557_cast_fp16")]; + tensor key_21_axes_0 = const()[name = tensor("key_21_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131548864)))]; + tensor encoder_layers_10_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131549952)))]; + tensor key_21_cast_fp16 = layer_norm(axes = key_21_axes_0, beta = encoder_layers_10_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_norm_self_att_weight_to_fp16, x = input_557_cast_fp16)[name = tensor("key_21_cast_fp16")]; + tensor input_559_interleave_0 = const()[name = tensor("input_559_interleave_0"), val = tensor(false)]; + tensor input_559_cast_fp16 = concat(axis = var_64, interleave = input_559_interleave_0, values = (cache_41_cast_fp16, key_21_cast_fp16))[name = tensor("input_559_cast_fp16")]; + tensor var_2366_begin_0 = const()[name = tensor("op_2366_begin_0"), val = tensor([0, 17, 0])]; + tensor var_2366_end_0 = const()[name = tensor("op_2366_end_0"), val = tensor([1, 70, 512])]; + tensor var_2366_end_mask_0 = const()[name = tensor("op_2366_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2366_cast_fp16 = slice_by_index(begin = var_2366_begin_0, end = var_2366_end_0, end_mask = var_2366_end_mask_0, x = cache_41_cast_fp16)[name = tensor("op_2366_cast_fp16")]; + tensor var_2372_interleave_0 = const()[name = tensor("op_2372_interleave_0"), val = tensor(false)]; + tensor var_2372_cast_fp16 = concat(axis = var_64, interleave = var_2372_interleave_0, values = (var_2366_cast_fp16, key_21_cast_fp16))[name = tensor("op_2372_cast_fp16")]; + tensor encoder_layers_10_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131551040)))]; + tensor linear_93_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_self_attn_linear_q_weight_to_fp16, x = key_21_cast_fp16)[name = tensor("linear_93_cast_fp16")]; + tensor var_2376 = const()[name = tensor("op_2376"), val = tensor([1, -1, 8, 64])]; + tensor q_61_cast_fp16 = reshape(shape = var_2376, x = linear_93_cast_fp16)[name = tensor("q_61_cast_fp16")]; + tensor encoder_layers_10_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132075392)))]; + tensor linear_94_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_self_attn_linear_k_weight_to_fp16, x = input_559_cast_fp16)[name = tensor("linear_94_cast_fp16")]; + tensor var_2380 = const()[name = tensor("op_2380"), val = tensor([1, -1, 8, 64])]; + tensor k_41_cast_fp16 = reshape(shape = var_2380, x = linear_94_cast_fp16)[name = tensor("k_41_cast_fp16")]; + tensor encoder_layers_10_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132599744)))]; + tensor linear_95_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_self_attn_linear_v_weight_to_fp16, x = input_559_cast_fp16)[name = tensor("linear_95_cast_fp16")]; + tensor var_2384 = const()[name = tensor("op_2384"), val = tensor([1, -1, 8, 64])]; + tensor v_21_cast_fp16 = reshape(shape = var_2384, x = linear_95_cast_fp16)[name = tensor("v_21_cast_fp16")]; + tensor value_23_perm_0 = const()[name = tensor("value_23_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_10_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133124096)))]; + tensor var_2396_cast_fp16 = add(x = q_61_cast_fp16, y = encoder_layers_10_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2396_cast_fp16")]; + tensor encoder_layers_10_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133125184)))]; + tensor var_2398_cast_fp16 = add(x = q_61_cast_fp16, y = encoder_layers_10_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2398_cast_fp16")]; + tensor q_with_bias_v_21_perm_0 = const()[name = tensor("q_with_bias_v_21_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_267_transpose_x_0 = const()[name = tensor("x_267_transpose_x_0"), val = tensor(false)]; + tensor x_267_transpose_y_0 = const()[name = tensor("x_267_transpose_y_0"), val = tensor(false)]; + tensor var_2400_to_fp16 = const()[name = tensor("op_2400_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133126272)))]; + tensor q_with_bias_v_21_cast_fp16 = transpose(perm = q_with_bias_v_21_perm_0, x = var_2398_cast_fp16)[name = tensor("transpose_147")]; + tensor x_267_cast_fp16 = matmul(transpose_x = x_267_transpose_x_0, transpose_y = x_267_transpose_y_0, x = q_with_bias_v_21_cast_fp16, y = var_2400_to_fp16)[name = tensor("x_267_cast_fp16")]; + tensor x_269_pad_0 = const()[name = tensor("x_269_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_269_mode_0 = const()[name = tensor("x_269_mode_0"), val = tensor("constant")]; + tensor const_153_to_fp16 = const()[name = tensor("const_153_to_fp16"), val = tensor(0x0p+0)]; + tensor x_269_cast_fp16 = pad(constant_val = const_153_to_fp16, mode = x_269_mode_0, pad = x_269_pad_0, x = x_267_cast_fp16)[name = tensor("x_269_cast_fp16")]; + tensor var_2408 = const()[name = tensor("op_2408"), val = tensor([1, 8, -1, 17])]; + tensor x_271_cast_fp16 = reshape(shape = var_2408, x = x_269_cast_fp16)[name = tensor("x_271_cast_fp16")]; + tensor var_2412_begin_0 = const()[name = tensor("op_2412_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2412_end_0 = const()[name = tensor("op_2412_end_0"), val = tensor([1, 8, 174, 17])]; + tensor var_2412_end_mask_0 = const()[name = tensor("op_2412_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2412_cast_fp16 = slice_by_index(begin = var_2412_begin_0, end = var_2412_end_0, end_mask = var_2412_end_mask_0, x = x_271_cast_fp16)[name = tensor("op_2412_cast_fp16")]; + tensor var_2413 = const()[name = tensor("op_2413"), val = tensor([1, 8, 17, 173])]; + tensor matrix_bd_41_cast_fp16 = reshape(shape = var_2413, x = var_2412_cast_fp16)[name = tensor("matrix_bd_41_cast_fp16")]; + tensor matrix_ac_21_transpose_x_0 = const()[name = tensor("matrix_ac_21_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_21_transpose_y_0 = const()[name = tensor("matrix_ac_21_transpose_y_0"), val = tensor(false)]; + tensor transpose_71_perm_0 = const()[name = tensor("transpose_71_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_72_perm_0 = const()[name = tensor("transpose_72_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_72 = transpose(perm = transpose_72_perm_0, x = k_41_cast_fp16)[name = tensor("transpose_145")]; + tensor transpose_71 = transpose(perm = transpose_71_perm_0, x = var_2396_cast_fp16)[name = tensor("transpose_146")]; + tensor matrix_ac_21_cast_fp16 = matmul(transpose_x = matrix_ac_21_transpose_x_0, transpose_y = matrix_ac_21_transpose_y_0, x = transpose_71, y = transpose_72)[name = tensor("matrix_ac_21_cast_fp16")]; + tensor matrix_bd_43_begin_0 = const()[name = tensor("matrix_bd_43_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_43_end_0 = const()[name = tensor("matrix_bd_43_end_0"), val = tensor([1, 8, 17, 87])]; + tensor matrix_bd_43_end_mask_0 = const()[name = tensor("matrix_bd_43_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_43_cast_fp16 = slice_by_index(begin = matrix_bd_43_begin_0, end = matrix_bd_43_end_0, end_mask = matrix_bd_43_end_mask_0, x = matrix_bd_41_cast_fp16)[name = tensor("matrix_bd_43_cast_fp16")]; + tensor var_2422_cast_fp16 = add(x = matrix_ac_21_cast_fp16, y = matrix_bd_43_cast_fp16)[name = tensor("op_2422_cast_fp16")]; + tensor _inversed_scores_41_y_0_to_fp16 = const()[name = tensor("_inversed_scores_41_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_41_cast_fp16 = mul(x = var_2422_cast_fp16, y = _inversed_scores_41_y_0_to_fp16)[name = tensor("_inversed_scores_41_cast_fp16")]; + tensor scores_43_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_41_cast_fp16, cond = mask_3)[name = tensor("scores_43_cast_fp16")]; + tensor var_2428_cast_fp16 = softmax(axis = var_62, x = scores_43_cast_fp16)[name = tensor("op_2428_cast_fp16")]; + tensor input_561_cast_fp16 = select(a = var_40_to_fp16, b = var_2428_cast_fp16, cond = mask_3)[name = tensor("input_561_cast_fp16")]; + tensor x_273_transpose_x_0 = const()[name = tensor("x_273_transpose_x_0"), val = tensor(false)]; + tensor x_273_transpose_y_0 = const()[name = tensor("x_273_transpose_y_0"), val = tensor(false)]; + tensor value_23_cast_fp16 = transpose(perm = value_23_perm_0, x = v_21_cast_fp16)[name = tensor("transpose_148")]; + tensor x_273_cast_fp16 = matmul(transpose_x = x_273_transpose_x_0, transpose_y = x_273_transpose_y_0, x = input_561_cast_fp16, y = value_23_cast_fp16)[name = tensor("x_273_cast_fp16")]; + tensor var_2432_perm_0 = const()[name = tensor("op_2432_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2433 = const()[name = tensor("op_2433"), val = tensor([1, -1, 512])]; + tensor var_2432_cast_fp16 = transpose(perm = var_2432_perm_0, x = x_273_cast_fp16)[name = tensor("transpose_144")]; + tensor input_563_cast_fp16 = reshape(shape = var_2433, x = var_2432_cast_fp16)[name = tensor("input_563_cast_fp16")]; + tensor encoder_layers_10_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133303488)))]; + tensor linear_97_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_self_attn_linear_out_weight_to_fp16, x = input_563_cast_fp16)[name = tensor("linear_97_cast_fp16")]; + tensor input_567_cast_fp16 = add(x = input_557_cast_fp16, y = linear_97_cast_fp16)[name = tensor("input_567_cast_fp16")]; + tensor x_277_axes_0 = const()[name = tensor("x_277_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133827840)))]; + tensor encoder_layers_10_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133828928)))]; + tensor x_277_cast_fp16 = layer_norm(axes = x_277_axes_0, beta = encoder_layers_10_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_norm_conv_weight_to_fp16, x = input_567_cast_fp16)[name = tensor("x_277_cast_fp16")]; + tensor input_569_perm_0 = const()[name = tensor("input_569_perm_0"), val = tensor([0, 2, 1])]; + tensor input_571_pad_type_0 = const()[name = tensor("input_571_pad_type_0"), val = tensor("valid")]; + tensor input_571_strides_0 = const()[name = tensor("input_571_strides_0"), val = tensor([1])]; + tensor input_571_pad_0 = const()[name = tensor("input_571_pad_0"), val = tensor([0, 0])]; + tensor input_571_dilations_0 = const()[name = tensor("input_571_dilations_0"), val = tensor([1])]; + tensor input_571_groups_0 = const()[name = tensor("input_571_groups_0"), val = tensor(1)]; + tensor encoder_layers_10_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_10_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133830016)))]; + tensor input_569_cast_fp16 = transpose(perm = input_569_perm_0, x = x_277_cast_fp16)[name = tensor("transpose_143")]; + tensor input_571_cast_fp16 = conv(dilations = input_571_dilations_0, groups = input_571_groups_0, pad = input_571_pad_0, pad_type = input_571_pad_type_0, strides = input_571_strides_0, weight = encoder_layers_10_conv_pointwise_conv1_weight_to_fp16, x = input_569_cast_fp16)[name = tensor("input_571_cast_fp16")]; + tensor x_279_split_num_splits_0 = const()[name = tensor("x_279_split_num_splits_0"), val = tensor(2)]; + tensor x_279_split_axis_0 = const()[name = tensor("x_279_split_axis_0"), val = tensor(1)]; + tensor x_279_split_cast_fp16_0, tensor x_279_split_cast_fp16_1 = split(axis = x_279_split_axis_0, num_splits = x_279_split_num_splits_0, x = input_571_cast_fp16)[name = tensor("x_279_split_cast_fp16")]; + tensor x_279_split_1_sigmoid_cast_fp16 = sigmoid(x = x_279_split_cast_fp16_1)[name = tensor("x_279_split_1_sigmoid_cast_fp16")]; + tensor x_279_cast_fp16 = mul(x = x_279_split_cast_fp16_0, y = x_279_split_1_sigmoid_cast_fp16)[name = tensor("x_279_cast_fp16")]; + tensor input_573_cast_fp16 = select(a = var_40_to_fp16, b = x_279_cast_fp16, cond = var_418)[name = tensor("input_573_cast_fp16")]; + tensor new_x_43_interleave_0 = const()[name = tensor("new_x_43_interleave_0"), val = tensor(false)]; + tensor new_x_43_cast_fp16 = concat(axis = var_62, interleave = new_x_43_interleave_0, values = (cache_43_cast_fp16, input_573_cast_fp16))[name = tensor("new_x_43_cast_fp16")]; + tensor var_2471_begin_0 = const()[name = tensor("op_2471_begin_0"), val = tensor([0, 0, 17])]; + tensor var_2471_end_0 = const()[name = tensor("op_2471_end_0"), val = tensor([1, 512, 25])]; + tensor var_2471_end_mask_0 = const()[name = tensor("op_2471_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2471_cast_fp16 = slice_by_index(begin = var_2471_begin_0, end = var_2471_end_0, end_mask = var_2471_end_mask_0, x = new_x_43_cast_fp16)[name = tensor("op_2471_cast_fp16")]; + tensor x_281_pad_type_0 = const()[name = tensor("x_281_pad_type_0"), val = tensor("valid")]; + tensor x_281_groups_0 = const()[name = tensor("x_281_groups_0"), val = tensor(512)]; + tensor x_281_strides_0 = const()[name = tensor("x_281_strides_0"), val = tensor([1])]; + tensor x_281_pad_0 = const()[name = tensor("x_281_pad_0"), val = tensor([0, 0])]; + tensor x_281_dilations_0 = const()[name = tensor("x_281_dilations_0"), val = tensor([1])]; + tensor encoder_layers_10_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_10_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134878656)))]; + tensor x_281_cast_fp16 = conv(dilations = x_281_dilations_0, groups = x_281_groups_0, pad = x_281_pad_0, pad_type = x_281_pad_type_0, strides = x_281_strides_0, weight = encoder_layers_10_conv_depthwise_conv_weight_to_fp16, x = new_x_43_cast_fp16)[name = tensor("x_281_cast_fp16")]; + tensor input_575_perm_0 = const()[name = tensor("input_575_perm_0"), val = tensor([0, 2, 1])]; + tensor x_283_axes_0 = const()[name = tensor("x_283_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_10_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134887936)))]; + tensor encoder_layers_10_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_10_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134889024)))]; + tensor input_575_cast_fp16 = transpose(perm = input_575_perm_0, x = x_281_cast_fp16)[name = tensor("transpose_142")]; + tensor x_283_cast_fp16 = layer_norm(axes = x_283_axes_0, beta = encoder_layers_10_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_conv_batch_norm_weight_to_fp16, x = input_575_cast_fp16)[name = tensor("x_283_cast_fp16")]; + tensor input_577_perm_0 = const()[name = tensor("input_577_perm_0"), val = tensor([0, 2, 1])]; + tensor input_577_cast_fp16 = transpose(perm = input_577_perm_0, x = x_283_cast_fp16)[name = tensor("transpose_141")]; + tensor input_579_cast_fp16 = silu(x = input_577_cast_fp16)[name = tensor("input_579_cast_fp16")]; + tensor x_285_pad_type_0 = const()[name = tensor("x_285_pad_type_0"), val = tensor("valid")]; + tensor x_285_strides_0 = const()[name = tensor("x_285_strides_0"), val = tensor([1])]; + tensor x_285_pad_0 = const()[name = tensor("x_285_pad_0"), val = tensor([0, 0])]; + tensor x_285_dilations_0 = const()[name = tensor("x_285_dilations_0"), val = tensor([1])]; + tensor x_285_groups_0 = const()[name = tensor("x_285_groups_0"), val = tensor(1)]; + tensor encoder_layers_10_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_10_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134890112)))]; + tensor x_285_cast_fp16 = conv(dilations = x_285_dilations_0, groups = x_285_groups_0, pad = x_285_pad_0, pad_type = x_285_pad_type_0, strides = x_285_strides_0, weight = encoder_layers_10_conv_pointwise_conv2_weight_to_fp16, x = input_579_cast_fp16)[name = tensor("x_285_cast_fp16")]; + tensor input_581_perm_0 = const()[name = tensor("input_581_perm_0"), val = tensor([0, 2, 1])]; + tensor input_581_cast_fp16 = transpose(perm = input_581_perm_0, x = x_285_cast_fp16)[name = tensor("transpose_140")]; + tensor input_583_cast_fp16 = add(x = input_567_cast_fp16, y = input_581_cast_fp16)[name = tensor("input_583_cast_fp16")]; + tensor input_585_axes_0 = const()[name = tensor("input_585_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135414464)))]; + tensor encoder_layers_10_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135415552)))]; + tensor input_585_cast_fp16 = layer_norm(axes = input_585_axes_0, beta = encoder_layers_10_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_norm_feed_forward2_weight_to_fp16, x = input_583_cast_fp16)[name = tensor("input_585_cast_fp16")]; + tensor encoder_layers_10_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_10_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135416640)))]; + tensor linear_98_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_10_feed_forward2_linear1_weight_to_fp16, x = input_585_cast_fp16)[name = tensor("linear_98_cast_fp16")]; + tensor input_589_cast_fp16 = silu(x = linear_98_cast_fp16)[name = tensor("input_589_cast_fp16")]; + tensor encoder_layers_10_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_10_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(137513856)))]; + tensor linear_99_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_feed_forward2_linear2_weight_to_fp16, x = input_589_cast_fp16)[name = tensor("linear_99_cast_fp16")]; + tensor var_2512_to_fp16 = const()[name = tensor("op_2512_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2513_cast_fp16 = mul(x = linear_99_cast_fp16, y = var_2512_to_fp16)[name = tensor("op_2513_cast_fp16")]; + tensor input_595_cast_fp16 = add(x = input_583_cast_fp16, y = var_2513_cast_fp16)[name = tensor("input_595_cast_fp16")]; + tensor input_597_axes_0 = const()[name = tensor("input_597_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139611072)))]; + tensor encoder_layers_10_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139612160)))]; + tensor input_597_cast_fp16 = layer_norm(axes = input_597_axes_0, beta = encoder_layers_10_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_norm_out_weight_to_fp16, x = input_595_cast_fp16)[name = tensor("input_597_cast_fp16")]; + tensor cache_45_begin_0 = const()[name = tensor("cache_45_begin_0"), val = tensor([11, 0, 0, 0])]; + tensor cache_45_end_0 = const()[name = tensor("cache_45_end_0"), val = tensor([12, 1, 70, 512])]; + tensor cache_45_end_mask_0 = const()[name = tensor("cache_45_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_45_squeeze_mask_0 = const()[name = tensor("cache_45_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_45_cast_fp16 = slice_by_index(begin = cache_45_begin_0, end = cache_45_end_0, end_mask = cache_45_end_mask_0, squeeze_mask = cache_45_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_45_cast_fp16")]; + tensor cache_47_begin_0 = const()[name = tensor("cache_47_begin_0"), val = tensor([11, 0, 0, 0])]; + tensor cache_47_end_0 = const()[name = tensor("cache_47_end_0"), val = tensor([12, 1, 512, 8])]; + tensor cache_47_end_mask_0 = const()[name = tensor("cache_47_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_47_squeeze_mask_0 = const()[name = tensor("cache_47_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_47_cast_fp16 = slice_by_index(begin = cache_47_begin_0, end = cache_47_end_0, end_mask = cache_47_end_mask_0, squeeze_mask = cache_47_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_47_cast_fp16")]; + tensor input_599_axes_0 = const()[name = tensor("input_599_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139613248)))]; + tensor encoder_layers_11_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139614336)))]; + tensor input_599_cast_fp16 = layer_norm(axes = input_599_axes_0, beta = encoder_layers_11_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_norm_feed_forward1_weight_to_fp16, x = input_597_cast_fp16)[name = tensor("input_599_cast_fp16")]; + tensor encoder_layers_11_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_11_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139615424)))]; + tensor linear_100_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_11_feed_forward1_linear1_weight_to_fp16, x = input_599_cast_fp16)[name = tensor("linear_100_cast_fp16")]; + tensor input_603_cast_fp16 = silu(x = linear_100_cast_fp16)[name = tensor("input_603_cast_fp16")]; + tensor encoder_layers_11_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_11_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141712640)))]; + tensor linear_101_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_feed_forward1_linear2_weight_to_fp16, x = input_603_cast_fp16)[name = tensor("linear_101_cast_fp16")]; + tensor var_2547_to_fp16 = const()[name = tensor("op_2547_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2548_cast_fp16 = mul(x = linear_101_cast_fp16, y = var_2547_to_fp16)[name = tensor("op_2548_cast_fp16")]; + tensor input_609_cast_fp16 = add(x = input_597_cast_fp16, y = var_2548_cast_fp16)[name = tensor("input_609_cast_fp16")]; + tensor key_23_axes_0 = const()[name = tensor("key_23_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143809856)))]; + tensor encoder_layers_11_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143810944)))]; + tensor key_23_cast_fp16 = layer_norm(axes = key_23_axes_0, beta = encoder_layers_11_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_norm_self_att_weight_to_fp16, x = input_609_cast_fp16)[name = tensor("key_23_cast_fp16")]; + tensor input_611_interleave_0 = const()[name = tensor("input_611_interleave_0"), val = tensor(false)]; + tensor input_611_cast_fp16 = concat(axis = var_64, interleave = input_611_interleave_0, values = (cache_45_cast_fp16, key_23_cast_fp16))[name = tensor("input_611_cast_fp16")]; + tensor var_2570_begin_0 = const()[name = tensor("op_2570_begin_0"), val = tensor([0, 17, 0])]; + tensor var_2570_end_0 = const()[name = tensor("op_2570_end_0"), val = tensor([1, 70, 512])]; + tensor var_2570_end_mask_0 = const()[name = tensor("op_2570_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2570_cast_fp16 = slice_by_index(begin = var_2570_begin_0, end = var_2570_end_0, end_mask = var_2570_end_mask_0, x = cache_45_cast_fp16)[name = tensor("op_2570_cast_fp16")]; + tensor var_2576_interleave_0 = const()[name = tensor("op_2576_interleave_0"), val = tensor(false)]; + tensor var_2576_cast_fp16 = concat(axis = var_64, interleave = var_2576_interleave_0, values = (var_2570_cast_fp16, key_23_cast_fp16))[name = tensor("op_2576_cast_fp16")]; + tensor encoder_layers_11_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143812032)))]; + tensor linear_102_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_self_attn_linear_q_weight_to_fp16, x = key_23_cast_fp16)[name = tensor("linear_102_cast_fp16")]; + tensor var_2580 = const()[name = tensor("op_2580"), val = tensor([1, -1, 8, 64])]; + tensor q_67_cast_fp16 = reshape(shape = var_2580, x = linear_102_cast_fp16)[name = tensor("q_67_cast_fp16")]; + tensor encoder_layers_11_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144336384)))]; + tensor linear_103_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_self_attn_linear_k_weight_to_fp16, x = input_611_cast_fp16)[name = tensor("linear_103_cast_fp16")]; + tensor var_2584 = const()[name = tensor("op_2584"), val = tensor([1, -1, 8, 64])]; + tensor k_45_cast_fp16 = reshape(shape = var_2584, x = linear_103_cast_fp16)[name = tensor("k_45_cast_fp16")]; + tensor encoder_layers_11_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144860736)))]; + tensor linear_104_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_self_attn_linear_v_weight_to_fp16, x = input_611_cast_fp16)[name = tensor("linear_104_cast_fp16")]; + tensor var_2588 = const()[name = tensor("op_2588"), val = tensor([1, -1, 8, 64])]; + tensor v_23_cast_fp16 = reshape(shape = var_2588, x = linear_104_cast_fp16)[name = tensor("v_23_cast_fp16")]; + tensor value_25_perm_0 = const()[name = tensor("value_25_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_11_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145385088)))]; + tensor var_2600_cast_fp16 = add(x = q_67_cast_fp16, y = encoder_layers_11_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2600_cast_fp16")]; + tensor encoder_layers_11_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145386176)))]; + tensor var_2602_cast_fp16 = add(x = q_67_cast_fp16, y = encoder_layers_11_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2602_cast_fp16")]; + tensor q_with_bias_v_23_perm_0 = const()[name = tensor("q_with_bias_v_23_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_293_transpose_x_0 = const()[name = tensor("x_293_transpose_x_0"), val = tensor(false)]; + tensor x_293_transpose_y_0 = const()[name = tensor("x_293_transpose_y_0"), val = tensor(false)]; + tensor var_2604_to_fp16 = const()[name = tensor("op_2604_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145387264)))]; + tensor q_with_bias_v_23_cast_fp16 = transpose(perm = q_with_bias_v_23_perm_0, x = var_2602_cast_fp16)[name = tensor("transpose_138")]; + tensor x_293_cast_fp16 = matmul(transpose_x = x_293_transpose_x_0, transpose_y = x_293_transpose_y_0, x = q_with_bias_v_23_cast_fp16, y = var_2604_to_fp16)[name = tensor("x_293_cast_fp16")]; + tensor x_295_pad_0 = const()[name = tensor("x_295_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_295_mode_0 = const()[name = tensor("x_295_mode_0"), val = tensor("constant")]; + tensor const_166_to_fp16 = const()[name = tensor("const_166_to_fp16"), val = tensor(0x0p+0)]; + tensor x_295_cast_fp16 = pad(constant_val = const_166_to_fp16, mode = x_295_mode_0, pad = x_295_pad_0, x = x_293_cast_fp16)[name = tensor("x_295_cast_fp16")]; + tensor var_2612 = const()[name = tensor("op_2612"), val = tensor([1, 8, -1, 17])]; + tensor x_297_cast_fp16 = reshape(shape = var_2612, x = x_295_cast_fp16)[name = tensor("x_297_cast_fp16")]; + tensor var_2616_begin_0 = const()[name = tensor("op_2616_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2616_end_0 = const()[name = tensor("op_2616_end_0"), val = tensor([1, 8, 174, 17])]; + tensor var_2616_end_mask_0 = const()[name = tensor("op_2616_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2616_cast_fp16 = slice_by_index(begin = var_2616_begin_0, end = var_2616_end_0, end_mask = var_2616_end_mask_0, x = x_297_cast_fp16)[name = tensor("op_2616_cast_fp16")]; + tensor var_2617 = const()[name = tensor("op_2617"), val = tensor([1, 8, 17, 173])]; + tensor matrix_bd_45_cast_fp16 = reshape(shape = var_2617, x = var_2616_cast_fp16)[name = tensor("matrix_bd_45_cast_fp16")]; + tensor matrix_ac_23_transpose_x_0 = const()[name = tensor("matrix_ac_23_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_23_transpose_y_0 = const()[name = tensor("matrix_ac_23_transpose_y_0"), val = tensor(false)]; + tensor transpose_73_perm_0 = const()[name = tensor("transpose_73_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_74_perm_0 = const()[name = tensor("transpose_74_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_74 = transpose(perm = transpose_74_perm_0, x = k_45_cast_fp16)[name = tensor("transpose_136")]; + tensor transpose_73 = transpose(perm = transpose_73_perm_0, x = var_2600_cast_fp16)[name = tensor("transpose_137")]; + tensor matrix_ac_23_cast_fp16 = matmul(transpose_x = matrix_ac_23_transpose_x_0, transpose_y = matrix_ac_23_transpose_y_0, x = transpose_73, y = transpose_74)[name = tensor("matrix_ac_23_cast_fp16")]; + tensor matrix_bd_47_begin_0 = const()[name = tensor("matrix_bd_47_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_47_end_0 = const()[name = tensor("matrix_bd_47_end_0"), val = tensor([1, 8, 17, 87])]; + tensor matrix_bd_47_end_mask_0 = const()[name = tensor("matrix_bd_47_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_47_cast_fp16 = slice_by_index(begin = matrix_bd_47_begin_0, end = matrix_bd_47_end_0, end_mask = matrix_bd_47_end_mask_0, x = matrix_bd_45_cast_fp16)[name = tensor("matrix_bd_47_cast_fp16")]; + tensor var_2626_cast_fp16 = add(x = matrix_ac_23_cast_fp16, y = matrix_bd_47_cast_fp16)[name = tensor("op_2626_cast_fp16")]; + tensor _inversed_scores_45_y_0_to_fp16 = const()[name = tensor("_inversed_scores_45_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_45_cast_fp16 = mul(x = var_2626_cast_fp16, y = _inversed_scores_45_y_0_to_fp16)[name = tensor("_inversed_scores_45_cast_fp16")]; + tensor scores_47_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_45_cast_fp16, cond = mask_3)[name = tensor("scores_47_cast_fp16")]; + tensor var_2632_cast_fp16 = softmax(axis = var_62, x = scores_47_cast_fp16)[name = tensor("op_2632_cast_fp16")]; + tensor input_613_cast_fp16 = select(a = var_40_to_fp16, b = var_2632_cast_fp16, cond = mask_3)[name = tensor("input_613_cast_fp16")]; + tensor x_299_transpose_x_0 = const()[name = tensor("x_299_transpose_x_0"), val = tensor(false)]; + tensor x_299_transpose_y_0 = const()[name = tensor("x_299_transpose_y_0"), val = tensor(false)]; + tensor value_25_cast_fp16 = transpose(perm = value_25_perm_0, x = v_23_cast_fp16)[name = tensor("transpose_139")]; + tensor x_299_cast_fp16 = matmul(transpose_x = x_299_transpose_x_0, transpose_y = x_299_transpose_y_0, x = input_613_cast_fp16, y = value_25_cast_fp16)[name = tensor("x_299_cast_fp16")]; + tensor var_2636_perm_0 = const()[name = tensor("op_2636_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2637 = const()[name = tensor("op_2637"), val = tensor([1, -1, 512])]; + tensor var_2636_cast_fp16 = transpose(perm = var_2636_perm_0, x = x_299_cast_fp16)[name = tensor("transpose_135")]; + tensor input_615_cast_fp16 = reshape(shape = var_2637, x = var_2636_cast_fp16)[name = tensor("input_615_cast_fp16")]; + tensor encoder_layers_11_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145564480)))]; + tensor linear_106_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_self_attn_linear_out_weight_to_fp16, x = input_615_cast_fp16)[name = tensor("linear_106_cast_fp16")]; + tensor input_619_cast_fp16 = add(x = input_609_cast_fp16, y = linear_106_cast_fp16)[name = tensor("input_619_cast_fp16")]; + tensor x_303_axes_0 = const()[name = tensor("x_303_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146088832)))]; + tensor encoder_layers_11_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146089920)))]; + tensor x_303_cast_fp16 = layer_norm(axes = x_303_axes_0, beta = encoder_layers_11_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_norm_conv_weight_to_fp16, x = input_619_cast_fp16)[name = tensor("x_303_cast_fp16")]; + tensor input_621_perm_0 = const()[name = tensor("input_621_perm_0"), val = tensor([0, 2, 1])]; + tensor input_623_pad_type_0 = const()[name = tensor("input_623_pad_type_0"), val = tensor("valid")]; + tensor input_623_strides_0 = const()[name = tensor("input_623_strides_0"), val = tensor([1])]; + tensor input_623_pad_0 = const()[name = tensor("input_623_pad_0"), val = tensor([0, 0])]; + tensor input_623_dilations_0 = const()[name = tensor("input_623_dilations_0"), val = tensor([1])]; + tensor input_623_groups_0 = const()[name = tensor("input_623_groups_0"), val = tensor(1)]; + tensor encoder_layers_11_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_11_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146091008)))]; + tensor input_621_cast_fp16 = transpose(perm = input_621_perm_0, x = x_303_cast_fp16)[name = tensor("transpose_134")]; + tensor input_623_cast_fp16 = conv(dilations = input_623_dilations_0, groups = input_623_groups_0, pad = input_623_pad_0, pad_type = input_623_pad_type_0, strides = input_623_strides_0, weight = encoder_layers_11_conv_pointwise_conv1_weight_to_fp16, x = input_621_cast_fp16)[name = tensor("input_623_cast_fp16")]; + tensor x_305_split_num_splits_0 = const()[name = tensor("x_305_split_num_splits_0"), val = tensor(2)]; + tensor x_305_split_axis_0 = const()[name = tensor("x_305_split_axis_0"), val = tensor(1)]; + tensor x_305_split_cast_fp16_0, tensor x_305_split_cast_fp16_1 = split(axis = x_305_split_axis_0, num_splits = x_305_split_num_splits_0, x = input_623_cast_fp16)[name = tensor("x_305_split_cast_fp16")]; + tensor x_305_split_1_sigmoid_cast_fp16 = sigmoid(x = x_305_split_cast_fp16_1)[name = tensor("x_305_split_1_sigmoid_cast_fp16")]; + tensor x_305_cast_fp16 = mul(x = x_305_split_cast_fp16_0, y = x_305_split_1_sigmoid_cast_fp16)[name = tensor("x_305_cast_fp16")]; + tensor input_625_cast_fp16 = select(a = var_40_to_fp16, b = x_305_cast_fp16, cond = var_418)[name = tensor("input_625_cast_fp16")]; + tensor new_x_47_interleave_0 = const()[name = tensor("new_x_47_interleave_0"), val = tensor(false)]; + tensor new_x_47_cast_fp16 = concat(axis = var_62, interleave = new_x_47_interleave_0, values = (cache_47_cast_fp16, input_625_cast_fp16))[name = tensor("new_x_47_cast_fp16")]; + tensor var_2675_begin_0 = const()[name = tensor("op_2675_begin_0"), val = tensor([0, 0, 17])]; + tensor var_2675_end_0 = const()[name = tensor("op_2675_end_0"), val = tensor([1, 512, 25])]; + tensor var_2675_end_mask_0 = const()[name = tensor("op_2675_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2675_cast_fp16 = slice_by_index(begin = var_2675_begin_0, end = var_2675_end_0, end_mask = var_2675_end_mask_0, x = new_x_47_cast_fp16)[name = tensor("op_2675_cast_fp16")]; + tensor x_307_pad_type_0 = const()[name = tensor("x_307_pad_type_0"), val = tensor("valid")]; + tensor x_307_groups_0 = const()[name = tensor("x_307_groups_0"), val = tensor(512)]; + tensor x_307_strides_0 = const()[name = tensor("x_307_strides_0"), val = tensor([1])]; + tensor x_307_pad_0 = const()[name = tensor("x_307_pad_0"), val = tensor([0, 0])]; + tensor x_307_dilations_0 = const()[name = tensor("x_307_dilations_0"), val = tensor([1])]; + tensor encoder_layers_11_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_11_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147139648)))]; + tensor x_307_cast_fp16 = conv(dilations = x_307_dilations_0, groups = x_307_groups_0, pad = x_307_pad_0, pad_type = x_307_pad_type_0, strides = x_307_strides_0, weight = encoder_layers_11_conv_depthwise_conv_weight_to_fp16, x = new_x_47_cast_fp16)[name = tensor("x_307_cast_fp16")]; + tensor input_627_perm_0 = const()[name = tensor("input_627_perm_0"), val = tensor([0, 2, 1])]; + tensor x_309_axes_0 = const()[name = tensor("x_309_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_11_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147148928)))]; + tensor encoder_layers_11_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_11_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147150016)))]; + tensor input_627_cast_fp16 = transpose(perm = input_627_perm_0, x = x_307_cast_fp16)[name = tensor("transpose_133")]; + tensor x_309_cast_fp16 = layer_norm(axes = x_309_axes_0, beta = encoder_layers_11_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_conv_batch_norm_weight_to_fp16, x = input_627_cast_fp16)[name = tensor("x_309_cast_fp16")]; + tensor input_629_perm_0 = const()[name = tensor("input_629_perm_0"), val = tensor([0, 2, 1])]; + tensor input_629_cast_fp16 = transpose(perm = input_629_perm_0, x = x_309_cast_fp16)[name = tensor("transpose_132")]; + tensor input_631_cast_fp16 = silu(x = input_629_cast_fp16)[name = tensor("input_631_cast_fp16")]; + tensor x_311_pad_type_0 = const()[name = tensor("x_311_pad_type_0"), val = tensor("valid")]; + tensor x_311_strides_0 = const()[name = tensor("x_311_strides_0"), val = tensor([1])]; + tensor x_311_pad_0 = const()[name = tensor("x_311_pad_0"), val = tensor([0, 0])]; + tensor x_311_dilations_0 = const()[name = tensor("x_311_dilations_0"), val = tensor([1])]; + tensor x_311_groups_0 = const()[name = tensor("x_311_groups_0"), val = tensor(1)]; + tensor encoder_layers_11_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_11_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147151104)))]; + tensor x_311_cast_fp16 = conv(dilations = x_311_dilations_0, groups = x_311_groups_0, pad = x_311_pad_0, pad_type = x_311_pad_type_0, strides = x_311_strides_0, weight = encoder_layers_11_conv_pointwise_conv2_weight_to_fp16, x = input_631_cast_fp16)[name = tensor("x_311_cast_fp16")]; + tensor input_633_perm_0 = const()[name = tensor("input_633_perm_0"), val = tensor([0, 2, 1])]; + tensor input_633_cast_fp16 = transpose(perm = input_633_perm_0, x = x_311_cast_fp16)[name = tensor("transpose_131")]; + tensor input_635_cast_fp16 = add(x = input_619_cast_fp16, y = input_633_cast_fp16)[name = tensor("input_635_cast_fp16")]; + tensor input_637_axes_0 = const()[name = tensor("input_637_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147675456)))]; + tensor encoder_layers_11_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147676544)))]; + tensor input_637_cast_fp16 = layer_norm(axes = input_637_axes_0, beta = encoder_layers_11_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_norm_feed_forward2_weight_to_fp16, x = input_635_cast_fp16)[name = tensor("input_637_cast_fp16")]; + tensor encoder_layers_11_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_11_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147677632)))]; + tensor linear_107_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_11_feed_forward2_linear1_weight_to_fp16, x = input_637_cast_fp16)[name = tensor("linear_107_cast_fp16")]; + tensor input_641_cast_fp16 = silu(x = linear_107_cast_fp16)[name = tensor("input_641_cast_fp16")]; + tensor encoder_layers_11_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_11_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(149774848)))]; + tensor linear_108_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_feed_forward2_linear2_weight_to_fp16, x = input_641_cast_fp16)[name = tensor("linear_108_cast_fp16")]; + tensor var_2716_to_fp16 = const()[name = tensor("op_2716_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2717_cast_fp16 = mul(x = linear_108_cast_fp16, y = var_2716_to_fp16)[name = tensor("op_2717_cast_fp16")]; + tensor input_647_cast_fp16 = add(x = input_635_cast_fp16, y = var_2717_cast_fp16)[name = tensor("input_647_cast_fp16")]; + tensor input_649_axes_0 = const()[name = tensor("input_649_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151872064)))]; + tensor encoder_layers_11_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151873152)))]; + tensor input_649_cast_fp16 = layer_norm(axes = input_649_axes_0, beta = encoder_layers_11_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_norm_out_weight_to_fp16, x = input_647_cast_fp16)[name = tensor("input_649_cast_fp16")]; + tensor cache_49_begin_0 = const()[name = tensor("cache_49_begin_0"), val = tensor([12, 0, 0, 0])]; + tensor cache_49_end_0 = const()[name = tensor("cache_49_end_0"), val = tensor([13, 1, 70, 512])]; + tensor cache_49_end_mask_0 = const()[name = tensor("cache_49_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_49_squeeze_mask_0 = const()[name = tensor("cache_49_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_49_cast_fp16 = slice_by_index(begin = cache_49_begin_0, end = cache_49_end_0, end_mask = cache_49_end_mask_0, squeeze_mask = cache_49_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_49_cast_fp16")]; + tensor cache_51_begin_0 = const()[name = tensor("cache_51_begin_0"), val = tensor([12, 0, 0, 0])]; + tensor cache_51_end_0 = const()[name = tensor("cache_51_end_0"), val = tensor([13, 1, 512, 8])]; + tensor cache_51_end_mask_0 = const()[name = tensor("cache_51_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_51_squeeze_mask_0 = const()[name = tensor("cache_51_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_51_cast_fp16 = slice_by_index(begin = cache_51_begin_0, end = cache_51_end_0, end_mask = cache_51_end_mask_0, squeeze_mask = cache_51_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_51_cast_fp16")]; + tensor input_651_axes_0 = const()[name = tensor("input_651_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151874240)))]; + tensor encoder_layers_12_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151875328)))]; + tensor input_651_cast_fp16 = layer_norm(axes = input_651_axes_0, beta = encoder_layers_12_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_norm_feed_forward1_weight_to_fp16, x = input_649_cast_fp16)[name = tensor("input_651_cast_fp16")]; + tensor encoder_layers_12_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_12_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151876416)))]; + tensor linear_109_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_12_feed_forward1_linear1_weight_to_fp16, x = input_651_cast_fp16)[name = tensor("linear_109_cast_fp16")]; + tensor input_655_cast_fp16 = silu(x = linear_109_cast_fp16)[name = tensor("input_655_cast_fp16")]; + tensor encoder_layers_12_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_12_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(153973632)))]; + tensor linear_110_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_feed_forward1_linear2_weight_to_fp16, x = input_655_cast_fp16)[name = tensor("linear_110_cast_fp16")]; + tensor var_2751_to_fp16 = const()[name = tensor("op_2751_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2752_cast_fp16 = mul(x = linear_110_cast_fp16, y = var_2751_to_fp16)[name = tensor("op_2752_cast_fp16")]; + tensor input_661_cast_fp16 = add(x = input_649_cast_fp16, y = var_2752_cast_fp16)[name = tensor("input_661_cast_fp16")]; + tensor key_25_axes_0 = const()[name = tensor("key_25_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156070848)))]; + tensor encoder_layers_12_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156071936)))]; + tensor key_25_cast_fp16 = layer_norm(axes = key_25_axes_0, beta = encoder_layers_12_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_norm_self_att_weight_to_fp16, x = input_661_cast_fp16)[name = tensor("key_25_cast_fp16")]; + tensor input_663_interleave_0 = const()[name = tensor("input_663_interleave_0"), val = tensor(false)]; + tensor input_663_cast_fp16 = concat(axis = var_64, interleave = input_663_interleave_0, values = (cache_49_cast_fp16, key_25_cast_fp16))[name = tensor("input_663_cast_fp16")]; + tensor var_2774_begin_0 = const()[name = tensor("op_2774_begin_0"), val = tensor([0, 17, 0])]; + tensor var_2774_end_0 = const()[name = tensor("op_2774_end_0"), val = tensor([1, 70, 512])]; + tensor var_2774_end_mask_0 = const()[name = tensor("op_2774_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2774_cast_fp16 = slice_by_index(begin = var_2774_begin_0, end = var_2774_end_0, end_mask = var_2774_end_mask_0, x = cache_49_cast_fp16)[name = tensor("op_2774_cast_fp16")]; + tensor var_2780_interleave_0 = const()[name = tensor("op_2780_interleave_0"), val = tensor(false)]; + tensor var_2780_cast_fp16 = concat(axis = var_64, interleave = var_2780_interleave_0, values = (var_2774_cast_fp16, key_25_cast_fp16))[name = tensor("op_2780_cast_fp16")]; + tensor encoder_layers_12_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156073024)))]; + tensor linear_111_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_self_attn_linear_q_weight_to_fp16, x = key_25_cast_fp16)[name = tensor("linear_111_cast_fp16")]; + tensor var_2784 = const()[name = tensor("op_2784"), val = tensor([1, -1, 8, 64])]; + tensor q_73_cast_fp16 = reshape(shape = var_2784, x = linear_111_cast_fp16)[name = tensor("q_73_cast_fp16")]; + tensor encoder_layers_12_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156597376)))]; + tensor linear_112_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_self_attn_linear_k_weight_to_fp16, x = input_663_cast_fp16)[name = tensor("linear_112_cast_fp16")]; + tensor var_2788 = const()[name = tensor("op_2788"), val = tensor([1, -1, 8, 64])]; + tensor k_49_cast_fp16 = reshape(shape = var_2788, x = linear_112_cast_fp16)[name = tensor("k_49_cast_fp16")]; + tensor encoder_layers_12_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157121728)))]; + tensor linear_113_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_self_attn_linear_v_weight_to_fp16, x = input_663_cast_fp16)[name = tensor("linear_113_cast_fp16")]; + tensor var_2792 = const()[name = tensor("op_2792"), val = tensor([1, -1, 8, 64])]; + tensor v_25_cast_fp16 = reshape(shape = var_2792, x = linear_113_cast_fp16)[name = tensor("v_25_cast_fp16")]; + tensor value_27_perm_0 = const()[name = tensor("value_27_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_12_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157646080)))]; + tensor var_2804_cast_fp16 = add(x = q_73_cast_fp16, y = encoder_layers_12_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2804_cast_fp16")]; + tensor encoder_layers_12_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157647168)))]; + tensor var_2806_cast_fp16 = add(x = q_73_cast_fp16, y = encoder_layers_12_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2806_cast_fp16")]; + tensor q_with_bias_v_25_perm_0 = const()[name = tensor("q_with_bias_v_25_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_319_transpose_x_0 = const()[name = tensor("x_319_transpose_x_0"), val = tensor(false)]; + tensor x_319_transpose_y_0 = const()[name = tensor("x_319_transpose_y_0"), val = tensor(false)]; + tensor var_2808_to_fp16 = const()[name = tensor("op_2808_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157648256)))]; + tensor q_with_bias_v_25_cast_fp16 = transpose(perm = q_with_bias_v_25_perm_0, x = var_2806_cast_fp16)[name = tensor("transpose_129")]; + tensor x_319_cast_fp16 = matmul(transpose_x = x_319_transpose_x_0, transpose_y = x_319_transpose_y_0, x = q_with_bias_v_25_cast_fp16, y = var_2808_to_fp16)[name = tensor("x_319_cast_fp16")]; + tensor x_321_pad_0 = const()[name = tensor("x_321_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_321_mode_0 = const()[name = tensor("x_321_mode_0"), val = tensor("constant")]; + tensor const_179_to_fp16 = const()[name = tensor("const_179_to_fp16"), val = tensor(0x0p+0)]; + tensor x_321_cast_fp16 = pad(constant_val = const_179_to_fp16, mode = x_321_mode_0, pad = x_321_pad_0, x = x_319_cast_fp16)[name = tensor("x_321_cast_fp16")]; + tensor var_2816 = const()[name = tensor("op_2816"), val = tensor([1, 8, -1, 17])]; + tensor x_323_cast_fp16 = reshape(shape = var_2816, x = x_321_cast_fp16)[name = tensor("x_323_cast_fp16")]; + tensor var_2820_begin_0 = const()[name = tensor("op_2820_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2820_end_0 = const()[name = tensor("op_2820_end_0"), val = tensor([1, 8, 174, 17])]; + tensor var_2820_end_mask_0 = const()[name = tensor("op_2820_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2820_cast_fp16 = slice_by_index(begin = var_2820_begin_0, end = var_2820_end_0, end_mask = var_2820_end_mask_0, x = x_323_cast_fp16)[name = tensor("op_2820_cast_fp16")]; + tensor var_2821 = const()[name = tensor("op_2821"), val = tensor([1, 8, 17, 173])]; + tensor matrix_bd_49_cast_fp16 = reshape(shape = var_2821, x = var_2820_cast_fp16)[name = tensor("matrix_bd_49_cast_fp16")]; + tensor matrix_ac_25_transpose_x_0 = const()[name = tensor("matrix_ac_25_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_25_transpose_y_0 = const()[name = tensor("matrix_ac_25_transpose_y_0"), val = tensor(false)]; + tensor transpose_75_perm_0 = const()[name = tensor("transpose_75_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_76_perm_0 = const()[name = tensor("transpose_76_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_76 = transpose(perm = transpose_76_perm_0, x = k_49_cast_fp16)[name = tensor("transpose_127")]; + tensor transpose_75 = transpose(perm = transpose_75_perm_0, x = var_2804_cast_fp16)[name = tensor("transpose_128")]; + tensor matrix_ac_25_cast_fp16 = matmul(transpose_x = matrix_ac_25_transpose_x_0, transpose_y = matrix_ac_25_transpose_y_0, x = transpose_75, y = transpose_76)[name = tensor("matrix_ac_25_cast_fp16")]; + tensor matrix_bd_51_begin_0 = const()[name = tensor("matrix_bd_51_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_51_end_0 = const()[name = tensor("matrix_bd_51_end_0"), val = tensor([1, 8, 17, 87])]; + tensor matrix_bd_51_end_mask_0 = const()[name = tensor("matrix_bd_51_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_51_cast_fp16 = slice_by_index(begin = matrix_bd_51_begin_0, end = matrix_bd_51_end_0, end_mask = matrix_bd_51_end_mask_0, x = matrix_bd_49_cast_fp16)[name = tensor("matrix_bd_51_cast_fp16")]; + tensor var_2830_cast_fp16 = add(x = matrix_ac_25_cast_fp16, y = matrix_bd_51_cast_fp16)[name = tensor("op_2830_cast_fp16")]; + tensor _inversed_scores_49_y_0_to_fp16 = const()[name = tensor("_inversed_scores_49_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_49_cast_fp16 = mul(x = var_2830_cast_fp16, y = _inversed_scores_49_y_0_to_fp16)[name = tensor("_inversed_scores_49_cast_fp16")]; + tensor scores_51_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_49_cast_fp16, cond = mask_3)[name = tensor("scores_51_cast_fp16")]; + tensor var_2836_cast_fp16 = softmax(axis = var_62, x = scores_51_cast_fp16)[name = tensor("op_2836_cast_fp16")]; + tensor input_665_cast_fp16 = select(a = var_40_to_fp16, b = var_2836_cast_fp16, cond = mask_3)[name = tensor("input_665_cast_fp16")]; + tensor x_325_transpose_x_0 = const()[name = tensor("x_325_transpose_x_0"), val = tensor(false)]; + tensor x_325_transpose_y_0 = const()[name = tensor("x_325_transpose_y_0"), val = tensor(false)]; + tensor value_27_cast_fp16 = transpose(perm = value_27_perm_0, x = v_25_cast_fp16)[name = tensor("transpose_130")]; + tensor x_325_cast_fp16 = matmul(transpose_x = x_325_transpose_x_0, transpose_y = x_325_transpose_y_0, x = input_665_cast_fp16, y = value_27_cast_fp16)[name = tensor("x_325_cast_fp16")]; + tensor var_2840_perm_0 = const()[name = tensor("op_2840_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2841 = const()[name = tensor("op_2841"), val = tensor([1, -1, 512])]; + tensor var_2840_cast_fp16 = transpose(perm = var_2840_perm_0, x = x_325_cast_fp16)[name = tensor("transpose_126")]; + tensor input_667_cast_fp16 = reshape(shape = var_2841, x = var_2840_cast_fp16)[name = tensor("input_667_cast_fp16")]; + tensor encoder_layers_12_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157825472)))]; + tensor linear_115_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_self_attn_linear_out_weight_to_fp16, x = input_667_cast_fp16)[name = tensor("linear_115_cast_fp16")]; + tensor input_671_cast_fp16 = add(x = input_661_cast_fp16, y = linear_115_cast_fp16)[name = tensor("input_671_cast_fp16")]; + tensor x_329_axes_0 = const()[name = tensor("x_329_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158349824)))]; + tensor encoder_layers_12_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158350912)))]; + tensor x_329_cast_fp16 = layer_norm(axes = x_329_axes_0, beta = encoder_layers_12_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_norm_conv_weight_to_fp16, x = input_671_cast_fp16)[name = tensor("x_329_cast_fp16")]; + tensor input_673_perm_0 = const()[name = tensor("input_673_perm_0"), val = tensor([0, 2, 1])]; + tensor input_675_pad_type_0 = const()[name = tensor("input_675_pad_type_0"), val = tensor("valid")]; + tensor input_675_strides_0 = const()[name = tensor("input_675_strides_0"), val = tensor([1])]; + tensor input_675_pad_0 = const()[name = tensor("input_675_pad_0"), val = tensor([0, 0])]; + tensor input_675_dilations_0 = const()[name = tensor("input_675_dilations_0"), val = tensor([1])]; + tensor input_675_groups_0 = const()[name = tensor("input_675_groups_0"), val = tensor(1)]; + tensor encoder_layers_12_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_12_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158352000)))]; + tensor input_673_cast_fp16 = transpose(perm = input_673_perm_0, x = x_329_cast_fp16)[name = tensor("transpose_125")]; + tensor input_675_cast_fp16 = conv(dilations = input_675_dilations_0, groups = input_675_groups_0, pad = input_675_pad_0, pad_type = input_675_pad_type_0, strides = input_675_strides_0, weight = encoder_layers_12_conv_pointwise_conv1_weight_to_fp16, x = input_673_cast_fp16)[name = tensor("input_675_cast_fp16")]; + tensor x_331_split_num_splits_0 = const()[name = tensor("x_331_split_num_splits_0"), val = tensor(2)]; + tensor x_331_split_axis_0 = const()[name = tensor("x_331_split_axis_0"), val = tensor(1)]; + tensor x_331_split_cast_fp16_0, tensor x_331_split_cast_fp16_1 = split(axis = x_331_split_axis_0, num_splits = x_331_split_num_splits_0, x = input_675_cast_fp16)[name = tensor("x_331_split_cast_fp16")]; + tensor x_331_split_1_sigmoid_cast_fp16 = sigmoid(x = x_331_split_cast_fp16_1)[name = tensor("x_331_split_1_sigmoid_cast_fp16")]; + tensor x_331_cast_fp16 = mul(x = x_331_split_cast_fp16_0, y = x_331_split_1_sigmoid_cast_fp16)[name = tensor("x_331_cast_fp16")]; + tensor input_677_cast_fp16 = select(a = var_40_to_fp16, b = x_331_cast_fp16, cond = var_418)[name = tensor("input_677_cast_fp16")]; + tensor new_x_51_interleave_0 = const()[name = tensor("new_x_51_interleave_0"), val = tensor(false)]; + tensor new_x_51_cast_fp16 = concat(axis = var_62, interleave = new_x_51_interleave_0, values = (cache_51_cast_fp16, input_677_cast_fp16))[name = tensor("new_x_51_cast_fp16")]; + tensor var_2879_begin_0 = const()[name = tensor("op_2879_begin_0"), val = tensor([0, 0, 17])]; + tensor var_2879_end_0 = const()[name = tensor("op_2879_end_0"), val = tensor([1, 512, 25])]; + tensor var_2879_end_mask_0 = const()[name = tensor("op_2879_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2879_cast_fp16 = slice_by_index(begin = var_2879_begin_0, end = var_2879_end_0, end_mask = var_2879_end_mask_0, x = new_x_51_cast_fp16)[name = tensor("op_2879_cast_fp16")]; + tensor x_333_pad_type_0 = const()[name = tensor("x_333_pad_type_0"), val = tensor("valid")]; + tensor x_333_groups_0 = const()[name = tensor("x_333_groups_0"), val = tensor(512)]; + tensor x_333_strides_0 = const()[name = tensor("x_333_strides_0"), val = tensor([1])]; + tensor x_333_pad_0 = const()[name = tensor("x_333_pad_0"), val = tensor([0, 0])]; + tensor x_333_dilations_0 = const()[name = tensor("x_333_dilations_0"), val = tensor([1])]; + tensor encoder_layers_12_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_12_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159400640)))]; + tensor x_333_cast_fp16 = conv(dilations = x_333_dilations_0, groups = x_333_groups_0, pad = x_333_pad_0, pad_type = x_333_pad_type_0, strides = x_333_strides_0, weight = encoder_layers_12_conv_depthwise_conv_weight_to_fp16, x = new_x_51_cast_fp16)[name = tensor("x_333_cast_fp16")]; + tensor input_679_perm_0 = const()[name = tensor("input_679_perm_0"), val = tensor([0, 2, 1])]; + tensor x_335_axes_0 = const()[name = tensor("x_335_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_12_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159409920)))]; + tensor encoder_layers_12_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_12_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159411008)))]; + tensor input_679_cast_fp16 = transpose(perm = input_679_perm_0, x = x_333_cast_fp16)[name = tensor("transpose_124")]; + tensor x_335_cast_fp16 = layer_norm(axes = x_335_axes_0, beta = encoder_layers_12_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_conv_batch_norm_weight_to_fp16, x = input_679_cast_fp16)[name = tensor("x_335_cast_fp16")]; + tensor input_681_perm_0 = const()[name = tensor("input_681_perm_0"), val = tensor([0, 2, 1])]; + tensor input_681_cast_fp16 = transpose(perm = input_681_perm_0, x = x_335_cast_fp16)[name = tensor("transpose_123")]; + tensor input_683_cast_fp16 = silu(x = input_681_cast_fp16)[name = tensor("input_683_cast_fp16")]; + tensor x_337_pad_type_0 = const()[name = tensor("x_337_pad_type_0"), val = tensor("valid")]; + tensor x_337_strides_0 = const()[name = tensor("x_337_strides_0"), val = tensor([1])]; + tensor x_337_pad_0 = const()[name = tensor("x_337_pad_0"), val = tensor([0, 0])]; + tensor x_337_dilations_0 = const()[name = tensor("x_337_dilations_0"), val = tensor([1])]; + tensor x_337_groups_0 = const()[name = tensor("x_337_groups_0"), val = tensor(1)]; + tensor encoder_layers_12_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_12_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159412096)))]; + tensor x_337_cast_fp16 = conv(dilations = x_337_dilations_0, groups = x_337_groups_0, pad = x_337_pad_0, pad_type = x_337_pad_type_0, strides = x_337_strides_0, weight = encoder_layers_12_conv_pointwise_conv2_weight_to_fp16, x = input_683_cast_fp16)[name = tensor("x_337_cast_fp16")]; + tensor input_685_perm_0 = const()[name = tensor("input_685_perm_0"), val = tensor([0, 2, 1])]; + tensor input_685_cast_fp16 = transpose(perm = input_685_perm_0, x = x_337_cast_fp16)[name = tensor("transpose_122")]; + tensor input_687_cast_fp16 = add(x = input_671_cast_fp16, y = input_685_cast_fp16)[name = tensor("input_687_cast_fp16")]; + tensor input_689_axes_0 = const()[name = tensor("input_689_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159936448)))]; + tensor encoder_layers_12_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159937536)))]; + tensor input_689_cast_fp16 = layer_norm(axes = input_689_axes_0, beta = encoder_layers_12_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_norm_feed_forward2_weight_to_fp16, x = input_687_cast_fp16)[name = tensor("input_689_cast_fp16")]; + tensor encoder_layers_12_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_12_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159938624)))]; + tensor linear_116_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_12_feed_forward2_linear1_weight_to_fp16, x = input_689_cast_fp16)[name = tensor("linear_116_cast_fp16")]; + tensor input_693_cast_fp16 = silu(x = linear_116_cast_fp16)[name = tensor("input_693_cast_fp16")]; + tensor encoder_layers_12_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_12_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162035840)))]; + tensor linear_117_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_feed_forward2_linear2_weight_to_fp16, x = input_693_cast_fp16)[name = tensor("linear_117_cast_fp16")]; + tensor var_2920_to_fp16 = const()[name = tensor("op_2920_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2921_cast_fp16 = mul(x = linear_117_cast_fp16, y = var_2920_to_fp16)[name = tensor("op_2921_cast_fp16")]; + tensor input_699_cast_fp16 = add(x = input_687_cast_fp16, y = var_2921_cast_fp16)[name = tensor("input_699_cast_fp16")]; + tensor input_701_axes_0 = const()[name = tensor("input_701_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(164133056)))]; + tensor encoder_layers_12_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(164134144)))]; + tensor input_701_cast_fp16 = layer_norm(axes = input_701_axes_0, beta = encoder_layers_12_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_norm_out_weight_to_fp16, x = input_699_cast_fp16)[name = tensor("input_701_cast_fp16")]; + tensor cache_53_begin_0 = const()[name = tensor("cache_53_begin_0"), val = tensor([13, 0, 0, 0])]; + tensor cache_53_end_0 = const()[name = tensor("cache_53_end_0"), val = tensor([14, 1, 70, 512])]; + tensor cache_53_end_mask_0 = const()[name = tensor("cache_53_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_53_squeeze_mask_0 = const()[name = tensor("cache_53_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_53_cast_fp16 = slice_by_index(begin = cache_53_begin_0, end = cache_53_end_0, end_mask = cache_53_end_mask_0, squeeze_mask = cache_53_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_53_cast_fp16")]; + tensor cache_55_begin_0 = const()[name = tensor("cache_55_begin_0"), val = tensor([13, 0, 0, 0])]; + tensor cache_55_end_0 = const()[name = tensor("cache_55_end_0"), val = tensor([14, 1, 512, 8])]; + tensor cache_55_end_mask_0 = const()[name = tensor("cache_55_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_55_squeeze_mask_0 = const()[name = tensor("cache_55_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_55_cast_fp16 = slice_by_index(begin = cache_55_begin_0, end = cache_55_end_0, end_mask = cache_55_end_mask_0, squeeze_mask = cache_55_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_55_cast_fp16")]; + tensor input_703_axes_0 = const()[name = tensor("input_703_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(164135232)))]; + tensor encoder_layers_13_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(164136320)))]; + tensor input_703_cast_fp16 = layer_norm(axes = input_703_axes_0, beta = encoder_layers_13_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_norm_feed_forward1_weight_to_fp16, x = input_701_cast_fp16)[name = tensor("input_703_cast_fp16")]; + tensor encoder_layers_13_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_13_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(164137408)))]; + tensor linear_118_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_13_feed_forward1_linear1_weight_to_fp16, x = input_703_cast_fp16)[name = tensor("linear_118_cast_fp16")]; + tensor input_707_cast_fp16 = silu(x = linear_118_cast_fp16)[name = tensor("input_707_cast_fp16")]; + tensor encoder_layers_13_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_13_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166234624)))]; + tensor linear_119_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_feed_forward1_linear2_weight_to_fp16, x = input_707_cast_fp16)[name = tensor("linear_119_cast_fp16")]; + tensor var_2955_to_fp16 = const()[name = tensor("op_2955_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2956_cast_fp16 = mul(x = linear_119_cast_fp16, y = var_2955_to_fp16)[name = tensor("op_2956_cast_fp16")]; + tensor input_713_cast_fp16 = add(x = input_701_cast_fp16, y = var_2956_cast_fp16)[name = tensor("input_713_cast_fp16")]; + tensor key_27_axes_0 = const()[name = tensor("key_27_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168331840)))]; + tensor encoder_layers_13_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168332928)))]; + tensor key_27_cast_fp16 = layer_norm(axes = key_27_axes_0, beta = encoder_layers_13_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_norm_self_att_weight_to_fp16, x = input_713_cast_fp16)[name = tensor("key_27_cast_fp16")]; + tensor input_715_interleave_0 = const()[name = tensor("input_715_interleave_0"), val = tensor(false)]; + tensor input_715_cast_fp16 = concat(axis = var_64, interleave = input_715_interleave_0, values = (cache_53_cast_fp16, key_27_cast_fp16))[name = tensor("input_715_cast_fp16")]; + tensor var_2978_begin_0 = const()[name = tensor("op_2978_begin_0"), val = tensor([0, 17, 0])]; + tensor var_2978_end_0 = const()[name = tensor("op_2978_end_0"), val = tensor([1, 70, 512])]; + tensor var_2978_end_mask_0 = const()[name = tensor("op_2978_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2978_cast_fp16 = slice_by_index(begin = var_2978_begin_0, end = var_2978_end_0, end_mask = var_2978_end_mask_0, x = cache_53_cast_fp16)[name = tensor("op_2978_cast_fp16")]; + tensor var_2984_interleave_0 = const()[name = tensor("op_2984_interleave_0"), val = tensor(false)]; + tensor var_2984_cast_fp16 = concat(axis = var_64, interleave = var_2984_interleave_0, values = (var_2978_cast_fp16, key_27_cast_fp16))[name = tensor("op_2984_cast_fp16")]; + tensor encoder_layers_13_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168334016)))]; + tensor linear_120_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_self_attn_linear_q_weight_to_fp16, x = key_27_cast_fp16)[name = tensor("linear_120_cast_fp16")]; + tensor var_2988 = const()[name = tensor("op_2988"), val = tensor([1, -1, 8, 64])]; + tensor q_79_cast_fp16 = reshape(shape = var_2988, x = linear_120_cast_fp16)[name = tensor("q_79_cast_fp16")]; + tensor encoder_layers_13_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168858368)))]; + tensor linear_121_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_self_attn_linear_k_weight_to_fp16, x = input_715_cast_fp16)[name = tensor("linear_121_cast_fp16")]; + tensor var_2992 = const()[name = tensor("op_2992"), val = tensor([1, -1, 8, 64])]; + tensor k_53_cast_fp16 = reshape(shape = var_2992, x = linear_121_cast_fp16)[name = tensor("k_53_cast_fp16")]; + tensor encoder_layers_13_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169382720)))]; + tensor linear_122_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_self_attn_linear_v_weight_to_fp16, x = input_715_cast_fp16)[name = tensor("linear_122_cast_fp16")]; + tensor var_2996 = const()[name = tensor("op_2996"), val = tensor([1, -1, 8, 64])]; + tensor v_27_cast_fp16 = reshape(shape = var_2996, x = linear_122_cast_fp16)[name = tensor("v_27_cast_fp16")]; + tensor value_29_perm_0 = const()[name = tensor("value_29_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_13_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169907072)))]; + tensor var_3008_cast_fp16 = add(x = q_79_cast_fp16, y = encoder_layers_13_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3008_cast_fp16")]; + tensor encoder_layers_13_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169908160)))]; + tensor var_3010_cast_fp16 = add(x = q_79_cast_fp16, y = encoder_layers_13_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3010_cast_fp16")]; + tensor q_with_bias_v_27_perm_0 = const()[name = tensor("q_with_bias_v_27_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_345_transpose_x_0 = const()[name = tensor("x_345_transpose_x_0"), val = tensor(false)]; + tensor x_345_transpose_y_0 = const()[name = tensor("x_345_transpose_y_0"), val = tensor(false)]; + tensor var_3012_to_fp16 = const()[name = tensor("op_3012_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169909248)))]; + tensor q_with_bias_v_27_cast_fp16 = transpose(perm = q_with_bias_v_27_perm_0, x = var_3010_cast_fp16)[name = tensor("transpose_120")]; + tensor x_345_cast_fp16 = matmul(transpose_x = x_345_transpose_x_0, transpose_y = x_345_transpose_y_0, x = q_with_bias_v_27_cast_fp16, y = var_3012_to_fp16)[name = tensor("x_345_cast_fp16")]; + tensor x_347_pad_0 = const()[name = tensor("x_347_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_347_mode_0 = const()[name = tensor("x_347_mode_0"), val = tensor("constant")]; + tensor const_192_to_fp16 = const()[name = tensor("const_192_to_fp16"), val = tensor(0x0p+0)]; + tensor x_347_cast_fp16 = pad(constant_val = const_192_to_fp16, mode = x_347_mode_0, pad = x_347_pad_0, x = x_345_cast_fp16)[name = tensor("x_347_cast_fp16")]; + tensor var_3020 = const()[name = tensor("op_3020"), val = tensor([1, 8, -1, 17])]; + tensor x_349_cast_fp16 = reshape(shape = var_3020, x = x_347_cast_fp16)[name = tensor("x_349_cast_fp16")]; + tensor var_3024_begin_0 = const()[name = tensor("op_3024_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_3024_end_0 = const()[name = tensor("op_3024_end_0"), val = tensor([1, 8, 174, 17])]; + tensor var_3024_end_mask_0 = const()[name = tensor("op_3024_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_3024_cast_fp16 = slice_by_index(begin = var_3024_begin_0, end = var_3024_end_0, end_mask = var_3024_end_mask_0, x = x_349_cast_fp16)[name = tensor("op_3024_cast_fp16")]; + tensor var_3025 = const()[name = tensor("op_3025"), val = tensor([1, 8, 17, 173])]; + tensor matrix_bd_53_cast_fp16 = reshape(shape = var_3025, x = var_3024_cast_fp16)[name = tensor("matrix_bd_53_cast_fp16")]; + tensor matrix_ac_27_transpose_x_0 = const()[name = tensor("matrix_ac_27_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_27_transpose_y_0 = const()[name = tensor("matrix_ac_27_transpose_y_0"), val = tensor(false)]; + tensor transpose_77_perm_0 = const()[name = tensor("transpose_77_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_78_perm_0 = const()[name = tensor("transpose_78_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_78 = transpose(perm = transpose_78_perm_0, x = k_53_cast_fp16)[name = tensor("transpose_118")]; + tensor transpose_77 = transpose(perm = transpose_77_perm_0, x = var_3008_cast_fp16)[name = tensor("transpose_119")]; + tensor matrix_ac_27_cast_fp16 = matmul(transpose_x = matrix_ac_27_transpose_x_0, transpose_y = matrix_ac_27_transpose_y_0, x = transpose_77, y = transpose_78)[name = tensor("matrix_ac_27_cast_fp16")]; + tensor matrix_bd_55_begin_0 = const()[name = tensor("matrix_bd_55_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_55_end_0 = const()[name = tensor("matrix_bd_55_end_0"), val = tensor([1, 8, 17, 87])]; + tensor matrix_bd_55_end_mask_0 = const()[name = tensor("matrix_bd_55_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_55_cast_fp16 = slice_by_index(begin = matrix_bd_55_begin_0, end = matrix_bd_55_end_0, end_mask = matrix_bd_55_end_mask_0, x = matrix_bd_53_cast_fp16)[name = tensor("matrix_bd_55_cast_fp16")]; + tensor var_3034_cast_fp16 = add(x = matrix_ac_27_cast_fp16, y = matrix_bd_55_cast_fp16)[name = tensor("op_3034_cast_fp16")]; + tensor _inversed_scores_53_y_0_to_fp16 = const()[name = tensor("_inversed_scores_53_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_53_cast_fp16 = mul(x = var_3034_cast_fp16, y = _inversed_scores_53_y_0_to_fp16)[name = tensor("_inversed_scores_53_cast_fp16")]; + tensor scores_55_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_53_cast_fp16, cond = mask_3)[name = tensor("scores_55_cast_fp16")]; + tensor var_3040_cast_fp16 = softmax(axis = var_62, x = scores_55_cast_fp16)[name = tensor("op_3040_cast_fp16")]; + tensor input_717_cast_fp16 = select(a = var_40_to_fp16, b = var_3040_cast_fp16, cond = mask_3)[name = tensor("input_717_cast_fp16")]; + tensor x_351_transpose_x_0 = const()[name = tensor("x_351_transpose_x_0"), val = tensor(false)]; + tensor x_351_transpose_y_0 = const()[name = tensor("x_351_transpose_y_0"), val = tensor(false)]; + tensor value_29_cast_fp16 = transpose(perm = value_29_perm_0, x = v_27_cast_fp16)[name = tensor("transpose_121")]; + tensor x_351_cast_fp16 = matmul(transpose_x = x_351_transpose_x_0, transpose_y = x_351_transpose_y_0, x = input_717_cast_fp16, y = value_29_cast_fp16)[name = tensor("x_351_cast_fp16")]; + tensor var_3044_perm_0 = const()[name = tensor("op_3044_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_3045 = const()[name = tensor("op_3045"), val = tensor([1, -1, 512])]; + tensor var_3044_cast_fp16 = transpose(perm = var_3044_perm_0, x = x_351_cast_fp16)[name = tensor("transpose_117")]; + tensor input_719_cast_fp16 = reshape(shape = var_3045, x = var_3044_cast_fp16)[name = tensor("input_719_cast_fp16")]; + tensor encoder_layers_13_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170086464)))]; + tensor linear_124_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_self_attn_linear_out_weight_to_fp16, x = input_719_cast_fp16)[name = tensor("linear_124_cast_fp16")]; + tensor input_723_cast_fp16 = add(x = input_713_cast_fp16, y = linear_124_cast_fp16)[name = tensor("input_723_cast_fp16")]; + tensor x_355_axes_0 = const()[name = tensor("x_355_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170610816)))]; + tensor encoder_layers_13_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170611904)))]; + tensor x_355_cast_fp16 = layer_norm(axes = x_355_axes_0, beta = encoder_layers_13_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_norm_conv_weight_to_fp16, x = input_723_cast_fp16)[name = tensor("x_355_cast_fp16")]; + tensor input_725_perm_0 = const()[name = tensor("input_725_perm_0"), val = tensor([0, 2, 1])]; + tensor input_727_pad_type_0 = const()[name = tensor("input_727_pad_type_0"), val = tensor("valid")]; + tensor input_727_strides_0 = const()[name = tensor("input_727_strides_0"), val = tensor([1])]; + tensor input_727_pad_0 = const()[name = tensor("input_727_pad_0"), val = tensor([0, 0])]; + tensor input_727_dilations_0 = const()[name = tensor("input_727_dilations_0"), val = tensor([1])]; + tensor input_727_groups_0 = const()[name = tensor("input_727_groups_0"), val = tensor(1)]; + tensor encoder_layers_13_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_13_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170612992)))]; + tensor input_725_cast_fp16 = transpose(perm = input_725_perm_0, x = x_355_cast_fp16)[name = tensor("transpose_116")]; + tensor input_727_cast_fp16 = conv(dilations = input_727_dilations_0, groups = input_727_groups_0, pad = input_727_pad_0, pad_type = input_727_pad_type_0, strides = input_727_strides_0, weight = encoder_layers_13_conv_pointwise_conv1_weight_to_fp16, x = input_725_cast_fp16)[name = tensor("input_727_cast_fp16")]; + tensor x_357_split_num_splits_0 = const()[name = tensor("x_357_split_num_splits_0"), val = tensor(2)]; + tensor x_357_split_axis_0 = const()[name = tensor("x_357_split_axis_0"), val = tensor(1)]; + tensor x_357_split_cast_fp16_0, tensor x_357_split_cast_fp16_1 = split(axis = x_357_split_axis_0, num_splits = x_357_split_num_splits_0, x = input_727_cast_fp16)[name = tensor("x_357_split_cast_fp16")]; + tensor x_357_split_1_sigmoid_cast_fp16 = sigmoid(x = x_357_split_cast_fp16_1)[name = tensor("x_357_split_1_sigmoid_cast_fp16")]; + tensor x_357_cast_fp16 = mul(x = x_357_split_cast_fp16_0, y = x_357_split_1_sigmoid_cast_fp16)[name = tensor("x_357_cast_fp16")]; + tensor input_729_cast_fp16 = select(a = var_40_to_fp16, b = x_357_cast_fp16, cond = var_418)[name = tensor("input_729_cast_fp16")]; + tensor new_x_55_interleave_0 = const()[name = tensor("new_x_55_interleave_0"), val = tensor(false)]; + tensor new_x_55_cast_fp16 = concat(axis = var_62, interleave = new_x_55_interleave_0, values = (cache_55_cast_fp16, input_729_cast_fp16))[name = tensor("new_x_55_cast_fp16")]; + tensor var_3083_begin_0 = const()[name = tensor("op_3083_begin_0"), val = tensor([0, 0, 17])]; + tensor var_3083_end_0 = const()[name = tensor("op_3083_end_0"), val = tensor([1, 512, 25])]; + tensor var_3083_end_mask_0 = const()[name = tensor("op_3083_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3083_cast_fp16 = slice_by_index(begin = var_3083_begin_0, end = var_3083_end_0, end_mask = var_3083_end_mask_0, x = new_x_55_cast_fp16)[name = tensor("op_3083_cast_fp16")]; + tensor x_359_pad_type_0 = const()[name = tensor("x_359_pad_type_0"), val = tensor("valid")]; + tensor x_359_groups_0 = const()[name = tensor("x_359_groups_0"), val = tensor(512)]; + tensor x_359_strides_0 = const()[name = tensor("x_359_strides_0"), val = tensor([1])]; + tensor x_359_pad_0 = const()[name = tensor("x_359_pad_0"), val = tensor([0, 0])]; + tensor x_359_dilations_0 = const()[name = tensor("x_359_dilations_0"), val = tensor([1])]; + tensor encoder_layers_13_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_13_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171661632)))]; + tensor x_359_cast_fp16 = conv(dilations = x_359_dilations_0, groups = x_359_groups_0, pad = x_359_pad_0, pad_type = x_359_pad_type_0, strides = x_359_strides_0, weight = encoder_layers_13_conv_depthwise_conv_weight_to_fp16, x = new_x_55_cast_fp16)[name = tensor("x_359_cast_fp16")]; + tensor input_731_perm_0 = const()[name = tensor("input_731_perm_0"), val = tensor([0, 2, 1])]; + tensor x_361_axes_0 = const()[name = tensor("x_361_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_13_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171670912)))]; + tensor encoder_layers_13_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_13_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171672000)))]; + tensor input_731_cast_fp16 = transpose(perm = input_731_perm_0, x = x_359_cast_fp16)[name = tensor("transpose_115")]; + tensor x_361_cast_fp16 = layer_norm(axes = x_361_axes_0, beta = encoder_layers_13_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_conv_batch_norm_weight_to_fp16, x = input_731_cast_fp16)[name = tensor("x_361_cast_fp16")]; + tensor input_733_perm_0 = const()[name = tensor("input_733_perm_0"), val = tensor([0, 2, 1])]; + tensor input_733_cast_fp16 = transpose(perm = input_733_perm_0, x = x_361_cast_fp16)[name = tensor("transpose_114")]; + tensor input_735_cast_fp16 = silu(x = input_733_cast_fp16)[name = tensor("input_735_cast_fp16")]; + tensor x_363_pad_type_0 = const()[name = tensor("x_363_pad_type_0"), val = tensor("valid")]; + tensor x_363_strides_0 = const()[name = tensor("x_363_strides_0"), val = tensor([1])]; + tensor x_363_pad_0 = const()[name = tensor("x_363_pad_0"), val = tensor([0, 0])]; + tensor x_363_dilations_0 = const()[name = tensor("x_363_dilations_0"), val = tensor([1])]; + tensor x_363_groups_0 = const()[name = tensor("x_363_groups_0"), val = tensor(1)]; + tensor encoder_layers_13_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_13_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171673088)))]; + tensor x_363_cast_fp16 = conv(dilations = x_363_dilations_0, groups = x_363_groups_0, pad = x_363_pad_0, pad_type = x_363_pad_type_0, strides = x_363_strides_0, weight = encoder_layers_13_conv_pointwise_conv2_weight_to_fp16, x = input_735_cast_fp16)[name = tensor("x_363_cast_fp16")]; + tensor input_737_perm_0 = const()[name = tensor("input_737_perm_0"), val = tensor([0, 2, 1])]; + tensor input_737_cast_fp16 = transpose(perm = input_737_perm_0, x = x_363_cast_fp16)[name = tensor("transpose_113")]; + tensor input_739_cast_fp16 = add(x = input_723_cast_fp16, y = input_737_cast_fp16)[name = tensor("input_739_cast_fp16")]; + tensor input_741_axes_0 = const()[name = tensor("input_741_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172197440)))]; + tensor encoder_layers_13_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172198528)))]; + tensor input_741_cast_fp16 = layer_norm(axes = input_741_axes_0, beta = encoder_layers_13_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_norm_feed_forward2_weight_to_fp16, x = input_739_cast_fp16)[name = tensor("input_741_cast_fp16")]; + tensor encoder_layers_13_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_13_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172199616)))]; + tensor linear_125_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_13_feed_forward2_linear1_weight_to_fp16, x = input_741_cast_fp16)[name = tensor("linear_125_cast_fp16")]; + tensor input_745_cast_fp16 = silu(x = linear_125_cast_fp16)[name = tensor("input_745_cast_fp16")]; + tensor encoder_layers_13_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_13_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174296832)))]; + tensor linear_126_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_feed_forward2_linear2_weight_to_fp16, x = input_745_cast_fp16)[name = tensor("linear_126_cast_fp16")]; + tensor var_3124_to_fp16 = const()[name = tensor("op_3124_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3125_cast_fp16 = mul(x = linear_126_cast_fp16, y = var_3124_to_fp16)[name = tensor("op_3125_cast_fp16")]; + tensor input_751_cast_fp16 = add(x = input_739_cast_fp16, y = var_3125_cast_fp16)[name = tensor("input_751_cast_fp16")]; + tensor input_753_axes_0 = const()[name = tensor("input_753_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176394048)))]; + tensor encoder_layers_13_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176395136)))]; + tensor input_753_cast_fp16 = layer_norm(axes = input_753_axes_0, beta = encoder_layers_13_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_norm_out_weight_to_fp16, x = input_751_cast_fp16)[name = tensor("input_753_cast_fp16")]; + tensor cache_57_begin_0 = const()[name = tensor("cache_57_begin_0"), val = tensor([14, 0, 0, 0])]; + tensor cache_57_end_0 = const()[name = tensor("cache_57_end_0"), val = tensor([15, 1, 70, 512])]; + tensor cache_57_end_mask_0 = const()[name = tensor("cache_57_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_57_squeeze_mask_0 = const()[name = tensor("cache_57_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_57_cast_fp16 = slice_by_index(begin = cache_57_begin_0, end = cache_57_end_0, end_mask = cache_57_end_mask_0, squeeze_mask = cache_57_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_57_cast_fp16")]; + tensor cache_59_begin_0 = const()[name = tensor("cache_59_begin_0"), val = tensor([14, 0, 0, 0])]; + tensor cache_59_end_0 = const()[name = tensor("cache_59_end_0"), val = tensor([15, 1, 512, 8])]; + tensor cache_59_end_mask_0 = const()[name = tensor("cache_59_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_59_squeeze_mask_0 = const()[name = tensor("cache_59_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_59_cast_fp16 = slice_by_index(begin = cache_59_begin_0, end = cache_59_end_0, end_mask = cache_59_end_mask_0, squeeze_mask = cache_59_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_59_cast_fp16")]; + tensor input_755_axes_0 = const()[name = tensor("input_755_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176396224)))]; + tensor encoder_layers_14_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176397312)))]; + tensor input_755_cast_fp16 = layer_norm(axes = input_755_axes_0, beta = encoder_layers_14_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_norm_feed_forward1_weight_to_fp16, x = input_753_cast_fp16)[name = tensor("input_755_cast_fp16")]; + tensor encoder_layers_14_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_14_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176398400)))]; + tensor linear_127_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_14_feed_forward1_linear1_weight_to_fp16, x = input_755_cast_fp16)[name = tensor("linear_127_cast_fp16")]; + tensor input_759_cast_fp16 = silu(x = linear_127_cast_fp16)[name = tensor("input_759_cast_fp16")]; + tensor encoder_layers_14_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_14_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178495616)))]; + tensor linear_128_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_feed_forward1_linear2_weight_to_fp16, x = input_759_cast_fp16)[name = tensor("linear_128_cast_fp16")]; + tensor var_3159_to_fp16 = const()[name = tensor("op_3159_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3160_cast_fp16 = mul(x = linear_128_cast_fp16, y = var_3159_to_fp16)[name = tensor("op_3160_cast_fp16")]; + tensor input_765_cast_fp16 = add(x = input_753_cast_fp16, y = var_3160_cast_fp16)[name = tensor("input_765_cast_fp16")]; + tensor key_29_axes_0 = const()[name = tensor("key_29_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180592832)))]; + tensor encoder_layers_14_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180593920)))]; + tensor key_29_cast_fp16 = layer_norm(axes = key_29_axes_0, beta = encoder_layers_14_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_norm_self_att_weight_to_fp16, x = input_765_cast_fp16)[name = tensor("key_29_cast_fp16")]; + tensor input_767_interleave_0 = const()[name = tensor("input_767_interleave_0"), val = tensor(false)]; + tensor input_767_cast_fp16 = concat(axis = var_64, interleave = input_767_interleave_0, values = (cache_57_cast_fp16, key_29_cast_fp16))[name = tensor("input_767_cast_fp16")]; + tensor var_3182_begin_0 = const()[name = tensor("op_3182_begin_0"), val = tensor([0, 17, 0])]; + tensor var_3182_end_0 = const()[name = tensor("op_3182_end_0"), val = tensor([1, 70, 512])]; + tensor var_3182_end_mask_0 = const()[name = tensor("op_3182_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3182_cast_fp16 = slice_by_index(begin = var_3182_begin_0, end = var_3182_end_0, end_mask = var_3182_end_mask_0, x = cache_57_cast_fp16)[name = tensor("op_3182_cast_fp16")]; + tensor var_3188_interleave_0 = const()[name = tensor("op_3188_interleave_0"), val = tensor(false)]; + tensor var_3188_cast_fp16 = concat(axis = var_64, interleave = var_3188_interleave_0, values = (var_3182_cast_fp16, key_29_cast_fp16))[name = tensor("op_3188_cast_fp16")]; + tensor encoder_layers_14_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180595008)))]; + tensor linear_129_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_self_attn_linear_q_weight_to_fp16, x = key_29_cast_fp16)[name = tensor("linear_129_cast_fp16")]; + tensor var_3192 = const()[name = tensor("op_3192"), val = tensor([1, -1, 8, 64])]; + tensor q_85_cast_fp16 = reshape(shape = var_3192, x = linear_129_cast_fp16)[name = tensor("q_85_cast_fp16")]; + tensor encoder_layers_14_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181119360)))]; + tensor linear_130_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_self_attn_linear_k_weight_to_fp16, x = input_767_cast_fp16)[name = tensor("linear_130_cast_fp16")]; + tensor var_3196 = const()[name = tensor("op_3196"), val = tensor([1, -1, 8, 64])]; + tensor k_57_cast_fp16 = reshape(shape = var_3196, x = linear_130_cast_fp16)[name = tensor("k_57_cast_fp16")]; + tensor encoder_layers_14_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181643712)))]; + tensor linear_131_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_self_attn_linear_v_weight_to_fp16, x = input_767_cast_fp16)[name = tensor("linear_131_cast_fp16")]; + tensor var_3200 = const()[name = tensor("op_3200"), val = tensor([1, -1, 8, 64])]; + tensor v_29_cast_fp16 = reshape(shape = var_3200, x = linear_131_cast_fp16)[name = tensor("v_29_cast_fp16")]; + tensor value_31_perm_0 = const()[name = tensor("value_31_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_14_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182168064)))]; + tensor var_3212_cast_fp16 = add(x = q_85_cast_fp16, y = encoder_layers_14_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3212_cast_fp16")]; + tensor encoder_layers_14_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182169152)))]; + tensor var_3214_cast_fp16 = add(x = q_85_cast_fp16, y = encoder_layers_14_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3214_cast_fp16")]; + tensor q_with_bias_v_29_perm_0 = const()[name = tensor("q_with_bias_v_29_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_371_transpose_x_0 = const()[name = tensor("x_371_transpose_x_0"), val = tensor(false)]; + tensor x_371_transpose_y_0 = const()[name = tensor("x_371_transpose_y_0"), val = tensor(false)]; + tensor var_3216_to_fp16 = const()[name = tensor("op_3216_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182170240)))]; + tensor q_with_bias_v_29_cast_fp16 = transpose(perm = q_with_bias_v_29_perm_0, x = var_3214_cast_fp16)[name = tensor("transpose_111")]; + tensor x_371_cast_fp16 = matmul(transpose_x = x_371_transpose_x_0, transpose_y = x_371_transpose_y_0, x = q_with_bias_v_29_cast_fp16, y = var_3216_to_fp16)[name = tensor("x_371_cast_fp16")]; + tensor x_373_pad_0 = const()[name = tensor("x_373_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_373_mode_0 = const()[name = tensor("x_373_mode_0"), val = tensor("constant")]; + tensor const_205_to_fp16 = const()[name = tensor("const_205_to_fp16"), val = tensor(0x0p+0)]; + tensor x_373_cast_fp16 = pad(constant_val = const_205_to_fp16, mode = x_373_mode_0, pad = x_373_pad_0, x = x_371_cast_fp16)[name = tensor("x_373_cast_fp16")]; + tensor var_3224 = const()[name = tensor("op_3224"), val = tensor([1, 8, -1, 17])]; + tensor x_375_cast_fp16 = reshape(shape = var_3224, x = x_373_cast_fp16)[name = tensor("x_375_cast_fp16")]; + tensor var_3228_begin_0 = const()[name = tensor("op_3228_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_3228_end_0 = const()[name = tensor("op_3228_end_0"), val = tensor([1, 8, 174, 17])]; + tensor var_3228_end_mask_0 = const()[name = tensor("op_3228_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_3228_cast_fp16 = slice_by_index(begin = var_3228_begin_0, end = var_3228_end_0, end_mask = var_3228_end_mask_0, x = x_375_cast_fp16)[name = tensor("op_3228_cast_fp16")]; + tensor var_3229 = const()[name = tensor("op_3229"), val = tensor([1, 8, 17, 173])]; + tensor matrix_bd_57_cast_fp16 = reshape(shape = var_3229, x = var_3228_cast_fp16)[name = tensor("matrix_bd_57_cast_fp16")]; + tensor matrix_ac_29_transpose_x_0 = const()[name = tensor("matrix_ac_29_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_29_transpose_y_0 = const()[name = tensor("matrix_ac_29_transpose_y_0"), val = tensor(false)]; + tensor transpose_79_perm_0 = const()[name = tensor("transpose_79_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_80_perm_0 = const()[name = tensor("transpose_80_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_80 = transpose(perm = transpose_80_perm_0, x = k_57_cast_fp16)[name = tensor("transpose_109")]; + tensor transpose_79 = transpose(perm = transpose_79_perm_0, x = var_3212_cast_fp16)[name = tensor("transpose_110")]; + tensor matrix_ac_29_cast_fp16 = matmul(transpose_x = matrix_ac_29_transpose_x_0, transpose_y = matrix_ac_29_transpose_y_0, x = transpose_79, y = transpose_80)[name = tensor("matrix_ac_29_cast_fp16")]; + tensor matrix_bd_59_begin_0 = const()[name = tensor("matrix_bd_59_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_59_end_0 = const()[name = tensor("matrix_bd_59_end_0"), val = tensor([1, 8, 17, 87])]; + tensor matrix_bd_59_end_mask_0 = const()[name = tensor("matrix_bd_59_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_59_cast_fp16 = slice_by_index(begin = matrix_bd_59_begin_0, end = matrix_bd_59_end_0, end_mask = matrix_bd_59_end_mask_0, x = matrix_bd_57_cast_fp16)[name = tensor("matrix_bd_59_cast_fp16")]; + tensor var_3238_cast_fp16 = add(x = matrix_ac_29_cast_fp16, y = matrix_bd_59_cast_fp16)[name = tensor("op_3238_cast_fp16")]; + tensor _inversed_scores_57_y_0_to_fp16 = const()[name = tensor("_inversed_scores_57_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_57_cast_fp16 = mul(x = var_3238_cast_fp16, y = _inversed_scores_57_y_0_to_fp16)[name = tensor("_inversed_scores_57_cast_fp16")]; + tensor scores_59_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_57_cast_fp16, cond = mask_3)[name = tensor("scores_59_cast_fp16")]; + tensor var_3244_cast_fp16 = softmax(axis = var_62, x = scores_59_cast_fp16)[name = tensor("op_3244_cast_fp16")]; + tensor input_769_cast_fp16 = select(a = var_40_to_fp16, b = var_3244_cast_fp16, cond = mask_3)[name = tensor("input_769_cast_fp16")]; + tensor x_377_transpose_x_0 = const()[name = tensor("x_377_transpose_x_0"), val = tensor(false)]; + tensor x_377_transpose_y_0 = const()[name = tensor("x_377_transpose_y_0"), val = tensor(false)]; + tensor value_31_cast_fp16 = transpose(perm = value_31_perm_0, x = v_29_cast_fp16)[name = tensor("transpose_112")]; + tensor x_377_cast_fp16 = matmul(transpose_x = x_377_transpose_x_0, transpose_y = x_377_transpose_y_0, x = input_769_cast_fp16, y = value_31_cast_fp16)[name = tensor("x_377_cast_fp16")]; + tensor var_3248_perm_0 = const()[name = tensor("op_3248_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_3249 = const()[name = tensor("op_3249"), val = tensor([1, -1, 512])]; + tensor var_3248_cast_fp16 = transpose(perm = var_3248_perm_0, x = x_377_cast_fp16)[name = tensor("transpose_108")]; + tensor input_771_cast_fp16 = reshape(shape = var_3249, x = var_3248_cast_fp16)[name = tensor("input_771_cast_fp16")]; + tensor encoder_layers_14_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182347456)))]; + tensor linear_133_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_self_attn_linear_out_weight_to_fp16, x = input_771_cast_fp16)[name = tensor("linear_133_cast_fp16")]; + tensor input_775_cast_fp16 = add(x = input_765_cast_fp16, y = linear_133_cast_fp16)[name = tensor("input_775_cast_fp16")]; + tensor x_381_axes_0 = const()[name = tensor("x_381_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182871808)))]; + tensor encoder_layers_14_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182872896)))]; + tensor x_381_cast_fp16 = layer_norm(axes = x_381_axes_0, beta = encoder_layers_14_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_norm_conv_weight_to_fp16, x = input_775_cast_fp16)[name = tensor("x_381_cast_fp16")]; + tensor input_777_perm_0 = const()[name = tensor("input_777_perm_0"), val = tensor([0, 2, 1])]; + tensor input_779_pad_type_0 = const()[name = tensor("input_779_pad_type_0"), val = tensor("valid")]; + tensor input_779_strides_0 = const()[name = tensor("input_779_strides_0"), val = tensor([1])]; + tensor input_779_pad_0 = const()[name = tensor("input_779_pad_0"), val = tensor([0, 0])]; + tensor input_779_dilations_0 = const()[name = tensor("input_779_dilations_0"), val = tensor([1])]; + tensor input_779_groups_0 = const()[name = tensor("input_779_groups_0"), val = tensor(1)]; + tensor encoder_layers_14_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_14_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182873984)))]; + tensor input_777_cast_fp16 = transpose(perm = input_777_perm_0, x = x_381_cast_fp16)[name = tensor("transpose_107")]; + tensor input_779_cast_fp16 = conv(dilations = input_779_dilations_0, groups = input_779_groups_0, pad = input_779_pad_0, pad_type = input_779_pad_type_0, strides = input_779_strides_0, weight = encoder_layers_14_conv_pointwise_conv1_weight_to_fp16, x = input_777_cast_fp16)[name = tensor("input_779_cast_fp16")]; + tensor x_383_split_num_splits_0 = const()[name = tensor("x_383_split_num_splits_0"), val = tensor(2)]; + tensor x_383_split_axis_0 = const()[name = tensor("x_383_split_axis_0"), val = tensor(1)]; + tensor x_383_split_cast_fp16_0, tensor x_383_split_cast_fp16_1 = split(axis = x_383_split_axis_0, num_splits = x_383_split_num_splits_0, x = input_779_cast_fp16)[name = tensor("x_383_split_cast_fp16")]; + tensor x_383_split_1_sigmoid_cast_fp16 = sigmoid(x = x_383_split_cast_fp16_1)[name = tensor("x_383_split_1_sigmoid_cast_fp16")]; + tensor x_383_cast_fp16 = mul(x = x_383_split_cast_fp16_0, y = x_383_split_1_sigmoid_cast_fp16)[name = tensor("x_383_cast_fp16")]; + tensor input_781_cast_fp16 = select(a = var_40_to_fp16, b = x_383_cast_fp16, cond = var_418)[name = tensor("input_781_cast_fp16")]; + tensor new_x_59_interleave_0 = const()[name = tensor("new_x_59_interleave_0"), val = tensor(false)]; + tensor new_x_59_cast_fp16 = concat(axis = var_62, interleave = new_x_59_interleave_0, values = (cache_59_cast_fp16, input_781_cast_fp16))[name = tensor("new_x_59_cast_fp16")]; + tensor var_3287_begin_0 = const()[name = tensor("op_3287_begin_0"), val = tensor([0, 0, 17])]; + tensor var_3287_end_0 = const()[name = tensor("op_3287_end_0"), val = tensor([1, 512, 25])]; + tensor var_3287_end_mask_0 = const()[name = tensor("op_3287_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3287_cast_fp16 = slice_by_index(begin = var_3287_begin_0, end = var_3287_end_0, end_mask = var_3287_end_mask_0, x = new_x_59_cast_fp16)[name = tensor("op_3287_cast_fp16")]; + tensor x_385_pad_type_0 = const()[name = tensor("x_385_pad_type_0"), val = tensor("valid")]; + tensor x_385_groups_0 = const()[name = tensor("x_385_groups_0"), val = tensor(512)]; + tensor x_385_strides_0 = const()[name = tensor("x_385_strides_0"), val = tensor([1])]; + tensor x_385_pad_0 = const()[name = tensor("x_385_pad_0"), val = tensor([0, 0])]; + tensor x_385_dilations_0 = const()[name = tensor("x_385_dilations_0"), val = tensor([1])]; + tensor encoder_layers_14_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_14_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183922624)))]; + tensor x_385_cast_fp16 = conv(dilations = x_385_dilations_0, groups = x_385_groups_0, pad = x_385_pad_0, pad_type = x_385_pad_type_0, strides = x_385_strides_0, weight = encoder_layers_14_conv_depthwise_conv_weight_to_fp16, x = new_x_59_cast_fp16)[name = tensor("x_385_cast_fp16")]; + tensor input_783_perm_0 = const()[name = tensor("input_783_perm_0"), val = tensor([0, 2, 1])]; + tensor x_387_axes_0 = const()[name = tensor("x_387_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_14_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183931904)))]; + tensor encoder_layers_14_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_14_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183932992)))]; + tensor input_783_cast_fp16 = transpose(perm = input_783_perm_0, x = x_385_cast_fp16)[name = tensor("transpose_106")]; + tensor x_387_cast_fp16 = layer_norm(axes = x_387_axes_0, beta = encoder_layers_14_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_conv_batch_norm_weight_to_fp16, x = input_783_cast_fp16)[name = tensor("x_387_cast_fp16")]; + tensor input_785_perm_0 = const()[name = tensor("input_785_perm_0"), val = tensor([0, 2, 1])]; + tensor input_785_cast_fp16 = transpose(perm = input_785_perm_0, x = x_387_cast_fp16)[name = tensor("transpose_105")]; + tensor input_787_cast_fp16 = silu(x = input_785_cast_fp16)[name = tensor("input_787_cast_fp16")]; + tensor x_389_pad_type_0 = const()[name = tensor("x_389_pad_type_0"), val = tensor("valid")]; + tensor x_389_strides_0 = const()[name = tensor("x_389_strides_0"), val = tensor([1])]; + tensor x_389_pad_0 = const()[name = tensor("x_389_pad_0"), val = tensor([0, 0])]; + tensor x_389_dilations_0 = const()[name = tensor("x_389_dilations_0"), val = tensor([1])]; + tensor x_389_groups_0 = const()[name = tensor("x_389_groups_0"), val = tensor(1)]; + tensor encoder_layers_14_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_14_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183934080)))]; + tensor x_389_cast_fp16 = conv(dilations = x_389_dilations_0, groups = x_389_groups_0, pad = x_389_pad_0, pad_type = x_389_pad_type_0, strides = x_389_strides_0, weight = encoder_layers_14_conv_pointwise_conv2_weight_to_fp16, x = input_787_cast_fp16)[name = tensor("x_389_cast_fp16")]; + tensor input_789_perm_0 = const()[name = tensor("input_789_perm_0"), val = tensor([0, 2, 1])]; + tensor input_789_cast_fp16 = transpose(perm = input_789_perm_0, x = x_389_cast_fp16)[name = tensor("transpose_104")]; + tensor input_791_cast_fp16 = add(x = input_775_cast_fp16, y = input_789_cast_fp16)[name = tensor("input_791_cast_fp16")]; + tensor input_793_axes_0 = const()[name = tensor("input_793_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184458432)))]; + tensor encoder_layers_14_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184459520)))]; + tensor input_793_cast_fp16 = layer_norm(axes = input_793_axes_0, beta = encoder_layers_14_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_norm_feed_forward2_weight_to_fp16, x = input_791_cast_fp16)[name = tensor("input_793_cast_fp16")]; + tensor encoder_layers_14_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_14_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184460608)))]; + tensor linear_134_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_14_feed_forward2_linear1_weight_to_fp16, x = input_793_cast_fp16)[name = tensor("linear_134_cast_fp16")]; + tensor input_797_cast_fp16 = silu(x = linear_134_cast_fp16)[name = tensor("input_797_cast_fp16")]; + tensor encoder_layers_14_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_14_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186557824)))]; + tensor linear_135_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_feed_forward2_linear2_weight_to_fp16, x = input_797_cast_fp16)[name = tensor("linear_135_cast_fp16")]; + tensor var_3328_to_fp16 = const()[name = tensor("op_3328_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3329_cast_fp16 = mul(x = linear_135_cast_fp16, y = var_3328_to_fp16)[name = tensor("op_3329_cast_fp16")]; + tensor input_803_cast_fp16 = add(x = input_791_cast_fp16, y = var_3329_cast_fp16)[name = tensor("input_803_cast_fp16")]; + tensor input_805_axes_0 = const()[name = tensor("input_805_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188655040)))]; + tensor encoder_layers_14_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188656128)))]; + tensor input_805_cast_fp16 = layer_norm(axes = input_805_axes_0, beta = encoder_layers_14_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_norm_out_weight_to_fp16, x = input_803_cast_fp16)[name = tensor("input_805_cast_fp16")]; + tensor cache_61_begin_0 = const()[name = tensor("cache_61_begin_0"), val = tensor([15, 0, 0, 0])]; + tensor cache_61_end_0 = const()[name = tensor("cache_61_end_0"), val = tensor([16, 1, 70, 512])]; + tensor cache_61_end_mask_0 = const()[name = tensor("cache_61_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_61_squeeze_mask_0 = const()[name = tensor("cache_61_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_61_cast_fp16 = slice_by_index(begin = cache_61_begin_0, end = cache_61_end_0, end_mask = cache_61_end_mask_0, squeeze_mask = cache_61_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_61_cast_fp16")]; + tensor cache_63_begin_0 = const()[name = tensor("cache_63_begin_0"), val = tensor([15, 0, 0, 0])]; + tensor cache_63_end_0 = const()[name = tensor("cache_63_end_0"), val = tensor([16, 1, 512, 8])]; + tensor cache_63_end_mask_0 = const()[name = tensor("cache_63_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_63_squeeze_mask_0 = const()[name = tensor("cache_63_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_63_cast_fp16 = slice_by_index(begin = cache_63_begin_0, end = cache_63_end_0, end_mask = cache_63_end_mask_0, squeeze_mask = cache_63_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_63_cast_fp16")]; + tensor input_807_axes_0 = const()[name = tensor("input_807_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188657216)))]; + tensor encoder_layers_15_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188658304)))]; + tensor input_807_cast_fp16 = layer_norm(axes = input_807_axes_0, beta = encoder_layers_15_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_norm_feed_forward1_weight_to_fp16, x = input_805_cast_fp16)[name = tensor("input_807_cast_fp16")]; + tensor encoder_layers_15_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_15_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188659392)))]; + tensor linear_136_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_15_feed_forward1_linear1_weight_to_fp16, x = input_807_cast_fp16)[name = tensor("linear_136_cast_fp16")]; + tensor input_811_cast_fp16 = silu(x = linear_136_cast_fp16)[name = tensor("input_811_cast_fp16")]; + tensor encoder_layers_15_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_15_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190756608)))]; + tensor linear_137_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_feed_forward1_linear2_weight_to_fp16, x = input_811_cast_fp16)[name = tensor("linear_137_cast_fp16")]; + tensor var_3363_to_fp16 = const()[name = tensor("op_3363_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3364_cast_fp16 = mul(x = linear_137_cast_fp16, y = var_3363_to_fp16)[name = tensor("op_3364_cast_fp16")]; + tensor input_817_cast_fp16 = add(x = input_805_cast_fp16, y = var_3364_cast_fp16)[name = tensor("input_817_cast_fp16")]; + tensor key_31_axes_0 = const()[name = tensor("key_31_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192853824)))]; + tensor encoder_layers_15_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192854912)))]; + tensor key_31_cast_fp16 = layer_norm(axes = key_31_axes_0, beta = encoder_layers_15_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_norm_self_att_weight_to_fp16, x = input_817_cast_fp16)[name = tensor("key_31_cast_fp16")]; + tensor input_819_interleave_0 = const()[name = tensor("input_819_interleave_0"), val = tensor(false)]; + tensor input_819_cast_fp16 = concat(axis = var_64, interleave = input_819_interleave_0, values = (cache_61_cast_fp16, key_31_cast_fp16))[name = tensor("input_819_cast_fp16")]; + tensor var_3386_begin_0 = const()[name = tensor("op_3386_begin_0"), val = tensor([0, 17, 0])]; + tensor var_3386_end_0 = const()[name = tensor("op_3386_end_0"), val = tensor([1, 70, 512])]; + tensor var_3386_end_mask_0 = const()[name = tensor("op_3386_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3386_cast_fp16 = slice_by_index(begin = var_3386_begin_0, end = var_3386_end_0, end_mask = var_3386_end_mask_0, x = cache_61_cast_fp16)[name = tensor("op_3386_cast_fp16")]; + tensor var_3392_interleave_0 = const()[name = tensor("op_3392_interleave_0"), val = tensor(false)]; + tensor var_3392_cast_fp16 = concat(axis = var_64, interleave = var_3392_interleave_0, values = (var_3386_cast_fp16, key_31_cast_fp16))[name = tensor("op_3392_cast_fp16")]; + tensor encoder_layers_15_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192856000)))]; + tensor linear_138_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_self_attn_linear_q_weight_to_fp16, x = key_31_cast_fp16)[name = tensor("linear_138_cast_fp16")]; + tensor var_3396 = const()[name = tensor("op_3396"), val = tensor([1, -1, 8, 64])]; + tensor q_91_cast_fp16 = reshape(shape = var_3396, x = linear_138_cast_fp16)[name = tensor("q_91_cast_fp16")]; + tensor encoder_layers_15_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(193380352)))]; + tensor linear_139_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_self_attn_linear_k_weight_to_fp16, x = input_819_cast_fp16)[name = tensor("linear_139_cast_fp16")]; + tensor var_3400 = const()[name = tensor("op_3400"), val = tensor([1, -1, 8, 64])]; + tensor k_61_cast_fp16 = reshape(shape = var_3400, x = linear_139_cast_fp16)[name = tensor("k_61_cast_fp16")]; + tensor encoder_layers_15_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(193904704)))]; + tensor linear_140_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_self_attn_linear_v_weight_to_fp16, x = input_819_cast_fp16)[name = tensor("linear_140_cast_fp16")]; + tensor var_3404 = const()[name = tensor("op_3404"), val = tensor([1, -1, 8, 64])]; + tensor v_31_cast_fp16 = reshape(shape = var_3404, x = linear_140_cast_fp16)[name = tensor("v_31_cast_fp16")]; + tensor value_33_perm_0 = const()[name = tensor("value_33_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_15_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194429056)))]; + tensor var_3416_cast_fp16 = add(x = q_91_cast_fp16, y = encoder_layers_15_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3416_cast_fp16")]; + tensor encoder_layers_15_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194430144)))]; + tensor var_3418_cast_fp16 = add(x = q_91_cast_fp16, y = encoder_layers_15_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3418_cast_fp16")]; + tensor q_with_bias_v_31_perm_0 = const()[name = tensor("q_with_bias_v_31_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_397_transpose_x_0 = const()[name = tensor("x_397_transpose_x_0"), val = tensor(false)]; + tensor x_397_transpose_y_0 = const()[name = tensor("x_397_transpose_y_0"), val = tensor(false)]; + tensor var_3420_to_fp16 = const()[name = tensor("op_3420_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194431232)))]; + tensor q_with_bias_v_31_cast_fp16 = transpose(perm = q_with_bias_v_31_perm_0, x = var_3418_cast_fp16)[name = tensor("transpose_102")]; + tensor x_397_cast_fp16 = matmul(transpose_x = x_397_transpose_x_0, transpose_y = x_397_transpose_y_0, x = q_with_bias_v_31_cast_fp16, y = var_3420_to_fp16)[name = tensor("x_397_cast_fp16")]; + tensor x_399_pad_0 = const()[name = tensor("x_399_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_399_mode_0 = const()[name = tensor("x_399_mode_0"), val = tensor("constant")]; + tensor const_218_to_fp16 = const()[name = tensor("const_218_to_fp16"), val = tensor(0x0p+0)]; + tensor x_399_cast_fp16 = pad(constant_val = const_218_to_fp16, mode = x_399_mode_0, pad = x_399_pad_0, x = x_397_cast_fp16)[name = tensor("x_399_cast_fp16")]; + tensor var_3428 = const()[name = tensor("op_3428"), val = tensor([1, 8, -1, 17])]; + tensor x_401_cast_fp16 = reshape(shape = var_3428, x = x_399_cast_fp16)[name = tensor("x_401_cast_fp16")]; + tensor var_3432_begin_0 = const()[name = tensor("op_3432_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_3432_end_0 = const()[name = tensor("op_3432_end_0"), val = tensor([1, 8, 174, 17])]; + tensor var_3432_end_mask_0 = const()[name = tensor("op_3432_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_3432_cast_fp16 = slice_by_index(begin = var_3432_begin_0, end = var_3432_end_0, end_mask = var_3432_end_mask_0, x = x_401_cast_fp16)[name = tensor("op_3432_cast_fp16")]; + tensor var_3433 = const()[name = tensor("op_3433"), val = tensor([1, 8, 17, 173])]; + tensor matrix_bd_61_cast_fp16 = reshape(shape = var_3433, x = var_3432_cast_fp16)[name = tensor("matrix_bd_61_cast_fp16")]; + tensor matrix_ac_31_transpose_x_0 = const()[name = tensor("matrix_ac_31_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_31_transpose_y_0 = const()[name = tensor("matrix_ac_31_transpose_y_0"), val = tensor(false)]; + tensor transpose_81_perm_0 = const()[name = tensor("transpose_81_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_82_perm_0 = const()[name = tensor("transpose_82_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_82 = transpose(perm = transpose_82_perm_0, x = k_61_cast_fp16)[name = tensor("transpose_100")]; + tensor transpose_81 = transpose(perm = transpose_81_perm_0, x = var_3416_cast_fp16)[name = tensor("transpose_101")]; + tensor matrix_ac_31_cast_fp16 = matmul(transpose_x = matrix_ac_31_transpose_x_0, transpose_y = matrix_ac_31_transpose_y_0, x = transpose_81, y = transpose_82)[name = tensor("matrix_ac_31_cast_fp16")]; + tensor matrix_bd_63_begin_0 = const()[name = tensor("matrix_bd_63_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_63_end_0 = const()[name = tensor("matrix_bd_63_end_0"), val = tensor([1, 8, 17, 87])]; + tensor matrix_bd_63_end_mask_0 = const()[name = tensor("matrix_bd_63_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_63_cast_fp16 = slice_by_index(begin = matrix_bd_63_begin_0, end = matrix_bd_63_end_0, end_mask = matrix_bd_63_end_mask_0, x = matrix_bd_61_cast_fp16)[name = tensor("matrix_bd_63_cast_fp16")]; + tensor var_3442_cast_fp16 = add(x = matrix_ac_31_cast_fp16, y = matrix_bd_63_cast_fp16)[name = tensor("op_3442_cast_fp16")]; + tensor _inversed_scores_61_y_0_to_fp16 = const()[name = tensor("_inversed_scores_61_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_61_cast_fp16 = mul(x = var_3442_cast_fp16, y = _inversed_scores_61_y_0_to_fp16)[name = tensor("_inversed_scores_61_cast_fp16")]; + tensor scores_63_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_61_cast_fp16, cond = mask_3)[name = tensor("scores_63_cast_fp16")]; + tensor var_3448_cast_fp16 = softmax(axis = var_62, x = scores_63_cast_fp16)[name = tensor("op_3448_cast_fp16")]; + tensor input_821_cast_fp16 = select(a = var_40_to_fp16, b = var_3448_cast_fp16, cond = mask_3)[name = tensor("input_821_cast_fp16")]; + tensor x_403_transpose_x_0 = const()[name = tensor("x_403_transpose_x_0"), val = tensor(false)]; + tensor x_403_transpose_y_0 = const()[name = tensor("x_403_transpose_y_0"), val = tensor(false)]; + tensor value_33_cast_fp16 = transpose(perm = value_33_perm_0, x = v_31_cast_fp16)[name = tensor("transpose_103")]; + tensor x_403_cast_fp16 = matmul(transpose_x = x_403_transpose_x_0, transpose_y = x_403_transpose_y_0, x = input_821_cast_fp16, y = value_33_cast_fp16)[name = tensor("x_403_cast_fp16")]; + tensor var_3452_perm_0 = const()[name = tensor("op_3452_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_3453 = const()[name = tensor("op_3453"), val = tensor([1, -1, 512])]; + tensor var_3452_cast_fp16 = transpose(perm = var_3452_perm_0, x = x_403_cast_fp16)[name = tensor("transpose_99")]; + tensor input_823_cast_fp16 = reshape(shape = var_3453, x = var_3452_cast_fp16)[name = tensor("input_823_cast_fp16")]; + tensor encoder_layers_15_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194608448)))]; + tensor linear_142_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_self_attn_linear_out_weight_to_fp16, x = input_823_cast_fp16)[name = tensor("linear_142_cast_fp16")]; + tensor input_827_cast_fp16 = add(x = input_817_cast_fp16, y = linear_142_cast_fp16)[name = tensor("input_827_cast_fp16")]; + tensor x_407_axes_0 = const()[name = tensor("x_407_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195132800)))]; + tensor encoder_layers_15_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195133888)))]; + tensor x_407_cast_fp16 = layer_norm(axes = x_407_axes_0, beta = encoder_layers_15_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_norm_conv_weight_to_fp16, x = input_827_cast_fp16)[name = tensor("x_407_cast_fp16")]; + tensor input_829_perm_0 = const()[name = tensor("input_829_perm_0"), val = tensor([0, 2, 1])]; + tensor input_831_pad_type_0 = const()[name = tensor("input_831_pad_type_0"), val = tensor("valid")]; + tensor input_831_strides_0 = const()[name = tensor("input_831_strides_0"), val = tensor([1])]; + tensor input_831_pad_0 = const()[name = tensor("input_831_pad_0"), val = tensor([0, 0])]; + tensor input_831_dilations_0 = const()[name = tensor("input_831_dilations_0"), val = tensor([1])]; + tensor input_831_groups_0 = const()[name = tensor("input_831_groups_0"), val = tensor(1)]; + tensor encoder_layers_15_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_15_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195134976)))]; + tensor input_829_cast_fp16 = transpose(perm = input_829_perm_0, x = x_407_cast_fp16)[name = tensor("transpose_98")]; + tensor input_831_cast_fp16 = conv(dilations = input_831_dilations_0, groups = input_831_groups_0, pad = input_831_pad_0, pad_type = input_831_pad_type_0, strides = input_831_strides_0, weight = encoder_layers_15_conv_pointwise_conv1_weight_to_fp16, x = input_829_cast_fp16)[name = tensor("input_831_cast_fp16")]; + tensor x_409_split_num_splits_0 = const()[name = tensor("x_409_split_num_splits_0"), val = tensor(2)]; + tensor x_409_split_axis_0 = const()[name = tensor("x_409_split_axis_0"), val = tensor(1)]; + tensor x_409_split_cast_fp16_0, tensor x_409_split_cast_fp16_1 = split(axis = x_409_split_axis_0, num_splits = x_409_split_num_splits_0, x = input_831_cast_fp16)[name = tensor("x_409_split_cast_fp16")]; + tensor x_409_split_1_sigmoid_cast_fp16 = sigmoid(x = x_409_split_cast_fp16_1)[name = tensor("x_409_split_1_sigmoid_cast_fp16")]; + tensor x_409_cast_fp16 = mul(x = x_409_split_cast_fp16_0, y = x_409_split_1_sigmoid_cast_fp16)[name = tensor("x_409_cast_fp16")]; + tensor input_833_cast_fp16 = select(a = var_40_to_fp16, b = x_409_cast_fp16, cond = var_418)[name = tensor("input_833_cast_fp16")]; + tensor new_x_63_interleave_0 = const()[name = tensor("new_x_63_interleave_0"), val = tensor(false)]; + tensor new_x_63_cast_fp16 = concat(axis = var_62, interleave = new_x_63_interleave_0, values = (cache_63_cast_fp16, input_833_cast_fp16))[name = tensor("new_x_63_cast_fp16")]; + tensor var_3491_begin_0 = const()[name = tensor("op_3491_begin_0"), val = tensor([0, 0, 17])]; + tensor var_3491_end_0 = const()[name = tensor("op_3491_end_0"), val = tensor([1, 512, 25])]; + tensor var_3491_end_mask_0 = const()[name = tensor("op_3491_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3491_cast_fp16 = slice_by_index(begin = var_3491_begin_0, end = var_3491_end_0, end_mask = var_3491_end_mask_0, x = new_x_63_cast_fp16)[name = tensor("op_3491_cast_fp16")]; + tensor x_411_pad_type_0 = const()[name = tensor("x_411_pad_type_0"), val = tensor("valid")]; + tensor x_411_groups_0 = const()[name = tensor("x_411_groups_0"), val = tensor(512)]; + tensor x_411_strides_0 = const()[name = tensor("x_411_strides_0"), val = tensor([1])]; + tensor x_411_pad_0 = const()[name = tensor("x_411_pad_0"), val = tensor([0, 0])]; + tensor x_411_dilations_0 = const()[name = tensor("x_411_dilations_0"), val = tensor([1])]; + tensor encoder_layers_15_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_15_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196183616)))]; + tensor x_411_cast_fp16 = conv(dilations = x_411_dilations_0, groups = x_411_groups_0, pad = x_411_pad_0, pad_type = x_411_pad_type_0, strides = x_411_strides_0, weight = encoder_layers_15_conv_depthwise_conv_weight_to_fp16, x = new_x_63_cast_fp16)[name = tensor("x_411_cast_fp16")]; + tensor input_835_perm_0 = const()[name = tensor("input_835_perm_0"), val = tensor([0, 2, 1])]; + tensor x_413_axes_0 = const()[name = tensor("x_413_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_15_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196192896)))]; + tensor encoder_layers_15_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_15_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196193984)))]; + tensor input_835_cast_fp16 = transpose(perm = input_835_perm_0, x = x_411_cast_fp16)[name = tensor("transpose_97")]; + tensor x_413_cast_fp16 = layer_norm(axes = x_413_axes_0, beta = encoder_layers_15_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_conv_batch_norm_weight_to_fp16, x = input_835_cast_fp16)[name = tensor("x_413_cast_fp16")]; + tensor input_837_perm_0 = const()[name = tensor("input_837_perm_0"), val = tensor([0, 2, 1])]; + tensor input_837_cast_fp16 = transpose(perm = input_837_perm_0, x = x_413_cast_fp16)[name = tensor("transpose_96")]; + tensor input_839_cast_fp16 = silu(x = input_837_cast_fp16)[name = tensor("input_839_cast_fp16")]; + tensor x_415_pad_type_0 = const()[name = tensor("x_415_pad_type_0"), val = tensor("valid")]; + tensor x_415_strides_0 = const()[name = tensor("x_415_strides_0"), val = tensor([1])]; + tensor x_415_pad_0 = const()[name = tensor("x_415_pad_0"), val = tensor([0, 0])]; + tensor x_415_dilations_0 = const()[name = tensor("x_415_dilations_0"), val = tensor([1])]; + tensor x_415_groups_0 = const()[name = tensor("x_415_groups_0"), val = tensor(1)]; + tensor encoder_layers_15_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_15_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196195072)))]; + tensor x_415_cast_fp16 = conv(dilations = x_415_dilations_0, groups = x_415_groups_0, pad = x_415_pad_0, pad_type = x_415_pad_type_0, strides = x_415_strides_0, weight = encoder_layers_15_conv_pointwise_conv2_weight_to_fp16, x = input_839_cast_fp16)[name = tensor("x_415_cast_fp16")]; + tensor input_841_perm_0 = const()[name = tensor("input_841_perm_0"), val = tensor([0, 2, 1])]; + tensor input_841_cast_fp16 = transpose(perm = input_841_perm_0, x = x_415_cast_fp16)[name = tensor("transpose_95")]; + tensor input_843_cast_fp16 = add(x = input_827_cast_fp16, y = input_841_cast_fp16)[name = tensor("input_843_cast_fp16")]; + tensor input_845_axes_0 = const()[name = tensor("input_845_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196719424)))]; + tensor encoder_layers_15_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196720512)))]; + tensor input_845_cast_fp16 = layer_norm(axes = input_845_axes_0, beta = encoder_layers_15_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_norm_feed_forward2_weight_to_fp16, x = input_843_cast_fp16)[name = tensor("input_845_cast_fp16")]; + tensor encoder_layers_15_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_15_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196721600)))]; + tensor linear_143_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_15_feed_forward2_linear1_weight_to_fp16, x = input_845_cast_fp16)[name = tensor("linear_143_cast_fp16")]; + tensor input_849_cast_fp16 = silu(x = linear_143_cast_fp16)[name = tensor("input_849_cast_fp16")]; + tensor encoder_layers_15_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_15_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198818816)))]; + tensor linear_144_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_feed_forward2_linear2_weight_to_fp16, x = input_849_cast_fp16)[name = tensor("linear_144_cast_fp16")]; + tensor var_3532_to_fp16 = const()[name = tensor("op_3532_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3533_cast_fp16 = mul(x = linear_144_cast_fp16, y = var_3532_to_fp16)[name = tensor("op_3533_cast_fp16")]; + tensor input_855_cast_fp16 = add(x = input_843_cast_fp16, y = var_3533_cast_fp16)[name = tensor("input_855_cast_fp16")]; + tensor input_857_axes_0 = const()[name = tensor("input_857_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200916032)))]; + tensor encoder_layers_15_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200917120)))]; + tensor input_857_cast_fp16 = layer_norm(axes = input_857_axes_0, beta = encoder_layers_15_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_norm_out_weight_to_fp16, x = input_855_cast_fp16)[name = tensor("input_857_cast_fp16")]; + tensor cache_65_begin_0 = const()[name = tensor("cache_65_begin_0"), val = tensor([16, 0, 0, 0])]; + tensor cache_65_end_0 = const()[name = tensor("cache_65_end_0"), val = tensor([17, 1, 70, 512])]; + tensor cache_65_end_mask_0 = const()[name = tensor("cache_65_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_65_squeeze_mask_0 = const()[name = tensor("cache_65_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_65_cast_fp16 = slice_by_index(begin = cache_65_begin_0, end = cache_65_end_0, end_mask = cache_65_end_mask_0, squeeze_mask = cache_65_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_65_cast_fp16")]; + tensor cache_begin_0 = const()[name = tensor("cache_begin_0"), val = tensor([16, 0, 0, 0])]; + tensor cache_end_0 = const()[name = tensor("cache_end_0"), val = tensor([17, 1, 512, 8])]; + tensor cache_end_mask_0 = const()[name = tensor("cache_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_squeeze_mask_0 = const()[name = tensor("cache_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_cast_fp16 = slice_by_index(begin = cache_begin_0, end = cache_end_0, end_mask = cache_end_mask_0, squeeze_mask = cache_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_cast_fp16")]; + tensor input_859_axes_0 = const()[name = tensor("input_859_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200918208)))]; + tensor encoder_layers_16_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200919296)))]; + tensor input_859_cast_fp16 = layer_norm(axes = input_859_axes_0, beta = encoder_layers_16_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_norm_feed_forward1_weight_to_fp16, x = input_857_cast_fp16)[name = tensor("input_859_cast_fp16")]; + tensor encoder_layers_16_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_16_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200920384)))]; + tensor linear_145_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_16_feed_forward1_linear1_weight_to_fp16, x = input_859_cast_fp16)[name = tensor("linear_145_cast_fp16")]; + tensor input_863_cast_fp16 = silu(x = linear_145_cast_fp16)[name = tensor("input_863_cast_fp16")]; + tensor encoder_layers_16_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_16_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(203017600)))]; + tensor linear_146_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_feed_forward1_linear2_weight_to_fp16, x = input_863_cast_fp16)[name = tensor("linear_146_cast_fp16")]; + tensor var_3567_to_fp16 = const()[name = tensor("op_3567_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3568_cast_fp16 = mul(x = linear_146_cast_fp16, y = var_3567_to_fp16)[name = tensor("op_3568_cast_fp16")]; + tensor input_869_cast_fp16 = add(x = input_857_cast_fp16, y = var_3568_cast_fp16)[name = tensor("input_869_cast_fp16")]; + tensor key_axes_0 = const()[name = tensor("key_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205114816)))]; + tensor encoder_layers_16_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205115904)))]; + tensor key_cast_fp16 = layer_norm(axes = key_axes_0, beta = encoder_layers_16_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_norm_self_att_weight_to_fp16, x = input_869_cast_fp16)[name = tensor("key_cast_fp16")]; + tensor input_871_interleave_0 = const()[name = tensor("input_871_interleave_0"), val = tensor(false)]; + tensor input_871_cast_fp16 = concat(axis = var_64, interleave = input_871_interleave_0, values = (cache_65_cast_fp16, key_cast_fp16))[name = tensor("input_871_cast_fp16")]; + tensor var_3590_begin_0 = const()[name = tensor("op_3590_begin_0"), val = tensor([0, 17, 0])]; + tensor var_3590_end_0 = const()[name = tensor("op_3590_end_0"), val = tensor([1, 70, 512])]; + tensor var_3590_end_mask_0 = const()[name = tensor("op_3590_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3590_cast_fp16 = slice_by_index(begin = var_3590_begin_0, end = var_3590_end_0, end_mask = var_3590_end_mask_0, x = cache_65_cast_fp16)[name = tensor("op_3590_cast_fp16")]; + tensor cache_last_channel_cur_interleave_0 = const()[name = tensor("cache_last_channel_cur_interleave_0"), val = tensor(false)]; + tensor cache_last_channel_cur_cast_fp16 = concat(axis = var_64, interleave = cache_last_channel_cur_interleave_0, values = (var_3590_cast_fp16, key_cast_fp16))[name = tensor("cache_last_channel_cur_cast_fp16")]; + tensor encoder_layers_16_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205116992)))]; + tensor linear_147_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_self_attn_linear_q_weight_to_fp16, x = key_cast_fp16)[name = tensor("linear_147_cast_fp16")]; + tensor var_3600 = const()[name = tensor("op_3600"), val = tensor([1, -1, 8, 64])]; + tensor q_97_cast_fp16 = reshape(shape = var_3600, x = linear_147_cast_fp16)[name = tensor("q_97_cast_fp16")]; + tensor encoder_layers_16_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205641344)))]; + tensor linear_148_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_self_attn_linear_k_weight_to_fp16, x = input_871_cast_fp16)[name = tensor("linear_148_cast_fp16")]; + tensor var_3604 = const()[name = tensor("op_3604"), val = tensor([1, -1, 8, 64])]; + tensor k_65_cast_fp16 = reshape(shape = var_3604, x = linear_148_cast_fp16)[name = tensor("k_65_cast_fp16")]; + tensor encoder_layers_16_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206165696)))]; + tensor linear_149_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_self_attn_linear_v_weight_to_fp16, x = input_871_cast_fp16)[name = tensor("linear_149_cast_fp16")]; + tensor var_3608 = const()[name = tensor("op_3608"), val = tensor([1, -1, 8, 64])]; + tensor v_cast_fp16 = reshape(shape = var_3608, x = linear_149_cast_fp16)[name = tensor("v_cast_fp16")]; + tensor value_perm_0 = const()[name = tensor("value_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_16_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206690048)))]; + tensor var_3620_cast_fp16 = add(x = q_97_cast_fp16, y = encoder_layers_16_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3620_cast_fp16")]; + tensor encoder_layers_16_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206691136)))]; + tensor var_3622_cast_fp16 = add(x = q_97_cast_fp16, y = encoder_layers_16_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3622_cast_fp16")]; + tensor q_with_bias_v_perm_0 = const()[name = tensor("q_with_bias_v_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_423_transpose_x_0 = const()[name = tensor("x_423_transpose_x_0"), val = tensor(false)]; + tensor x_423_transpose_y_0 = const()[name = tensor("x_423_transpose_y_0"), val = tensor(false)]; + tensor var_3624_to_fp16 = const()[name = tensor("op_3624_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206692224)))]; + tensor q_with_bias_v_cast_fp16 = transpose(perm = q_with_bias_v_perm_0, x = var_3622_cast_fp16)[name = tensor("transpose_93")]; + tensor x_423_cast_fp16 = matmul(transpose_x = x_423_transpose_x_0, transpose_y = x_423_transpose_y_0, x = q_with_bias_v_cast_fp16, y = var_3624_to_fp16)[name = tensor("x_423_cast_fp16")]; + tensor x_425_pad_0 = const()[name = tensor("x_425_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_425_mode_0 = const()[name = tensor("x_425_mode_0"), val = tensor("constant")]; + tensor const_231_to_fp16 = const()[name = tensor("const_231_to_fp16"), val = tensor(0x0p+0)]; + tensor x_425_cast_fp16 = pad(constant_val = const_231_to_fp16, mode = x_425_mode_0, pad = x_425_pad_0, x = x_423_cast_fp16)[name = tensor("x_425_cast_fp16")]; + tensor var_3632 = const()[name = tensor("op_3632"), val = tensor([1, 8, -1, 17])]; + tensor x_427_cast_fp16 = reshape(shape = var_3632, x = x_425_cast_fp16)[name = tensor("x_427_cast_fp16")]; + tensor var_3636_begin_0 = const()[name = tensor("op_3636_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_3636_end_0 = const()[name = tensor("op_3636_end_0"), val = tensor([1, 8, 174, 17])]; + tensor var_3636_end_mask_0 = const()[name = tensor("op_3636_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_3636_cast_fp16 = slice_by_index(begin = var_3636_begin_0, end = var_3636_end_0, end_mask = var_3636_end_mask_0, x = x_427_cast_fp16)[name = tensor("op_3636_cast_fp16")]; + tensor var_3637 = const()[name = tensor("op_3637"), val = tensor([1, 8, 17, 173])]; + tensor matrix_bd_65_cast_fp16 = reshape(shape = var_3637, x = var_3636_cast_fp16)[name = tensor("matrix_bd_65_cast_fp16")]; + tensor matrix_ac_transpose_x_0 = const()[name = tensor("matrix_ac_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_transpose_y_0 = const()[name = tensor("matrix_ac_transpose_y_0"), val = tensor(false)]; + tensor transpose_83_perm_0 = const()[name = tensor("transpose_83_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_84_perm_0 = const()[name = tensor("transpose_84_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_84 = transpose(perm = transpose_84_perm_0, x = k_65_cast_fp16)[name = tensor("transpose_91")]; + tensor transpose_83 = transpose(perm = transpose_83_perm_0, x = var_3620_cast_fp16)[name = tensor("transpose_92")]; + tensor matrix_ac_cast_fp16 = matmul(transpose_x = matrix_ac_transpose_x_0, transpose_y = matrix_ac_transpose_y_0, x = transpose_83, y = transpose_84)[name = tensor("matrix_ac_cast_fp16")]; + tensor matrix_bd_begin_0 = const()[name = tensor("matrix_bd_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_end_0 = const()[name = tensor("matrix_bd_end_0"), val = tensor([1, 8, 17, 87])]; + tensor matrix_bd_end_mask_0 = const()[name = tensor("matrix_bd_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_cast_fp16 = slice_by_index(begin = matrix_bd_begin_0, end = matrix_bd_end_0, end_mask = matrix_bd_end_mask_0, x = matrix_bd_65_cast_fp16)[name = tensor("matrix_bd_cast_fp16")]; + tensor var_3646_cast_fp16 = add(x = matrix_ac_cast_fp16, y = matrix_bd_cast_fp16)[name = tensor("op_3646_cast_fp16")]; + tensor _inversed_scores_65_y_0_to_fp16 = const()[name = tensor("_inversed_scores_65_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_65_cast_fp16 = mul(x = var_3646_cast_fp16, y = _inversed_scores_65_y_0_to_fp16)[name = tensor("_inversed_scores_65_cast_fp16")]; + tensor scores_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_65_cast_fp16, cond = mask_3)[name = tensor("scores_cast_fp16")]; + tensor var_3652_cast_fp16 = softmax(axis = var_62, x = scores_cast_fp16)[name = tensor("op_3652_cast_fp16")]; + tensor input_873_cast_fp16 = select(a = var_40_to_fp16, b = var_3652_cast_fp16, cond = mask_3)[name = tensor("input_873_cast_fp16")]; + tensor x_429_transpose_x_0 = const()[name = tensor("x_429_transpose_x_0"), val = tensor(false)]; + tensor x_429_transpose_y_0 = const()[name = tensor("x_429_transpose_y_0"), val = tensor(false)]; + tensor value_cast_fp16 = transpose(perm = value_perm_0, x = v_cast_fp16)[name = tensor("transpose_94")]; + tensor x_429_cast_fp16 = matmul(transpose_x = x_429_transpose_x_0, transpose_y = x_429_transpose_y_0, x = input_873_cast_fp16, y = value_cast_fp16)[name = tensor("x_429_cast_fp16")]; + tensor var_3656_perm_0 = const()[name = tensor("op_3656_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_3657 = const()[name = tensor("op_3657"), val = tensor([1, -1, 512])]; + tensor var_3656_cast_fp16 = transpose(perm = var_3656_perm_0, x = x_429_cast_fp16)[name = tensor("transpose_90")]; + tensor input_875_cast_fp16 = reshape(shape = var_3657, x = var_3656_cast_fp16)[name = tensor("input_875_cast_fp16")]; + tensor encoder_layers_16_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206869440)))]; + tensor linear_151_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_self_attn_linear_out_weight_to_fp16, x = input_875_cast_fp16)[name = tensor("linear_151_cast_fp16")]; + tensor input_879_cast_fp16 = add(x = input_869_cast_fp16, y = linear_151_cast_fp16)[name = tensor("input_879_cast_fp16")]; + tensor x_433_axes_0 = const()[name = tensor("x_433_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207393792)))]; + tensor encoder_layers_16_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207394880)))]; + tensor x_433_cast_fp16 = layer_norm(axes = x_433_axes_0, beta = encoder_layers_16_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_norm_conv_weight_to_fp16, x = input_879_cast_fp16)[name = tensor("x_433_cast_fp16")]; + tensor input_881_perm_0 = const()[name = tensor("input_881_perm_0"), val = tensor([0, 2, 1])]; + tensor input_883_pad_type_0 = const()[name = tensor("input_883_pad_type_0"), val = tensor("valid")]; + tensor input_883_strides_0 = const()[name = tensor("input_883_strides_0"), val = tensor([1])]; + tensor input_883_pad_0 = const()[name = tensor("input_883_pad_0"), val = tensor([0, 0])]; + tensor input_883_dilations_0 = const()[name = tensor("input_883_dilations_0"), val = tensor([1])]; + tensor input_883_groups_0 = const()[name = tensor("input_883_groups_0"), val = tensor(1)]; + tensor encoder_layers_16_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_16_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207395968)))]; + tensor input_881_cast_fp16 = transpose(perm = input_881_perm_0, x = x_433_cast_fp16)[name = tensor("transpose_89")]; + tensor input_883_cast_fp16 = conv(dilations = input_883_dilations_0, groups = input_883_groups_0, pad = input_883_pad_0, pad_type = input_883_pad_type_0, strides = input_883_strides_0, weight = encoder_layers_16_conv_pointwise_conv1_weight_to_fp16, x = input_881_cast_fp16)[name = tensor("input_883_cast_fp16")]; + tensor x_435_split_num_splits_0 = const()[name = tensor("x_435_split_num_splits_0"), val = tensor(2)]; + tensor x_435_split_axis_0 = const()[name = tensor("x_435_split_axis_0"), val = tensor(1)]; + tensor x_435_split_cast_fp16_0, tensor x_435_split_cast_fp16_1 = split(axis = x_435_split_axis_0, num_splits = x_435_split_num_splits_0, x = input_883_cast_fp16)[name = tensor("x_435_split_cast_fp16")]; + tensor x_435_split_1_sigmoid_cast_fp16 = sigmoid(x = x_435_split_cast_fp16_1)[name = tensor("x_435_split_1_sigmoid_cast_fp16")]; + tensor x_435_cast_fp16 = mul(x = x_435_split_cast_fp16_0, y = x_435_split_1_sigmoid_cast_fp16)[name = tensor("x_435_cast_fp16")]; + tensor input_885_cast_fp16 = select(a = var_40_to_fp16, b = x_435_cast_fp16, cond = var_418)[name = tensor("input_885_cast_fp16")]; + tensor new_x_interleave_0 = const()[name = tensor("new_x_interleave_0"), val = tensor(false)]; + tensor new_x_cast_fp16 = concat(axis = var_62, interleave = new_x_interleave_0, values = (cache_cast_fp16, input_885_cast_fp16))[name = tensor("new_x_cast_fp16")]; + tensor cache_last_time_cur_begin_0 = const()[name = tensor("cache_last_time_cur_begin_0"), val = tensor([0, 0, 17])]; + tensor cache_last_time_cur_end_0 = const()[name = tensor("cache_last_time_cur_end_0"), val = tensor([1, 512, 25])]; + tensor cache_last_time_cur_end_mask_0 = const()[name = tensor("cache_last_time_cur_end_mask_0"), val = tensor([true, true, true])]; + tensor cache_last_time_cur_cast_fp16 = slice_by_index(begin = cache_last_time_cur_begin_0, end = cache_last_time_cur_end_0, end_mask = cache_last_time_cur_end_mask_0, x = new_x_cast_fp16)[name = tensor("cache_last_time_cur_cast_fp16")]; + tensor x_437_pad_type_0 = const()[name = tensor("x_437_pad_type_0"), val = tensor("valid")]; + tensor x_437_groups_0 = const()[name = tensor("x_437_groups_0"), val = tensor(512)]; + tensor x_437_strides_0 = const()[name = tensor("x_437_strides_0"), val = tensor([1])]; + tensor x_437_pad_0 = const()[name = tensor("x_437_pad_0"), val = tensor([0, 0])]; + tensor x_437_dilations_0 = const()[name = tensor("x_437_dilations_0"), val = tensor([1])]; + tensor encoder_layers_16_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_16_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208444608)))]; + tensor x_437_cast_fp16 = conv(dilations = x_437_dilations_0, groups = x_437_groups_0, pad = x_437_pad_0, pad_type = x_437_pad_type_0, strides = x_437_strides_0, weight = encoder_layers_16_conv_depthwise_conv_weight_to_fp16, x = new_x_cast_fp16)[name = tensor("x_437_cast_fp16")]; + tensor input_887_perm_0 = const()[name = tensor("input_887_perm_0"), val = tensor([0, 2, 1])]; + tensor x_439_axes_0 = const()[name = tensor("x_439_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_16_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208453888)))]; + tensor encoder_layers_16_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_16_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208454976)))]; + tensor input_887_cast_fp16 = transpose(perm = input_887_perm_0, x = x_437_cast_fp16)[name = tensor("transpose_88")]; + tensor x_439_cast_fp16 = layer_norm(axes = x_439_axes_0, beta = encoder_layers_16_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_conv_batch_norm_weight_to_fp16, x = input_887_cast_fp16)[name = tensor("x_439_cast_fp16")]; + tensor input_889_perm_0 = const()[name = tensor("input_889_perm_0"), val = tensor([0, 2, 1])]; + tensor input_889_cast_fp16 = transpose(perm = input_889_perm_0, x = x_439_cast_fp16)[name = tensor("transpose_87")]; + tensor input_891_cast_fp16 = silu(x = input_889_cast_fp16)[name = tensor("input_891_cast_fp16")]; + tensor x_441_pad_type_0 = const()[name = tensor("x_441_pad_type_0"), val = tensor("valid")]; + tensor x_441_strides_0 = const()[name = tensor("x_441_strides_0"), val = tensor([1])]; + tensor x_441_pad_0 = const()[name = tensor("x_441_pad_0"), val = tensor([0, 0])]; + tensor x_441_dilations_0 = const()[name = tensor("x_441_dilations_0"), val = tensor([1])]; + tensor x_441_groups_0 = const()[name = tensor("x_441_groups_0"), val = tensor(1)]; + tensor encoder_layers_16_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_16_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208456064)))]; + tensor x_441_cast_fp16 = conv(dilations = x_441_dilations_0, groups = x_441_groups_0, pad = x_441_pad_0, pad_type = x_441_pad_type_0, strides = x_441_strides_0, weight = encoder_layers_16_conv_pointwise_conv2_weight_to_fp16, x = input_891_cast_fp16)[name = tensor("x_441_cast_fp16")]; + tensor input_893_perm_0 = const()[name = tensor("input_893_perm_0"), val = tensor([0, 2, 1])]; + tensor input_893_cast_fp16 = transpose(perm = input_893_perm_0, x = x_441_cast_fp16)[name = tensor("transpose_86")]; + tensor input_895_cast_fp16 = add(x = input_879_cast_fp16, y = input_893_cast_fp16)[name = tensor("input_895_cast_fp16")]; + tensor input_897_axes_0 = const()[name = tensor("input_897_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208980416)))]; + tensor encoder_layers_16_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208981504)))]; + tensor input_897_cast_fp16 = layer_norm(axes = input_897_axes_0, beta = encoder_layers_16_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_norm_feed_forward2_weight_to_fp16, x = input_895_cast_fp16)[name = tensor("input_897_cast_fp16")]; + tensor encoder_layers_16_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_16_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208982592)))]; + tensor linear_152_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_16_feed_forward2_linear1_weight_to_fp16, x = input_897_cast_fp16)[name = tensor("linear_152_cast_fp16")]; + tensor input_901_cast_fp16 = silu(x = linear_152_cast_fp16)[name = tensor("input_901_cast_fp16")]; + tensor encoder_layers_16_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_16_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211079808)))]; + tensor linear_153_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_feed_forward2_linear2_weight_to_fp16, x = input_901_cast_fp16)[name = tensor("linear_153_cast_fp16")]; + tensor var_3736_to_fp16 = const()[name = tensor("op_3736_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3737_cast_fp16 = mul(x = linear_153_cast_fp16, y = var_3736_to_fp16)[name = tensor("op_3737_cast_fp16")]; + tensor input_cast_fp16 = add(x = input_895_cast_fp16, y = var_3737_cast_fp16)[name = tensor("input_cast_fp16")]; + tensor audio_signal_axes_0 = const()[name = tensor("audio_signal_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213177024)))]; + tensor encoder_layers_16_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213178112)))]; + tensor audio_signal_cast_fp16 = layer_norm(axes = audio_signal_axes_0, beta = encoder_layers_16_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_norm_out_weight_to_fp16, x = input_cast_fp16)[name = tensor("audio_signal_cast_fp16")]; + tensor obj_1_perm_0 = const()[name = tensor("obj_1_perm_0"), val = tensor([0, 2, 1])]; + tensor obj_1_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_1_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor obj_5_axis_0 = const()[name = tensor("obj_5_axis_0"), val = tensor(0)]; + tensor obj_5_cast_fp16 = stack(axis = obj_5_axis_0, values = (var_332_cast_fp16, var_536_cast_fp16, var_740_cast_fp16, var_944_cast_fp16, var_1148_cast_fp16, var_1352_cast_fp16, var_1556_cast_fp16, var_1760_cast_fp16, var_1964_cast_fp16, var_2168_cast_fp16, var_2372_cast_fp16, var_2576_cast_fp16, var_2780_cast_fp16, var_2984_cast_fp16, var_3188_cast_fp16, var_3392_cast_fp16, cache_last_channel_cur_cast_fp16))[name = tensor("obj_5_cast_fp16")]; + tensor obj_7_axis_0 = const()[name = tensor("obj_7_axis_0"), val = tensor(0)]; + tensor obj_7_cast_fp16 = stack(axis = obj_7_axis_0, values = (var_431_cast_fp16, var_635_cast_fp16, var_839_cast_fp16, var_1043_cast_fp16, var_1247_cast_fp16, var_1451_cast_fp16, var_1655_cast_fp16, var_1859_cast_fp16, var_2063_cast_fp16, var_2267_cast_fp16, var_2471_cast_fp16, var_2675_cast_fp16, var_2879_cast_fp16, var_3083_cast_fp16, var_3287_cast_fp16, var_3491_cast_fp16, cache_last_time_cur_cast_fp16))[name = tensor("obj_7_cast_fp16")]; + tensor obj_7_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_7_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor var_3753 = add(x = cache_last_channel_len, y = max_audio_length_1)[name = tensor("op_3753")]; + tensor var_3753_promoted_to_fp16_dtype_0 = const()[name = tensor("op_3753_promoted_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor const_237_to_fp16 = const()[name = tensor("const_237_to_fp16"), val = tensor(-inf)]; + tensor var_45_promoted_to_fp16 = const()[name = tensor("op_45_promoted_to_fp16"), val = tensor(0x1.18p+6)]; + tensor var_3753_to_fp16 = cast(dtype = var_3753_promoted_to_fp16_dtype_0, x = var_3753)[name = tensor("cast_184")]; + tensor clip_1_cast_fp16 = clip(alpha = const_237_to_fp16, beta = var_45_promoted_to_fp16, x = var_3753_to_fp16)[name = tensor("clip_1_cast_fp16")]; + tensor var_3780_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_3780_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor cast_179_dtype_0 = const()[name = tensor("cast_179_dtype_0"), val = tensor("int32")]; + tensor cast_180_dtype_0 = const()[name = tensor("cast_180_dtype_0"), val = tensor("int32")]; + tensor new_cache_last_channel_len = cast(dtype = cast_180_dtype_0, x = clip_1_cast_fp16)[name = tensor("cast_181")]; + tensor encoded_length = cast(dtype = cast_179_dtype_0, x = clip_0_cast_fp16)[name = tensor("cast_182")]; + tensor new_cache_last_channel = cast(dtype = var_3780_cast_fp16_to_fp32_dtype_0, x = obj_5_cast_fp16)[name = tensor("cast_183")]; + tensor new_cache_last_time = cast(dtype = obj_7_cast_fp16_to_fp32_dtype_0, x = obj_7_cast_fp16)[name = tensor("cast_185")]; + tensor obj_1_cast_fp16 = transpose(perm = obj_1_perm_0, x = audio_signal_cast_fp16)[name = tensor("transpose_85")]; + tensor encoded_output = cast(dtype = obj_1_cast_fp16_to_fp32_dtype_0, x = obj_1_cast_fp16)[name = tensor("cast_186")]; + tensor new_pre_cache = cast(dtype = var_28_cast_fp16_to_fp32_dtype_0, x = var_28_cast_fp16)[name = tensor("cast_192")]; + } -> (encoded_output, encoded_length, new_pre_cache, new_cache_last_channel, new_cache_last_time, new_cache_last_channel_len); +} \ No newline at end of file diff --git a/1280ms/streaming_encoder.mlmodelc/weights/weight.bin b/1280ms/streaming_encoder.mlmodelc/weights/weight.bin new file mode 100644 index 0000000000000000000000000000000000000000..6fc415d6e82641770c5d78f17d76d1e71d028e06 --- /dev/null +++ b/1280ms/streaming_encoder.mlmodelc/weights/weight.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c71acb590ceb2af449de5c7e3516e76057eaf4589d1f16edba774831db74b17 +size 213179200 diff --git a/1280ms/streaming_encoder.mlpackage/Data/com.apple.CoreML/model.mlmodel b/1280ms/streaming_encoder.mlpackage/Data/com.apple.CoreML/model.mlmodel new file mode 100644 index 0000000000000000000000000000000000000000..7e0bb55b1d00510e94e0c0473b9231362e92c892 --- /dev/null +++ b/1280ms/streaming_encoder.mlpackage/Data/com.apple.CoreML/model.mlmodel @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9181d223091aa592cb656d49346e640a38ec2426de5ec2d06edbc14e92b8968b +size 508252 diff --git a/1280ms/streaming_encoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/1280ms/streaming_encoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin new file mode 100644 index 0000000000000000000000000000000000000000..6fc415d6e82641770c5d78f17d76d1e71d028e06 --- /dev/null +++ b/1280ms/streaming_encoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c71acb590ceb2af449de5c7e3516e76057eaf4589d1f16edba774831db74b17 +size 213179200 diff --git a/1280ms/streaming_encoder.mlpackage/Manifest.json b/1280ms/streaming_encoder.mlpackage/Manifest.json new file mode 100644 index 0000000000000000000000000000000000000000..e1f379019d051455a60312bf6f5664348c61be66 --- /dev/null +++ b/1280ms/streaming_encoder.mlpackage/Manifest.json @@ -0,0 +1,18 @@ +{ + "fileFormatVersion": "1.0.0", + "itemInfoEntries": { + "468B5E19-6BA9-478C-8D2A-23953ACBD5E3": { + "author": "com.apple.CoreML", + "description": "CoreML Model Specification", + "name": "model.mlmodel", + "path": "com.apple.CoreML/model.mlmodel" + }, + "F22DE286-FE1A-4BBD-A7ED-B0130595DAF3": { + "author": "com.apple.CoreML", + "description": "CoreML Model Weights", + "name": "weights", + "path": "com.apple.CoreML/weights" + } + }, + "rootModelIdentifier": "468B5E19-6BA9-478C-8D2A-23953ACBD5E3" +} diff --git a/1280ms/vocab.json b/1280ms/vocab.json new file mode 100644 index 0000000000000000000000000000000000000000..462081190dbb331a0dde13e99d5cfa99b302f0a7 --- /dev/null +++ b/1280ms/vocab.json @@ -0,0 +1,1028 @@ +{ + "0": "", + "1": "▁t", + "2": "▁th", + "3": "▁a", + "4": "▁i", + "5": "▁the", + "6": "▁s", + "7": "re", + "8": "▁w", + "9": "▁o", + "10": "in", + "11": "at", + "12": "er", + "13": "nd", + "14": "ou", + "15": "▁c", + "16": "▁b", + "17": "▁h", + "18": "en", + "19": "on", + "20": "▁m", + "21": "▁f", + "22": "ing", + "23": "▁p", + "24": "▁to", + "25": "▁and", + "26": "▁d", + "27": "an", + "28": "or", + "29": "es", + "30": "▁y", + "31": "▁l", + "32": "▁of", + "33": "ll", + "34": "▁in", + "35": "ed", + "36": "it", + "37": "▁g", + "38": "is", + "39": "▁you", + "40": "▁n", + "41": "ar", + "42": "om", + "43": "as", + "44": "ve", + "45": "▁e", + "46": "ic", + "47": "▁it", + "48": "al", + "49": "us", + "50": "▁wh", + "51": "▁we", + "52": "▁be", + "53": "ion", + "54": "ow", + "55": "le", + "56": "▁is", + "57": "et", + "58": "ent", + "59": "ot", + "60": "ut", + "61": "▁re", + "62": "▁on", + "63": "ay", + "64": "▁ha", + "65": "ig", + "66": "▁so", + "67": "ct", + "68": "▁he", + "69": "▁for", + "70": "ver", + "71": "ke", + "72": "ro", + "73": "▁st", + "74": "id", + "75": "▁go", + "76": "all", + "77": "se", + "78": "ly", + "79": "▁u", + "80": "ch", + "81": "st", + "82": "ld", + "83": "▁k", + "84": "ce", + "85": "ur", + "86": "▁li", + "87": "am", + "88": "▁r", + "89": "ht", + "90": "▁j", + "91": "ith", + "92": "▁se", + "93": "ir", + "94": "▁as", + "95": "▁an", + "96": "im", + "97": "▁do", + "98": "ad", + "99": "▁was", + "100": "ight", + "101": "th", + "102": "▁are", + "103": "▁but", + "104": "▁sh", + "105": "ust", + "106": "ally", + "107": "▁not", + "108": "▁or", + "109": "▁com", + "110": "▁can", + "111": "▁me", + "112": "op", + "113": "▁mo", + "114": "▁at", + "115": "ill", + "116": "▁ch", + "117": "▁ne", + "118": "ant", + "119": "▁de", + "120": "▁kn", + "121": "▁one", + "122": "il", + "123": "ol", + "124": "▁con", + "125": "ter", + "126": "▁ab", + "127": "▁fr", + "128": "ere", + "129": "ck", + "130": "▁al", + "131": "▁all", + "132": "qu", + "133": "▁pro", + "134": "▁som", + "135": "ould", + "136": "▁tw", + "137": "ul", + "138": "ra", + "139": "od", + "140": "ers", + "141": "▁su", + "142": "ive", + "143": "▁v", + "144": "use", + "145": "ate", + "146": "ge", + "147": "if", + "148": "▁ex", + "149": "ess", + "150": "pp", + "151": "▁lo", + "152": "out", + "153": "▁if", + "154": "est", + "155": "ain", + "156": "ist", + "157": "and", + "158": "ea", + "159": "very", + "160": "art", + "161": "▁wor", + "162": "▁my", + "163": "ab", + "164": "ment", + "165": "▁bec", + "166": "un", + "167": "ity", + "168": "ri", + "169": "pe", + "170": "ions", + "171": "▁by", + "172": "ok", + "173": "our", + "174": "ort", + "175": "ind", + "176": "ink", + "177": "nt", + "178": "▁up", + "179": "um", + "180": "▁don", + "181": "▁get", + "182": "red", + "183": "▁out", + "184": "el", + "185": "ause", + "186": "res", + "187": "▁ma", + "188": "ich", + "189": "▁us", + "190": "rou", + "191": "▁int", + "192": "em", + "193": "os", + "194": "ies", + "195": "ie", + "196": "▁pl", + "197": "▁tr", + "198": "ven", + "199": "ous", + "200": "▁le", + "201": "▁two", + "202": "ard", + "203": "ine", + "204": "▁co", + "205": "een", + "206": "▁now", + "207": "ty", + "208": "her", + "209": "ack", + "210": "▁pe", + "211": "ame", + "212": "▁how", + "213": "▁who", + "214": "▁see", + "215": "▁tim", + "216": "ect", + "217": "ast", + "218": "▁our", + "219": "ci", + "220": "ree", + "221": "ople", + "222": "gh", + "223": "▁no", + "224": "▁had", + "225": "▁man", + "226": "▁qu", + "227": "▁en", + "228": "ide", + "229": "ure", + "230": "ud", + "231": "so", + "232": "▁his", + "233": "▁sa", + "234": "▁sp", + "235": "▁say", + "236": "ose", + "237": "ther", + "238": "▁act", + "239": "▁ta", + "240": "▁cl", + "241": "ings", + "242": "pt", + "243": "king", + "244": "▁any", + "245": "▁has", + "246": "▁un", + "247": "iv", + "248": "▁im", + "249": "▁ag", + "250": "▁te", + "251": "▁fe", + "252": "one", + "253": "per", + "254": "ong", + "255": "▁po", + "256": "▁ad", + "257": "ff", + "258": "ore", + "259": "itt", + "260": "ans", + "261": "iz", + "262": "eah", + "263": "reat", + "264": "act", + "265": "own", + "266": "hing", + "267": "enty", + "268": "age", + "269": "ber", + "270": "ice", + "271": "▁am", + "272": "ple", + "273": "are", + "274": "▁per", + "275": "und", + "276": "ite", + "277": "ix", + "278": "pl", + "279": "▁way", + "280": "▁did", + "281": "▁pr", + "282": "▁got", + "283": "ars", + "284": "▁she", + "285": "▁let", + "286": "ag", + "287": "▁ac", + "288": "int", + "289": "▁ar", + "290": "ry", + "291": "ign", + "292": "ish", + "293": "▁fir", + "294": "ace", + "295": "ble", + "296": "og", + "297": "ue", + "298": "▁ye", + "299": "ap", + "300": "iff", + "301": "▁ro", + "302": "▁her", + "303": "nder", + "304": "▁ok", + "305": "▁res", + "306": "▁gu", + "307": "ence", + "308": "▁may", + "309": "ated", + "310": "ip", + "311": "▁bo", + "312": "▁him", + "313": "way", + "314": "ac", + "315": "ical", + "316": "ass", + "317": "ase", + "318": "▁dis", + "319": "able", + "320": "ick", + "321": "▁app", + "322": "ance", + "323": "▁pre", + "324": "▁six", + "325": "▁off", + "326": "▁new", + "327": "ia", + "328": "orm", + "329": "ank", + "330": "▁lot", + "331": "ach", + "332": "▁fo", + "333": "inet", + "334": "ire", + "335": "ary", + "336": "ult", + "337": "▁tal", + "338": "▁mu", + "339": "▁bl", + "340": "ount", + "341": "sel", + "342": "vel", + "343": "▁br", + "344": "▁imp", + "345": "ep", + "346": "cess", + "347": "ord", + "348": "▁sc", + "349": "▁inc", + "350": "ound", + "351": "ang", + "352": "be", + "353": "ress", + "354": "uct", + "355": "▁ind", + "356": "▁af", + "357": "ving", + "358": "▁oh", + "359": "▁bet", + "360": "▁use", + "361": "ome", + "362": "ens", + "363": "ys", + "364": "▁bu", + "365": "co", + "366": "ory", + "367": "ater", + "368": "ild", + "369": "ght", + "370": "ial", + "371": "▁day", + "372": "ning", + "373": "na", + "374": "ile", + "375": "▁spe", + "376": "▁mar", + "377": "ody", + "378": "ough", + "379": "ade", + "380": "vers", + "381": "xt", + "382": "▁fl", + "383": "▁ke", + "384": "ian", + "385": "▁sy", + "386": "▁put", + "387": "fore", + "388": "ub", + "389": "▁ph", + "390": "fe", + "391": "▁em", + "392": "▁ser", + "393": "form", + "394": "ting", + "395": "te", + "396": "av", + "397": "ious", + "398": "▁rec", + "399": "ks", + "400": "▁gr", + "401": "ces", + "402": "wn", + "403": "ors", + "404": "▁jo", + "405": "ents", + "406": "▁des", + "407": "▁try", + "408": "▁equ", + "409": "▁z", + "410": "▁rem", + "411": "▁str", + "412": "self", + "413": "▁bit", + "414": "ph", + "415": "ved", + "416": "▁why", + "417": "▁bas", + "418": "▁hel", + "419": "▁rel", + "420": "ath", + "421": "ject", + "422": "ail", + "423": "▁la", + "424": "ual", + "425": "▁god", + "426": "▁nat", + "427": "erm", + "428": "day", + "429": "▁id", + "430": "ft", + "431": "▁wr", + "432": "▁min", + "433": "ates", + "434": "▁gen", + "435": "tain", + "436": "▁ob", + "437": "ull", + "438": "ict", + "439": "▁tra", + "440": "▁end", + "441": "▁hig", + "442": "▁fif", + "443": "oth", + "444": "tern", + "445": "▁its", + "446": "vent", + "447": "▁sm", + "448": "ons", + "449": "▁add", + "450": "iss", + "451": "▁bel", + "452": "ful", + "453": "get", + "454": "▁ele", + "455": "▁rep", + "456": "ak", + "457": "▁ho", + "458": "▁pos", + "459": "▁num", + "460": "ange", + "461": "ves", + "462": "ific", + "463": "urn", + "464": "ise", + "465": "▁cr", + "466": "▁um", + "467": "ward", + "468": "▁reg", + "469": "ady", + "470": "ower", + "471": "uc", + "472": "▁dec", + "473": "lic", + "474": "▁set", + "475": "▁gon", + "476": "▁op", + "477": "▁ear", + "478": "▁sub", + "479": "▁sl", + "480": "les", + "481": "stem", + "482": "cial", + "483": "olog", + "484": "atch", + "485": "ily", + "486": "body", + "487": "nds", + "488": "ular", + "489": "ren", + "490": "▁own", + "491": "▁too", + "492": "cent", + "493": "ible", + "494": "pect", + "495": "ered", + "496": "ways", + "497": "teen", + "498": "▁uh", + "499": "▁big", + "500": "▁mod", + "501": "▁att", + "502": "▁car", + "503": "gr", + "504": "▁acc", + "505": "ied", + "506": "mun", + "507": "ib", + "508": "▁mon", + "509": "▁sch", + "510": "▁pol", + "511": "▁dat", + "512": "▁fin", + "513": "▁sim", + "514": "▁inv", + "515": "▁def", + "516": "ked", + "517": "▁ent", + "518": "▁yes", + "519": "ows", + "520": "ics", + "521": "ited", + "522": "ute", + "523": "ism", + "524": "ps", + "525": "▁ed", + "526": "▁el", + "527": "ably", + "528": "ppen", + "529": "als", + "530": "▁ten", + "531": "ract", + "532": "ss", + "533": "▁ass", + "534": "▁met", + "535": "gan", + "536": "▁eng", + "537": "▁stu", + "538": "ween", + "539": "arch", + "540": "▁gl", + "541": "▁cor", + "542": "▁dr", + "543": "vern", + "544": "▁ty", + "545": "▁run", + "546": "hip", + "547": "cus", + "548": "cond", + "549": "▁ins", + "550": "irty", + "551": "▁pub", + "552": "lud", + "553": "llow", + "554": "▁cou", + "555": "ew", + "556": "iew", + "557": "▁sur", + "558": "ero", + "559": "ood", + "560": "ness", + "561": "▁fun", + "562": "▁eff", + "563": "cept", + "564": "▁ca", + "565": "▁exp", + "566": "duct", + "567": "▁sw", + "568": "ize", + "569": "ope", + "570": "▁par", + "571": "kes", + "572": "cy", + "573": "▁ev", + "574": "▁ref", + "575": "ell", + "576": "▁bus", + "577": "ug", + "578": "rib", + "579": "▁cur", + "580": "mo", + "581": "ock", + "582": "ures", + "583": "air", + "584": "▁war", + "585": "str", + "586": "▁med", + "587": "▁wa", + "588": "▁val", + "589": "▁sin", + "590": "blem", + "591": "▁fam", + "592": "li", + "593": "▁far", + "594": "▁cle", + "595": "▁col", + "596": "mon", + "597": "▁gra", + "598": "led", + "599": "ense", + "600": "tin", + "601": "ues", + "602": "its", + "603": "▁mem", + "604": "▁inf", + "605": "▁eas", + "606": "ideo", + "607": "▁top", + "608": "io", + "609": "pan", + "610": "▁hum", + "611": "▁old", + "612": "ead", + "613": "▁ord", + "614": "ric", + "615": "ants", + "616": "oy", + "617": "esn", + "618": "uck", + "619": "ason", + "620": "ced", + "621": "ool", + "622": "rat", + "623": "ouse", + "624": "▁lar", + "625": "▁art", + "626": "▁wee", + "627": "▁cer", + "628": "ized", + "629": "▁mat", + "630": "con", + "631": "erg", + "632": "land", + "633": "ines", + "634": "▁chr", + "635": "▁aut", + "636": "▁lea", + "637": "▁sou", + "638": "oney", + "639": "tty", + "640": "▁ple", + "641": "ulat", + "642": "oks", + "643": "▁few", + "644": "▁sol", + "645": "▁che", + "646": "chn", + "647": "ird", + "648": "▁bre", + "649": "▁dur", + "650": "▁wom", + "651": "me", + "652": "izat", + "653": "eric", + "654": "ote", + "655": "▁uni", + "656": "eren", + "657": "arn", + "658": "ross", + "659": "ices", + "660": "ten", + "661": "eral", + "662": "ever", + "663": "ieve", + "664": "lish", + "665": "ash", + "666": "▁opp", + "667": "alth", + "668": "ger", + "669": "▁sk", + "670": "▁red", + "671": "peri", + "672": "▁det", + "673": "▁ext", + "674": "ner", + "675": "ah", + "676": "▁var", + "677": "▁loc", + "678": "gram", + "679": "ists", + "680": "ives", + "681": "▁es", + "682": "▁nor", + "683": "tro", + "684": "ale", + "685": "▁iss", + "686": "▁pri", + "687": "gin", + "688": "az", + "689": "oc", + "690": "▁pop", + "691": "ern", + "692": "▁sit", + "693": "ket", + "694": "▁pa", + "695": "▁law", + "696": "ages", + "697": "br", + "698": "▁cam", + "699": "▁mom", + "700": "osed", + "701": "▁bro", + "702": "ne", + "703": "bs", + "704": "▁cre", + "705": "erat", + "706": "▁sec", + "707": "▁cap", + "708": "▁vis", + "709": "▁pat", + "710": "ield", + "711": "iet", + "712": "▁tri", + "713": "up", + "714": "▁bra", + "715": "ts", + "716": "▁mot", + "717": "▁unt", + "718": "put", + "719": "bo", + "720": "ork", + "721": "mer", + "722": "ital", + "723": "▁air", + "724": "ined", + "725": "▁beh", + "726": "▁adv", + "727": "▁ret", + "728": "imes", + "729": "▁tea", + "730": "ural", + "731": "sid", + "732": "ters", + "733": "▁pur", + "734": "▁sci", + "735": "bers", + "736": "ient", + "737": "ier", + "738": "cc", + "739": "sw", + "740": "▁av", + "741": "reen", + "742": "ode", + "743": "ont", + "744": "▁dra", + "745": "ann", + "746": "nect", + "747": "▁x", + "748": "▁eu", + "749": "ton", + "750": "inat", + "751": "ene", + "752": "ared", + "753": "els", + "754": "▁mor", + "755": "▁rat", + "756": "cri", + "757": "▁men", + "758": "▁ah", + "759": "ames", + "760": "▁arm", + "761": "eak", + "762": "▁pay", + "763": "▁hal", + "764": "ins", + "765": "ilit", + "766": "stit", + "767": "▁ra", + "768": "▁leg", + "769": "cl", + "770": "pr", + "771": "▁wal", + "772": "▁bad", + "773": "▁ge", + "774": "roup", + "775": "▁mus", + "776": "man", + "777": "▁gi", + "778": "eds", + "779": "▁aw", + "780": "po", + "781": "ark", + "782": "row", + "783": "▁dep", + "784": "ully", + "785": "ral", + "786": "lect", + "787": "pend", + "788": "▁sev", + "789": "ime", + "790": "gest", + "791": "here", + "792": "▁yet", + "793": "ted", + "794": "▁rev", + "795": "ds", + "796": "▁ask", + "797": "less", + "798": "▁di", + "799": "ets", + "800": "line", + "801": "▁aff", + "802": "ired", + "803": "▁est", + "804": "ken", + "805": "vid", + "806": "most", + "807": "ivid", + "808": "unch", + "809": "par", + "810": "med", + "811": "rop", + "812": "ased", + "813": "eone", + "814": "▁ve", + "815": "▁abs", + "816": "ergy", + "817": "ret", + "818": "▁saw", + "819": "▁ey", + "820": "▁cal", + "821": "uat", + "822": "▁mid", + "823": "vat", + "824": "ream", + "825": "vice", + "826": "ians", + "827": "rent", + "828": "ctor", + "829": "err", + "830": "ush", + "831": "ases", + "832": "▁suc", + "833": "erms", + "834": "ave", + "835": "angu", + "836": "ries", + "837": "▁wo", + "838": "arts", + "839": "▁fil", + "840": "▁fat", + "841": "▁cho", + "842": "orts", + "843": "▁fre", + "844": "ee", + "845": "ught", + "846": "eng", + "847": "ump", + "848": "▁bar", + "849": "ying", + "850": "ane", + "851": "▁tem", + "852": "anks", + "853": "ury", + "854": "iat", + "855": "mit", + "856": "trol", + "857": "▁net", + "858": "▁maj", + "859": "▁cra", + "860": "ling", + "861": "▁fig", + "862": "orn", + "863": "icat", + "864": "pany", + "865": "▁occ", + "866": "ott", + "867": "ands", + "868": "▁exc", + "869": "▁mr", + "870": "ency", + "871": "rope", + "872": "itch", + "873": "▁lit", + "874": "abil", + "875": "not", + "876": "ma", + "877": "▁typ", + "878": "▁opt", + "879": "ob", + "880": "ser", + "881": "ety", + "882": "ms", + "883": "peci", + "884": "aces", + "885": "aut", + "886": "▁hon", + "887": "cuss", + "888": "▁sal", + "889": "▁sor", + "890": "att", + "891": "▁lab", + "892": "▁har", + "893": "urch", + "894": "nded", + "895": "uce", + "896": "ids", + "897": "▁hy", + "898": "▁fut", + "899": "▁ste", + "900": "ours", + "901": "ems", + "902": "utes", + "903": "ng", + "904": "ta", + "905": "▁won", + "906": "▁fa", + "907": "▁env", + "908": "ards", + "909": "▁job", + "910": "ium", + "911": "▁dot", + "912": "▁obv", + "913": "ina", + "914": "side", + "915": "elve", + "916": "cu", + "917": "▁jes", + "918": "▁pot", + "919": "▁pie", + "920": "▁tre", + "921": "▁hey", + "922": "▁mag", + "923": "ron", + "924": "▁key", + "925": "swer", + "926": "▁win", + "927": "ucat", + "928": "work", + "929": "ides", + "930": "▁low", + "931": "▁vol", + "932": "▁oth", + "933": "atic", + "934": "lf", + "935": "ads", + "936": "inds", + "937": "com", + "938": "ths", + "939": "▁ver", + "940": "ised", + "941": "lo", + "942": "▁squ", + "943": "▁cut", + "944": "oked", + "945": "irit", + "946": "ateg", + "947": "ppy", + "948": "mitt", + "949": "come", + "950": "hn", + "951": "igin", + "952": "mand", + "953": "▁dam", + "954": "ho", + "955": "▁da", + "956": "▁fur", + "957": "iron", + "958": "ilar", + "959": "▁fac", + "960": "▁neg", + "961": "▁ago", + "962": "ged", + "963": "miss", + "964": "enth", + "965": "▁dou", + "966": "▁hit", + "967": "▁guy", + "968": "▁bi", + "969": "ove", + "970": "fess", + "971": "ples", + "972": "owed", + "973": "ured", + "974": "▁ris", + "975": "ints", + "976": "rew", + "977": "▁sum", + "978": "▁hu", + "979": "ploy", + "980": "ude", + "981": "ried", + "982": "▁cir", + "983": "▁dev", + "984": "ear", + "985": "▁tot", + "986": "▁ann", + "987": "duc", + "988": "ik", + "989": "pon", + "990": "sted", + "991": "▁ide", + "992": "▁'", + "993": "ipp", + "994": "▁eat", + "995": "▁dom", + "996": "▁", + "997": "e", + "998": "t", + "999": "o", + "1000": "a", + "1001": "i", + "1002": "n", + "1003": "s", + "1004": "r", + "1005": "h", + "1006": "l", + "1007": "d", + "1008": "u", + "1009": "c", + "1010": "m", + "1011": "y", + "1012": "g", + "1013": "w", + "1014": "f", + "1015": "p", + "1016": "b", + "1017": "v", + "1018": "k", + "1019": "'", + "1020": "j", + "1021": "x", + "1022": "q", + "1023": "z", + "1024": "", + "1025": "" +} \ No newline at end of file diff --git a/160ms/.DS_Store b/160ms/.DS_Store new file mode 100644 index 0000000000000000000000000000000000000000..142c5d02ad1e55affd5c417a2d8abf266ebbfc38 Binary files /dev/null and b/160ms/.DS_Store differ diff --git a/160ms/convert_parakeet_eou.py b/160ms/convert_parakeet_eou.py new file mode 100644 index 0000000000000000000000000000000000000000..9c865f90c8132840b5755dcd62fcfac15e3f0352 --- /dev/null +++ b/160ms/convert_parakeet_eou.py @@ -0,0 +1,740 @@ +#!/usr/bin/env python3 +"""CLI for exporting Parakeet Realtime EOU 120M components to CoreML. + +This model is a cache-aware streaming FastConformer-RNNT model optimized for +low-latency speech recognition with end-of-utterance detection. + +Key differences from Parakeet TDT v3: +- Smaller model (120M vs 600M params) +- No duration outputs (standard RNNT, not TDT) +- Cache-aware streaming encoder (17 layers, attention context [70,1]) +- Special token for end-of-utterance detection +- Optimized for 80-160ms latency + +Reference: https://huggingface.co/nvidia/parakeet_realtime_eou_120m-v1 +""" +from __future__ import annotations + +import json +from dataclasses import asdict +from pathlib import Path +from typing import Dict, Optional, Tuple + +import coremltools as ct +import numpy as np +import soundfile as sf +import torch +import typer + +import nemo.collections.asr as nemo_asr + +from individual_components import ( + DecoderWrapper, + EncoderWrapper, + ExportSettings, + JointWrapper, + JointDecisionWrapper, + JointDecisionSingleStep, + PreprocessorWrapper, + MelEncoderWrapper, + _coreml_convert, +) + +def apply_stft_patch(): + # Monkey patch coremltools.stft to handle extra arguments from newer torch versions + try: + import coremltools.converters.mil.frontend.torch.ops as torch_ops + _original_stft = torch_ops.stft + + def patched_stft(context, node): + if len(node.inputs) > 8: + node.inputs = node.inputs[:8] + return _original_stft(context, node) + + torch_ops.stft = patched_stft + if "stft" in torch_ops._TORCH_OPS_REGISTRY: + torch_ops._TORCH_OPS_REGISTRY["stft"] = patched_stft + print("Monkey patched coremltools.stft for compatibility.") + except Exception as e: + print(f"Warning: Could not monkey patch stft: {e}") + +DEFAULT_MODEL_ID = "nvidia/parakeet_realtime_eou_120m-v1" +AUTHOR = "Fluid Inference" + + +def _compute_length(seconds: float, sample_rate: int) -> int: + return int(round(seconds * sample_rate)) + + +def _prepare_audio( + validation_audio: Optional[Path], + sample_rate: int, + max_samples: int, + seed: Optional[int], +) -> torch.Tensor: + if validation_audio is None: + if seed is not None: + torch.manual_seed(seed) + audio = torch.randn(1, max_samples, dtype=torch.float32) + return audio + + data, sr = sf.read(str(validation_audio), dtype="float32") + if sr != sample_rate: + raise typer.BadParameter( + f"Validation audio sample rate {sr} does not match model rate {sample_rate}" + ) + + if data.ndim > 1: + data = data[:, 0] + + if data.size == 0: + raise typer.BadParameter("Validation audio is empty") + + if data.size < max_samples: + pad_width = max_samples - data.size + data = np.pad(data, (0, pad_width)) + elif data.size > max_samples: + data = data[:max_samples] + + audio = torch.from_numpy(data).unsqueeze(0).to(dtype=torch.float32) + return audio + + +def _save_mlpackage(model: ct.models.MLModel, path: Path, description: str) -> None: + try: + model.minimum_deployment_target = ct.target.iOS17 + except Exception: + pass + model.short_description = description + model.author = AUTHOR + path.parent.mkdir(parents=True, exist_ok=True) + model.save(str(path)) + + +def _tensor_shape(tensor: torch.Tensor) -> Tuple[int, ...]: + return tuple(int(dim) for dim in tensor.shape) + + +def _parse_compute_units(name: str) -> ct.ComputeUnit: + """Parse a human-friendly compute units string into ct.ComputeUnit.""" + normalized = str(name).strip().upper() + mapping = { + "ALL": ct.ComputeUnit.ALL, + "CPU_ONLY": ct.ComputeUnit.CPU_ONLY, + "CPU_AND_GPU": ct.ComputeUnit.CPU_AND_GPU, + "CPU_AND_NE": ct.ComputeUnit.CPU_AND_NE, + "CPU_AND_NEURALENGINE": ct.ComputeUnit.CPU_AND_NE, + } + if normalized not in mapping: + raise typer.BadParameter( + f"Unknown compute units '{name}'. Choose from: " + ", ".join(mapping.keys()) + ) + return mapping[normalized] + + +def _parse_compute_precision(name: Optional[str]) -> Optional[ct.precision]: + """Parse compute precision string into ct.precision or None.""" + if name is None: + return None + normalized = str(name).strip().upper() + if normalized == "": + return None + mapping = { + "FLOAT32": ct.precision.FLOAT32, + "FLOAT16": ct.precision.FLOAT16, + } + if normalized not in mapping: + raise typer.BadParameter( + f"Unknown compute precision '{name}'. Choose from: " + + ", ".join(mapping.keys()) + ) + return mapping[normalized] + + +app = typer.Typer(add_completion=False, pretty_exceptions_show_locals=False) + + +@app.command() +def convert( + nemo_path: Optional[Path] = typer.Option( + None, + "--nemo-path", + exists=True, + resolve_path=True, + help="Path to parakeet_realtime_eou_120m-v1.nemo checkpoint (skip to auto-download)", + ), + model_id: str = typer.Option( + DEFAULT_MODEL_ID, + "--model-id", + help="Model identifier to download when --nemo-path is omitted", + ), + output_dir: Path = typer.Option( + Path("parakeet_eou_coreml"), + help="Directory where mlpackages and metadata will be written", + ), + preprocessor_cu: str = typer.Option( + "CPU_ONLY", + "--preprocessor-cu", + help="Compute units for preprocessor (default CPU_ONLY)", + ), + mel_encoder_cu: str = typer.Option( + "CPU_ONLY", + "--mel-encoder-cu", + help="Compute units for fused mel+encoder (default CPU_ONLY)", + ), + compute_precision: Optional[str] = typer.Option( + None, + "--compute-precision", + help="Export precision: FLOAT32 (default) or FLOAT16 to shrink non-quantized weights.", + ), + max_audio_seconds: float = typer.Option( + 15.0, + "--max-audio-seconds", + help="Maximum audio duration in seconds for the fixed window export", + ), + validation_audio: Optional[Path] = typer.Option( + None, + "--validation-audio", + exists=True, + resolve_path=True, + help="Path to a 16kHz WAV file for tracing (uses random if not provided)", + ), +) -> None: + """Export all Parakeet Realtime EOU sub-modules to CoreML. + + This exports the cache-aware streaming FastConformer-RNNT model for + low-latency speech recognition with end-of-utterance detection. + """ + export_settings = ExportSettings( + output_dir=output_dir, + compute_units=ct.ComputeUnit.CPU_ONLY, + deployment_target=ct.target.iOS17, + compute_precision=_parse_compute_precision(compute_precision), + max_audio_seconds=max_audio_seconds, + max_symbol_steps=1, + ) + + typer.echo("Export configuration:") + typer.echo(asdict(export_settings)) + + output_dir.mkdir(parents=True, exist_ok=True) + pre_cu = _parse_compute_units(preprocessor_cu) + melenc_cu = _parse_compute_units(mel_encoder_cu) + + if nemo_path is not None: + typer.echo(f"Loading NeMo model from {nemo_path}…") + # Try loading as generic ASRModel first, then specific class + try: + asr_model = nemo_asr.models.ASRModel.restore_from( + str(nemo_path), map_location="cpu" + ) + except Exception: + # Fallback to EncDecRNNTBPEModel + asr_model = nemo_asr.models.EncDecRNNTBPEModel.restore_from( + str(nemo_path), map_location="cpu" + ) + checkpoint_meta = { + "type": "file", + "path": str(nemo_path), + } + else: + typer.echo(f"Downloading NeMo model via {model_id}…") + # Use ASRModel.from_pretrained as recommended for this model + try: + asr_model = nemo_asr.models.ASRModel.from_pretrained( + model_id, map_location="cpu" + ) + except Exception: + # Fallback to EncDecRNNTBPEModel + asr_model = nemo_asr.models.EncDecRNNTBPEModel.from_pretrained( + model_id, map_location="cpu" + ) + checkpoint_meta = { + "type": "pretrained", + "model_id": model_id, + } + asr_model.eval() + + # Print model info + typer.echo(f"Model class: {type(asr_model).__name__}") + typer.echo(f"Encoder class: {type(asr_model.encoder).__name__}") + + sample_rate = int(asr_model.cfg.preprocessor.sample_rate) + max_samples = _compute_length(export_settings.max_audio_seconds, sample_rate) + + # Prepare audio for tracing + if validation_audio is not None: + typer.echo(f"Using validation audio: {validation_audio}") + audio_tensor = _prepare_audio(validation_audio, sample_rate, max_samples, seed=None) + else: + typer.echo("Using random audio for tracing (seed=42)") + audio_tensor = _prepare_audio(None, sample_rate, max_samples, seed=42) + + audio_length = torch.tensor([max_samples], dtype=torch.int32) + + preprocessor = PreprocessorWrapper(asr_model.preprocessor.eval()) + encoder = EncoderWrapper(asr_model.encoder.eval()) + decoder = DecoderWrapper(asr_model.decoder.eval()) + joint = JointWrapper(asr_model.joint.eval()) + + decoder_export_flag = getattr(asr_model.decoder, "_rnnt_export", False) + asr_model.decoder._rnnt_export = True + + try: + with torch.no_grad(): + mel_ref, mel_length_ref = preprocessor(audio_tensor, audio_length) + mel_length_ref = mel_length_ref.to(dtype=torch.int32) + encoder_ref, encoder_length_ref, frame_times_ref = encoder( + mel_ref, mel_length_ref + ) + encoder_length_ref = encoder_length_ref.to(dtype=torch.int32) + + # Clone tensors to drop inference flags + mel_ref = mel_ref.clone().detach() + mel_length_ref = mel_length_ref.clone().detach() + encoder_ref = encoder_ref.clone().detach() + encoder_length_ref = encoder_length_ref.clone().detach() + frame_times_ref = frame_times_ref.clone().detach() + + vocab_size = int(asr_model.tokenizer.vocab_size) + decoder_hidden = int(asr_model.decoder.pred_hidden) + decoder_layers = int(asr_model.decoder.pred_rnn_layers) + + # Check if model has extra outputs (TDT-style duration) + num_extra = getattr(asr_model.joint, "num_extra_outputs", 0) + typer.echo(f"Vocab size: {vocab_size}, num_extra_outputs: {num_extra}") + + targets = torch.full( + (1, export_settings.max_symbol_steps), + fill_value=asr_model.decoder.blank_idx, + dtype=torch.int32, + ) + target_lengths = torch.tensor( + [export_settings.max_symbol_steps], dtype=torch.int32 + ) + zero_state = torch.zeros( + decoder_layers, + 1, + decoder_hidden, + dtype=torch.float32, + ) + + with torch.no_grad(): + decoder_ref, h_ref, c_ref = decoder( + targets, target_lengths, zero_state, zero_state + ) + joint_ref = joint(encoder_ref, decoder_ref) + + decoder_ref = decoder_ref.clone() + h_ref = h_ref.clone() + c_ref = c_ref.clone() + joint_ref = joint_ref.clone() + + typer.echo(f"Encoder output shape: {encoder_ref.shape}") + typer.echo(f"Decoder output shape: {decoder_ref.shape}") + typer.echo(f"Joint output shape: {joint_ref.shape}") + + # === Export Preprocessor === + typer.echo("Tracing and converting preprocessor…") + preprocessor = preprocessor.cpu() + audio_tensor = audio_tensor.cpu() + audio_length = audio_length.cpu() + traced_preprocessor = torch.jit.trace( + preprocessor, (audio_tensor, audio_length), strict=False + ) + traced_preprocessor.eval() + preprocessor_inputs = [ + ct.TensorType( + name="audio_signal", + shape=(1, ct.RangeDim(1, max_samples)), + dtype=np.float32, + ), + ct.TensorType(name="audio_length", shape=(1,), dtype=np.int32), + ] + preprocessor_outputs = [ + ct.TensorType(name="mel", dtype=np.float32), + ct.TensorType(name="mel_length", dtype=np.int32), + ] + preprocessor_model = _coreml_convert( + traced_preprocessor, + preprocessor_inputs, + preprocessor_outputs, + export_settings, + compute_units_override=pre_cu, + ) + preprocessor_path = output_dir / "parakeet_eou_preprocessor.mlpackage" + _save_mlpackage( + preprocessor_model, + preprocessor_path, + f"Parakeet EOU preprocessor ({max_audio_seconds}s window)", + ) + + # === Export Encoder === + typer.echo("Tracing and converting encoder…") + traced_encoder = torch.jit.trace( + encoder, (mel_ref, mel_length_ref), strict=False + ) + traced_encoder.eval() + encoder_inputs = [ + ct.TensorType( + name="mel", shape=_tensor_shape(mel_ref), dtype=np.float32 + ), + ct.TensorType(name="mel_length", shape=(1,), dtype=np.int32), + ] + encoder_outputs = [ + ct.TensorType(name="encoder", dtype=np.float32), + ct.TensorType(name="encoder_length", dtype=np.int32), + ct.TensorType(name="frame_times", dtype=np.float32), + ] + encoder_model = _coreml_convert( + traced_encoder, + encoder_inputs, + encoder_outputs, + export_settings, + compute_units_override=ct.ComputeUnit.CPU_ONLY, + ) + encoder_path = output_dir / "parakeet_eou_encoder.mlpackage" + _save_mlpackage( + encoder_model, + encoder_path, + f"Parakeet EOU encoder ({max_audio_seconds}s window)", + ) + + # === Export Fused Mel+Encoder === + typer.echo("Tracing and converting fused mel+encoder…") + mel_encoder = MelEncoderWrapper(preprocessor, encoder) + traced_mel_encoder = torch.jit.trace( + mel_encoder, (audio_tensor, audio_length), strict=False + ) + traced_mel_encoder.eval() + mel_encoder_inputs = [ + ct.TensorType( + name="audio_signal", shape=(1, max_samples), dtype=np.float32 + ), + ct.TensorType(name="audio_length", shape=(1,), dtype=np.int32), + ] + mel_encoder_outputs = [ + ct.TensorType(name="encoder", dtype=np.float32), + ct.TensorType(name="encoder_length", dtype=np.int32), + ct.TensorType(name="frame_times", dtype=np.float32), + ] + mel_encoder_model = _coreml_convert( + traced_mel_encoder, + mel_encoder_inputs, + mel_encoder_outputs, + export_settings, + compute_units_override=melenc_cu, + ) + mel_encoder_path = output_dir / "parakeet_eou_mel_encoder.mlpackage" + _save_mlpackage( + mel_encoder_model, + mel_encoder_path, + f"Parakeet EOU fused Mel+Encoder ({max_audio_seconds}s window)", + ) + + # === Export Decoder === + typer.echo("Tracing and converting decoder…") + traced_decoder = torch.jit.trace( + decoder, + (targets, target_lengths, zero_state, zero_state), + strict=False, + ) + traced_decoder.eval() + decoder_inputs = [ + ct.TensorType( + name="targets", shape=_tensor_shape(targets), dtype=np.int32 + ), + ct.TensorType(name="target_length", shape=(1,), dtype=np.int32), + ct.TensorType( + name="h_in", shape=_tensor_shape(zero_state), dtype=np.float32 + ), + ct.TensorType( + name="c_in", shape=_tensor_shape(zero_state), dtype=np.float32 + ), + ] + decoder_outputs = [ + ct.TensorType(name="decoder", dtype=np.float32), + ct.TensorType(name="h_out", dtype=np.float32), + ct.TensorType(name="c_out", dtype=np.float32), + ] + decoder_model = _coreml_convert( + traced_decoder, + decoder_inputs, + decoder_outputs, + export_settings, + compute_units_override=ct.ComputeUnit.CPU_ONLY, + ) + decoder_path = output_dir / "parakeet_eou_decoder.mlpackage" + _save_mlpackage( + decoder_model, + decoder_path, + "Parakeet EOU decoder (RNNT prediction network)", + ) + + # === Export Joint === + typer.echo("Tracing and converting joint…") + traced_joint = torch.jit.trace( + joint, + (encoder_ref, decoder_ref), + strict=False, + ) + traced_joint.eval() + joint_inputs = [ + ct.TensorType( + name="encoder", shape=_tensor_shape(encoder_ref), dtype=np.float32 + ), + ct.TensorType( + name="decoder", shape=_tensor_shape(decoder_ref), dtype=np.float32 + ), + ] + joint_outputs = [ + ct.TensorType(name="logits", dtype=np.float32), + ] + joint_model = _coreml_convert( + traced_joint, + joint_inputs, + joint_outputs, + export_settings, + compute_units_override=ct.ComputeUnit.CPU_ONLY, + ) + joint_path = output_dir / "parakeet_eou_joint.mlpackage" + _save_mlpackage( + joint_model, + joint_path, + "Parakeet EOU joint network (RNNT)", + ) + + # === Export Joint Decision Head === + typer.echo("Tracing and converting joint decision head…") + joint_decision = JointDecisionWrapper(joint, vocab_size=vocab_size) + traced_joint_decision = torch.jit.trace( + joint_decision, + (encoder_ref, decoder_ref), + strict=False, + ) + traced_joint_decision.eval() + joint_decision_inputs = [ + ct.TensorType( + name="encoder", shape=_tensor_shape(encoder_ref), dtype=np.float32 + ), + ct.TensorType( + name="decoder", shape=_tensor_shape(decoder_ref), dtype=np.float32 + ), + ] + joint_decision_outputs = [ + ct.TensorType(name="token_id", dtype=np.int32), + ct.TensorType(name="token_prob", dtype=np.float32), + ] + joint_decision_model = _coreml_convert( + traced_joint_decision, + joint_decision_inputs, + joint_decision_outputs, + export_settings, + compute_units_override=ct.ComputeUnit.CPU_ONLY, + ) + joint_decision_path = output_dir / "parakeet_eou_joint_decision.mlpackage" + _save_mlpackage( + joint_decision_model, + joint_decision_path, + "Parakeet EOU joint + decision head (softmax, argmax)", + ) + + # === Export Single-Step Joint Decision === + typer.echo("Tracing and converting single-step joint decision…") + jd_single = JointDecisionSingleStep(joint, vocab_size=vocab_size) + # Create single-step slices from refs + enc_step = encoder_ref[:, :, :1].contiguous() + dec_step = decoder_ref[:, :, :1].contiguous() + traced_jd_single = torch.jit.trace( + jd_single, + (enc_step, dec_step), + strict=False, + ) + traced_jd_single.eval() + jd_single_inputs = [ + ct.TensorType( + name="encoder_step", + shape=(1, enc_step.shape[1], 1), + dtype=np.float32, + ), + ct.TensorType( + name="decoder_step", + shape=(1, dec_step.shape[1], 1), + dtype=np.float32, + ), + ] + jd_single_outputs = [ + ct.TensorType(name="token_id", dtype=np.int32), + ct.TensorType(name="token_prob", dtype=np.float32), + ct.TensorType(name="top_k_ids", dtype=np.int32), + ct.TensorType(name="top_k_logits", dtype=np.float32), + ] + jd_single_model = _coreml_convert( + traced_jd_single, + jd_single_inputs, + jd_single_outputs, + export_settings, + compute_units_override=ct.ComputeUnit.CPU_ONLY, + ) + jd_single_path = output_dir / "parakeet_eou_joint_decision_single_step.mlpackage" + _save_mlpackage( + jd_single_model, + jd_single_path, + "Parakeet EOU single-step joint decision (current frame)", + ) + + # === Save Metadata === + metadata: Dict[str, object] = { + "model_id": model_id, + "model_name": "parakeet_realtime_eou_120m-v1", + "model_class": type(asr_model).__name__, + "encoder_class": type(asr_model.encoder).__name__, + "sample_rate": sample_rate, + "max_audio_seconds": export_settings.max_audio_seconds, + "max_audio_samples": max_samples, + "max_symbol_steps": export_settings.max_symbol_steps, + "vocab_size": vocab_size, + "vocab_with_blank": vocab_size + 1, + "decoder_hidden": decoder_hidden, + "decoder_layers": decoder_layers, + "num_extra_outputs": num_extra, + "has_eou_token": True, + "checkpoint": checkpoint_meta, + "coreml": { + "compute_units": export_settings.compute_units.name, + "compute_precision": ( + export_settings.compute_precision.name + if export_settings.compute_precision is not None + else "FLOAT32" + ), + }, + "components": { + "preprocessor": { + "inputs": { + "audio_signal": [1, max_samples], + "audio_length": [1], + }, + "outputs": { + "mel": list(_tensor_shape(mel_ref)), + "mel_length": [1], + }, + "path": preprocessor_path.name, + }, + "encoder": { + "inputs": { + "mel": list(_tensor_shape(mel_ref)), + "mel_length": [1], + }, + "outputs": { + "encoder": list(_tensor_shape(encoder_ref)), + "encoder_length": [1], + "frame_times": [1, _tensor_shape(encoder_ref)[2]], + }, + "path": encoder_path.name, + }, + "mel_encoder": { + "inputs": { + "audio_signal": [1, max_samples], + "audio_length": [1], + }, + "outputs": { + "encoder": list(_tensor_shape(encoder_ref)), + "encoder_length": [1], + "frame_times": [1, _tensor_shape(encoder_ref)[2]], + }, + "path": mel_encoder_path.name, + }, + "decoder": { + "inputs": { + "targets": list(_tensor_shape(targets)), + "target_length": [1], + "h_in": list(_tensor_shape(zero_state)), + "c_in": list(_tensor_shape(zero_state)), + }, + "outputs": { + "decoder": list(_tensor_shape(decoder_ref)), + "h_out": list(_tensor_shape(h_ref)), + "c_out": list(_tensor_shape(c_ref)), + }, + "path": decoder_path.name, + }, + "joint": { + "inputs": { + "encoder": list(_tensor_shape(encoder_ref)), + "decoder": list(_tensor_shape(decoder_ref)), + }, + "outputs": { + "logits": list(_tensor_shape(joint_ref)), + }, + "path": joint_path.name, + }, + "joint_decision": { + "inputs": { + "encoder": list(_tensor_shape(encoder_ref)), + "decoder": list(_tensor_shape(decoder_ref)), + }, + "outputs": { + "token_id": [ + _tensor_shape(encoder_ref)[0], + _tensor_shape(encoder_ref)[2], + _tensor_shape(decoder_ref)[2], + ], + "token_prob": [ + _tensor_shape(encoder_ref)[0], + _tensor_shape(encoder_ref)[2], + _tensor_shape(decoder_ref)[2], + ], + }, + "path": joint_decision_path.name, + }, + "joint_decision_single_step": { + "inputs": { + "encoder_step": [1, _tensor_shape(encoder_ref)[1], 1], + "decoder_step": [1, _tensor_shape(decoder_ref)[1], 1], + }, + "outputs": { + "token_id": [1, 1, 1], + "token_prob": [1, 1, 1], + "top_k_ids": [1, 1, 1, 64], + "top_k_logits": [1, 1, 1, 64], + }, + "path": jd_single_path.name, + }, + }, + } + + # Export tokenizer vocab if available + try: + tokenizer = asr_model.tokenizer + vocab = { + "blank_id": int(asr_model.decoder.blank_idx), + "vocab_size": vocab_size, + } + # Try to get special tokens + if hasattr(tokenizer, "tokenizer"): + inner_tokenizer = tokenizer.tokenizer + if hasattr(inner_tokenizer, "get_vocab"): + full_vocab = inner_tokenizer.get_vocab() + # Find EOU token + eou_token = None + for token, idx in full_vocab.items(): + if "" in token.upper() or "eou" in token.lower(): + eou_token = {"token": token, "id": idx} + break + if eou_token: + vocab["eou_token"] = eou_token + metadata["tokenizer"] = vocab + except Exception as e: + typer.echo(f"Warning: Could not export tokenizer info: {e}") + + metadata_path = output_dir / "metadata.json" + metadata_path.write_text(json.dumps(metadata, indent=2)) + typer.echo(f"\nExport complete. Metadata written to {metadata_path}") + typer.echo(f"Output directory: {output_dir}") + + finally: + asr_model.decoder._rnnt_export = decoder_export_flag + + +if __name__ == "__main__": + app() diff --git a/160ms/convert_streaming_encoder.py b/160ms/convert_streaming_encoder.py new file mode 100644 index 0000000000000000000000000000000000000000..a1c199c8a6c540dab48b973ab4232db2901b8e72 --- /dev/null +++ b/160ms/convert_streaming_encoder.py @@ -0,0 +1,193 @@ + +import torch +import torch.nn as nn +import coremltools as ct +import numpy as np +import typer +from pathlib import Path +from typing import Tuple, List, Optional +import json +import shutil + +# Iimport torch +import coremltools as ct +import numpy as np +import argparse +from nemo.collections.asr.models import EncDecRNNTBPEModel + +app = typer.Typer() + +class LoopbackEncoderWrapper(nn.Module): + """ + Wraps the entire Parakeet Encoder (PreEncode + Conformer) for CoreML Loopback Streaming. + + Inputs: + - audio_signal: [B, D, T] (Mel spectrogram chunk) + - audio_length: [B] + - pre_cache: [B, D, pre_cache_size] (Previous audio context) + - cache_last_channel: [layers, B, cache_size, hidden] + - cache_last_time: [layers, B, hidden, time_cache] + - cache_last_channel_len: [B] + + Outputs: + - encoded_output: [B, D_out, T_out] + - encoded_length: [B] + - new_pre_cache: [B, D, pre_cache_size] + - new_cache_last_channel + - new_cache_last_time + - new_cache_last_channel_len + """ + def __init__(self, encoder, pre_cache_size=16): + super().__init__() + self.encoder = encoder + self.pre_cache_size = pre_cache_size + + def forward( + self, + audio_signal: torch.Tensor, + audio_length: torch.Tensor, + pre_cache: torch.Tensor, + cache_last_channel: torch.Tensor, + cache_last_time: torch.Tensor, + cache_last_channel_len: torch.Tensor + ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor]: + + # 1. Prepend pre_cache to audio_signal + # audio_signal: [B, D, T] + # pre_cache: [B, D, T_cache] + full_input = torch.cat([pre_cache, audio_signal], dim=2) + full_length = audio_length + self.pre_cache_size + + # 2. Extract NEW pre_cache (last N frames of full_input) + # Note: We do this BEFORE processing because we want the raw audio context + new_pre_cache = full_input[:, :, -self.pre_cache_size:] + + # 3. Process with Encoder + # Reconstruct NeMo cache object + current_cache = [cache_last_channel, cache_last_time, cache_last_channel_len] + + encoded, encoded_len, new_cache_channel, new_cache_time, new_cache_len = self.encoder.cache_aware_stream_step( + processed_signal=full_input, + processed_signal_length=full_length, + cache_last_channel=cache_last_channel, + cache_last_time=cache_last_time, + cache_last_channel_len=cache_last_channel_len + ) + + # 4. Drop the first few frames corresponding to pre_cache? + # NeMo's cache_aware_stream_step usually handles the "valid" output frames. + # But since we manually prepended, we might get extra output frames. + # However, for streaming, we usually want the model to see the context but only output the new tokens. + # Let's trust NeMo's streaming logic for now, or check if we need to slice. + # Given we are using 'cache_aware_stream_step', it expects the full context window? + # Actually, standard usage is: input IS the new chunk, but internal convolution looks at past. + # But since we are stateless, we MUST provide the past. + # So passing (pre_cache + chunk) is correct. + + # Cast lengths to Int32 for CoreML + encoded_len_32 = encoded_len.to(dtype=torch.int32) + new_channel_len_32 = new_cache_len.to(dtype=torch.int32) + + return encoded, encoded_len_32, new_pre_cache, new_cache_channel, new_cache_time, new_channel_len_32 + +def _coreml_convert( + traced_model, + inputs, + outputs, + compute_units=ct.ComputeUnit.CPU_ONLY +): + return ct.convert( + traced_model, + inputs=inputs, + outputs=outputs, + compute_units=compute_units, + minimum_deployment_target=ct.target.macOS14, + ) + +def main(): + model_id: str = "nvidia/parakeet_realtime_eou_120m-v1" + output_dir: str = "temp_swift_models/StreamingLoopback" + output_path = Path(output_dir) + output_path.mkdir(parents=True, exist_ok=True) + + print(f"Loading model: {model_id}...") + asr_model = EncDecRNNTBPEModel.from_pretrained(model_name=model_id) + asr_model.eval() + + parser = argparse.ArgumentParser() + parser.add_argument("--chunk-frames", type=int, default=17, help="Number of frames in the input chunk (e.g. 17 for 160ms, 129 for 1.28s)") + args = parser.parse_args() + + encoder = asr_model.encoder + + # --- Configuration --- + # 160ms chunk = 16 frames (but preprocessor produces 17 with padding/centering) + # 1.28s chunk = 128 frames (preprocessor produces 129) + chunk_size_in = args.chunk_frames + mel_dim = 128 + hidden_dim = encoder.d_model # 512 + num_layers = len(encoder.layers) # 17 + + # Cache sizes + cache_channel_size = 70 + cache_time_size = 8 + pre_cache_size = 16 + + print(f"Config: Chunk={chunk_size_in}, Mel={mel_dim}, Hidden={hidden_dim}, Layers={num_layers}") + print(f"Cache: Channel={cache_channel_size}, Time={cache_time_size}, Pre={pre_cache_size}") + + # --- Wrapper --- + wrapper = LoopbackEncoderWrapper(encoder, pre_cache_size=pre_cache_size) + wrapper.eval() + + # --- Test Inputs (for Tracing) --- + batch_size = 1 + test_mel = torch.randn(batch_size, mel_dim, chunk_size_in) + test_mel_len = torch.tensor([chunk_size_in], dtype=torch.int32) + test_pre_cache = torch.zeros(batch_size, mel_dim, pre_cache_size) + + # Initial Cache (Zeros) + test_cache_channel = torch.zeros(num_layers, batch_size, cache_channel_size, hidden_dim) + test_cache_time = torch.zeros(num_layers, batch_size, hidden_dim, cache_time_size) + test_cache_len = torch.zeros(batch_size, dtype=torch.int32) + + print("Tracing model...") + traced_model = torch.jit.trace( + wrapper, + (test_mel, test_mel_len, test_pre_cache, test_cache_channel, test_cache_time, test_cache_len), + strict=False + ) + + # --- CoreML Conversion --- + print("Converting to CoreML...") + + inputs = [ + ct.TensorType(name="audio_signal", shape=(1, 128, chunk_size_in), dtype=np.float32), + ct.TensorType(name="audio_length", shape=(1,), dtype=np.int32), + ct.TensorType(name="pre_cache", shape=(1, 128, pre_cache_size), dtype=np.float32), + ct.TensorType(name="cache_last_channel", shape=(num_layers, 1, cache_channel_size, hidden_dim), dtype=np.float32), + ct.TensorType(name="cache_last_time", shape=(num_layers, 1, hidden_dim, cache_time_size), dtype=np.float32), + ct.TensorType(name="cache_last_channel_len", shape=(1,), dtype=np.int32), + ] + + outputs = [ + ct.TensorType(name="encoded_output", dtype=np.float32), + ct.TensorType(name="encoded_length", dtype=np.int32), + ct.TensorType(name="new_pre_cache", dtype=np.float32), + ct.TensorType(name="new_cache_last_channel", dtype=np.float32), + ct.TensorType(name="new_cache_last_time", dtype=np.float32), + ct.TensorType(name="new_cache_last_channel_len", dtype=np.int32), + ] + + mlmodel = _coreml_convert(traced_model, inputs, outputs) + + save_path = output_path / "streaming_encoder.mlpackage" + mlmodel.save(str(save_path)) + print(f"Saved: {save_path}") + + # Also export Preprocessor, Decoder, Joint for completeness? + # For now, let's assume we reuse the existing ones or export them separately if needed. + # But the user asked specifically for the Encoder loopback. + +if __name__ == "__main__": + main() diff --git a/160ms/decoder.mlmodelc/analytics/coremldata.bin b/160ms/decoder.mlmodelc/analytics/coremldata.bin new file mode 100644 index 0000000000000000000000000000000000000000..c6f51dedd2638184c5a2d71512339ceea086abc0 --- /dev/null +++ b/160ms/decoder.mlmodelc/analytics/coremldata.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3996975a8cbc1949159c55605b3132b39b2484f51acbd55d796d93c70de02b49 +size 243 diff --git a/160ms/decoder.mlmodelc/coremldata.bin b/160ms/decoder.mlmodelc/coremldata.bin new file mode 100644 index 0000000000000000000000000000000000000000..71936e61ffe7697e4af55bae8870d95c1fbb66c5 --- /dev/null +++ b/160ms/decoder.mlmodelc/coremldata.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3ccbff963d8cf07e2be2bd56ea3384a89ea49628922c6bd95ff62e2ae57dc34 +size 497 diff --git a/160ms/decoder.mlmodelc/metadata.json b/160ms/decoder.mlmodelc/metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..3581673bcb062f7e0e1e0efa9f59edc5f218d50e --- /dev/null +++ b/160ms/decoder.mlmodelc/metadata.json @@ -0,0 +1,118 @@ +[ + { + "metadataOutputVersion" : "3.0", + "shortDescription" : "Parakeet EOU decoder (RNNT prediction network)", + "outputSchema" : [ + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 640 × 1)", + "shortDescription" : "", + "shape" : "[1, 640, 1]", + "name" : "decoder", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 1 × 640)", + "shortDescription" : "", + "shape" : "[1, 1, 640]", + "name" : "h_out", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 1 × 640)", + "shortDescription" : "", + "shape" : "[1, 1, 640]", + "name" : "c_out", + "type" : "MultiArray" + } + ], + "storagePrecision" : "Float16", + "modelParameters" : [ + + ], + "author" : "Fluid Inference", + "specificationVersion" : 8, + "mlProgramOperationTypeHistogram" : { + "Ios17.squeeze" : 2, + "Ios17.gather" : 1, + "Ios17.cast" : 6, + "Ios17.lstm" : 1, + "Ios17.transpose" : 2, + "Identity" : 1, + "Ios17.expandDims" : 2 + }, + "computePrecision" : "Mixed (Float16, Float32, Int16, Int32)", + "isUpdatable" : "0", + "stateSchema" : [ + + ], + "availability" : { + "macOS" : "14.0", + "tvOS" : "17.0", + "visionOS" : "1.0", + "watchOS" : "10.0", + "iOS" : "17.0", + "macCatalyst" : "17.0" + }, + "modelType" : { + "name" : "MLModelType_mlProgram" + }, + "inputSchema" : [ + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1 × 1)", + "shortDescription" : "", + "shape" : "[1, 1]", + "name" : "targets", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1)", + "shortDescription" : "", + "shape" : "[1]", + "name" : "target_length", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 1 × 640)", + "shortDescription" : "", + "shape" : "[1, 1, 640]", + "name" : "h_in", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 1 × 640)", + "shortDescription" : "", + "shape" : "[1, 1, 640]", + "name" : "c_in", + "type" : "MultiArray" + } + ], + "userDefinedMetadata" : { + "com.github.apple.coremltools.version" : "8.3.0", + "com.github.apple.coremltools.source" : "torch==2.4.0", + "com.github.apple.coremltools.source_dialect" : "TorchScript" + }, + "generatedClassName" : "parakeet_eou_decoder", + "method" : "predict" + } +] \ No newline at end of file diff --git a/160ms/decoder.mlmodelc/model.mil b/160ms/decoder.mlmodelc/model.mil new file mode 100644 index 0000000000000000000000000000000000000000..5ead417fbdc13f8bf0c05a26039b2828503d4eca --- /dev/null +++ b/160ms/decoder.mlmodelc/model.mil @@ -0,0 +1,45 @@ +program(1.0) +[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.4.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] +{ + func main(tensor c_in, tensor h_in, tensor target_length, tensor targets) { + tensor y_axis_0 = const()[name = tensor("y_axis_0"), val = tensor(0)]; + tensor y_batch_dims_0 = const()[name = tensor("y_batch_dims_0"), val = tensor(0)]; + tensor y_validate_indices_0 = const()[name = tensor("y_validate_indices_0"), val = tensor(false)]; + tensor module_prediction_embed_weight_to_fp16 = const()[name = tensor("module_prediction_embed_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; + tensor targets_to_int16_dtype_0 = const()[name = tensor("targets_to_int16_dtype_0"), val = tensor("int16")]; + tensor targets_to_int16 = cast(dtype = targets_to_int16_dtype_0, x = targets)[name = tensor("cast_8")]; + tensor y_cast_fp16_cast_uint16 = gather(axis = y_axis_0, batch_dims = y_batch_dims_0, indices = targets_to_int16, validate_indices = y_validate_indices_0, x = module_prediction_embed_weight_to_fp16)[name = tensor("y_cast_fp16_cast_uint16")]; + tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([1, 0, 2])]; + tensor input_lstm_h0_squeeze_axes_0 = const()[name = tensor("input_lstm_h0_squeeze_axes_0"), val = tensor([0])]; + tensor h_in_to_fp16_dtype_0 = const()[name = tensor("h_in_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor h_in_to_fp16 = cast(dtype = h_in_to_fp16_dtype_0, x = h_in)[name = tensor("cast_7")]; + tensor input_lstm_h0_squeeze_cast_fp16 = squeeze(axes = input_lstm_h0_squeeze_axes_0, x = h_in_to_fp16)[name = tensor("input_lstm_h0_squeeze_cast_fp16")]; + tensor input_lstm_c0_squeeze_axes_0 = const()[name = tensor("input_lstm_c0_squeeze_axes_0"), val = tensor([0])]; + tensor c_in_to_fp16_dtype_0 = const()[name = tensor("c_in_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor c_in_to_fp16 = cast(dtype = c_in_to_fp16_dtype_0, x = c_in)[name = tensor("cast_6")]; + tensor input_lstm_c0_squeeze_cast_fp16 = squeeze(axes = input_lstm_c0_squeeze_axes_0, x = c_in_to_fp16)[name = tensor("input_lstm_c0_squeeze_cast_fp16")]; + tensor input_direction_0 = const()[name = tensor("input_direction_0"), val = tensor("forward")]; + tensor input_output_sequence_0 = const()[name = tensor("input_output_sequence_0"), val = tensor(true)]; + tensor input_recurrent_activation_0 = const()[name = tensor("input_recurrent_activation_0"), val = tensor("sigmoid")]; + tensor input_cell_activation_0 = const()[name = tensor("input_cell_activation_0"), val = tensor("tanh")]; + tensor input_activation_0 = const()[name = tensor("input_activation_0"), val = tensor("tanh")]; + tensor concat_1_to_fp16 = const()[name = tensor("concat_1_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1314688)))]; + tensor concat_2_to_fp16 = const()[name = tensor("concat_2_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4591552)))]; + tensor concat_0_to_fp16 = const()[name = tensor("concat_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7868416)))]; + tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = y_cast_fp16_cast_uint16)[name = tensor("transpose_2")]; + tensor input_cast_fp16_0, tensor input_cast_fp16_1, tensor input_cast_fp16_2 = lstm(activation = input_activation_0, bias = concat_0_to_fp16, cell_activation = input_cell_activation_0, direction = input_direction_0, initial_c = input_lstm_c0_squeeze_cast_fp16, initial_h = input_lstm_h0_squeeze_cast_fp16, output_sequence = input_output_sequence_0, recurrent_activation = input_recurrent_activation_0, weight_hh = concat_2_to_fp16, weight_ih = concat_1_to_fp16, x = input_3_cast_fp16)[name = tensor("input_cast_fp16")]; + tensor obj_3_axes_0 = const()[name = tensor("obj_3_axes_0"), val = tensor([0])]; + tensor obj_3_cast_fp16 = expand_dims(axes = obj_3_axes_0, x = input_cast_fp16_1)[name = tensor("obj_3_cast_fp16")]; + tensor obj_3_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_3_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor obj_axes_0 = const()[name = tensor("obj_axes_0"), val = tensor([0])]; + tensor obj_cast_fp16 = expand_dims(axes = obj_axes_0, x = input_cast_fp16_2)[name = tensor("obj_cast_fp16")]; + tensor obj_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor transpose_0_perm_0 = const()[name = tensor("transpose_0_perm_0"), val = tensor([1, 2, 0])]; + tensor transpose_0_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("transpose_0_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor transpose_0_cast_fp16 = transpose(perm = transpose_0_perm_0, x = input_cast_fp16_0)[name = tensor("transpose_1")]; + tensor decoder = cast(dtype = transpose_0_cast_fp16_to_fp32_dtype_0, x = transpose_0_cast_fp16)[name = tensor("cast_3")]; + tensor c_out = cast(dtype = obj_cast_fp16_to_fp32_dtype_0, x = obj_cast_fp16)[name = tensor("cast_4")]; + tensor h_out = cast(dtype = obj_3_cast_fp16_to_fp32_dtype_0, x = obj_3_cast_fp16)[name = tensor("cast_5")]; + tensor target_length_tmp = identity(x = target_length)[name = tensor("target_length_tmp")]; + } -> (decoder, h_out, c_out); +} \ No newline at end of file diff --git a/160ms/decoder.mlmodelc/weights/weight.bin b/160ms/decoder.mlmodelc/weights/weight.bin new file mode 100644 index 0000000000000000000000000000000000000000..cde1618cbed235b421984a302289e3bdd7e3df02 --- /dev/null +++ b/160ms/decoder.mlmodelc/weights/weight.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b4cacecdcd9df79ab1e56de67230baf5a8664d2afe0bb8f3408eefa972cb2f4 +size 7873600 diff --git a/160ms/decoder.mlpackage/Data/com.apple.CoreML/model.mlmodel b/160ms/decoder.mlpackage/Data/com.apple.CoreML/model.mlmodel new file mode 100644 index 0000000000000000000000000000000000000000..0631f4ecb012ab42d13312be15475aff506055c5 --- /dev/null +++ b/160ms/decoder.mlpackage/Data/com.apple.CoreML/model.mlmodel @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:09f2dbd1f6a06faa6995f71d4b25d7c446996b6059cfac5ecc889853bdc7c6e5 +size 6728 diff --git a/160ms/decoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/160ms/decoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin new file mode 100644 index 0000000000000000000000000000000000000000..cde1618cbed235b421984a302289e3bdd7e3df02 --- /dev/null +++ b/160ms/decoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b4cacecdcd9df79ab1e56de67230baf5a8664d2afe0bb8f3408eefa972cb2f4 +size 7873600 diff --git a/160ms/decoder.mlpackage/Manifest.json b/160ms/decoder.mlpackage/Manifest.json new file mode 100644 index 0000000000000000000000000000000000000000..4e498499a0a4026095bbc46d4e1a47ecbc39b0ca --- /dev/null +++ b/160ms/decoder.mlpackage/Manifest.json @@ -0,0 +1,18 @@ +{ + "fileFormatVersion": "1.0.0", + "itemInfoEntries": { + "8201D73A-2B5D-488C-9C2B-7E2E75DF700D": { + "author": "com.apple.CoreML", + "description": "CoreML Model Weights", + "name": "weights", + "path": "com.apple.CoreML/weights" + }, + "F8EEBE8D-F17D-4556-B8DF-9BC11701B36D": { + "author": "com.apple.CoreML", + "description": "CoreML Model Specification", + "name": "model.mlmodel", + "path": "com.apple.CoreML/model.mlmodel" + } + }, + "rootModelIdentifier": "F8EEBE8D-F17D-4556-B8DF-9BC11701B36D" +} diff --git a/160ms/individual_components.py b/160ms/individual_components.py new file mode 100644 index 0000000000000000000000000000000000000000..47271397bc8d9d17cc0fabcf4bb63be7e7c2109c --- /dev/null +++ b/160ms/individual_components.py @@ -0,0 +1,250 @@ +#!/usr/bin/env python3 +"""Export Parakeet Realtime EOU 120M RNNT components into CoreML.""" +from __future__ import annotations + +from dataclasses import dataclass +from pathlib import Path +from typing import Optional, Tuple + +import coremltools as ct +import torch + + +@dataclass +class ExportSettings: + output_dir: Path + compute_units: ct.ComputeUnit + deployment_target: Optional[ct.target] + compute_precision: Optional[ct.precision] + max_audio_seconds: float + max_symbol_steps: int + + +class PreprocessorWrapper(torch.nn.Module): + """Wrapper for the audio preprocessor (mel spectrogram extraction).""" + + def __init__(self, module: torch.nn.Module) -> None: + super().__init__() + self.module = module + + def forward( + self, audio_signal: torch.Tensor, length: torch.Tensor + ) -> Tuple[torch.Tensor, torch.Tensor]: + mel, mel_length = self.module( + input_signal=audio_signal, length=length.to(dtype=torch.long) + ) + return mel, mel_length + + +class EncoderWrapper(torch.nn.Module): + """Wrapper for the cache-aware FastConformer encoder. + + Note: For the realtime EOU model, the encoder is cache-aware which means + it can operate in a streaming fashion. For CoreML export, we export + without cache state for simplicity (full-context mode). + """ + + def __init__(self, module: torch.nn.Module) -> None: + super().__init__() + self.module = module + + def forward( + self, features: torch.Tensor, length: torch.Tensor + ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: + encoded, encoded_lengths = self.module( + audio_signal=features, length=length.to(dtype=torch.long) + ) + # Synthesize per-frame timestamps (seconds) using the 80 ms encoder stride. + # Shape: [B, T_enc] + frame_times = ( + torch.arange(encoded.shape[-1], device=encoded.device, dtype=torch.float32) + * 0.08 + ) + return encoded, encoded_lengths, frame_times + + +class DecoderWrapper(torch.nn.Module): + """Wrapper for the RNNT prediction network (decoder).""" + + def __init__(self, module: torch.nn.Module) -> None: + super().__init__() + self.module = module + + def forward( + self, + targets: torch.Tensor, + target_lengths: torch.Tensor, + h_in: torch.Tensor, + c_in: torch.Tensor, + ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: + state = [h_in, c_in] + decoder_output, _, new_state = self.module( + targets=targets.to(dtype=torch.long), + target_length=target_lengths.to(dtype=torch.long), + states=state, + ) + return decoder_output, new_state[0], new_state[1] + + +class JointWrapper(torch.nn.Module): + """Wrapper for the RNNT joint network. + + Note: Unlike Parakeet TDT v3, the realtime EOU model does NOT have + duration outputs (num_extra_outputs). The joint network outputs only + token logits over the vocabulary + blank. + """ + + def __init__(self, module: torch.nn.Module) -> None: + super().__init__() + self.module = module + + def forward( + self, encoder_outputs: torch.Tensor, decoder_outputs: torch.Tensor + ) -> torch.Tensor: + # Input: encoder_outputs [B, D, T], decoder_outputs [B, D, U] + # Transpose to match what projection layers expect + encoder_outputs = encoder_outputs.transpose(1, 2) # [B, T, D] + decoder_outputs = decoder_outputs.transpose(1, 2) # [B, U, D] + + # Apply projections + enc_proj = self.module.enc(encoder_outputs) # [B, T, joint_hidden] + dec_proj = self.module.pred(decoder_outputs) # [B, U, joint_hidden] + + # Explicit broadcasting along T and U to avoid converter ambiguity + x = enc_proj.unsqueeze(2) + dec_proj.unsqueeze(1) # [B, T, U, joint_hidden] + x = self.module.joint_net[0](x) # ReLU + x = self.module.joint_net[1](x) # Dropout (no-op in eval) + out = self.module.joint_net[2](x) # Linear -> logits [B, T, U, vocab+blank] + return out + + +class MelEncoderWrapper(torch.nn.Module): + """Fused wrapper: waveform -> mel -> encoder. + + Inputs: + - audio_signal: [B, S] + - audio_length: [B] + + Outputs: + - encoder: [B, D, T_enc] + - encoder_length: [B] + - frame_times: [T_enc] + """ + + def __init__( + self, preprocessor: PreprocessorWrapper, encoder: EncoderWrapper + ) -> None: + super().__init__() + self.preprocessor = preprocessor + self.encoder = encoder + + def forward( + self, audio_signal: torch.Tensor, audio_length: torch.Tensor + ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: + mel, mel_length = self.preprocessor(audio_signal, audio_length) + encoded, enc_len, frame_times = self.encoder(mel, mel_length.to(dtype=torch.int32)) + return encoded, enc_len, frame_times + + +class JointDecisionWrapper(torch.nn.Module): + """Joint + decision head: outputs label id and label prob. + + Unlike Parakeet TDT v3, this model does NOT have duration outputs. + + Inputs: + - encoder_outputs: [B, D, T] + - decoder_outputs: [B, D, U] + + Returns: + - token_id: [B, T, U] int32 + - token_prob: [B, T, U] float32 + """ + + def __init__(self, joint: JointWrapper, vocab_size: int) -> None: + super().__init__() + self.joint = joint + self.vocab_with_blank = int(vocab_size) + 1 + + def forward(self, encoder_outputs: torch.Tensor, decoder_outputs: torch.Tensor): + logits = self.joint(encoder_outputs, decoder_outputs) + + # Token selection + token_ids = torch.argmax(logits, dim=-1).to(dtype=torch.int32) + token_probs_all = torch.softmax(logits, dim=-1) + # gather expects int64 (long) indices; cast only for gather + token_prob = torch.gather( + token_probs_all, dim=-1, index=token_ids.long().unsqueeze(-1) + ).squeeze(-1) + + return token_ids, token_prob + + +class JointDecisionSingleStep(torch.nn.Module): + """Single-step variant for streaming: encoder_step -> token decision. + + Inputs: + - encoder_step: [B=1, D, T=1] + - decoder_step: [B=1, D, U=1] + + Returns: + - token_id: [1, 1, 1] int32 + - token_prob: [1, 1, 1] float32 + - top_k_ids: [1, 1, 1, K] int32 + - top_k_logits: [1, 1, 1, K] float32 + """ + + def __init__(self, joint: JointWrapper, vocab_size: int, top_k: int = 64) -> None: + super().__init__() + self.joint = joint + self.vocab_with_blank = int(vocab_size) + 1 + self.top_k = int(top_k) + + def forward(self, encoder_step: torch.Tensor, decoder_step: torch.Tensor): + # Reuse JointWrapper which expects [B, D, T] and [B, D, U] + logits = self.joint(encoder_step, decoder_step) # [1, 1, 1, V+blank] + + token_ids = torch.argmax(logits, dim=-1, keepdim=False).to(dtype=torch.int32) + token_probs_all = torch.softmax(logits, dim=-1) + token_prob = torch.gather( + token_probs_all, dim=-1, index=token_ids.long().unsqueeze(-1) + ).squeeze(-1) + + # Also expose top-K candidates for host-side processing + topk_logits, topk_ids_long = torch.topk( + logits, k=min(self.top_k, logits.shape[-1]), dim=-1 + ) + topk_ids = topk_ids_long.to(dtype=torch.int32) + return token_ids, token_prob, topk_ids, topk_logits + + +def _coreml_convert( + traced: torch.jit.ScriptModule, + inputs, + outputs, + settings: ExportSettings, + compute_units_override: Optional[ct.ComputeUnit] = None, + compute_precision: Optional[ct.precision] = None, +) -> ct.models.MLModel: + cu = ( + compute_units_override + if compute_units_override is not None + else settings.compute_units + ) + kwargs = { + "convert_to": "mlprogram", + "inputs": inputs, + "outputs": outputs, + "compute_units": cu, + } + print("Converting:", traced.__class__.__name__) + print("Conversion kwargs:", kwargs) + if settings.deployment_target is not None: + kwargs["minimum_deployment_target"] = settings.deployment_target + + # Priority: explicit argument > settings + if compute_precision is not None: + kwargs["compute_precision"] = compute_precision + elif settings.compute_precision is not None: + kwargs["compute_precision"] = settings.compute_precision + + return ct.convert(traced, **kwargs) diff --git a/160ms/joint_decision.mlmodelc/analytics/coremldata.bin b/160ms/joint_decision.mlmodelc/analytics/coremldata.bin new file mode 100644 index 0000000000000000000000000000000000000000..ff9929e12bae25c509655bda0ebda6b9fe13fade --- /dev/null +++ b/160ms/joint_decision.mlmodelc/analytics/coremldata.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5bca32ad130dcad6605cc00044c752aa5b45ef57d14c17f2d1a2fa49d6cf55b5 +size 243 diff --git a/160ms/joint_decision.mlmodelc/coremldata.bin b/160ms/joint_decision.mlmodelc/coremldata.bin new file mode 100644 index 0000000000000000000000000000000000000000..1aa5afe7c02e93f9baba2757cad5c21f563157bf --- /dev/null +++ b/160ms/joint_decision.mlmodelc/coremldata.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22d4abc4625b935ee035b5f8ce7cb28d1041b9b01c12173e287bf4b5f5d99625 +size 493 diff --git a/160ms/joint_decision.mlmodelc/metadata.json b/160ms/joint_decision.mlmodelc/metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..a0a319857f429f5ec4b106397cd52ec354ef2e96 --- /dev/null +++ b/160ms/joint_decision.mlmodelc/metadata.json @@ -0,0 +1,112 @@ +[ + { + "metadataOutputVersion" : "3.0", + "shortDescription" : "Parakeet EOU single-step joint decision", + "outputSchema" : [ + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1 × 1 × 1)", + "shortDescription" : "", + "shape" : "[1, 1, 1]", + "name" : "token_id", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 1 × 1)", + "shortDescription" : "", + "shape" : "[1, 1, 1]", + "name" : "token_prob", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1 × 1 × 1 × 64)", + "shortDescription" : "", + "shape" : "[1, 1, 1, 64]", + "name" : "top_k_ids", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 1 × 1 × 64)", + "shortDescription" : "", + "shape" : "[1, 1, 1, 64]", + "name" : "top_k_logits", + "type" : "MultiArray" + } + ], + "storagePrecision" : "Float16", + "modelParameters" : [ + + ], + "author" : "Fluid Inference", + "specificationVersion" : 8, + "mlProgramOperationTypeHistogram" : { + "Ios17.reduceArgmax" : 1, + "Ios17.squeeze" : 1, + "Ios17.cast" : 6, + "Ios17.linear" : 3, + "Ios17.transpose" : 2, + "Ios17.add" : 1, + "Ios16.relu" : 1, + "Ios16.softmax" : 1, + "Ios17.gatherAlongAxis" : 1, + "Ios17.topk" : 1, + "Ios17.expandDims" : 3 + }, + "computePrecision" : "Mixed (Float16, Float32, Int16, Int32, UInt16)", + "isUpdatable" : "0", + "stateSchema" : [ + + ], + "availability" : { + "macOS" : "14.0", + "tvOS" : "17.0", + "visionOS" : "1.0", + "watchOS" : "10.0", + "iOS" : "17.0", + "macCatalyst" : "17.0" + }, + "modelType" : { + "name" : "MLModelType_mlProgram" + }, + "inputSchema" : [ + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 512 × 1)", + "shortDescription" : "", + "shape" : "[1, 512, 1]", + "name" : "encoder_step", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 640 × 1)", + "shortDescription" : "", + "shape" : "[1, 640, 1]", + "name" : "decoder_step", + "type" : "MultiArray" + } + ], + "userDefinedMetadata" : { + "com.github.apple.coremltools.source_dialect" : "TorchScript", + "com.github.apple.coremltools.version" : "8.3.0", + "com.github.apple.coremltools.source" : "torch==2.4.0" + }, + "generatedClassName" : "parakeet_eou_joint_decision_single_step", + "method" : "predict" + } +] \ No newline at end of file diff --git a/160ms/joint_decision.mlmodelc/model.mil b/160ms/joint_decision.mlmodelc/model.mil new file mode 100644 index 0000000000000000000000000000000000000000..172a7579866c84407b32b5f746b7f1ae132599d8 --- /dev/null +++ b/160ms/joint_decision.mlmodelc/model.mil @@ -0,0 +1,57 @@ +program(1.0) +[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.4.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] +{ + func main(tensor decoder_step, tensor encoder_step) { + tensor input_1_perm_0 = const()[name = tensor("input_1_perm_0"), val = tensor([0, 2, 1])]; + tensor encoder_step_to_fp16_dtype_0 = const()[name = tensor("encoder_step_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([0, 2, 1])]; + tensor decoder_step_to_fp16_dtype_0 = const()[name = tensor("decoder_step_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor joint_module_enc_weight_to_fp16 = const()[name = tensor("joint_module_enc_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; + tensor joint_module_enc_bias_to_fp16 = const()[name = tensor("joint_module_enc_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(655488)))]; + tensor encoder_step_to_fp16 = cast(dtype = encoder_step_to_fp16_dtype_0, x = encoder_step)[name = tensor("cast_8")]; + tensor input_1_cast_fp16 = transpose(perm = input_1_perm_0, x = encoder_step_to_fp16)[name = tensor("transpose_1")]; + tensor linear_0_cast_fp16 = linear(bias = joint_module_enc_bias_to_fp16, weight = joint_module_enc_weight_to_fp16, x = input_1_cast_fp16)[name = tensor("linear_0_cast_fp16")]; + tensor joint_module_pred_weight_to_fp16 = const()[name = tensor("joint_module_pred_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(656832)))]; + tensor joint_module_pred_bias_to_fp16 = const()[name = tensor("joint_module_pred_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1476096)))]; + tensor decoder_step_to_fp16 = cast(dtype = decoder_step_to_fp16_dtype_0, x = decoder_step)[name = tensor("cast_7")]; + tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = decoder_step_to_fp16)[name = tensor("transpose_0")]; + tensor linear_1_cast_fp16 = linear(bias = joint_module_pred_bias_to_fp16, weight = joint_module_pred_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("linear_1_cast_fp16")]; + tensor var_23_axes_0 = const()[name = tensor("op_23_axes_0"), val = tensor([2])]; + tensor var_23_cast_fp16 = expand_dims(axes = var_23_axes_0, x = linear_0_cast_fp16)[name = tensor("op_23_cast_fp16")]; + tensor var_24_axes_0 = const()[name = tensor("op_24_axes_0"), val = tensor([1])]; + tensor var_24_cast_fp16 = expand_dims(axes = var_24_axes_0, x = linear_1_cast_fp16)[name = tensor("op_24_cast_fp16")]; + tensor input_5_cast_fp16 = add(x = var_23_cast_fp16, y = var_24_cast_fp16)[name = tensor("input_5_cast_fp16")]; + tensor input_7_cast_fp16 = relu(x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; + tensor joint_module_joint_net_2_weight_to_fp16 = const()[name = tensor("joint_module_joint_net_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1477440)))]; + tensor joint_module_joint_net_2_bias_to_fp16 = const()[name = tensor("joint_module_joint_net_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2792064)))]; + tensor linear_2_cast_fp16 = linear(bias = joint_module_joint_net_2_bias_to_fp16, weight = joint_module_joint_net_2_weight_to_fp16, x = input_7_cast_fp16)[name = tensor("linear_2_cast_fp16")]; + tensor var_38_axis_0 = const()[name = tensor("op_38_axis_0"), val = tensor(-1)]; + tensor var_38_keep_dims_0 = const()[name = tensor("op_38_keep_dims_0"), val = tensor(false)]; + tensor var_38_output_dtype_0 = const()[name = tensor("op_38_output_dtype_0"), val = tensor("int32")]; + tensor token_id = reduce_argmax(axis = var_38_axis_0, keep_dims = var_38_keep_dims_0, output_dtype = var_38_output_dtype_0, x = linear_2_cast_fp16)[name = tensor("op_38_cast_fp16")]; + tensor var_44 = const()[name = tensor("op_44"), val = tensor(-1)]; + tensor token_probs_all_cast_fp16 = softmax(axis = var_44, x = linear_2_cast_fp16)[name = tensor("token_probs_all_cast_fp16")]; + tensor var_53_axes_0 = const()[name = tensor("op_53_axes_0"), val = tensor([-1])]; + tensor var_53 = expand_dims(axes = var_53_axes_0, x = token_id)[name = tensor("op_53")]; + tensor var_54 = const()[name = tensor("op_54"), val = tensor(-1)]; + tensor var_56_validate_indices_0 = const()[name = tensor("op_56_validate_indices_0"), val = tensor(false)]; + tensor var_53_to_int16_dtype_0 = const()[name = tensor("op_53_to_int16_dtype_0"), val = tensor("int16")]; + tensor var_53_to_int16 = cast(dtype = var_53_to_int16_dtype_0, x = var_53)[name = tensor("cast_6")]; + tensor var_56_cast_fp16_cast_int16 = gather_along_axis(axis = var_54, indices = var_53_to_int16, validate_indices = var_56_validate_indices_0, x = token_probs_all_cast_fp16)[name = tensor("op_56_cast_fp16_cast_int16")]; + tensor var_58_axes_0 = const()[name = tensor("op_58_axes_0"), val = tensor([-1])]; + tensor var_58_cast_fp16 = squeeze(axes = var_58_axes_0, x = var_56_cast_fp16_cast_int16)[name = tensor("op_58_cast_fp16")]; + tensor var_58_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_58_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor var_59 = const()[name = tensor("op_59"), val = tensor(64)]; + tensor var_63_axis_0 = const()[name = tensor("op_63_axis_0"), val = tensor(-1)]; + tensor var_63_ascending_0 = const()[name = tensor("op_63_ascending_0"), val = tensor(false)]; + tensor var_63_sort_0 = const()[name = tensor("op_63_sort_0"), val = tensor(true)]; + tensor var_63_return_indices_0 = const()[name = tensor("op_63_return_indices_0"), val = tensor(true)]; + tensor var_63_cast_fp16_cast_int16_output_indices_dtype_0 = const()[name = tensor("op_63_cast_fp16_cast_int16_output_indices_dtype_0"), val = tensor("uint16")]; + tensor var_63_cast_fp16_cast_int16_0, tensor var_63_cast_fp16_cast_int16_1 = topk(ascending = var_63_ascending_0, axis = var_63_axis_0, k = var_59, output_indices_dtype = var_63_cast_fp16_cast_int16_output_indices_dtype_0, return_indices = var_63_return_indices_0, sort = var_63_sort_0, x = linear_2_cast_fp16)[name = tensor("op_63_cast_fp16_cast_int16")]; + tensor var_63_cast_fp16_cast_int16_1_to_int32_dtype_0 = const()[name = tensor("op_63_cast_fp16_cast_int16_1_to_int32_dtype_0"), val = tensor("int32")]; + tensor var_63_cast_fp16_0_to_fp32_dtype_0 = const()[name = tensor("op_63_cast_fp16_0_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor top_k_logits = cast(dtype = var_63_cast_fp16_0_to_fp32_dtype_0, x = var_63_cast_fp16_cast_int16_0)[name = tensor("cast_3")]; + tensor top_k_ids = cast(dtype = var_63_cast_fp16_cast_int16_1_to_int32_dtype_0, x = var_63_cast_fp16_cast_int16_1)[name = tensor("cast_4")]; + tensor token_prob = cast(dtype = var_58_cast_fp16_to_fp32_dtype_0, x = var_58_cast_fp16)[name = tensor("cast_5")]; + } -> (token_id, token_prob, top_k_ids, top_k_logits); +} \ No newline at end of file diff --git a/160ms/joint_decision.mlmodelc/weights/weight.bin b/160ms/joint_decision.mlmodelc/weights/weight.bin new file mode 100644 index 0000000000000000000000000000000000000000..ec3d2c1dea18b332fa1a3cd8d31981bae4e3f649 --- /dev/null +++ b/160ms/joint_decision.mlmodelc/weights/weight.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7039b2010a269153f5a96edf28637f921a86ef8822f248f2d6712f7a6bce84b4 +size 2794182 diff --git a/160ms/joint_decision.mlpackage/Data/com.apple.CoreML/model.mlmodel b/160ms/joint_decision.mlpackage/Data/com.apple.CoreML/model.mlmodel new file mode 100644 index 0000000000000000000000000000000000000000..678be6fae0d594ff4bb923f973b0459139c8b370 --- /dev/null +++ b/160ms/joint_decision.mlpackage/Data/com.apple.CoreML/model.mlmodel @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25d4d7be6eeb60c7de1d3a1278a5a4700cbe34017e1a8c1cab33204ddb2e4d5e +size 8701 diff --git a/160ms/joint_decision.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/160ms/joint_decision.mlpackage/Data/com.apple.CoreML/weights/weight.bin new file mode 100644 index 0000000000000000000000000000000000000000..ec3d2c1dea18b332fa1a3cd8d31981bae4e3f649 --- /dev/null +++ b/160ms/joint_decision.mlpackage/Data/com.apple.CoreML/weights/weight.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7039b2010a269153f5a96edf28637f921a86ef8822f248f2d6712f7a6bce84b4 +size 2794182 diff --git a/160ms/joint_decision.mlpackage/Manifest.json b/160ms/joint_decision.mlpackage/Manifest.json new file mode 100644 index 0000000000000000000000000000000000000000..fffd90a77691d335c8153fd9c38d49adb725c949 --- /dev/null +++ b/160ms/joint_decision.mlpackage/Manifest.json @@ -0,0 +1,18 @@ +{ + "fileFormatVersion": "1.0.0", + "itemInfoEntries": { + "634E266B-4447-41D3-879E-F3611888F54B": { + "author": "com.apple.CoreML", + "description": "CoreML Model Weights", + "name": "weights", + "path": "com.apple.CoreML/weights" + }, + "C7F40527-180B-45CD-BC12-4F054F2E5D9A": { + "author": "com.apple.CoreML", + "description": "CoreML Model Specification", + "name": "model.mlmodel", + "path": "com.apple.CoreML/model.mlmodel" + } + }, + "rootModelIdentifier": "C7F40527-180B-45CD-BC12-4F054F2E5D9A" +} diff --git a/160ms/parakeet_eou_preprocessor.mlmodelc/analytics/coremldata.bin b/160ms/parakeet_eou_preprocessor.mlmodelc/analytics/coremldata.bin new file mode 100644 index 0000000000000000000000000000000000000000..ca93efecf9fbf12e37943f7e192cd324a39e9f94 --- /dev/null +++ b/160ms/parakeet_eou_preprocessor.mlmodelc/analytics/coremldata.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4ada8b0b99ac1d2ba7acbffacfbbf1a06cb69d30e9410d237ee0aa4c2b0ad63 +size 243 diff --git a/160ms/parakeet_eou_preprocessor.mlmodelc/coremldata.bin b/160ms/parakeet_eou_preprocessor.mlmodelc/coremldata.bin new file mode 100644 index 0000000000000000000000000000000000000000..ca69399b73b039b35bb1b4967b25b7a7a12e8178 --- /dev/null +++ b/160ms/parakeet_eou_preprocessor.mlmodelc/coremldata.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc7252fa47622fe39577361233627062019a3bb740fdbb5366a7bae09df0ec5e +size 422 diff --git a/160ms/parakeet_eou_preprocessor.mlmodelc/metadata.json b/160ms/parakeet_eou_preprocessor.mlmodelc/metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..98b70ffcae507184b68f87d7adaad706e5c68c9a --- /dev/null +++ b/160ms/parakeet_eou_preprocessor.mlmodelc/metadata.json @@ -0,0 +1,105 @@ +[ + { + "metadataOutputVersion" : "3.0", + "shortDescription" : "Parakeet EOU preprocessor", + "outputSchema" : [ + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32)", + "shortDescription" : "", + "shape" : "[]", + "name" : "mel", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1)", + "shortDescription" : "", + "shape" : "[1]", + "name" : "mel_length", + "type" : "MultiArray" + } + ], + "storagePrecision" : "Float32", + "modelParameters" : [ + + ], + "author" : "Fluid Inference", + "specificationVersion" : 8, + "mlProgramOperationTypeHistogram" : { + "Range1d" : 1, + "Ios17.reshape" : 2, + "Identity" : 1, + "Ios17.matmul" : 1, + "Ios17.expandDims" : 5, + "Select" : 1, + "Ios17.add" : 3, + "Ios17.sliceByIndex" : 3, + "Ios16.reduceSum" : 1, + "Shape" : 1, + "Ios17.gather" : 1, + "Pad" : 1, + "Ios17.log" : 1, + "Ios17.conv" : 2, + "Ios17.sub" : 2, + "Ios17.pow" : 1, + "Ios17.cast" : 2, + "Stack" : 1, + "Ios17.concat" : 1, + "Ios17.floorDiv" : 1, + "Ios17.greaterEqual" : 1, + "Ios17.mul" : 1 + }, + "computePrecision" : "Mixed (Float32, Int32)", + "isUpdatable" : "0", + "stateSchema" : [ + + ], + "availability" : { + "macOS" : "14.0", + "tvOS" : "17.0", + "visionOS" : "1.0", + "watchOS" : "10.0", + "iOS" : "17.0", + "macCatalyst" : "17.0" + }, + "modelType" : { + "name" : "MLModelType_mlProgram" + }, + "inputSchema" : [ + { + "dataType" : "Float32", + "hasShapeFlexibility" : "1", + "isOptional" : "0", + "shapeFlexibility" : "1 × 1...32000", + "shapeRange" : "[[1, 1], [1, 32000]]", + "formattedType" : "MultiArray (Float32 1 × 1)", + "type" : "MultiArray", + "shape" : "[1, 1]", + "name" : "audio_signal", + "shortDescription" : "" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1)", + "shortDescription" : "", + "shape" : "[1]", + "name" : "audio_length", + "type" : "MultiArray" + } + ], + "userDefinedMetadata" : { + "com.github.apple.coremltools.source_dialect" : "TorchScript", + "com.github.apple.coremltools.source" : "torch==2.4.0", + "com.github.apple.coremltools.version" : "8.3.0" + }, + "generatedClassName" : "parakeet_eou_preprocessor", + "method" : "predict" + } +] \ No newline at end of file diff --git a/160ms/parakeet_eou_preprocessor.mlmodelc/model.mil b/160ms/parakeet_eou_preprocessor.mlmodelc/model.mil new file mode 100644 index 0000000000000000000000000000000000000000..25db6a1374c1718340459f374949bfba2bff85bb --- /dev/null +++ b/160ms/parakeet_eou_preprocessor.mlmodelc/model.mil @@ -0,0 +1,96 @@ +program(1.0) +[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.4.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] +{ + func main(tensor audio_length, tensor audio_signal) [FlexibleShapeInformation = tuple, dict, tensor>>, tuple, dict, list, ?>>>>((("DefaultShapes", {{"audio_signal", [1, 1]}}), ("RangeDims", {{"audio_signal", [[1, 1], [1, 32000]]}})))] { + tensor var_9 = const()[name = tensor("op_9"), val = tensor(1)]; + tensor var_10 = const()[name = tensor("op_10"), val = tensor(160)]; + tensor var_32 = const()[name = tensor("op_32"), val = tensor(512)]; + tensor var_33 = add(x = audio_length, y = var_32)[name = tensor("op_33")]; + tensor var_34 = const()[name = tensor("op_34"), val = tensor(512)]; + tensor var_35 = sub(x = var_33, y = var_34)[name = tensor("op_35")]; + tensor floor_div_0 = floor_div(x = var_35, y = var_10)[name = tensor("floor_div_0")]; + tensor var_36_dtype_0 = const()[name = tensor("op_36_dtype_0"), val = tensor("fp32")]; + tensor var_37_promoted = const()[name = tensor("op_37_promoted"), val = tensor(0x1p+0)]; + tensor var_36 = cast(dtype = var_36_dtype_0, x = floor_div_0)[name = tensor("cast_11")]; + tensor seq_len_1 = add(x = var_36, y = var_37_promoted)[name = tensor("seq_len_1")]; + tensor cast_2_dtype_0 = const()[name = tensor("cast_2_dtype_0"), val = tensor("int32")]; + tensor var_41_begin_0 = const()[name = tensor("op_41_begin_0"), val = tensor([0, 0])]; + tensor var_41_end_0 = const()[name = tensor("op_41_end_0"), val = tensor([1, 1])]; + tensor var_41_end_mask_0 = const()[name = tensor("op_41_end_mask_0"), val = tensor([true, false])]; + tensor var_41_squeeze_mask_0 = const()[name = tensor("op_41_squeeze_mask_0"), val = tensor([false, true])]; + tensor var_41 = slice_by_index(begin = var_41_begin_0, end = var_41_end_0, end_mask = var_41_end_mask_0, squeeze_mask = var_41_squeeze_mask_0, x = audio_signal)[name = tensor("op_41")]; + tensor var_42_axes_0 = const()[name = tensor("op_42_axes_0"), val = tensor([1])]; + tensor var_42 = expand_dims(axes = var_42_axes_0, x = var_41)[name = tensor("op_42")]; + tensor var_44_begin_0 = const()[name = tensor("op_44_begin_0"), val = tensor([0, 1])]; + tensor var_44_end_0 = const()[name = tensor("op_44_end_0"), val = tensor([1, 0])]; + tensor var_44_end_mask_0 = const()[name = tensor("op_44_end_mask_0"), val = tensor([true, true])]; + tensor var_44 = slice_by_index(begin = var_44_begin_0, end = var_44_end_0, end_mask = var_44_end_mask_0, x = audio_signal)[name = tensor("op_44")]; + tensor var_46_begin_0 = const()[name = tensor("op_46_begin_0"), val = tensor([0, 0])]; + tensor var_46_end_0 = const()[name = tensor("op_46_end_0"), val = tensor([1, -1])]; + tensor var_46_end_mask_0 = const()[name = tensor("op_46_end_mask_0"), val = tensor([true, false])]; + tensor var_46 = slice_by_index(begin = var_46_begin_0, end = var_46_end_0, end_mask = var_46_end_mask_0, x = audio_signal)[name = tensor("op_46")]; + tensor var_47 = const()[name = tensor("op_47"), val = tensor(0x1.f0a3d8p-1)]; + tensor var_48 = mul(x = var_46, y = var_47)[name = tensor("op_48")]; + tensor var_49 = sub(x = var_44, y = var_48)[name = tensor("op_49")]; + tensor input_1_interleave_0 = const()[name = tensor("input_1_interleave_0"), val = tensor(false)]; + tensor input_1 = concat(axis = var_9, interleave = input_1_interleave_0, values = (var_42, var_49))[name = tensor("input_1")]; + tensor concat_0x = const()[name = tensor("concat_0x"), val = tensor([1, 1, -1])]; + tensor input_3 = reshape(shape = concat_0x, x = input_1)[name = tensor("input_3")]; + tensor const_1 = const()[name = tensor("const_1"), val = tensor(0x0p+0)]; + tensor input_5_pad_0 = const()[name = tensor("input_5_pad_0"), val = tensor([0, 0, 0, 0, 256, 256])]; + tensor input_5_mode_0 = const()[name = tensor("input_5_mode_0"), val = tensor("reflect")]; + tensor input_5 = pad(constant_val = const_1, mode = input_5_mode_0, pad = input_5_pad_0, x = input_3)[name = tensor("input_5")]; + tensor concat_1x = const()[name = tensor("concat_1x"), val = tensor([1, -1])]; + tensor input = reshape(shape = concat_1x, x = input_5)[name = tensor("input")]; + tensor expand_dims_1 = const()[name = tensor("expand_dims_1"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; + tensor expand_dims_2 = const()[name = tensor("expand_dims_2"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526464)))]; + tensor expand_dims_3 = const()[name = tensor("expand_dims_3"), val = tensor([160])]; + tensor expand_dims_4_axes_0 = const()[name = tensor("expand_dims_4_axes_0"), val = tensor([1])]; + tensor expand_dims_4 = expand_dims(axes = expand_dims_4_axes_0, x = input)[name = tensor("expand_dims_4")]; + tensor conv_0_pad_type_0 = const()[name = tensor("conv_0_pad_type_0"), val = tensor("valid")]; + tensor conv_0_pad_0 = const()[name = tensor("conv_0_pad_0"), val = tensor([0, 0])]; + tensor conv_0_dilations_0 = const()[name = tensor("conv_0_dilations_0"), val = tensor([1])]; + tensor conv_0_groups_0 = const()[name = tensor("conv_0_groups_0"), val = tensor(1)]; + tensor conv_0 = conv(dilations = conv_0_dilations_0, groups = conv_0_groups_0, pad = conv_0_pad_0, pad_type = conv_0_pad_type_0, strides = expand_dims_3, weight = expand_dims_1, x = expand_dims_4)[name = tensor("conv_0")]; + tensor conv_1_pad_type_0 = const()[name = tensor("conv_1_pad_type_0"), val = tensor("valid")]; + tensor conv_1_pad_0 = const()[name = tensor("conv_1_pad_0"), val = tensor([0, 0])]; + tensor conv_1_dilations_0 = const()[name = tensor("conv_1_dilations_0"), val = tensor([1])]; + tensor conv_1_groups_0 = const()[name = tensor("conv_1_groups_0"), val = tensor(1)]; + tensor conv_1 = conv(dilations = conv_1_dilations_0, groups = conv_1_groups_0, pad = conv_1_pad_0, pad_type = conv_1_pad_type_0, strides = expand_dims_3, weight = expand_dims_2, x = expand_dims_4)[name = tensor("conv_1")]; + tensor stack_0_axis_0 = const()[name = tensor("stack_0_axis_0"), val = tensor(-1)]; + tensor stack_0 = stack(axis = stack_0_axis_0, values = (conv_0, conv_1))[name = tensor("stack_0")]; + tensor var_17_promoted = const()[name = tensor("op_17_promoted"), val = tensor(0x1p+1)]; + tensor var_65 = pow(x = stack_0, y = var_17_promoted)[name = tensor("op_65")]; + tensor var_67_axes_0 = const()[name = tensor("op_67_axes_0"), val = tensor([-1])]; + tensor var_67_keep_dims_0 = const()[name = tensor("op_67_keep_dims_0"), val = tensor(false)]; + tensor var_67 = reduce_sum(axes = var_67_axes_0, keep_dims = var_67_keep_dims_0, x = var_65)[name = tensor("op_67")]; + tensor x_9 = identity(x = var_67)[name = tensor("x_9")]; + tensor const_2 = const()[name = tensor("const_2"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1052864)))]; + tensor x_11_transpose_x_0 = const()[name = tensor("x_11_transpose_x_0"), val = tensor(false)]; + tensor x_11_transpose_y_0 = const()[name = tensor("x_11_transpose_y_0"), val = tensor(false)]; + tensor x_11 = matmul(transpose_x = x_11_transpose_x_0, transpose_y = x_11_transpose_y_0, x = const_2, y = x_9)[name = tensor("x_11")]; + tensor var_74 = const()[name = tensor("op_74"), val = tensor(0x1p-24)]; + tensor var_75 = add(x = x_11, y = var_74)[name = tensor("op_75")]; + tensor x_epsilon_0 = const()[name = tensor("x_epsilon_0"), val = tensor(0x1p-149)]; + tensor x = log(epsilon = x_epsilon_0, x = var_75)[name = tensor("x")]; + tensor var_77_shape = shape(x = x)[name = tensor("op_77_shape")]; + tensor select_4 = const()[name = tensor("select_4"), val = tensor(2)]; + tensor gather_4_axis_0 = const()[name = tensor("gather_4_axis_0"), val = tensor(0)]; + tensor gather_4_batch_dims_0 = const()[name = tensor("gather_4_batch_dims_0"), val = tensor(0)]; + tensor gather_4_validate_indices_0 = const()[name = tensor("gather_4_validate_indices_0"), val = tensor(false)]; + tensor gather_4 = gather(axis = gather_4_axis_0, batch_dims = gather_4_batch_dims_0, indices = select_4, validate_indices = gather_4_validate_indices_0, x = var_77_shape)[name = tensor("gather_4")]; + tensor const_3 = const()[name = tensor("const_3"), val = tensor(0)]; + tensor const_4 = const()[name = tensor("const_4"), val = tensor(1)]; + tensor mask_1 = range_1d(end = gather_4, start = const_3, step = const_4)[name = tensor("mask_1")]; + tensor expand_dims_0_axes_0 = const()[name = tensor("expand_dims_0_axes_0"), val = tensor([0])]; + tensor expand_dims_0 = expand_dims(axes = expand_dims_0_axes_0, x = mask_1)[name = tensor("expand_dims_0")]; + tensor var_82_axes_0 = const()[name = tensor("op_82_axes_0"), val = tensor([1])]; + tensor mel_length = cast(dtype = cast_2_dtype_0, x = seq_len_1)[name = tensor("cast_10")]; + tensor var_82 = expand_dims(axes = var_82_axes_0, x = mel_length)[name = tensor("op_82")]; + tensor mask = greater_equal(x = expand_dims_0, y = var_82)[name = tensor("mask")]; + tensor var_84_axes_0 = const()[name = tensor("op_84_axes_0"), val = tensor([1])]; + tensor var_84 = expand_dims(axes = var_84_axes_0, x = mask)[name = tensor("op_84")]; + tensor cast_7 = const()[name = tensor("cast_7"), val = tensor(0x0p+0)]; + tensor mel = select(a = cast_7, b = x, cond = var_84)[name = tensor("processed_signal")]; + } -> (mel, mel_length); +} \ No newline at end of file diff --git a/160ms/parakeet_eou_preprocessor.mlmodelc/weights/weight.bin b/160ms/parakeet_eou_preprocessor.mlmodelc/weights/weight.bin new file mode 100644 index 0000000000000000000000000000000000000000..99f6476377d3e7474aa570db365b7429f9d7b5b8 --- /dev/null +++ b/160ms/parakeet_eou_preprocessor.mlmodelc/weights/weight.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:009bba4fde82dc55db9b55d77cf3ba5f791ce366c49f079285fe25a3b6e2291d +size 1184512 diff --git a/160ms/parakeet_eou_preprocessor.mlpackage/Data/com.apple.CoreML/model.mlmodel b/160ms/parakeet_eou_preprocessor.mlpackage/Data/com.apple.CoreML/model.mlmodel new file mode 100644 index 0000000000000000000000000000000000000000..3e2bc094c7b810cced59b083c5713339db85e068 --- /dev/null +++ b/160ms/parakeet_eou_preprocessor.mlpackage/Data/com.apple.CoreML/model.mlmodel @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0849137cc7b828554ddf85d6bc7923b76e8ade3a8f111e72dcedecc9889b21ba +size 16111 diff --git a/160ms/parakeet_eou_preprocessor.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/160ms/parakeet_eou_preprocessor.mlpackage/Data/com.apple.CoreML/weights/weight.bin new file mode 100644 index 0000000000000000000000000000000000000000..117eb770f48b9e96d96462009988c22143d0c615 --- /dev/null +++ b/160ms/parakeet_eou_preprocessor.mlpackage/Data/com.apple.CoreML/weights/weight.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f257ad1ac11575d73a6ffda555319b2c96b0a224f0dc03ddd8c62950e9b18e53 +size 592384 diff --git a/160ms/parakeet_eou_preprocessor.mlpackage/Manifest.json b/160ms/parakeet_eou_preprocessor.mlpackage/Manifest.json new file mode 100644 index 0000000000000000000000000000000000000000..13fb6413b0fb5c9e552cea893a70ea84bb6ffefa --- /dev/null +++ b/160ms/parakeet_eou_preprocessor.mlpackage/Manifest.json @@ -0,0 +1,18 @@ +{ + "fileFormatVersion": "1.0.0", + "itemInfoEntries": { + "07C051EE-C19D-4B6F-A62D-88BF3ADE779A": { + "author": "com.apple.CoreML", + "description": "CoreML Model Specification", + "name": "model.mlmodel", + "path": "com.apple.CoreML/model.mlmodel" + }, + "CBBDA0B6-FD16-43BF-914C-B518684DCC1B": { + "author": "com.apple.CoreML", + "description": "CoreML Model Weights", + "name": "weights", + "path": "com.apple.CoreML/weights" + } + }, + "rootModelIdentifier": "07C051EE-C19D-4B6F-A62D-88BF3ADE779A" +} diff --git a/160ms/streaming_encoder.mlmodelc/analytics/coremldata.bin b/160ms/streaming_encoder.mlmodelc/analytics/coremldata.bin new file mode 100644 index 0000000000000000000000000000000000000000..042b43c0828ad617d3510bdaea4d7e87e0d9d2d8 --- /dev/null +++ b/160ms/streaming_encoder.mlmodelc/analytics/coremldata.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a981b257db79b4f86e6fa06a92562160a0ae71554746c24af24d8634b85f0356 +size 243 diff --git a/160ms/streaming_encoder.mlmodelc/coremldata.bin b/160ms/streaming_encoder.mlmodelc/coremldata.bin new file mode 100644 index 0000000000000000000000000000000000000000..7c32b6d8b90a4cfb081aa7a8531b4a8dc5765a4b --- /dev/null +++ b/160ms/streaming_encoder.mlmodelc/coremldata.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e762abc60d999bcd10aab985b68191a602f2e8e03165cf08671c60f93936037a +size 670 diff --git a/160ms/streaming_encoder.mlmodelc/metadata.json b/160ms/streaming_encoder.mlmodelc/metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..32148a71ed327b665c0b11778846e2378749683d --- /dev/null +++ b/160ms/streaming_encoder.mlmodelc/metadata.json @@ -0,0 +1,187 @@ +[ + { + "metadataOutputVersion" : "3.0", + "storagePrecision" : "Float16", + "outputSchema" : [ + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 512 × 2)", + "shortDescription" : "", + "shape" : "[1, 512, 2]", + "name" : "encoded_output", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1)", + "shortDescription" : "", + "shape" : "[1]", + "name" : "encoded_length", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 128 × 16)", + "shortDescription" : "", + "shape" : "[1, 128, 16]", + "name" : "new_pre_cache", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 17 × 1 × 70 × 512)", + "shortDescription" : "", + "shape" : "[17, 1, 70, 512]", + "name" : "new_cache_last_channel", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 17 × 1 × 512 × 8)", + "shortDescription" : "", + "shape" : "[17, 1, 512, 8]", + "name" : "new_cache_last_time", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1)", + "shortDescription" : "", + "shape" : "[1]", + "name" : "new_cache_last_channel_len", + "type" : "MultiArray" + } + ], + "modelParameters" : [ + + ], + "specificationVersion" : 8, + "mlProgramOperationTypeHistogram" : { + "Ios17.floor" : 3, + "Ios17.logicalAnd" : 3, + "Ios17.reshape" : 103, + "Ios16.softmax" : 17, + "Ios17.matmul" : 51, + "Ios17.transpose" : 157, + "Split" : 17, + "Ios17.expandDims" : 6, + "Select" : 51, + "Ios17.add" : 126, + "Tile" : 1, + "Ios17.sliceByIndex" : 141, + "Ios16.sigmoid" : 17, + "Pad" : 20, + "Ios17.logicalNot" : 2, + "Ios17.layerNorm" : 102, + "Ios17.less" : 1, + "Ios17.sub" : 1, + "Ios17.conv" : 56, + "Ios17.clip" : 3, + "Ios16.relu" : 3, + "Ios17.linear" : 137, + "Ios17.concat" : 52, + "Ios17.greaterEqual" : 1, + "Ios17.cast" : 16, + "Ios16.silu" : 51, + "Stack" : 2, + "Ios17.mul" : 72 + }, + "computePrecision" : "Mixed (Float16, Float32, Int32)", + "isUpdatable" : "0", + "stateSchema" : [ + + ], + "availability" : { + "macOS" : "14.0", + "tvOS" : "17.0", + "visionOS" : "1.0", + "watchOS" : "10.0", + "iOS" : "17.0", + "macCatalyst" : "17.0" + }, + "modelType" : { + "name" : "MLModelType_mlProgram" + }, + "userDefinedMetadata" : { + "com.github.apple.coremltools.source_dialect" : "TorchScript", + "com.github.apple.coremltools.version" : "8.3.0", + "com.github.apple.coremltools.source" : "torch==2.4.0" + }, + "inputSchema" : [ + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 128 × 17)", + "shortDescription" : "", + "shape" : "[1, 128, 17]", + "name" : "audio_signal", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1)", + "shortDescription" : "", + "shape" : "[1]", + "name" : "audio_length", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 128 × 16)", + "shortDescription" : "", + "shape" : "[1, 128, 16]", + "name" : "pre_cache", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 17 × 1 × 70 × 512)", + "shortDescription" : "", + "shape" : "[17, 1, 70, 512]", + "name" : "cache_last_channel", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 17 × 1 × 512 × 8)", + "shortDescription" : "", + "shape" : "[17, 1, 512, 8]", + "name" : "cache_last_time", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1)", + "shortDescription" : "", + "shape" : "[1]", + "name" : "cache_last_channel_len", + "type" : "MultiArray" + } + ], + "generatedClassName" : "streaming_encoder", + "method" : "predict" + } +] \ No newline at end of file diff --git a/160ms/streaming_encoder.mlmodelc/model.mil b/160ms/streaming_encoder.mlmodelc/model.mil new file mode 100644 index 0000000000000000000000000000000000000000..52cb37c36ef8f9be82cd6744ed43a2e13a911b40 --- /dev/null +++ b/160ms/streaming_encoder.mlmodelc/model.mil @@ -0,0 +1,3124 @@ +program(1.0) +[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.4.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] +{ + func main(tensor audio_length, tensor audio_signal, tensor cache_last_channel, tensor cache_last_channel_len, tensor cache_last_time, tensor pre_cache) { + tensor var_9 = const()[name = tensor("op_9"), val = tensor(2)]; + tensor full_input_interleave_0 = const()[name = tensor("full_input_interleave_0"), val = tensor(false)]; + tensor pre_cache_to_fp16_dtype_0 = const()[name = tensor("pre_cache_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor audio_signal_to_fp16_dtype_0 = const()[name = tensor("audio_signal_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor audio_signal_to_fp16 = cast(dtype = audio_signal_to_fp16_dtype_0, x = audio_signal)[name = tensor("cast_195")]; + tensor pre_cache_to_fp16 = cast(dtype = pre_cache_to_fp16_dtype_0, x = pre_cache)[name = tensor("cast_196")]; + tensor full_input_cast_fp16 = concat(axis = var_9, interleave = full_input_interleave_0, values = (pre_cache_to_fp16, audio_signal_to_fp16))[name = tensor("full_input_cast_fp16")]; + tensor var_12 = const()[name = tensor("op_12"), val = tensor(16)]; + tensor value_1 = add(x = audio_length, y = var_12)[name = tensor("value_1")]; + tensor var_28_begin_0 = const()[name = tensor("op_28_begin_0"), val = tensor([0, 0, 17])]; + tensor var_28_end_0 = const()[name = tensor("op_28_end_0"), val = tensor([1, 128, 33])]; + tensor var_28_end_mask_0 = const()[name = tensor("op_28_end_mask_0"), val = tensor([true, true, true])]; + tensor var_28_cast_fp16 = slice_by_index(begin = var_28_begin_0, end = var_28_end_0, end_mask = var_28_end_mask_0, x = full_input_cast_fp16)[name = tensor("op_28_cast_fp16")]; + tensor var_28_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_28_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor var_62 = const()[name = tensor("op_62"), val = tensor(-1)]; + tensor var_64 = const()[name = tensor("op_64"), val = tensor(1)]; + tensor x_1_perm_0 = const()[name = tensor("x_1_perm_0"), val = tensor([0, 2, 1])]; + tensor cast_0_to_fp16_dtype_0 = const()[name = tensor("cast_0_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor _inversed_108_y_0_to_fp16 = const()[name = tensor("_inversed_108_y_0_to_fp16"), val = tensor(0x1p-1)]; + tensor value_1_to_fp16 = cast(dtype = cast_0_to_fp16_dtype_0, x = value_1)[name = tensor("cast_193")]; + tensor _inversed_108_cast_fp16 = mul(x = value_1_to_fp16, y = _inversed_108_y_0_to_fp16)[name = tensor("_inversed_108_cast_fp16")]; + tensor var_109_to_fp16 = const()[name = tensor("op_109_to_fp16"), val = tensor(0x1p+0)]; + tensor lengths_1_cast_fp16 = add(x = _inversed_108_cast_fp16, y = var_109_to_fp16)[name = tensor("lengths_1_cast_fp16")]; + tensor lengths_3_cast_fp16 = floor(x = lengths_1_cast_fp16)[name = tensor("lengths_3_cast_fp16")]; + tensor _inversed_116_y_0_to_fp16 = const()[name = tensor("_inversed_116_y_0_to_fp16"), val = tensor(0x1p-1)]; + tensor _inversed_116_cast_fp16 = mul(x = lengths_3_cast_fp16, y = _inversed_116_y_0_to_fp16)[name = tensor("_inversed_116_cast_fp16")]; + tensor var_117_to_fp16 = const()[name = tensor("op_117_to_fp16"), val = tensor(0x1p+0)]; + tensor lengths_7_cast_fp16 = add(x = _inversed_116_cast_fp16, y = var_117_to_fp16)[name = tensor("lengths_7_cast_fp16")]; + tensor lengths_9_cast_fp16 = floor(x = lengths_7_cast_fp16)[name = tensor("lengths_9_cast_fp16")]; + tensor _inversed_124_y_0_to_fp16 = const()[name = tensor("_inversed_124_y_0_to_fp16"), val = tensor(0x1p-1)]; + tensor _inversed_124_cast_fp16 = mul(x = lengths_9_cast_fp16, y = _inversed_124_y_0_to_fp16)[name = tensor("_inversed_124_cast_fp16")]; + tensor var_125_to_fp16 = const()[name = tensor("op_125_to_fp16"), val = tensor(0x1p+0)]; + tensor lengths_13_cast_fp16 = add(x = _inversed_124_cast_fp16, y = var_125_to_fp16)[name = tensor("lengths_13_cast_fp16")]; + tensor lengths_cast_fp16 = floor(x = lengths_13_cast_fp16)[name = tensor("lengths_cast_fp16")]; + tensor cast_9_dtype_0 = const()[name = tensor("cast_9_dtype_0"), val = tensor("int32")]; + tensor input_1_axes_0 = const()[name = tensor("input_1_axes_0"), val = tensor([1])]; + tensor x_1_cast_fp16 = transpose(perm = x_1_perm_0, x = full_input_cast_fp16)[name = tensor("transpose_241")]; + tensor input_1_cast_fp16 = expand_dims(axes = input_1_axes_0, x = x_1_cast_fp16)[name = tensor("input_1_cast_fp16")]; + tensor input_3_pad_0 = const()[name = tensor("input_3_pad_0"), val = tensor([0, 0, 0, 0, 2, 1, 2, 1])]; + tensor input_3_mode_0 = const()[name = tensor("input_3_mode_0"), val = tensor("constant")]; + tensor const_0_to_fp16 = const()[name = tensor("const_0_to_fp16"), val = tensor(0x0p+0)]; + tensor input_3_cast_fp16 = pad(constant_val = const_0_to_fp16, mode = input_3_mode_0, pad = input_3_pad_0, x = input_1_cast_fp16)[name = tensor("input_3_cast_fp16")]; + tensor input_5_pad_type_0 = const()[name = tensor("input_5_pad_type_0"), val = tensor("valid")]; + tensor input_5_strides_0 = const()[name = tensor("input_5_strides_0"), val = tensor([2, 2])]; + tensor input_5_pad_0 = const()[name = tensor("input_5_pad_0"), val = tensor([0, 0, 0, 0])]; + tensor input_5_dilations_0 = const()[name = tensor("input_5_dilations_0"), val = tensor([1, 1])]; + tensor input_5_groups_0 = const()[name = tensor("input_5_groups_0"), val = tensor(1)]; + tensor encoder_pre_encode_conv_0_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; + tensor encoder_pre_encode_conv_0_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4736)))]; + tensor input_5_cast_fp16 = conv(bias = encoder_pre_encode_conv_0_bias_to_fp16, dilations = input_5_dilations_0, groups = input_5_groups_0, pad = input_5_pad_0, pad_type = input_5_pad_type_0, strides = input_5_strides_0, weight = encoder_pre_encode_conv_0_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; + tensor input_7_cast_fp16 = relu(x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; + tensor input_9_pad_0 = const()[name = tensor("input_9_pad_0"), val = tensor([0, 0, 0, 0, 2, 1, 2, 1])]; + tensor input_9_mode_0 = const()[name = tensor("input_9_mode_0"), val = tensor("constant")]; + tensor const_1_to_fp16 = const()[name = tensor("const_1_to_fp16"), val = tensor(0x0p+0)]; + tensor input_9_cast_fp16 = pad(constant_val = const_1_to_fp16, mode = input_9_mode_0, pad = input_9_pad_0, x = input_7_cast_fp16)[name = tensor("input_9_cast_fp16")]; + tensor input_11_pad_type_0 = const()[name = tensor("input_11_pad_type_0"), val = tensor("valid")]; + tensor input_11_strides_0 = const()[name = tensor("input_11_strides_0"), val = tensor([2, 2])]; + tensor input_11_groups_0 = const()[name = tensor("input_11_groups_0"), val = tensor(256)]; + tensor input_11_pad_0 = const()[name = tensor("input_11_pad_0"), val = tensor([0, 0, 0, 0])]; + tensor input_11_dilations_0 = const()[name = tensor("input_11_dilations_0"), val = tensor([1, 1])]; + tensor encoder_pre_encode_conv_2_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5312)))]; + tensor encoder_pre_encode_conv_2_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9984)))]; + tensor input_11_cast_fp16 = conv(bias = encoder_pre_encode_conv_2_bias_to_fp16, dilations = input_11_dilations_0, groups = input_11_groups_0, pad = input_11_pad_0, pad_type = input_11_pad_type_0, strides = input_11_strides_0, weight = encoder_pre_encode_conv_2_weight_to_fp16, x = input_9_cast_fp16)[name = tensor("input_11_cast_fp16")]; + tensor input_13_pad_type_0 = const()[name = tensor("input_13_pad_type_0"), val = tensor("valid")]; + tensor input_13_strides_0 = const()[name = tensor("input_13_strides_0"), val = tensor([1, 1])]; + tensor input_13_pad_0 = const()[name = tensor("input_13_pad_0"), val = tensor([0, 0, 0, 0])]; + tensor input_13_dilations_0 = const()[name = tensor("input_13_dilations_0"), val = tensor([1, 1])]; + tensor input_13_groups_0 = const()[name = tensor("input_13_groups_0"), val = tensor(1)]; + tensor encoder_pre_encode_conv_3_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_3_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10560)))]; + tensor encoder_pre_encode_conv_3_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_3_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141696)))]; + tensor input_13_cast_fp16 = conv(bias = encoder_pre_encode_conv_3_bias_to_fp16, dilations = input_13_dilations_0, groups = input_13_groups_0, pad = input_13_pad_0, pad_type = input_13_pad_type_0, strides = input_13_strides_0, weight = encoder_pre_encode_conv_3_weight_to_fp16, x = input_11_cast_fp16)[name = tensor("input_13_cast_fp16")]; + tensor input_15_cast_fp16 = relu(x = input_13_cast_fp16)[name = tensor("input_15_cast_fp16")]; + tensor input_17_pad_0 = const()[name = tensor("input_17_pad_0"), val = tensor([0, 0, 0, 0, 2, 1, 2, 1])]; + tensor input_17_mode_0 = const()[name = tensor("input_17_mode_0"), val = tensor("constant")]; + tensor const_2_to_fp16 = const()[name = tensor("const_2_to_fp16"), val = tensor(0x0p+0)]; + tensor input_17_cast_fp16 = pad(constant_val = const_2_to_fp16, mode = input_17_mode_0, pad = input_17_pad_0, x = input_15_cast_fp16)[name = tensor("input_17_cast_fp16")]; + tensor input_19_pad_type_0 = const()[name = tensor("input_19_pad_type_0"), val = tensor("valid")]; + tensor input_19_strides_0 = const()[name = tensor("input_19_strides_0"), val = tensor([2, 2])]; + tensor input_19_groups_0 = const()[name = tensor("input_19_groups_0"), val = tensor(256)]; + tensor input_19_pad_0 = const()[name = tensor("input_19_pad_0"), val = tensor([0, 0, 0, 0])]; + tensor input_19_dilations_0 = const()[name = tensor("input_19_dilations_0"), val = tensor([1, 1])]; + tensor encoder_pre_encode_conv_5_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_5_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142272)))]; + tensor encoder_pre_encode_conv_5_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_5_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146944)))]; + tensor input_19_cast_fp16 = conv(bias = encoder_pre_encode_conv_5_bias_to_fp16, dilations = input_19_dilations_0, groups = input_19_groups_0, pad = input_19_pad_0, pad_type = input_19_pad_type_0, strides = input_19_strides_0, weight = encoder_pre_encode_conv_5_weight_to_fp16, x = input_17_cast_fp16)[name = tensor("input_19_cast_fp16")]; + tensor input_21_pad_type_0 = const()[name = tensor("input_21_pad_type_0"), val = tensor("valid")]; + tensor input_21_strides_0 = const()[name = tensor("input_21_strides_0"), val = tensor([1, 1])]; + tensor input_21_pad_0 = const()[name = tensor("input_21_pad_0"), val = tensor([0, 0, 0, 0])]; + tensor input_21_dilations_0 = const()[name = tensor("input_21_dilations_0"), val = tensor([1, 1])]; + tensor input_21_groups_0 = const()[name = tensor("input_21_groups_0"), val = tensor(1)]; + tensor encoder_pre_encode_conv_6_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_6_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147520)))]; + tensor encoder_pre_encode_conv_6_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_6_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(278656)))]; + tensor input_21_cast_fp16 = conv(bias = encoder_pre_encode_conv_6_bias_to_fp16, dilations = input_21_dilations_0, groups = input_21_groups_0, pad = input_21_pad_0, pad_type = input_21_pad_type_0, strides = input_21_strides_0, weight = encoder_pre_encode_conv_6_weight_to_fp16, x = input_19_cast_fp16)[name = tensor("input_21_cast_fp16")]; + tensor x_3_cast_fp16 = relu(x = input_21_cast_fp16)[name = tensor("x_3_cast_fp16")]; + tensor var_181_perm_0 = const()[name = tensor("op_181_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_182 = const()[name = tensor("op_182"), val = tensor([1, 5, -1])]; + tensor var_181_cast_fp16 = transpose(perm = var_181_perm_0, x = x_3_cast_fp16)[name = tensor("transpose_240")]; + tensor input_23_cast_fp16 = reshape(shape = var_182, x = var_181_cast_fp16)[name = tensor("input_23_cast_fp16")]; + tensor encoder_pre_encode_out_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279232)))]; + tensor encoder_pre_encode_out_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4735744)))]; + tensor linear_0_cast_fp16 = linear(bias = encoder_pre_encode_out_bias_to_fp16, weight = encoder_pre_encode_out_weight_to_fp16, x = input_23_cast_fp16)[name = tensor("linear_0_cast_fp16")]; + tensor var_192_begin_0 = const()[name = tensor("op_192_begin_0"), val = tensor([0, 2, 0])]; + tensor var_192_end_0 = const()[name = tensor("op_192_end_0"), val = tensor([1, 5, 512])]; + tensor var_192_end_mask_0 = const()[name = tensor("op_192_end_mask_0"), val = tensor([true, true, true])]; + tensor var_192_cast_fp16 = slice_by_index(begin = var_192_begin_0, end = var_192_end_0, end_mask = var_192_end_mask_0, x = linear_0_cast_fp16)[name = tensor("op_192_cast_fp16")]; + tensor var_194 = const()[name = tensor("op_194"), val = tensor(2)]; + tensor lengths_cast_fp16_to_int32 = cast(dtype = cast_9_dtype_0, x = lengths_cast_fp16)[name = tensor("cast_192")]; + tensor var_195 = sub(x = lengths_cast_fp16_to_int32, y = var_194)[name = tensor("op_195")]; + tensor var_195_promoted_to_fp16_dtype_0 = const()[name = tensor("op_195_promoted_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor var_60_promoted_to_fp16 = const()[name = tensor("op_60_promoted_to_fp16"), val = tensor(0x0p+0)]; + tensor const_5_to_fp16 = const()[name = tensor("const_5_to_fp16"), val = tensor(inf)]; + tensor var_195_to_fp16 = cast(dtype = var_195_promoted_to_fp16_dtype_0, x = var_195)[name = tensor("cast_191")]; + tensor clip_0_cast_fp16 = clip(alpha = var_60_promoted_to_fp16, beta = const_5_to_fp16, x = var_195_to_fp16)[name = tensor("clip_0_cast_fp16")]; + tensor cache_keep_size = const()[name = tensor("cache_keep_size"), val = tensor([1])]; + tensor var_211_promoted_to_fp16 = const()[name = tensor("op_211_promoted_to_fp16"), val = tensor(0x1.18p+6)]; + tensor padding_length_cast_fp16 = add(x = clip_0_cast_fp16, y = var_211_promoted_to_fp16)[name = tensor("padding_length_cast_fp16")]; + tensor const_7 = const()[name = tensor("const_7"), val = tensor(-1)]; + tensor var_213 = mul(x = cache_last_channel_len, y = const_7)[name = tensor("op_213")]; + tensor var_214 = const()[name = tensor("op_214"), val = tensor(70)]; + tensor offset = add(x = var_213, y = var_214)[name = tensor("offset")]; + tensor var_254_axes_0 = const()[name = tensor("op_254_axes_0"), val = tensor([-1])]; + tensor var_254_cast_fp16 = expand_dims(axes = var_254_axes_0, x = padding_length_cast_fp16)[name = tensor("op_254_cast_fp16")]; + tensor var_253_promoted_to_fp16 = const()[name = tensor("op_253_promoted_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4736832)))]; + tensor pad_mask_1_cast_fp16 = less(x = var_253_promoted_to_fp16, y = var_254_cast_fp16)[name = tensor("pad_mask_1_cast_fp16")]; + tensor expand_dims_1 = const()[name = tensor("expand_dims_1"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72]])]; + tensor var_260_axes_0 = const()[name = tensor("op_260_axes_0"), val = tensor([-1])]; + tensor var_260 = expand_dims(axes = var_260_axes_0, x = offset)[name = tensor("op_260")]; + tensor pad_mask_off = greater_equal(x = expand_dims_1, y = var_260)[name = tensor("pad_mask_off")]; + tensor pad_mask_3 = logical_and(x = pad_mask_off, y = pad_mask_1_cast_fp16)[name = tensor("pad_mask_3")]; + tensor var_263_axes_0 = const()[name = tensor("op_263_axes_0"), val = tensor([1])]; + tensor var_263 = expand_dims(axes = var_263_axes_0, x = pad_mask_3)[name = tensor("op_263")]; + tensor var_264 = const()[name = tensor("op_264"), val = tensor([1, 73, 1])]; + tensor pad_mask_for_att_mask_1 = tile(reps = var_264, x = var_263)[name = tensor("pad_mask_for_att_mask_1")]; + tensor var_266_perm_0 = const()[name = tensor("op_266_perm_0"), val = tensor([0, 2, 1])]; + tensor var_266 = transpose(perm = var_266_perm_0, x = pad_mask_for_att_mask_1)[name = tensor("transpose_239")]; + tensor pad_mask_for_att_mask = logical_and(x = pad_mask_for_att_mask_1, y = var_266)[name = tensor("pad_mask_for_att_mask")]; + tensor const_15 = const()[name = tensor("const_15"), val = tensor([[[true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false], [false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true]]])]; + tensor att_mask_9 = logical_and(x = pad_mask_for_att_mask, y = const_15)[name = tensor("att_mask_9")]; + tensor att_mask = logical_not(x = att_mask_9)[name = tensor("att_mask")]; + tensor pad_mask_5 = logical_not(x = pad_mask_3)[name = tensor("pad_mask_5")]; + tensor pad_mask_begin_0 = const()[name = tensor("pad_mask_begin_0"), val = tensor([0, 70])]; + tensor pad_mask_end_0 = const()[name = tensor("pad_mask_end_0"), val = tensor([1, 73])]; + tensor pad_mask_end_mask_0 = const()[name = tensor("pad_mask_end_mask_0"), val = tensor([true, true])]; + tensor pad_mask = slice_by_index(begin = pad_mask_begin_0, end = pad_mask_end_0, end_mask = pad_mask_end_mask_0, x = pad_mask_5)[name = tensor("pad_mask")]; + tensor mask_1_begin_0 = const()[name = tensor("mask_1_begin_0"), val = tensor([0, 70, 0])]; + tensor mask_1_end_0 = const()[name = tensor("mask_1_end_0"), val = tensor([1, 73, 73])]; + tensor mask_1_end_mask_0 = const()[name = tensor("mask_1_end_mask_0"), val = tensor([true, true, true])]; + tensor mask_1 = slice_by_index(begin = mask_1_begin_0, end = mask_1_end_0, end_mask = mask_1_end_mask_0, x = att_mask)[name = tensor("mask_1")]; + tensor cache_1_begin_0 = const()[name = tensor("cache_1_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor cache_1_end_0 = const()[name = tensor("cache_1_end_0"), val = tensor([1, 1, 70, 512])]; + tensor cache_1_end_mask_0 = const()[name = tensor("cache_1_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_1_squeeze_mask_0 = const()[name = tensor("cache_1_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_last_channel_to_fp16_dtype_0 = const()[name = tensor("cache_last_channel_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor cache_last_channel_to_fp16 = cast(dtype = cache_last_channel_to_fp16_dtype_0, x = cache_last_channel)[name = tensor("cast_190")]; + tensor cache_1_cast_fp16 = slice_by_index(begin = cache_1_begin_0, end = cache_1_end_0, end_mask = cache_1_end_mask_0, squeeze_mask = cache_1_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_1_cast_fp16")]; + tensor cache_3_begin_0 = const()[name = tensor("cache_3_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor cache_3_end_0 = const()[name = tensor("cache_3_end_0"), val = tensor([1, 1, 512, 8])]; + tensor cache_3_end_mask_0 = const()[name = tensor("cache_3_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_3_squeeze_mask_0 = const()[name = tensor("cache_3_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_last_time_to_fp16_dtype_0 = const()[name = tensor("cache_last_time_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor cache_last_time_to_fp16 = cast(dtype = cache_last_time_to_fp16_dtype_0, x = cache_last_time)[name = tensor("cast_189")]; + tensor cache_3_cast_fp16 = slice_by_index(begin = cache_3_begin_0, end = cache_3_end_0, end_mask = cache_3_end_mask_0, squeeze_mask = cache_3_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_3_cast_fp16")]; + tensor input_27_axes_0 = const()[name = tensor("input_27_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4737088)))]; + tensor encoder_layers_0_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4738176)))]; + tensor var_38_to_fp16 = const()[name = tensor("op_38_to_fp16"), val = tensor(0x1.5p-17)]; + tensor input_27_cast_fp16 = layer_norm(axes = input_27_axes_0, beta = encoder_layers_0_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_norm_feed_forward1_weight_to_fp16, x = var_192_cast_fp16)[name = tensor("input_27_cast_fp16")]; + tensor encoder_layers_0_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_0_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4739264)))]; + tensor linear_1_bias_0_to_fp16 = const()[name = tensor("linear_1_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6836480)))]; + tensor linear_1_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_0_feed_forward1_linear1_weight_to_fp16, x = input_27_cast_fp16)[name = tensor("linear_1_cast_fp16")]; + tensor input_31_cast_fp16 = silu(x = linear_1_cast_fp16)[name = tensor("input_31_cast_fp16")]; + tensor encoder_layers_0_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_0_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6840640)))]; + tensor linear_2_bias_0_to_fp16 = const()[name = tensor("linear_2_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8937856)))]; + tensor linear_2_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_feed_forward1_linear2_weight_to_fp16, x = input_31_cast_fp16)[name = tensor("linear_2_cast_fp16")]; + tensor var_303_to_fp16 = const()[name = tensor("op_303_to_fp16"), val = tensor(0x1p-1)]; + tensor var_304_cast_fp16 = mul(x = linear_2_cast_fp16, y = var_303_to_fp16)[name = tensor("op_304_cast_fp16")]; + tensor input_37_cast_fp16 = add(x = var_192_cast_fp16, y = var_304_cast_fp16)[name = tensor("input_37_cast_fp16")]; + tensor key_1_axes_0 = const()[name = tensor("key_1_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8938944)))]; + tensor encoder_layers_0_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8940032)))]; + tensor key_1_cast_fp16 = layer_norm(axes = key_1_axes_0, beta = encoder_layers_0_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_norm_self_att_weight_to_fp16, x = input_37_cast_fp16)[name = tensor("key_1_cast_fp16")]; + tensor input_39_interleave_0 = const()[name = tensor("input_39_interleave_0"), val = tensor(false)]; + tensor input_39_cast_fp16 = concat(axis = var_64, interleave = input_39_interleave_0, values = (cache_1_cast_fp16, key_1_cast_fp16))[name = tensor("input_39_cast_fp16")]; + tensor var_326_begin_0 = const()[name = tensor("op_326_begin_0"), val = tensor([0, 1, 0])]; + tensor var_326_end_0 = const()[name = tensor("op_326_end_0"), val = tensor([1, 70, 512])]; + tensor var_326_end_mask_0 = const()[name = tensor("op_326_end_mask_0"), val = tensor([true, true, true])]; + tensor var_326_cast_fp16 = slice_by_index(begin = var_326_begin_0, end = var_326_end_0, end_mask = var_326_end_mask_0, x = cache_1_cast_fp16)[name = tensor("op_326_cast_fp16")]; + tensor var_329_begin_0 = const()[name = tensor("op_329_begin_0"), val = tensor([0, 0, 0])]; + tensor var_329_end_0 = const()[name = tensor("op_329_end_0"), val = tensor([1, 1, 512])]; + tensor var_329_end_mask_0 = const()[name = tensor("op_329_end_mask_0"), val = tensor([true, false, true])]; + tensor var_329_cast_fp16 = slice_by_index(begin = var_329_begin_0, end = var_329_end_0, end_mask = var_329_end_mask_0, x = key_1_cast_fp16)[name = tensor("op_329_cast_fp16")]; + tensor var_332_interleave_0 = const()[name = tensor("op_332_interleave_0"), val = tensor(false)]; + tensor var_332_cast_fp16 = concat(axis = var_64, interleave = var_332_interleave_0, values = (var_326_cast_fp16, var_329_cast_fp16))[name = tensor("op_332_cast_fp16")]; + tensor encoder_layers_0_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8941120)))]; + tensor linear_3_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_self_attn_linear_q_weight_to_fp16, x = key_1_cast_fp16)[name = tensor("linear_3_cast_fp16")]; + tensor var_336 = const()[name = tensor("op_336"), val = tensor([1, -1, 8, 64])]; + tensor q_1_cast_fp16 = reshape(shape = var_336, x = linear_3_cast_fp16)[name = tensor("q_1_cast_fp16")]; + tensor encoder_layers_0_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9465472)))]; + tensor linear_4_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_self_attn_linear_k_weight_to_fp16, x = input_39_cast_fp16)[name = tensor("linear_4_cast_fp16")]; + tensor var_340 = const()[name = tensor("op_340"), val = tensor([1, -1, 8, 64])]; + tensor k_1_cast_fp16 = reshape(shape = var_340, x = linear_4_cast_fp16)[name = tensor("k_1_cast_fp16")]; + tensor encoder_layers_0_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9989824)))]; + tensor linear_5_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_self_attn_linear_v_weight_to_fp16, x = input_39_cast_fp16)[name = tensor("linear_5_cast_fp16")]; + tensor var_344 = const()[name = tensor("op_344"), val = tensor([1, -1, 8, 64])]; + tensor v_1_cast_fp16 = reshape(shape = var_344, x = linear_5_cast_fp16)[name = tensor("v_1_cast_fp16")]; + tensor value_3_perm_0 = const()[name = tensor("value_3_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_0_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10514176)))]; + tensor var_356_cast_fp16 = add(x = q_1_cast_fp16, y = encoder_layers_0_self_attn_pos_bias_u_to_fp16)[name = tensor("op_356_cast_fp16")]; + tensor encoder_layers_0_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10515264)))]; + tensor var_358_cast_fp16 = add(x = q_1_cast_fp16, y = encoder_layers_0_self_attn_pos_bias_v_to_fp16)[name = tensor("op_358_cast_fp16")]; + tensor q_with_bias_v_1_perm_0 = const()[name = tensor("q_with_bias_v_1_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_7_transpose_x_0 = const()[name = tensor("x_7_transpose_x_0"), val = tensor(false)]; + tensor x_7_transpose_y_0 = const()[name = tensor("x_7_transpose_y_0"), val = tensor(false)]; + tensor var_360_to_fp16 = const()[name = tensor("op_360_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10516352)))]; + tensor q_with_bias_v_1_cast_fp16 = transpose(perm = q_with_bias_v_1_perm_0, x = var_358_cast_fp16)[name = tensor("transpose_237")]; + tensor x_7_cast_fp16 = matmul(transpose_x = x_7_transpose_x_0, transpose_y = x_7_transpose_y_0, x = q_with_bias_v_1_cast_fp16, y = var_360_to_fp16)[name = tensor("x_7_cast_fp16")]; + tensor x_9_pad_0 = const()[name = tensor("x_9_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_9_mode_0 = const()[name = tensor("x_9_mode_0"), val = tensor("constant")]; + tensor const_23_to_fp16 = const()[name = tensor("const_23_to_fp16"), val = tensor(0x0p+0)]; + tensor x_9_cast_fp16 = pad(constant_val = const_23_to_fp16, mode = x_9_mode_0, pad = x_9_pad_0, x = x_7_cast_fp16)[name = tensor("x_9_cast_fp16")]; + tensor var_368 = const()[name = tensor("op_368"), val = tensor([1, 8, -1, 3])]; + tensor x_11_cast_fp16 = reshape(shape = var_368, x = x_9_cast_fp16)[name = tensor("x_11_cast_fp16")]; + tensor var_372_begin_0 = const()[name = tensor("op_372_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_372_end_0 = const()[name = tensor("op_372_end_0"), val = tensor([1, 8, 146, 3])]; + tensor var_372_end_mask_0 = const()[name = tensor("op_372_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_372_cast_fp16 = slice_by_index(begin = var_372_begin_0, end = var_372_end_0, end_mask = var_372_end_mask_0, x = x_11_cast_fp16)[name = tensor("op_372_cast_fp16")]; + tensor var_373 = const()[name = tensor("op_373"), val = tensor([1, 8, 3, 145])]; + tensor matrix_bd_1_cast_fp16 = reshape(shape = var_373, x = var_372_cast_fp16)[name = tensor("matrix_bd_1_cast_fp16")]; + tensor matrix_ac_1_transpose_x_0 = const()[name = tensor("matrix_ac_1_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_1_transpose_y_0 = const()[name = tensor("matrix_ac_1_transpose_y_0"), val = tensor(false)]; + tensor transpose_51_perm_0 = const()[name = tensor("transpose_51_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_52_perm_0 = const()[name = tensor("transpose_52_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_52 = transpose(perm = transpose_52_perm_0, x = k_1_cast_fp16)[name = tensor("transpose_235")]; + tensor transpose_51 = transpose(perm = transpose_51_perm_0, x = var_356_cast_fp16)[name = tensor("transpose_236")]; + tensor matrix_ac_1_cast_fp16 = matmul(transpose_x = matrix_ac_1_transpose_x_0, transpose_y = matrix_ac_1_transpose_y_0, x = transpose_51, y = transpose_52)[name = tensor("matrix_ac_1_cast_fp16")]; + tensor matrix_bd_3_begin_0 = const()[name = tensor("matrix_bd_3_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_3_end_0 = const()[name = tensor("matrix_bd_3_end_0"), val = tensor([1, 8, 3, 73])]; + tensor matrix_bd_3_end_mask_0 = const()[name = tensor("matrix_bd_3_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_3_cast_fp16 = slice_by_index(begin = matrix_bd_3_begin_0, end = matrix_bd_3_end_0, end_mask = matrix_bd_3_end_mask_0, x = matrix_bd_1_cast_fp16)[name = tensor("matrix_bd_3_cast_fp16")]; + tensor var_382_cast_fp16 = add(x = matrix_ac_1_cast_fp16, y = matrix_bd_3_cast_fp16)[name = tensor("op_382_cast_fp16")]; + tensor _inversed_scores_1_y_0_to_fp16 = const()[name = tensor("_inversed_scores_1_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_1_cast_fp16 = mul(x = var_382_cast_fp16, y = _inversed_scores_1_y_0_to_fp16)[name = tensor("_inversed_scores_1_cast_fp16")]; + tensor mask_3_axes_0 = const()[name = tensor("mask_3_axes_0"), val = tensor([1])]; + tensor mask_3 = expand_dims(axes = mask_3_axes_0, x = mask_1)[name = tensor("mask_3")]; + tensor var_41_to_fp16 = const()[name = tensor("op_41_to_fp16"), val = tensor(-0x1.388p+13)]; + tensor scores_3_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_1_cast_fp16, cond = mask_3)[name = tensor("scores_3_cast_fp16")]; + tensor var_388_cast_fp16 = softmax(axis = var_62, x = scores_3_cast_fp16)[name = tensor("op_388_cast_fp16")]; + tensor var_40_to_fp16 = const()[name = tensor("op_40_to_fp16"), val = tensor(0x0p+0)]; + tensor input_41_cast_fp16 = select(a = var_40_to_fp16, b = var_388_cast_fp16, cond = mask_3)[name = tensor("input_41_cast_fp16")]; + tensor x_13_transpose_x_0 = const()[name = tensor("x_13_transpose_x_0"), val = tensor(false)]; + tensor x_13_transpose_y_0 = const()[name = tensor("x_13_transpose_y_0"), val = tensor(false)]; + tensor value_3_cast_fp16 = transpose(perm = value_3_perm_0, x = v_1_cast_fp16)[name = tensor("transpose_238")]; + tensor x_13_cast_fp16 = matmul(transpose_x = x_13_transpose_x_0, transpose_y = x_13_transpose_y_0, x = input_41_cast_fp16, y = value_3_cast_fp16)[name = tensor("x_13_cast_fp16")]; + tensor var_392_perm_0 = const()[name = tensor("op_392_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_393 = const()[name = tensor("op_393"), val = tensor([1, -1, 512])]; + tensor var_392_cast_fp16 = transpose(perm = var_392_perm_0, x = x_13_cast_fp16)[name = tensor("transpose_234")]; + tensor input_43_cast_fp16 = reshape(shape = var_393, x = var_392_cast_fp16)[name = tensor("input_43_cast_fp16")]; + tensor encoder_layers_0_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10664896)))]; + tensor linear_7_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_self_attn_linear_out_weight_to_fp16, x = input_43_cast_fp16)[name = tensor("linear_7_cast_fp16")]; + tensor input_47_cast_fp16 = add(x = input_37_cast_fp16, y = linear_7_cast_fp16)[name = tensor("input_47_cast_fp16")]; + tensor x_17_axes_0 = const()[name = tensor("x_17_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11189248)))]; + tensor encoder_layers_0_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11190336)))]; + tensor x_17_cast_fp16 = layer_norm(axes = x_17_axes_0, beta = encoder_layers_0_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_norm_conv_weight_to_fp16, x = input_47_cast_fp16)[name = tensor("x_17_cast_fp16")]; + tensor input_49_perm_0 = const()[name = tensor("input_49_perm_0"), val = tensor([0, 2, 1])]; + tensor input_51_pad_type_0 = const()[name = tensor("input_51_pad_type_0"), val = tensor("valid")]; + tensor input_51_strides_0 = const()[name = tensor("input_51_strides_0"), val = tensor([1])]; + tensor input_51_pad_0 = const()[name = tensor("input_51_pad_0"), val = tensor([0, 0])]; + tensor input_51_dilations_0 = const()[name = tensor("input_51_dilations_0"), val = tensor([1])]; + tensor input_51_groups_0 = const()[name = tensor("input_51_groups_0"), val = tensor(1)]; + tensor encoder_layers_0_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_0_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11191424)))]; + tensor input_49_cast_fp16 = transpose(perm = input_49_perm_0, x = x_17_cast_fp16)[name = tensor("transpose_233")]; + tensor input_51_cast_fp16 = conv(dilations = input_51_dilations_0, groups = input_51_groups_0, pad = input_51_pad_0, pad_type = input_51_pad_type_0, strides = input_51_strides_0, weight = encoder_layers_0_conv_pointwise_conv1_weight_to_fp16, x = input_49_cast_fp16)[name = tensor("input_51_cast_fp16")]; + tensor x_19_split_num_splits_0 = const()[name = tensor("x_19_split_num_splits_0"), val = tensor(2)]; + tensor x_19_split_axis_0 = const()[name = tensor("x_19_split_axis_0"), val = tensor(1)]; + tensor x_19_split_cast_fp16_0, tensor x_19_split_cast_fp16_1 = split(axis = x_19_split_axis_0, num_splits = x_19_split_num_splits_0, x = input_51_cast_fp16)[name = tensor("x_19_split_cast_fp16")]; + tensor x_19_split_1_sigmoid_cast_fp16 = sigmoid(x = x_19_split_cast_fp16_1)[name = tensor("x_19_split_1_sigmoid_cast_fp16")]; + tensor x_19_cast_fp16 = mul(x = x_19_split_cast_fp16_0, y = x_19_split_1_sigmoid_cast_fp16)[name = tensor("x_19_cast_fp16")]; + tensor var_418_axes_0 = const()[name = tensor("op_418_axes_0"), val = tensor([1])]; + tensor var_418 = expand_dims(axes = var_418_axes_0, x = pad_mask)[name = tensor("op_418")]; + tensor input_53_cast_fp16 = select(a = var_40_to_fp16, b = x_19_cast_fp16, cond = var_418)[name = tensor("input_53_cast_fp16")]; + tensor new_x_3_interleave_0 = const()[name = tensor("new_x_3_interleave_0"), val = tensor(false)]; + tensor new_x_3_cast_fp16 = concat(axis = var_62, interleave = new_x_3_interleave_0, values = (cache_3_cast_fp16, input_53_cast_fp16))[name = tensor("new_x_3_cast_fp16")]; + tensor next_cache_1_begin_0 = const()[name = tensor("next_cache_1_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_1_end_0 = const()[name = tensor("next_cache_1_end_0"), val = tensor([1, 512, 9])]; + tensor next_cache_1_end_mask_0 = const()[name = tensor("next_cache_1_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_1_cast_fp16 = slice_by_index(begin = next_cache_1_begin_0, end = next_cache_1_end_0, end_mask = next_cache_1_end_mask_0, x = new_x_3_cast_fp16)[name = tensor("next_cache_1_cast_fp16")]; + tensor var_434_begin_0 = const()[name = tensor("op_434_begin_0"), val = tensor([0, 0, 1])]; + tensor var_434_end_0 = const()[name = tensor("op_434_end_0"), val = tensor([1, 512, 9])]; + tensor var_434_end_mask_0 = const()[name = tensor("op_434_end_mask_0"), val = tensor([true, true, true])]; + tensor var_434_cast_fp16 = slice_by_index(begin = var_434_begin_0, end = var_434_end_0, end_mask = var_434_end_mask_0, x = next_cache_1_cast_fp16)[name = tensor("op_434_cast_fp16")]; + tensor x_21_pad_type_0 = const()[name = tensor("x_21_pad_type_0"), val = tensor("valid")]; + tensor x_21_groups_0 = const()[name = tensor("x_21_groups_0"), val = tensor(512)]; + tensor x_21_strides_0 = const()[name = tensor("x_21_strides_0"), val = tensor([1])]; + tensor x_21_pad_0 = const()[name = tensor("x_21_pad_0"), val = tensor([0, 0])]; + tensor x_21_dilations_0 = const()[name = tensor("x_21_dilations_0"), val = tensor([1])]; + tensor encoder_layers_0_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_0_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12240064)))]; + tensor x_21_cast_fp16 = conv(dilations = x_21_dilations_0, groups = x_21_groups_0, pad = x_21_pad_0, pad_type = x_21_pad_type_0, strides = x_21_strides_0, weight = encoder_layers_0_conv_depthwise_conv_weight_to_fp16, x = new_x_3_cast_fp16)[name = tensor("x_21_cast_fp16")]; + tensor input_55_perm_0 = const()[name = tensor("input_55_perm_0"), val = tensor([0, 2, 1])]; + tensor x_23_axes_0 = const()[name = tensor("x_23_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_0_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12249344)))]; + tensor encoder_layers_0_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_0_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12250432)))]; + tensor input_55_cast_fp16 = transpose(perm = input_55_perm_0, x = x_21_cast_fp16)[name = tensor("transpose_232")]; + tensor x_23_cast_fp16 = layer_norm(axes = x_23_axes_0, beta = encoder_layers_0_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_conv_batch_norm_weight_to_fp16, x = input_55_cast_fp16)[name = tensor("x_23_cast_fp16")]; + tensor input_57_perm_0 = const()[name = tensor("input_57_perm_0"), val = tensor([0, 2, 1])]; + tensor input_57_cast_fp16 = transpose(perm = input_57_perm_0, x = x_23_cast_fp16)[name = tensor("transpose_231")]; + tensor input_59_cast_fp16 = silu(x = input_57_cast_fp16)[name = tensor("input_59_cast_fp16")]; + tensor x_25_pad_type_0 = const()[name = tensor("x_25_pad_type_0"), val = tensor("valid")]; + tensor x_25_strides_0 = const()[name = tensor("x_25_strides_0"), val = tensor([1])]; + tensor x_25_pad_0 = const()[name = tensor("x_25_pad_0"), val = tensor([0, 0])]; + tensor x_25_dilations_0 = const()[name = tensor("x_25_dilations_0"), val = tensor([1])]; + tensor x_25_groups_0 = const()[name = tensor("x_25_groups_0"), val = tensor(1)]; + tensor encoder_layers_0_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_0_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12251520)))]; + tensor x_25_cast_fp16 = conv(dilations = x_25_dilations_0, groups = x_25_groups_0, pad = x_25_pad_0, pad_type = x_25_pad_type_0, strides = x_25_strides_0, weight = encoder_layers_0_conv_pointwise_conv2_weight_to_fp16, x = input_59_cast_fp16)[name = tensor("x_25_cast_fp16")]; + tensor input_61_perm_0 = const()[name = tensor("input_61_perm_0"), val = tensor([0, 2, 1])]; + tensor input_61_cast_fp16 = transpose(perm = input_61_perm_0, x = x_25_cast_fp16)[name = tensor("transpose_230")]; + tensor input_63_cast_fp16 = add(x = input_47_cast_fp16, y = input_61_cast_fp16)[name = tensor("input_63_cast_fp16")]; + tensor input_65_axes_0 = const()[name = tensor("input_65_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12775872)))]; + tensor encoder_layers_0_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12776960)))]; + tensor input_65_cast_fp16 = layer_norm(axes = input_65_axes_0, beta = encoder_layers_0_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_norm_feed_forward2_weight_to_fp16, x = input_63_cast_fp16)[name = tensor("input_65_cast_fp16")]; + tensor encoder_layers_0_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_0_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12778048)))]; + tensor linear_8_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_0_feed_forward2_linear1_weight_to_fp16, x = input_65_cast_fp16)[name = tensor("linear_8_cast_fp16")]; + tensor input_69_cast_fp16 = silu(x = linear_8_cast_fp16)[name = tensor("input_69_cast_fp16")]; + tensor encoder_layers_0_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_0_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14875264)))]; + tensor linear_9_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_feed_forward2_linear2_weight_to_fp16, x = input_69_cast_fp16)[name = tensor("linear_9_cast_fp16")]; + tensor var_475_to_fp16 = const()[name = tensor("op_475_to_fp16"), val = tensor(0x1p-1)]; + tensor var_476_cast_fp16 = mul(x = linear_9_cast_fp16, y = var_475_to_fp16)[name = tensor("op_476_cast_fp16")]; + tensor input_75_cast_fp16 = add(x = input_63_cast_fp16, y = var_476_cast_fp16)[name = tensor("input_75_cast_fp16")]; + tensor input_77_axes_0 = const()[name = tensor("input_77_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16972480)))]; + tensor encoder_layers_0_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16973568)))]; + tensor input_77_cast_fp16 = layer_norm(axes = input_77_axes_0, beta = encoder_layers_0_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_norm_out_weight_to_fp16, x = input_75_cast_fp16)[name = tensor("input_77_cast_fp16")]; + tensor cache_5_begin_0 = const()[name = tensor("cache_5_begin_0"), val = tensor([1, 0, 0, 0])]; + tensor cache_5_end_0 = const()[name = tensor("cache_5_end_0"), val = tensor([2, 1, 70, 512])]; + tensor cache_5_end_mask_0 = const()[name = tensor("cache_5_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_5_squeeze_mask_0 = const()[name = tensor("cache_5_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_5_cast_fp16 = slice_by_index(begin = cache_5_begin_0, end = cache_5_end_0, end_mask = cache_5_end_mask_0, squeeze_mask = cache_5_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_5_cast_fp16")]; + tensor cache_7_begin_0 = const()[name = tensor("cache_7_begin_0"), val = tensor([1, 0, 0, 0])]; + tensor cache_7_end_0 = const()[name = tensor("cache_7_end_0"), val = tensor([2, 1, 512, 8])]; + tensor cache_7_end_mask_0 = const()[name = tensor("cache_7_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_7_squeeze_mask_0 = const()[name = tensor("cache_7_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_7_cast_fp16 = slice_by_index(begin = cache_7_begin_0, end = cache_7_end_0, end_mask = cache_7_end_mask_0, squeeze_mask = cache_7_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_7_cast_fp16")]; + tensor input_79_axes_0 = const()[name = tensor("input_79_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16974656)))]; + tensor encoder_layers_1_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16975744)))]; + tensor input_79_cast_fp16 = layer_norm(axes = input_79_axes_0, beta = encoder_layers_1_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_norm_feed_forward1_weight_to_fp16, x = input_77_cast_fp16)[name = tensor("input_79_cast_fp16")]; + tensor encoder_layers_1_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_1_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16976832)))]; + tensor linear_10_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_1_feed_forward1_linear1_weight_to_fp16, x = input_79_cast_fp16)[name = tensor("linear_10_cast_fp16")]; + tensor input_83_cast_fp16 = silu(x = linear_10_cast_fp16)[name = tensor("input_83_cast_fp16")]; + tensor encoder_layers_1_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_1_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19074048)))]; + tensor linear_11_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_feed_forward1_linear2_weight_to_fp16, x = input_83_cast_fp16)[name = tensor("linear_11_cast_fp16")]; + tensor var_510_to_fp16 = const()[name = tensor("op_510_to_fp16"), val = tensor(0x1p-1)]; + tensor var_511_cast_fp16 = mul(x = linear_11_cast_fp16, y = var_510_to_fp16)[name = tensor("op_511_cast_fp16")]; + tensor input_89_cast_fp16 = add(x = input_77_cast_fp16, y = var_511_cast_fp16)[name = tensor("input_89_cast_fp16")]; + tensor key_3_axes_0 = const()[name = tensor("key_3_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21171264)))]; + tensor encoder_layers_1_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21172352)))]; + tensor key_3_cast_fp16 = layer_norm(axes = key_3_axes_0, beta = encoder_layers_1_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_norm_self_att_weight_to_fp16, x = input_89_cast_fp16)[name = tensor("key_3_cast_fp16")]; + tensor input_91_interleave_0 = const()[name = tensor("input_91_interleave_0"), val = tensor(false)]; + tensor input_91_cast_fp16 = concat(axis = var_64, interleave = input_91_interleave_0, values = (cache_5_cast_fp16, key_3_cast_fp16))[name = tensor("input_91_cast_fp16")]; + tensor var_533_begin_0 = const()[name = tensor("op_533_begin_0"), val = tensor([0, 1, 0])]; + tensor var_533_end_0 = const()[name = tensor("op_533_end_0"), val = tensor([1, 70, 512])]; + tensor var_533_end_mask_0 = const()[name = tensor("op_533_end_mask_0"), val = tensor([true, true, true])]; + tensor var_533_cast_fp16 = slice_by_index(begin = var_533_begin_0, end = var_533_end_0, end_mask = var_533_end_mask_0, x = cache_5_cast_fp16)[name = tensor("op_533_cast_fp16")]; + tensor var_536_begin_0 = const()[name = tensor("op_536_begin_0"), val = tensor([0, 0, 0])]; + tensor var_536_end_0 = const()[name = tensor("op_536_end_0"), val = tensor([1, 1, 512])]; + tensor var_536_end_mask_0 = const()[name = tensor("op_536_end_mask_0"), val = tensor([true, false, true])]; + tensor var_536_cast_fp16 = slice_by_index(begin = var_536_begin_0, end = var_536_end_0, end_mask = var_536_end_mask_0, x = key_3_cast_fp16)[name = tensor("op_536_cast_fp16")]; + tensor var_539_interleave_0 = const()[name = tensor("op_539_interleave_0"), val = tensor(false)]; + tensor var_539_cast_fp16 = concat(axis = var_64, interleave = var_539_interleave_0, values = (var_533_cast_fp16, var_536_cast_fp16))[name = tensor("op_539_cast_fp16")]; + tensor encoder_layers_1_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21173440)))]; + tensor linear_12_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_self_attn_linear_q_weight_to_fp16, x = key_3_cast_fp16)[name = tensor("linear_12_cast_fp16")]; + tensor var_543 = const()[name = tensor("op_543"), val = tensor([1, -1, 8, 64])]; + tensor q_7_cast_fp16 = reshape(shape = var_543, x = linear_12_cast_fp16)[name = tensor("q_7_cast_fp16")]; + tensor encoder_layers_1_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21697792)))]; + tensor linear_13_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_self_attn_linear_k_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("linear_13_cast_fp16")]; + tensor var_547 = const()[name = tensor("op_547"), val = tensor([1, -1, 8, 64])]; + tensor k_5_cast_fp16 = reshape(shape = var_547, x = linear_13_cast_fp16)[name = tensor("k_5_cast_fp16")]; + tensor encoder_layers_1_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22222144)))]; + tensor linear_14_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_self_attn_linear_v_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("linear_14_cast_fp16")]; + tensor var_551 = const()[name = tensor("op_551"), val = tensor([1, -1, 8, 64])]; + tensor v_3_cast_fp16 = reshape(shape = var_551, x = linear_14_cast_fp16)[name = tensor("v_3_cast_fp16")]; + tensor value_5_perm_0 = const()[name = tensor("value_5_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_1_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22746496)))]; + tensor var_563_cast_fp16 = add(x = q_7_cast_fp16, y = encoder_layers_1_self_attn_pos_bias_u_to_fp16)[name = tensor("op_563_cast_fp16")]; + tensor encoder_layers_1_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22747584)))]; + tensor var_565_cast_fp16 = add(x = q_7_cast_fp16, y = encoder_layers_1_self_attn_pos_bias_v_to_fp16)[name = tensor("op_565_cast_fp16")]; + tensor q_with_bias_v_3_perm_0 = const()[name = tensor("q_with_bias_v_3_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_33_transpose_x_0 = const()[name = tensor("x_33_transpose_x_0"), val = tensor(false)]; + tensor x_33_transpose_y_0 = const()[name = tensor("x_33_transpose_y_0"), val = tensor(false)]; + tensor var_567_to_fp16 = const()[name = tensor("op_567_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22748672)))]; + tensor q_with_bias_v_3_cast_fp16 = transpose(perm = q_with_bias_v_3_perm_0, x = var_565_cast_fp16)[name = tensor("transpose_228")]; + tensor x_33_cast_fp16 = matmul(transpose_x = x_33_transpose_x_0, transpose_y = x_33_transpose_y_0, x = q_with_bias_v_3_cast_fp16, y = var_567_to_fp16)[name = tensor("x_33_cast_fp16")]; + tensor x_35_pad_0 = const()[name = tensor("x_35_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_35_mode_0 = const()[name = tensor("x_35_mode_0"), val = tensor("constant")]; + tensor const_36_to_fp16 = const()[name = tensor("const_36_to_fp16"), val = tensor(0x0p+0)]; + tensor x_35_cast_fp16 = pad(constant_val = const_36_to_fp16, mode = x_35_mode_0, pad = x_35_pad_0, x = x_33_cast_fp16)[name = tensor("x_35_cast_fp16")]; + tensor var_575 = const()[name = tensor("op_575"), val = tensor([1, 8, -1, 3])]; + tensor x_37_cast_fp16 = reshape(shape = var_575, x = x_35_cast_fp16)[name = tensor("x_37_cast_fp16")]; + tensor var_579_begin_0 = const()[name = tensor("op_579_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_579_end_0 = const()[name = tensor("op_579_end_0"), val = tensor([1, 8, 146, 3])]; + tensor var_579_end_mask_0 = const()[name = tensor("op_579_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_579_cast_fp16 = slice_by_index(begin = var_579_begin_0, end = var_579_end_0, end_mask = var_579_end_mask_0, x = x_37_cast_fp16)[name = tensor("op_579_cast_fp16")]; + tensor var_580 = const()[name = tensor("op_580"), val = tensor([1, 8, 3, 145])]; + tensor matrix_bd_5_cast_fp16 = reshape(shape = var_580, x = var_579_cast_fp16)[name = tensor("matrix_bd_5_cast_fp16")]; + tensor matrix_ac_3_transpose_x_0 = const()[name = tensor("matrix_ac_3_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_3_transpose_y_0 = const()[name = tensor("matrix_ac_3_transpose_y_0"), val = tensor(false)]; + tensor transpose_53_perm_0 = const()[name = tensor("transpose_53_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_54_perm_0 = const()[name = tensor("transpose_54_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_54 = transpose(perm = transpose_54_perm_0, x = k_5_cast_fp16)[name = tensor("transpose_226")]; + tensor transpose_53 = transpose(perm = transpose_53_perm_0, x = var_563_cast_fp16)[name = tensor("transpose_227")]; + tensor matrix_ac_3_cast_fp16 = matmul(transpose_x = matrix_ac_3_transpose_x_0, transpose_y = matrix_ac_3_transpose_y_0, x = transpose_53, y = transpose_54)[name = tensor("matrix_ac_3_cast_fp16")]; + tensor matrix_bd_7_begin_0 = const()[name = tensor("matrix_bd_7_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_7_end_0 = const()[name = tensor("matrix_bd_7_end_0"), val = tensor([1, 8, 3, 73])]; + tensor matrix_bd_7_end_mask_0 = const()[name = tensor("matrix_bd_7_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_7_cast_fp16 = slice_by_index(begin = matrix_bd_7_begin_0, end = matrix_bd_7_end_0, end_mask = matrix_bd_7_end_mask_0, x = matrix_bd_5_cast_fp16)[name = tensor("matrix_bd_7_cast_fp16")]; + tensor var_589_cast_fp16 = add(x = matrix_ac_3_cast_fp16, y = matrix_bd_7_cast_fp16)[name = tensor("op_589_cast_fp16")]; + tensor _inversed_scores_5_y_0_to_fp16 = const()[name = tensor("_inversed_scores_5_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_5_cast_fp16 = mul(x = var_589_cast_fp16, y = _inversed_scores_5_y_0_to_fp16)[name = tensor("_inversed_scores_5_cast_fp16")]; + tensor scores_7_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_5_cast_fp16, cond = mask_3)[name = tensor("scores_7_cast_fp16")]; + tensor var_595_cast_fp16 = softmax(axis = var_62, x = scores_7_cast_fp16)[name = tensor("op_595_cast_fp16")]; + tensor input_93_cast_fp16 = select(a = var_40_to_fp16, b = var_595_cast_fp16, cond = mask_3)[name = tensor("input_93_cast_fp16")]; + tensor x_39_transpose_x_0 = const()[name = tensor("x_39_transpose_x_0"), val = tensor(false)]; + tensor x_39_transpose_y_0 = const()[name = tensor("x_39_transpose_y_0"), val = tensor(false)]; + tensor value_5_cast_fp16 = transpose(perm = value_5_perm_0, x = v_3_cast_fp16)[name = tensor("transpose_229")]; + tensor x_39_cast_fp16 = matmul(transpose_x = x_39_transpose_x_0, transpose_y = x_39_transpose_y_0, x = input_93_cast_fp16, y = value_5_cast_fp16)[name = tensor("x_39_cast_fp16")]; + tensor var_599_perm_0 = const()[name = tensor("op_599_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_600 = const()[name = tensor("op_600"), val = tensor([1, -1, 512])]; + tensor var_599_cast_fp16 = transpose(perm = var_599_perm_0, x = x_39_cast_fp16)[name = tensor("transpose_225")]; + tensor input_95_cast_fp16 = reshape(shape = var_600, x = var_599_cast_fp16)[name = tensor("input_95_cast_fp16")]; + tensor encoder_layers_1_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22897216)))]; + tensor linear_16_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_self_attn_linear_out_weight_to_fp16, x = input_95_cast_fp16)[name = tensor("linear_16_cast_fp16")]; + tensor input_99_cast_fp16 = add(x = input_89_cast_fp16, y = linear_16_cast_fp16)[name = tensor("input_99_cast_fp16")]; + tensor x_43_axes_0 = const()[name = tensor("x_43_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23421568)))]; + tensor encoder_layers_1_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23422656)))]; + tensor x_43_cast_fp16 = layer_norm(axes = x_43_axes_0, beta = encoder_layers_1_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_norm_conv_weight_to_fp16, x = input_99_cast_fp16)[name = tensor("x_43_cast_fp16")]; + tensor input_101_perm_0 = const()[name = tensor("input_101_perm_0"), val = tensor([0, 2, 1])]; + tensor input_103_pad_type_0 = const()[name = tensor("input_103_pad_type_0"), val = tensor("valid")]; + tensor input_103_strides_0 = const()[name = tensor("input_103_strides_0"), val = tensor([1])]; + tensor input_103_pad_0 = const()[name = tensor("input_103_pad_0"), val = tensor([0, 0])]; + tensor input_103_dilations_0 = const()[name = tensor("input_103_dilations_0"), val = tensor([1])]; + tensor input_103_groups_0 = const()[name = tensor("input_103_groups_0"), val = tensor(1)]; + tensor encoder_layers_1_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_1_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23423744)))]; + tensor input_101_cast_fp16 = transpose(perm = input_101_perm_0, x = x_43_cast_fp16)[name = tensor("transpose_224")]; + tensor input_103_cast_fp16 = conv(dilations = input_103_dilations_0, groups = input_103_groups_0, pad = input_103_pad_0, pad_type = input_103_pad_type_0, strides = input_103_strides_0, weight = encoder_layers_1_conv_pointwise_conv1_weight_to_fp16, x = input_101_cast_fp16)[name = tensor("input_103_cast_fp16")]; + tensor x_45_split_num_splits_0 = const()[name = tensor("x_45_split_num_splits_0"), val = tensor(2)]; + tensor x_45_split_axis_0 = const()[name = tensor("x_45_split_axis_0"), val = tensor(1)]; + tensor x_45_split_cast_fp16_0, tensor x_45_split_cast_fp16_1 = split(axis = x_45_split_axis_0, num_splits = x_45_split_num_splits_0, x = input_103_cast_fp16)[name = tensor("x_45_split_cast_fp16")]; + tensor x_45_split_1_sigmoid_cast_fp16 = sigmoid(x = x_45_split_cast_fp16_1)[name = tensor("x_45_split_1_sigmoid_cast_fp16")]; + tensor x_45_cast_fp16 = mul(x = x_45_split_cast_fp16_0, y = x_45_split_1_sigmoid_cast_fp16)[name = tensor("x_45_cast_fp16")]; + tensor input_105_cast_fp16 = select(a = var_40_to_fp16, b = x_45_cast_fp16, cond = var_418)[name = tensor("input_105_cast_fp16")]; + tensor new_x_7_interleave_0 = const()[name = tensor("new_x_7_interleave_0"), val = tensor(false)]; + tensor new_x_7_cast_fp16 = concat(axis = var_62, interleave = new_x_7_interleave_0, values = (cache_7_cast_fp16, input_105_cast_fp16))[name = tensor("new_x_7_cast_fp16")]; + tensor next_cache_3_begin_0 = const()[name = tensor("next_cache_3_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_3_end_0 = const()[name = tensor("next_cache_3_end_0"), val = tensor([1, 512, 9])]; + tensor next_cache_3_end_mask_0 = const()[name = tensor("next_cache_3_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_3_cast_fp16 = slice_by_index(begin = next_cache_3_begin_0, end = next_cache_3_end_0, end_mask = next_cache_3_end_mask_0, x = new_x_7_cast_fp16)[name = tensor("next_cache_3_cast_fp16")]; + tensor var_641_begin_0 = const()[name = tensor("op_641_begin_0"), val = tensor([0, 0, 1])]; + tensor var_641_end_0 = const()[name = tensor("op_641_end_0"), val = tensor([1, 512, 9])]; + tensor var_641_end_mask_0 = const()[name = tensor("op_641_end_mask_0"), val = tensor([true, true, true])]; + tensor var_641_cast_fp16 = slice_by_index(begin = var_641_begin_0, end = var_641_end_0, end_mask = var_641_end_mask_0, x = next_cache_3_cast_fp16)[name = tensor("op_641_cast_fp16")]; + tensor x_47_pad_type_0 = const()[name = tensor("x_47_pad_type_0"), val = tensor("valid")]; + tensor x_47_groups_0 = const()[name = tensor("x_47_groups_0"), val = tensor(512)]; + tensor x_47_strides_0 = const()[name = tensor("x_47_strides_0"), val = tensor([1])]; + tensor x_47_pad_0 = const()[name = tensor("x_47_pad_0"), val = tensor([0, 0])]; + tensor x_47_dilations_0 = const()[name = tensor("x_47_dilations_0"), val = tensor([1])]; + tensor encoder_layers_1_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_1_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24472384)))]; + tensor x_47_cast_fp16 = conv(dilations = x_47_dilations_0, groups = x_47_groups_0, pad = x_47_pad_0, pad_type = x_47_pad_type_0, strides = x_47_strides_0, weight = encoder_layers_1_conv_depthwise_conv_weight_to_fp16, x = new_x_7_cast_fp16)[name = tensor("x_47_cast_fp16")]; + tensor input_107_perm_0 = const()[name = tensor("input_107_perm_0"), val = tensor([0, 2, 1])]; + tensor x_49_axes_0 = const()[name = tensor("x_49_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_1_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24481664)))]; + tensor encoder_layers_1_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_1_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24482752)))]; + tensor input_107_cast_fp16 = transpose(perm = input_107_perm_0, x = x_47_cast_fp16)[name = tensor("transpose_223")]; + tensor x_49_cast_fp16 = layer_norm(axes = x_49_axes_0, beta = encoder_layers_1_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_conv_batch_norm_weight_to_fp16, x = input_107_cast_fp16)[name = tensor("x_49_cast_fp16")]; + tensor input_109_perm_0 = const()[name = tensor("input_109_perm_0"), val = tensor([0, 2, 1])]; + tensor input_109_cast_fp16 = transpose(perm = input_109_perm_0, x = x_49_cast_fp16)[name = tensor("transpose_222")]; + tensor input_111_cast_fp16 = silu(x = input_109_cast_fp16)[name = tensor("input_111_cast_fp16")]; + tensor x_51_pad_type_0 = const()[name = tensor("x_51_pad_type_0"), val = tensor("valid")]; + tensor x_51_strides_0 = const()[name = tensor("x_51_strides_0"), val = tensor([1])]; + tensor x_51_pad_0 = const()[name = tensor("x_51_pad_0"), val = tensor([0, 0])]; + tensor x_51_dilations_0 = const()[name = tensor("x_51_dilations_0"), val = tensor([1])]; + tensor x_51_groups_0 = const()[name = tensor("x_51_groups_0"), val = tensor(1)]; + tensor encoder_layers_1_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_1_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24483840)))]; + tensor x_51_cast_fp16 = conv(dilations = x_51_dilations_0, groups = x_51_groups_0, pad = x_51_pad_0, pad_type = x_51_pad_type_0, strides = x_51_strides_0, weight = encoder_layers_1_conv_pointwise_conv2_weight_to_fp16, x = input_111_cast_fp16)[name = tensor("x_51_cast_fp16")]; + tensor input_113_perm_0 = const()[name = tensor("input_113_perm_0"), val = tensor([0, 2, 1])]; + tensor input_113_cast_fp16 = transpose(perm = input_113_perm_0, x = x_51_cast_fp16)[name = tensor("transpose_221")]; + tensor input_115_cast_fp16 = add(x = input_99_cast_fp16, y = input_113_cast_fp16)[name = tensor("input_115_cast_fp16")]; + tensor input_117_axes_0 = const()[name = tensor("input_117_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25008192)))]; + tensor encoder_layers_1_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25009280)))]; + tensor input_117_cast_fp16 = layer_norm(axes = input_117_axes_0, beta = encoder_layers_1_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_norm_feed_forward2_weight_to_fp16, x = input_115_cast_fp16)[name = tensor("input_117_cast_fp16")]; + tensor encoder_layers_1_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_1_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25010368)))]; + tensor linear_17_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_1_feed_forward2_linear1_weight_to_fp16, x = input_117_cast_fp16)[name = tensor("linear_17_cast_fp16")]; + tensor input_121_cast_fp16 = silu(x = linear_17_cast_fp16)[name = tensor("input_121_cast_fp16")]; + tensor encoder_layers_1_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_1_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27107584)))]; + tensor linear_18_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_feed_forward2_linear2_weight_to_fp16, x = input_121_cast_fp16)[name = tensor("linear_18_cast_fp16")]; + tensor var_682_to_fp16 = const()[name = tensor("op_682_to_fp16"), val = tensor(0x1p-1)]; + tensor var_683_cast_fp16 = mul(x = linear_18_cast_fp16, y = var_682_to_fp16)[name = tensor("op_683_cast_fp16")]; + tensor input_127_cast_fp16 = add(x = input_115_cast_fp16, y = var_683_cast_fp16)[name = tensor("input_127_cast_fp16")]; + tensor input_129_axes_0 = const()[name = tensor("input_129_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29204800)))]; + tensor encoder_layers_1_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29205888)))]; + tensor input_129_cast_fp16 = layer_norm(axes = input_129_axes_0, beta = encoder_layers_1_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_norm_out_weight_to_fp16, x = input_127_cast_fp16)[name = tensor("input_129_cast_fp16")]; + tensor cache_9_begin_0 = const()[name = tensor("cache_9_begin_0"), val = tensor([2, 0, 0, 0])]; + tensor cache_9_end_0 = const()[name = tensor("cache_9_end_0"), val = tensor([3, 1, 70, 512])]; + tensor cache_9_end_mask_0 = const()[name = tensor("cache_9_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_9_squeeze_mask_0 = const()[name = tensor("cache_9_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_9_cast_fp16 = slice_by_index(begin = cache_9_begin_0, end = cache_9_end_0, end_mask = cache_9_end_mask_0, squeeze_mask = cache_9_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_9_cast_fp16")]; + tensor cache_11_begin_0 = const()[name = tensor("cache_11_begin_0"), val = tensor([2, 0, 0, 0])]; + tensor cache_11_end_0 = const()[name = tensor("cache_11_end_0"), val = tensor([3, 1, 512, 8])]; + tensor cache_11_end_mask_0 = const()[name = tensor("cache_11_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_11_squeeze_mask_0 = const()[name = tensor("cache_11_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_11_cast_fp16 = slice_by_index(begin = cache_11_begin_0, end = cache_11_end_0, end_mask = cache_11_end_mask_0, squeeze_mask = cache_11_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_11_cast_fp16")]; + tensor input_131_axes_0 = const()[name = tensor("input_131_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29206976)))]; + tensor encoder_layers_2_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29208064)))]; + tensor input_131_cast_fp16 = layer_norm(axes = input_131_axes_0, beta = encoder_layers_2_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_norm_feed_forward1_weight_to_fp16, x = input_129_cast_fp16)[name = tensor("input_131_cast_fp16")]; + tensor encoder_layers_2_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_2_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29209152)))]; + tensor linear_19_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_2_feed_forward1_linear1_weight_to_fp16, x = input_131_cast_fp16)[name = tensor("linear_19_cast_fp16")]; + tensor input_135_cast_fp16 = silu(x = linear_19_cast_fp16)[name = tensor("input_135_cast_fp16")]; + tensor encoder_layers_2_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_2_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31306368)))]; + tensor linear_20_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_feed_forward1_linear2_weight_to_fp16, x = input_135_cast_fp16)[name = tensor("linear_20_cast_fp16")]; + tensor var_717_to_fp16 = const()[name = tensor("op_717_to_fp16"), val = tensor(0x1p-1)]; + tensor var_718_cast_fp16 = mul(x = linear_20_cast_fp16, y = var_717_to_fp16)[name = tensor("op_718_cast_fp16")]; + tensor input_141_cast_fp16 = add(x = input_129_cast_fp16, y = var_718_cast_fp16)[name = tensor("input_141_cast_fp16")]; + tensor key_5_axes_0 = const()[name = tensor("key_5_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33403584)))]; + tensor encoder_layers_2_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33404672)))]; + tensor key_5_cast_fp16 = layer_norm(axes = key_5_axes_0, beta = encoder_layers_2_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_norm_self_att_weight_to_fp16, x = input_141_cast_fp16)[name = tensor("key_5_cast_fp16")]; + tensor input_143_interleave_0 = const()[name = tensor("input_143_interleave_0"), val = tensor(false)]; + tensor input_143_cast_fp16 = concat(axis = var_64, interleave = input_143_interleave_0, values = (cache_9_cast_fp16, key_5_cast_fp16))[name = tensor("input_143_cast_fp16")]; + tensor var_740_begin_0 = const()[name = tensor("op_740_begin_0"), val = tensor([0, 1, 0])]; + tensor var_740_end_0 = const()[name = tensor("op_740_end_0"), val = tensor([1, 70, 512])]; + tensor var_740_end_mask_0 = const()[name = tensor("op_740_end_mask_0"), val = tensor([true, true, true])]; + tensor var_740_cast_fp16 = slice_by_index(begin = var_740_begin_0, end = var_740_end_0, end_mask = var_740_end_mask_0, x = cache_9_cast_fp16)[name = tensor("op_740_cast_fp16")]; + tensor var_743_begin_0 = const()[name = tensor("op_743_begin_0"), val = tensor([0, 0, 0])]; + tensor var_743_end_0 = const()[name = tensor("op_743_end_0"), val = tensor([1, 1, 512])]; + tensor var_743_end_mask_0 = const()[name = tensor("op_743_end_mask_0"), val = tensor([true, false, true])]; + tensor var_743_cast_fp16 = slice_by_index(begin = var_743_begin_0, end = var_743_end_0, end_mask = var_743_end_mask_0, x = key_5_cast_fp16)[name = tensor("op_743_cast_fp16")]; + tensor var_746_interleave_0 = const()[name = tensor("op_746_interleave_0"), val = tensor(false)]; + tensor var_746_cast_fp16 = concat(axis = var_64, interleave = var_746_interleave_0, values = (var_740_cast_fp16, var_743_cast_fp16))[name = tensor("op_746_cast_fp16")]; + tensor encoder_layers_2_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33405760)))]; + tensor linear_21_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_self_attn_linear_q_weight_to_fp16, x = key_5_cast_fp16)[name = tensor("linear_21_cast_fp16")]; + tensor var_750 = const()[name = tensor("op_750"), val = tensor([1, -1, 8, 64])]; + tensor q_13_cast_fp16 = reshape(shape = var_750, x = linear_21_cast_fp16)[name = tensor("q_13_cast_fp16")]; + tensor encoder_layers_2_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33930112)))]; + tensor linear_22_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_self_attn_linear_k_weight_to_fp16, x = input_143_cast_fp16)[name = tensor("linear_22_cast_fp16")]; + tensor var_754 = const()[name = tensor("op_754"), val = tensor([1, -1, 8, 64])]; + tensor k_9_cast_fp16 = reshape(shape = var_754, x = linear_22_cast_fp16)[name = tensor("k_9_cast_fp16")]; + tensor encoder_layers_2_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34454464)))]; + tensor linear_23_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_self_attn_linear_v_weight_to_fp16, x = input_143_cast_fp16)[name = tensor("linear_23_cast_fp16")]; + tensor var_758 = const()[name = tensor("op_758"), val = tensor([1, -1, 8, 64])]; + tensor v_5_cast_fp16 = reshape(shape = var_758, x = linear_23_cast_fp16)[name = tensor("v_5_cast_fp16")]; + tensor value_7_perm_0 = const()[name = tensor("value_7_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_2_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34978816)))]; + tensor var_770_cast_fp16 = add(x = q_13_cast_fp16, y = encoder_layers_2_self_attn_pos_bias_u_to_fp16)[name = tensor("op_770_cast_fp16")]; + tensor encoder_layers_2_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34979904)))]; + tensor var_772_cast_fp16 = add(x = q_13_cast_fp16, y = encoder_layers_2_self_attn_pos_bias_v_to_fp16)[name = tensor("op_772_cast_fp16")]; + tensor q_with_bias_v_5_perm_0 = const()[name = tensor("q_with_bias_v_5_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_59_transpose_x_0 = const()[name = tensor("x_59_transpose_x_0"), val = tensor(false)]; + tensor x_59_transpose_y_0 = const()[name = tensor("x_59_transpose_y_0"), val = tensor(false)]; + tensor var_774_to_fp16 = const()[name = tensor("op_774_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34980992)))]; + tensor q_with_bias_v_5_cast_fp16 = transpose(perm = q_with_bias_v_5_perm_0, x = var_772_cast_fp16)[name = tensor("transpose_219")]; + tensor x_59_cast_fp16 = matmul(transpose_x = x_59_transpose_x_0, transpose_y = x_59_transpose_y_0, x = q_with_bias_v_5_cast_fp16, y = var_774_to_fp16)[name = tensor("x_59_cast_fp16")]; + tensor x_61_pad_0 = const()[name = tensor("x_61_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_61_mode_0 = const()[name = tensor("x_61_mode_0"), val = tensor("constant")]; + tensor const_49_to_fp16 = const()[name = tensor("const_49_to_fp16"), val = tensor(0x0p+0)]; + tensor x_61_cast_fp16 = pad(constant_val = const_49_to_fp16, mode = x_61_mode_0, pad = x_61_pad_0, x = x_59_cast_fp16)[name = tensor("x_61_cast_fp16")]; + tensor var_782 = const()[name = tensor("op_782"), val = tensor([1, 8, -1, 3])]; + tensor x_63_cast_fp16 = reshape(shape = var_782, x = x_61_cast_fp16)[name = tensor("x_63_cast_fp16")]; + tensor var_786_begin_0 = const()[name = tensor("op_786_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_786_end_0 = const()[name = tensor("op_786_end_0"), val = tensor([1, 8, 146, 3])]; + tensor var_786_end_mask_0 = const()[name = tensor("op_786_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_786_cast_fp16 = slice_by_index(begin = var_786_begin_0, end = var_786_end_0, end_mask = var_786_end_mask_0, x = x_63_cast_fp16)[name = tensor("op_786_cast_fp16")]; + tensor var_787 = const()[name = tensor("op_787"), val = tensor([1, 8, 3, 145])]; + tensor matrix_bd_9_cast_fp16 = reshape(shape = var_787, x = var_786_cast_fp16)[name = tensor("matrix_bd_9_cast_fp16")]; + tensor matrix_ac_5_transpose_x_0 = const()[name = tensor("matrix_ac_5_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_5_transpose_y_0 = const()[name = tensor("matrix_ac_5_transpose_y_0"), val = tensor(false)]; + tensor transpose_55_perm_0 = const()[name = tensor("transpose_55_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_56_perm_0 = const()[name = tensor("transpose_56_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_56 = transpose(perm = transpose_56_perm_0, x = k_9_cast_fp16)[name = tensor("transpose_217")]; + tensor transpose_55 = transpose(perm = transpose_55_perm_0, x = var_770_cast_fp16)[name = tensor("transpose_218")]; + tensor matrix_ac_5_cast_fp16 = matmul(transpose_x = matrix_ac_5_transpose_x_0, transpose_y = matrix_ac_5_transpose_y_0, x = transpose_55, y = transpose_56)[name = tensor("matrix_ac_5_cast_fp16")]; + tensor matrix_bd_11_begin_0 = const()[name = tensor("matrix_bd_11_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_11_end_0 = const()[name = tensor("matrix_bd_11_end_0"), val = tensor([1, 8, 3, 73])]; + tensor matrix_bd_11_end_mask_0 = const()[name = tensor("matrix_bd_11_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_11_cast_fp16 = slice_by_index(begin = matrix_bd_11_begin_0, end = matrix_bd_11_end_0, end_mask = matrix_bd_11_end_mask_0, x = matrix_bd_9_cast_fp16)[name = tensor("matrix_bd_11_cast_fp16")]; + tensor var_796_cast_fp16 = add(x = matrix_ac_5_cast_fp16, y = matrix_bd_11_cast_fp16)[name = tensor("op_796_cast_fp16")]; + tensor _inversed_scores_9_y_0_to_fp16 = const()[name = tensor("_inversed_scores_9_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_9_cast_fp16 = mul(x = var_796_cast_fp16, y = _inversed_scores_9_y_0_to_fp16)[name = tensor("_inversed_scores_9_cast_fp16")]; + tensor scores_11_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_9_cast_fp16, cond = mask_3)[name = tensor("scores_11_cast_fp16")]; + tensor var_802_cast_fp16 = softmax(axis = var_62, x = scores_11_cast_fp16)[name = tensor("op_802_cast_fp16")]; + tensor input_145_cast_fp16 = select(a = var_40_to_fp16, b = var_802_cast_fp16, cond = mask_3)[name = tensor("input_145_cast_fp16")]; + tensor x_65_transpose_x_0 = const()[name = tensor("x_65_transpose_x_0"), val = tensor(false)]; + tensor x_65_transpose_y_0 = const()[name = tensor("x_65_transpose_y_0"), val = tensor(false)]; + tensor value_7_cast_fp16 = transpose(perm = value_7_perm_0, x = v_5_cast_fp16)[name = tensor("transpose_220")]; + tensor x_65_cast_fp16 = matmul(transpose_x = x_65_transpose_x_0, transpose_y = x_65_transpose_y_0, x = input_145_cast_fp16, y = value_7_cast_fp16)[name = tensor("x_65_cast_fp16")]; + tensor var_806_perm_0 = const()[name = tensor("op_806_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_807 = const()[name = tensor("op_807"), val = tensor([1, -1, 512])]; + tensor var_806_cast_fp16 = transpose(perm = var_806_perm_0, x = x_65_cast_fp16)[name = tensor("transpose_216")]; + tensor input_147_cast_fp16 = reshape(shape = var_807, x = var_806_cast_fp16)[name = tensor("input_147_cast_fp16")]; + tensor encoder_layers_2_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35129536)))]; + tensor linear_25_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_self_attn_linear_out_weight_to_fp16, x = input_147_cast_fp16)[name = tensor("linear_25_cast_fp16")]; + tensor input_151_cast_fp16 = add(x = input_141_cast_fp16, y = linear_25_cast_fp16)[name = tensor("input_151_cast_fp16")]; + tensor x_69_axes_0 = const()[name = tensor("x_69_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35653888)))]; + tensor encoder_layers_2_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35654976)))]; + tensor x_69_cast_fp16 = layer_norm(axes = x_69_axes_0, beta = encoder_layers_2_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_norm_conv_weight_to_fp16, x = input_151_cast_fp16)[name = tensor("x_69_cast_fp16")]; + tensor input_153_perm_0 = const()[name = tensor("input_153_perm_0"), val = tensor([0, 2, 1])]; + tensor input_155_pad_type_0 = const()[name = tensor("input_155_pad_type_0"), val = tensor("valid")]; + tensor input_155_strides_0 = const()[name = tensor("input_155_strides_0"), val = tensor([1])]; + tensor input_155_pad_0 = const()[name = tensor("input_155_pad_0"), val = tensor([0, 0])]; + tensor input_155_dilations_0 = const()[name = tensor("input_155_dilations_0"), val = tensor([1])]; + tensor input_155_groups_0 = const()[name = tensor("input_155_groups_0"), val = tensor(1)]; + tensor encoder_layers_2_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_2_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35656064)))]; + tensor input_153_cast_fp16 = transpose(perm = input_153_perm_0, x = x_69_cast_fp16)[name = tensor("transpose_215")]; + tensor input_155_cast_fp16 = conv(dilations = input_155_dilations_0, groups = input_155_groups_0, pad = input_155_pad_0, pad_type = input_155_pad_type_0, strides = input_155_strides_0, weight = encoder_layers_2_conv_pointwise_conv1_weight_to_fp16, x = input_153_cast_fp16)[name = tensor("input_155_cast_fp16")]; + tensor x_71_split_num_splits_0 = const()[name = tensor("x_71_split_num_splits_0"), val = tensor(2)]; + tensor x_71_split_axis_0 = const()[name = tensor("x_71_split_axis_0"), val = tensor(1)]; + tensor x_71_split_cast_fp16_0, tensor x_71_split_cast_fp16_1 = split(axis = x_71_split_axis_0, num_splits = x_71_split_num_splits_0, x = input_155_cast_fp16)[name = tensor("x_71_split_cast_fp16")]; + tensor x_71_split_1_sigmoid_cast_fp16 = sigmoid(x = x_71_split_cast_fp16_1)[name = tensor("x_71_split_1_sigmoid_cast_fp16")]; + tensor x_71_cast_fp16 = mul(x = x_71_split_cast_fp16_0, y = x_71_split_1_sigmoid_cast_fp16)[name = tensor("x_71_cast_fp16")]; + tensor input_157_cast_fp16 = select(a = var_40_to_fp16, b = x_71_cast_fp16, cond = var_418)[name = tensor("input_157_cast_fp16")]; + tensor new_x_11_interleave_0 = const()[name = tensor("new_x_11_interleave_0"), val = tensor(false)]; + tensor new_x_11_cast_fp16 = concat(axis = var_62, interleave = new_x_11_interleave_0, values = (cache_11_cast_fp16, input_157_cast_fp16))[name = tensor("new_x_11_cast_fp16")]; + tensor next_cache_5_begin_0 = const()[name = tensor("next_cache_5_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_5_end_0 = const()[name = tensor("next_cache_5_end_0"), val = tensor([1, 512, 9])]; + tensor next_cache_5_end_mask_0 = const()[name = tensor("next_cache_5_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_5_cast_fp16 = slice_by_index(begin = next_cache_5_begin_0, end = next_cache_5_end_0, end_mask = next_cache_5_end_mask_0, x = new_x_11_cast_fp16)[name = tensor("next_cache_5_cast_fp16")]; + tensor var_848_begin_0 = const()[name = tensor("op_848_begin_0"), val = tensor([0, 0, 1])]; + tensor var_848_end_0 = const()[name = tensor("op_848_end_0"), val = tensor([1, 512, 9])]; + tensor var_848_end_mask_0 = const()[name = tensor("op_848_end_mask_0"), val = tensor([true, true, true])]; + tensor var_848_cast_fp16 = slice_by_index(begin = var_848_begin_0, end = var_848_end_0, end_mask = var_848_end_mask_0, x = next_cache_5_cast_fp16)[name = tensor("op_848_cast_fp16")]; + tensor x_73_pad_type_0 = const()[name = tensor("x_73_pad_type_0"), val = tensor("valid")]; + tensor x_73_groups_0 = const()[name = tensor("x_73_groups_0"), val = tensor(512)]; + tensor x_73_strides_0 = const()[name = tensor("x_73_strides_0"), val = tensor([1])]; + tensor x_73_pad_0 = const()[name = tensor("x_73_pad_0"), val = tensor([0, 0])]; + tensor x_73_dilations_0 = const()[name = tensor("x_73_dilations_0"), val = tensor([1])]; + tensor encoder_layers_2_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_2_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36704704)))]; + tensor x_73_cast_fp16 = conv(dilations = x_73_dilations_0, groups = x_73_groups_0, pad = x_73_pad_0, pad_type = x_73_pad_type_0, strides = x_73_strides_0, weight = encoder_layers_2_conv_depthwise_conv_weight_to_fp16, x = new_x_11_cast_fp16)[name = tensor("x_73_cast_fp16")]; + tensor input_159_perm_0 = const()[name = tensor("input_159_perm_0"), val = tensor([0, 2, 1])]; + tensor x_75_axes_0 = const()[name = tensor("x_75_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_2_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36713984)))]; + tensor encoder_layers_2_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_2_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36715072)))]; + tensor input_159_cast_fp16 = transpose(perm = input_159_perm_0, x = x_73_cast_fp16)[name = tensor("transpose_214")]; + tensor x_75_cast_fp16 = layer_norm(axes = x_75_axes_0, beta = encoder_layers_2_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_conv_batch_norm_weight_to_fp16, x = input_159_cast_fp16)[name = tensor("x_75_cast_fp16")]; + tensor input_161_perm_0 = const()[name = tensor("input_161_perm_0"), val = tensor([0, 2, 1])]; + tensor input_161_cast_fp16 = transpose(perm = input_161_perm_0, x = x_75_cast_fp16)[name = tensor("transpose_213")]; + tensor input_163_cast_fp16 = silu(x = input_161_cast_fp16)[name = tensor("input_163_cast_fp16")]; + tensor x_77_pad_type_0 = const()[name = tensor("x_77_pad_type_0"), val = tensor("valid")]; + tensor x_77_strides_0 = const()[name = tensor("x_77_strides_0"), val = tensor([1])]; + tensor x_77_pad_0 = const()[name = tensor("x_77_pad_0"), val = tensor([0, 0])]; + tensor x_77_dilations_0 = const()[name = tensor("x_77_dilations_0"), val = tensor([1])]; + tensor x_77_groups_0 = const()[name = tensor("x_77_groups_0"), val = tensor(1)]; + tensor encoder_layers_2_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_2_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36716160)))]; + tensor x_77_cast_fp16 = conv(dilations = x_77_dilations_0, groups = x_77_groups_0, pad = x_77_pad_0, pad_type = x_77_pad_type_0, strides = x_77_strides_0, weight = encoder_layers_2_conv_pointwise_conv2_weight_to_fp16, x = input_163_cast_fp16)[name = tensor("x_77_cast_fp16")]; + tensor input_165_perm_0 = const()[name = tensor("input_165_perm_0"), val = tensor([0, 2, 1])]; + tensor input_165_cast_fp16 = transpose(perm = input_165_perm_0, x = x_77_cast_fp16)[name = tensor("transpose_212")]; + tensor input_167_cast_fp16 = add(x = input_151_cast_fp16, y = input_165_cast_fp16)[name = tensor("input_167_cast_fp16")]; + tensor input_169_axes_0 = const()[name = tensor("input_169_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37240512)))]; + tensor encoder_layers_2_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37241600)))]; + tensor input_169_cast_fp16 = layer_norm(axes = input_169_axes_0, beta = encoder_layers_2_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_norm_feed_forward2_weight_to_fp16, x = input_167_cast_fp16)[name = tensor("input_169_cast_fp16")]; + tensor encoder_layers_2_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_2_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37242688)))]; + tensor linear_26_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_2_feed_forward2_linear1_weight_to_fp16, x = input_169_cast_fp16)[name = tensor("linear_26_cast_fp16")]; + tensor input_173_cast_fp16 = silu(x = linear_26_cast_fp16)[name = tensor("input_173_cast_fp16")]; + tensor encoder_layers_2_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_2_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39339904)))]; + tensor linear_27_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_feed_forward2_linear2_weight_to_fp16, x = input_173_cast_fp16)[name = tensor("linear_27_cast_fp16")]; + tensor var_889_to_fp16 = const()[name = tensor("op_889_to_fp16"), val = tensor(0x1p-1)]; + tensor var_890_cast_fp16 = mul(x = linear_27_cast_fp16, y = var_889_to_fp16)[name = tensor("op_890_cast_fp16")]; + tensor input_179_cast_fp16 = add(x = input_167_cast_fp16, y = var_890_cast_fp16)[name = tensor("input_179_cast_fp16")]; + tensor input_181_axes_0 = const()[name = tensor("input_181_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41437120)))]; + tensor encoder_layers_2_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41438208)))]; + tensor input_181_cast_fp16 = layer_norm(axes = input_181_axes_0, beta = encoder_layers_2_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_norm_out_weight_to_fp16, x = input_179_cast_fp16)[name = tensor("input_181_cast_fp16")]; + tensor cache_13_begin_0 = const()[name = tensor("cache_13_begin_0"), val = tensor([3, 0, 0, 0])]; + tensor cache_13_end_0 = const()[name = tensor("cache_13_end_0"), val = tensor([4, 1, 70, 512])]; + tensor cache_13_end_mask_0 = const()[name = tensor("cache_13_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_13_squeeze_mask_0 = const()[name = tensor("cache_13_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_13_cast_fp16 = slice_by_index(begin = cache_13_begin_0, end = cache_13_end_0, end_mask = cache_13_end_mask_0, squeeze_mask = cache_13_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_13_cast_fp16")]; + tensor cache_15_begin_0 = const()[name = tensor("cache_15_begin_0"), val = tensor([3, 0, 0, 0])]; + tensor cache_15_end_0 = const()[name = tensor("cache_15_end_0"), val = tensor([4, 1, 512, 8])]; + tensor cache_15_end_mask_0 = const()[name = tensor("cache_15_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_15_squeeze_mask_0 = const()[name = tensor("cache_15_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_15_cast_fp16 = slice_by_index(begin = cache_15_begin_0, end = cache_15_end_0, end_mask = cache_15_end_mask_0, squeeze_mask = cache_15_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_15_cast_fp16")]; + tensor input_183_axes_0 = const()[name = tensor("input_183_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41439296)))]; + tensor encoder_layers_3_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41440384)))]; + tensor input_183_cast_fp16 = layer_norm(axes = input_183_axes_0, beta = encoder_layers_3_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_norm_feed_forward1_weight_to_fp16, x = input_181_cast_fp16)[name = tensor("input_183_cast_fp16")]; + tensor encoder_layers_3_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_3_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41441472)))]; + tensor linear_28_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_3_feed_forward1_linear1_weight_to_fp16, x = input_183_cast_fp16)[name = tensor("linear_28_cast_fp16")]; + tensor input_187_cast_fp16 = silu(x = linear_28_cast_fp16)[name = tensor("input_187_cast_fp16")]; + tensor encoder_layers_3_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_3_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43538688)))]; + tensor linear_29_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_feed_forward1_linear2_weight_to_fp16, x = input_187_cast_fp16)[name = tensor("linear_29_cast_fp16")]; + tensor var_924_to_fp16 = const()[name = tensor("op_924_to_fp16"), val = tensor(0x1p-1)]; + tensor var_925_cast_fp16 = mul(x = linear_29_cast_fp16, y = var_924_to_fp16)[name = tensor("op_925_cast_fp16")]; + tensor input_193_cast_fp16 = add(x = input_181_cast_fp16, y = var_925_cast_fp16)[name = tensor("input_193_cast_fp16")]; + tensor key_7_axes_0 = const()[name = tensor("key_7_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45635904)))]; + tensor encoder_layers_3_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45636992)))]; + tensor key_7_cast_fp16 = layer_norm(axes = key_7_axes_0, beta = encoder_layers_3_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_norm_self_att_weight_to_fp16, x = input_193_cast_fp16)[name = tensor("key_7_cast_fp16")]; + tensor input_195_interleave_0 = const()[name = tensor("input_195_interleave_0"), val = tensor(false)]; + tensor input_195_cast_fp16 = concat(axis = var_64, interleave = input_195_interleave_0, values = (cache_13_cast_fp16, key_7_cast_fp16))[name = tensor("input_195_cast_fp16")]; + tensor var_947_begin_0 = const()[name = tensor("op_947_begin_0"), val = tensor([0, 1, 0])]; + tensor var_947_end_0 = const()[name = tensor("op_947_end_0"), val = tensor([1, 70, 512])]; + tensor var_947_end_mask_0 = const()[name = tensor("op_947_end_mask_0"), val = tensor([true, true, true])]; + tensor var_947_cast_fp16 = slice_by_index(begin = var_947_begin_0, end = var_947_end_0, end_mask = var_947_end_mask_0, x = cache_13_cast_fp16)[name = tensor("op_947_cast_fp16")]; + tensor var_950_begin_0 = const()[name = tensor("op_950_begin_0"), val = tensor([0, 0, 0])]; + tensor var_950_end_0 = const()[name = tensor("op_950_end_0"), val = tensor([1, 1, 512])]; + tensor var_950_end_mask_0 = const()[name = tensor("op_950_end_mask_0"), val = tensor([true, false, true])]; + tensor var_950_cast_fp16 = slice_by_index(begin = var_950_begin_0, end = var_950_end_0, end_mask = var_950_end_mask_0, x = key_7_cast_fp16)[name = tensor("op_950_cast_fp16")]; + tensor var_953_interleave_0 = const()[name = tensor("op_953_interleave_0"), val = tensor(false)]; + tensor var_953_cast_fp16 = concat(axis = var_64, interleave = var_953_interleave_0, values = (var_947_cast_fp16, var_950_cast_fp16))[name = tensor("op_953_cast_fp16")]; + tensor encoder_layers_3_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45638080)))]; + tensor linear_30_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_self_attn_linear_q_weight_to_fp16, x = key_7_cast_fp16)[name = tensor("linear_30_cast_fp16")]; + tensor var_957 = const()[name = tensor("op_957"), val = tensor([1, -1, 8, 64])]; + tensor q_19_cast_fp16 = reshape(shape = var_957, x = linear_30_cast_fp16)[name = tensor("q_19_cast_fp16")]; + tensor encoder_layers_3_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46162432)))]; + tensor linear_31_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_self_attn_linear_k_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("linear_31_cast_fp16")]; + tensor var_961 = const()[name = tensor("op_961"), val = tensor([1, -1, 8, 64])]; + tensor k_13_cast_fp16 = reshape(shape = var_961, x = linear_31_cast_fp16)[name = tensor("k_13_cast_fp16")]; + tensor encoder_layers_3_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46686784)))]; + tensor linear_32_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_self_attn_linear_v_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("linear_32_cast_fp16")]; + tensor var_965 = const()[name = tensor("op_965"), val = tensor([1, -1, 8, 64])]; + tensor v_7_cast_fp16 = reshape(shape = var_965, x = linear_32_cast_fp16)[name = tensor("v_7_cast_fp16")]; + tensor value_9_perm_0 = const()[name = tensor("value_9_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_3_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47211136)))]; + tensor var_977_cast_fp16 = add(x = q_19_cast_fp16, y = encoder_layers_3_self_attn_pos_bias_u_to_fp16)[name = tensor("op_977_cast_fp16")]; + tensor encoder_layers_3_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47212224)))]; + tensor var_979_cast_fp16 = add(x = q_19_cast_fp16, y = encoder_layers_3_self_attn_pos_bias_v_to_fp16)[name = tensor("op_979_cast_fp16")]; + tensor q_with_bias_v_7_perm_0 = const()[name = tensor("q_with_bias_v_7_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_85_transpose_x_0 = const()[name = tensor("x_85_transpose_x_0"), val = tensor(false)]; + tensor x_85_transpose_y_0 = const()[name = tensor("x_85_transpose_y_0"), val = tensor(false)]; + tensor var_981_to_fp16 = const()[name = tensor("op_981_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47213312)))]; + tensor q_with_bias_v_7_cast_fp16 = transpose(perm = q_with_bias_v_7_perm_0, x = var_979_cast_fp16)[name = tensor("transpose_210")]; + tensor x_85_cast_fp16 = matmul(transpose_x = x_85_transpose_x_0, transpose_y = x_85_transpose_y_0, x = q_with_bias_v_7_cast_fp16, y = var_981_to_fp16)[name = tensor("x_85_cast_fp16")]; + tensor x_87_pad_0 = const()[name = tensor("x_87_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_87_mode_0 = const()[name = tensor("x_87_mode_0"), val = tensor("constant")]; + tensor const_62_to_fp16 = const()[name = tensor("const_62_to_fp16"), val = tensor(0x0p+0)]; + tensor x_87_cast_fp16 = pad(constant_val = const_62_to_fp16, mode = x_87_mode_0, pad = x_87_pad_0, x = x_85_cast_fp16)[name = tensor("x_87_cast_fp16")]; + tensor var_989 = const()[name = tensor("op_989"), val = tensor([1, 8, -1, 3])]; + tensor x_89_cast_fp16 = reshape(shape = var_989, x = x_87_cast_fp16)[name = tensor("x_89_cast_fp16")]; + tensor var_993_begin_0 = const()[name = tensor("op_993_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_993_end_0 = const()[name = tensor("op_993_end_0"), val = tensor([1, 8, 146, 3])]; + tensor var_993_end_mask_0 = const()[name = tensor("op_993_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_993_cast_fp16 = slice_by_index(begin = var_993_begin_0, end = var_993_end_0, end_mask = var_993_end_mask_0, x = x_89_cast_fp16)[name = tensor("op_993_cast_fp16")]; + tensor var_994 = const()[name = tensor("op_994"), val = tensor([1, 8, 3, 145])]; + tensor matrix_bd_13_cast_fp16 = reshape(shape = var_994, x = var_993_cast_fp16)[name = tensor("matrix_bd_13_cast_fp16")]; + tensor matrix_ac_7_transpose_x_0 = const()[name = tensor("matrix_ac_7_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_7_transpose_y_0 = const()[name = tensor("matrix_ac_7_transpose_y_0"), val = tensor(false)]; + tensor transpose_57_perm_0 = const()[name = tensor("transpose_57_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_58_perm_0 = const()[name = tensor("transpose_58_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_58 = transpose(perm = transpose_58_perm_0, x = k_13_cast_fp16)[name = tensor("transpose_208")]; + tensor transpose_57 = transpose(perm = transpose_57_perm_0, x = var_977_cast_fp16)[name = tensor("transpose_209")]; + tensor matrix_ac_7_cast_fp16 = matmul(transpose_x = matrix_ac_7_transpose_x_0, transpose_y = matrix_ac_7_transpose_y_0, x = transpose_57, y = transpose_58)[name = tensor("matrix_ac_7_cast_fp16")]; + tensor matrix_bd_15_begin_0 = const()[name = tensor("matrix_bd_15_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_15_end_0 = const()[name = tensor("matrix_bd_15_end_0"), val = tensor([1, 8, 3, 73])]; + tensor matrix_bd_15_end_mask_0 = const()[name = tensor("matrix_bd_15_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_15_cast_fp16 = slice_by_index(begin = matrix_bd_15_begin_0, end = matrix_bd_15_end_0, end_mask = matrix_bd_15_end_mask_0, x = matrix_bd_13_cast_fp16)[name = tensor("matrix_bd_15_cast_fp16")]; + tensor var_1003_cast_fp16 = add(x = matrix_ac_7_cast_fp16, y = matrix_bd_15_cast_fp16)[name = tensor("op_1003_cast_fp16")]; + tensor _inversed_scores_13_y_0_to_fp16 = const()[name = tensor("_inversed_scores_13_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_13_cast_fp16 = mul(x = var_1003_cast_fp16, y = _inversed_scores_13_y_0_to_fp16)[name = tensor("_inversed_scores_13_cast_fp16")]; + tensor scores_15_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_13_cast_fp16, cond = mask_3)[name = tensor("scores_15_cast_fp16")]; + tensor var_1009_cast_fp16 = softmax(axis = var_62, x = scores_15_cast_fp16)[name = tensor("op_1009_cast_fp16")]; + tensor input_197_cast_fp16 = select(a = var_40_to_fp16, b = var_1009_cast_fp16, cond = mask_3)[name = tensor("input_197_cast_fp16")]; + tensor x_91_transpose_x_0 = const()[name = tensor("x_91_transpose_x_0"), val = tensor(false)]; + tensor x_91_transpose_y_0 = const()[name = tensor("x_91_transpose_y_0"), val = tensor(false)]; + tensor value_9_cast_fp16 = transpose(perm = value_9_perm_0, x = v_7_cast_fp16)[name = tensor("transpose_211")]; + tensor x_91_cast_fp16 = matmul(transpose_x = x_91_transpose_x_0, transpose_y = x_91_transpose_y_0, x = input_197_cast_fp16, y = value_9_cast_fp16)[name = tensor("x_91_cast_fp16")]; + tensor var_1013_perm_0 = const()[name = tensor("op_1013_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1014 = const()[name = tensor("op_1014"), val = tensor([1, -1, 512])]; + tensor var_1013_cast_fp16 = transpose(perm = var_1013_perm_0, x = x_91_cast_fp16)[name = tensor("transpose_207")]; + tensor input_199_cast_fp16 = reshape(shape = var_1014, x = var_1013_cast_fp16)[name = tensor("input_199_cast_fp16")]; + tensor encoder_layers_3_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47361856)))]; + tensor linear_34_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_self_attn_linear_out_weight_to_fp16, x = input_199_cast_fp16)[name = tensor("linear_34_cast_fp16")]; + tensor input_203_cast_fp16 = add(x = input_193_cast_fp16, y = linear_34_cast_fp16)[name = tensor("input_203_cast_fp16")]; + tensor x_95_axes_0 = const()[name = tensor("x_95_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47886208)))]; + tensor encoder_layers_3_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47887296)))]; + tensor x_95_cast_fp16 = layer_norm(axes = x_95_axes_0, beta = encoder_layers_3_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_norm_conv_weight_to_fp16, x = input_203_cast_fp16)[name = tensor("x_95_cast_fp16")]; + tensor input_205_perm_0 = const()[name = tensor("input_205_perm_0"), val = tensor([0, 2, 1])]; + tensor input_207_pad_type_0 = const()[name = tensor("input_207_pad_type_0"), val = tensor("valid")]; + tensor input_207_strides_0 = const()[name = tensor("input_207_strides_0"), val = tensor([1])]; + tensor input_207_pad_0 = const()[name = tensor("input_207_pad_0"), val = tensor([0, 0])]; + tensor input_207_dilations_0 = const()[name = tensor("input_207_dilations_0"), val = tensor([1])]; + tensor input_207_groups_0 = const()[name = tensor("input_207_groups_0"), val = tensor(1)]; + tensor encoder_layers_3_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_3_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47888384)))]; + tensor input_205_cast_fp16 = transpose(perm = input_205_perm_0, x = x_95_cast_fp16)[name = tensor("transpose_206")]; + tensor input_207_cast_fp16 = conv(dilations = input_207_dilations_0, groups = input_207_groups_0, pad = input_207_pad_0, pad_type = input_207_pad_type_0, strides = input_207_strides_0, weight = encoder_layers_3_conv_pointwise_conv1_weight_to_fp16, x = input_205_cast_fp16)[name = tensor("input_207_cast_fp16")]; + tensor x_97_split_num_splits_0 = const()[name = tensor("x_97_split_num_splits_0"), val = tensor(2)]; + tensor x_97_split_axis_0 = const()[name = tensor("x_97_split_axis_0"), val = tensor(1)]; + tensor x_97_split_cast_fp16_0, tensor x_97_split_cast_fp16_1 = split(axis = x_97_split_axis_0, num_splits = x_97_split_num_splits_0, x = input_207_cast_fp16)[name = tensor("x_97_split_cast_fp16")]; + tensor x_97_split_1_sigmoid_cast_fp16 = sigmoid(x = x_97_split_cast_fp16_1)[name = tensor("x_97_split_1_sigmoid_cast_fp16")]; + tensor x_97_cast_fp16 = mul(x = x_97_split_cast_fp16_0, y = x_97_split_1_sigmoid_cast_fp16)[name = tensor("x_97_cast_fp16")]; + tensor input_209_cast_fp16 = select(a = var_40_to_fp16, b = x_97_cast_fp16, cond = var_418)[name = tensor("input_209_cast_fp16")]; + tensor new_x_15_interleave_0 = const()[name = tensor("new_x_15_interleave_0"), val = tensor(false)]; + tensor new_x_15_cast_fp16 = concat(axis = var_62, interleave = new_x_15_interleave_0, values = (cache_15_cast_fp16, input_209_cast_fp16))[name = tensor("new_x_15_cast_fp16")]; + tensor next_cache_7_begin_0 = const()[name = tensor("next_cache_7_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_7_end_0 = const()[name = tensor("next_cache_7_end_0"), val = tensor([1, 512, 9])]; + tensor next_cache_7_end_mask_0 = const()[name = tensor("next_cache_7_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_7_cast_fp16 = slice_by_index(begin = next_cache_7_begin_0, end = next_cache_7_end_0, end_mask = next_cache_7_end_mask_0, x = new_x_15_cast_fp16)[name = tensor("next_cache_7_cast_fp16")]; + tensor var_1055_begin_0 = const()[name = tensor("op_1055_begin_0"), val = tensor([0, 0, 1])]; + tensor var_1055_end_0 = const()[name = tensor("op_1055_end_0"), val = tensor([1, 512, 9])]; + tensor var_1055_end_mask_0 = const()[name = tensor("op_1055_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1055_cast_fp16 = slice_by_index(begin = var_1055_begin_0, end = var_1055_end_0, end_mask = var_1055_end_mask_0, x = next_cache_7_cast_fp16)[name = tensor("op_1055_cast_fp16")]; + tensor x_99_pad_type_0 = const()[name = tensor("x_99_pad_type_0"), val = tensor("valid")]; + tensor x_99_groups_0 = const()[name = tensor("x_99_groups_0"), val = tensor(512)]; + tensor x_99_strides_0 = const()[name = tensor("x_99_strides_0"), val = tensor([1])]; + tensor x_99_pad_0 = const()[name = tensor("x_99_pad_0"), val = tensor([0, 0])]; + tensor x_99_dilations_0 = const()[name = tensor("x_99_dilations_0"), val = tensor([1])]; + tensor encoder_layers_3_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_3_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48937024)))]; + tensor x_99_cast_fp16 = conv(dilations = x_99_dilations_0, groups = x_99_groups_0, pad = x_99_pad_0, pad_type = x_99_pad_type_0, strides = x_99_strides_0, weight = encoder_layers_3_conv_depthwise_conv_weight_to_fp16, x = new_x_15_cast_fp16)[name = tensor("x_99_cast_fp16")]; + tensor input_211_perm_0 = const()[name = tensor("input_211_perm_0"), val = tensor([0, 2, 1])]; + tensor x_101_axes_0 = const()[name = tensor("x_101_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_3_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48946304)))]; + tensor encoder_layers_3_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_3_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48947392)))]; + tensor input_211_cast_fp16 = transpose(perm = input_211_perm_0, x = x_99_cast_fp16)[name = tensor("transpose_205")]; + tensor x_101_cast_fp16 = layer_norm(axes = x_101_axes_0, beta = encoder_layers_3_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_conv_batch_norm_weight_to_fp16, x = input_211_cast_fp16)[name = tensor("x_101_cast_fp16")]; + tensor input_213_perm_0 = const()[name = tensor("input_213_perm_0"), val = tensor([0, 2, 1])]; + tensor input_213_cast_fp16 = transpose(perm = input_213_perm_0, x = x_101_cast_fp16)[name = tensor("transpose_204")]; + tensor input_215_cast_fp16 = silu(x = input_213_cast_fp16)[name = tensor("input_215_cast_fp16")]; + tensor x_103_pad_type_0 = const()[name = tensor("x_103_pad_type_0"), val = tensor("valid")]; + tensor x_103_strides_0 = const()[name = tensor("x_103_strides_0"), val = tensor([1])]; + tensor x_103_pad_0 = const()[name = tensor("x_103_pad_0"), val = tensor([0, 0])]; + tensor x_103_dilations_0 = const()[name = tensor("x_103_dilations_0"), val = tensor([1])]; + tensor x_103_groups_0 = const()[name = tensor("x_103_groups_0"), val = tensor(1)]; + tensor encoder_layers_3_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_3_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48948480)))]; + tensor x_103_cast_fp16 = conv(dilations = x_103_dilations_0, groups = x_103_groups_0, pad = x_103_pad_0, pad_type = x_103_pad_type_0, strides = x_103_strides_0, weight = encoder_layers_3_conv_pointwise_conv2_weight_to_fp16, x = input_215_cast_fp16)[name = tensor("x_103_cast_fp16")]; + tensor input_217_perm_0 = const()[name = tensor("input_217_perm_0"), val = tensor([0, 2, 1])]; + tensor input_217_cast_fp16 = transpose(perm = input_217_perm_0, x = x_103_cast_fp16)[name = tensor("transpose_203")]; + tensor input_219_cast_fp16 = add(x = input_203_cast_fp16, y = input_217_cast_fp16)[name = tensor("input_219_cast_fp16")]; + tensor input_221_axes_0 = const()[name = tensor("input_221_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49472832)))]; + tensor encoder_layers_3_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49473920)))]; + tensor input_221_cast_fp16 = layer_norm(axes = input_221_axes_0, beta = encoder_layers_3_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_norm_feed_forward2_weight_to_fp16, x = input_219_cast_fp16)[name = tensor("input_221_cast_fp16")]; + tensor encoder_layers_3_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_3_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49475008)))]; + tensor linear_35_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_3_feed_forward2_linear1_weight_to_fp16, x = input_221_cast_fp16)[name = tensor("linear_35_cast_fp16")]; + tensor input_225_cast_fp16 = silu(x = linear_35_cast_fp16)[name = tensor("input_225_cast_fp16")]; + tensor encoder_layers_3_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_3_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51572224)))]; + tensor linear_36_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_feed_forward2_linear2_weight_to_fp16, x = input_225_cast_fp16)[name = tensor("linear_36_cast_fp16")]; + tensor var_1096_to_fp16 = const()[name = tensor("op_1096_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1097_cast_fp16 = mul(x = linear_36_cast_fp16, y = var_1096_to_fp16)[name = tensor("op_1097_cast_fp16")]; + tensor input_231_cast_fp16 = add(x = input_219_cast_fp16, y = var_1097_cast_fp16)[name = tensor("input_231_cast_fp16")]; + tensor input_233_axes_0 = const()[name = tensor("input_233_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53669440)))]; + tensor encoder_layers_3_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53670528)))]; + tensor input_233_cast_fp16 = layer_norm(axes = input_233_axes_0, beta = encoder_layers_3_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_norm_out_weight_to_fp16, x = input_231_cast_fp16)[name = tensor("input_233_cast_fp16")]; + tensor cache_17_begin_0 = const()[name = tensor("cache_17_begin_0"), val = tensor([4, 0, 0, 0])]; + tensor cache_17_end_0 = const()[name = tensor("cache_17_end_0"), val = tensor([5, 1, 70, 512])]; + tensor cache_17_end_mask_0 = const()[name = tensor("cache_17_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_17_squeeze_mask_0 = const()[name = tensor("cache_17_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_17_cast_fp16 = slice_by_index(begin = cache_17_begin_0, end = cache_17_end_0, end_mask = cache_17_end_mask_0, squeeze_mask = cache_17_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_17_cast_fp16")]; + tensor cache_19_begin_0 = const()[name = tensor("cache_19_begin_0"), val = tensor([4, 0, 0, 0])]; + tensor cache_19_end_0 = const()[name = tensor("cache_19_end_0"), val = tensor([5, 1, 512, 8])]; + tensor cache_19_end_mask_0 = const()[name = tensor("cache_19_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_19_squeeze_mask_0 = const()[name = tensor("cache_19_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_19_cast_fp16 = slice_by_index(begin = cache_19_begin_0, end = cache_19_end_0, end_mask = cache_19_end_mask_0, squeeze_mask = cache_19_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_19_cast_fp16")]; + tensor input_235_axes_0 = const()[name = tensor("input_235_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53671616)))]; + tensor encoder_layers_4_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53672704)))]; + tensor input_235_cast_fp16 = layer_norm(axes = input_235_axes_0, beta = encoder_layers_4_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_norm_feed_forward1_weight_to_fp16, x = input_233_cast_fp16)[name = tensor("input_235_cast_fp16")]; + tensor encoder_layers_4_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_4_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53673792)))]; + tensor linear_37_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_4_feed_forward1_linear1_weight_to_fp16, x = input_235_cast_fp16)[name = tensor("linear_37_cast_fp16")]; + tensor input_239_cast_fp16 = silu(x = linear_37_cast_fp16)[name = tensor("input_239_cast_fp16")]; + tensor encoder_layers_4_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_4_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55771008)))]; + tensor linear_38_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_feed_forward1_linear2_weight_to_fp16, x = input_239_cast_fp16)[name = tensor("linear_38_cast_fp16")]; + tensor var_1131_to_fp16 = const()[name = tensor("op_1131_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1132_cast_fp16 = mul(x = linear_38_cast_fp16, y = var_1131_to_fp16)[name = tensor("op_1132_cast_fp16")]; + tensor input_245_cast_fp16 = add(x = input_233_cast_fp16, y = var_1132_cast_fp16)[name = tensor("input_245_cast_fp16")]; + tensor key_9_axes_0 = const()[name = tensor("key_9_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57868224)))]; + tensor encoder_layers_4_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57869312)))]; + tensor key_9_cast_fp16 = layer_norm(axes = key_9_axes_0, beta = encoder_layers_4_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_norm_self_att_weight_to_fp16, x = input_245_cast_fp16)[name = tensor("key_9_cast_fp16")]; + tensor input_247_interleave_0 = const()[name = tensor("input_247_interleave_0"), val = tensor(false)]; + tensor input_247_cast_fp16 = concat(axis = var_64, interleave = input_247_interleave_0, values = (cache_17_cast_fp16, key_9_cast_fp16))[name = tensor("input_247_cast_fp16")]; + tensor var_1154_begin_0 = const()[name = tensor("op_1154_begin_0"), val = tensor([0, 1, 0])]; + tensor var_1154_end_0 = const()[name = tensor("op_1154_end_0"), val = tensor([1, 70, 512])]; + tensor var_1154_end_mask_0 = const()[name = tensor("op_1154_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1154_cast_fp16 = slice_by_index(begin = var_1154_begin_0, end = var_1154_end_0, end_mask = var_1154_end_mask_0, x = cache_17_cast_fp16)[name = tensor("op_1154_cast_fp16")]; + tensor var_1157_begin_0 = const()[name = tensor("op_1157_begin_0"), val = tensor([0, 0, 0])]; + tensor var_1157_end_0 = const()[name = tensor("op_1157_end_0"), val = tensor([1, 1, 512])]; + tensor var_1157_end_mask_0 = const()[name = tensor("op_1157_end_mask_0"), val = tensor([true, false, true])]; + tensor var_1157_cast_fp16 = slice_by_index(begin = var_1157_begin_0, end = var_1157_end_0, end_mask = var_1157_end_mask_0, x = key_9_cast_fp16)[name = tensor("op_1157_cast_fp16")]; + tensor var_1160_interleave_0 = const()[name = tensor("op_1160_interleave_0"), val = tensor(false)]; + tensor var_1160_cast_fp16 = concat(axis = var_64, interleave = var_1160_interleave_0, values = (var_1154_cast_fp16, var_1157_cast_fp16))[name = tensor("op_1160_cast_fp16")]; + tensor encoder_layers_4_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57870400)))]; + tensor linear_39_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_self_attn_linear_q_weight_to_fp16, x = key_9_cast_fp16)[name = tensor("linear_39_cast_fp16")]; + tensor var_1164 = const()[name = tensor("op_1164"), val = tensor([1, -1, 8, 64])]; + tensor q_25_cast_fp16 = reshape(shape = var_1164, x = linear_39_cast_fp16)[name = tensor("q_25_cast_fp16")]; + tensor encoder_layers_4_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58394752)))]; + tensor linear_40_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_self_attn_linear_k_weight_to_fp16, x = input_247_cast_fp16)[name = tensor("linear_40_cast_fp16")]; + tensor var_1168 = const()[name = tensor("op_1168"), val = tensor([1, -1, 8, 64])]; + tensor k_17_cast_fp16 = reshape(shape = var_1168, x = linear_40_cast_fp16)[name = tensor("k_17_cast_fp16")]; + tensor encoder_layers_4_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58919104)))]; + tensor linear_41_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_self_attn_linear_v_weight_to_fp16, x = input_247_cast_fp16)[name = tensor("linear_41_cast_fp16")]; + tensor var_1172 = const()[name = tensor("op_1172"), val = tensor([1, -1, 8, 64])]; + tensor v_9_cast_fp16 = reshape(shape = var_1172, x = linear_41_cast_fp16)[name = tensor("v_9_cast_fp16")]; + tensor value_11_perm_0 = const()[name = tensor("value_11_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_4_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59443456)))]; + tensor var_1184_cast_fp16 = add(x = q_25_cast_fp16, y = encoder_layers_4_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1184_cast_fp16")]; + tensor encoder_layers_4_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59444544)))]; + tensor var_1186_cast_fp16 = add(x = q_25_cast_fp16, y = encoder_layers_4_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1186_cast_fp16")]; + tensor q_with_bias_v_9_perm_0 = const()[name = tensor("q_with_bias_v_9_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_111_transpose_x_0 = const()[name = tensor("x_111_transpose_x_0"), val = tensor(false)]; + tensor x_111_transpose_y_0 = const()[name = tensor("x_111_transpose_y_0"), val = tensor(false)]; + tensor var_1188_to_fp16 = const()[name = tensor("op_1188_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59445632)))]; + tensor q_with_bias_v_9_cast_fp16 = transpose(perm = q_with_bias_v_9_perm_0, x = var_1186_cast_fp16)[name = tensor("transpose_201")]; + tensor x_111_cast_fp16 = matmul(transpose_x = x_111_transpose_x_0, transpose_y = x_111_transpose_y_0, x = q_with_bias_v_9_cast_fp16, y = var_1188_to_fp16)[name = tensor("x_111_cast_fp16")]; + tensor x_113_pad_0 = const()[name = tensor("x_113_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_113_mode_0 = const()[name = tensor("x_113_mode_0"), val = tensor("constant")]; + tensor const_75_to_fp16 = const()[name = tensor("const_75_to_fp16"), val = tensor(0x0p+0)]; + tensor x_113_cast_fp16 = pad(constant_val = const_75_to_fp16, mode = x_113_mode_0, pad = x_113_pad_0, x = x_111_cast_fp16)[name = tensor("x_113_cast_fp16")]; + tensor var_1196 = const()[name = tensor("op_1196"), val = tensor([1, 8, -1, 3])]; + tensor x_115_cast_fp16 = reshape(shape = var_1196, x = x_113_cast_fp16)[name = tensor("x_115_cast_fp16")]; + tensor var_1200_begin_0 = const()[name = tensor("op_1200_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1200_end_0 = const()[name = tensor("op_1200_end_0"), val = tensor([1, 8, 146, 3])]; + tensor var_1200_end_mask_0 = const()[name = tensor("op_1200_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1200_cast_fp16 = slice_by_index(begin = var_1200_begin_0, end = var_1200_end_0, end_mask = var_1200_end_mask_0, x = x_115_cast_fp16)[name = tensor("op_1200_cast_fp16")]; + tensor var_1201 = const()[name = tensor("op_1201"), val = tensor([1, 8, 3, 145])]; + tensor matrix_bd_17_cast_fp16 = reshape(shape = var_1201, x = var_1200_cast_fp16)[name = tensor("matrix_bd_17_cast_fp16")]; + tensor matrix_ac_9_transpose_x_0 = const()[name = tensor("matrix_ac_9_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_9_transpose_y_0 = const()[name = tensor("matrix_ac_9_transpose_y_0"), val = tensor(false)]; + tensor transpose_59_perm_0 = const()[name = tensor("transpose_59_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_60_perm_0 = const()[name = tensor("transpose_60_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_60 = transpose(perm = transpose_60_perm_0, x = k_17_cast_fp16)[name = tensor("transpose_199")]; + tensor transpose_59 = transpose(perm = transpose_59_perm_0, x = var_1184_cast_fp16)[name = tensor("transpose_200")]; + tensor matrix_ac_9_cast_fp16 = matmul(transpose_x = matrix_ac_9_transpose_x_0, transpose_y = matrix_ac_9_transpose_y_0, x = transpose_59, y = transpose_60)[name = tensor("matrix_ac_9_cast_fp16")]; + tensor matrix_bd_19_begin_0 = const()[name = tensor("matrix_bd_19_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_19_end_0 = const()[name = tensor("matrix_bd_19_end_0"), val = tensor([1, 8, 3, 73])]; + tensor matrix_bd_19_end_mask_0 = const()[name = tensor("matrix_bd_19_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_19_cast_fp16 = slice_by_index(begin = matrix_bd_19_begin_0, end = matrix_bd_19_end_0, end_mask = matrix_bd_19_end_mask_0, x = matrix_bd_17_cast_fp16)[name = tensor("matrix_bd_19_cast_fp16")]; + tensor var_1210_cast_fp16 = add(x = matrix_ac_9_cast_fp16, y = matrix_bd_19_cast_fp16)[name = tensor("op_1210_cast_fp16")]; + tensor _inversed_scores_17_y_0_to_fp16 = const()[name = tensor("_inversed_scores_17_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_17_cast_fp16 = mul(x = var_1210_cast_fp16, y = _inversed_scores_17_y_0_to_fp16)[name = tensor("_inversed_scores_17_cast_fp16")]; + tensor scores_19_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_17_cast_fp16, cond = mask_3)[name = tensor("scores_19_cast_fp16")]; + tensor var_1216_cast_fp16 = softmax(axis = var_62, x = scores_19_cast_fp16)[name = tensor("op_1216_cast_fp16")]; + tensor input_249_cast_fp16 = select(a = var_40_to_fp16, b = var_1216_cast_fp16, cond = mask_3)[name = tensor("input_249_cast_fp16")]; + tensor x_117_transpose_x_0 = const()[name = tensor("x_117_transpose_x_0"), val = tensor(false)]; + tensor x_117_transpose_y_0 = const()[name = tensor("x_117_transpose_y_0"), val = tensor(false)]; + tensor value_11_cast_fp16 = transpose(perm = value_11_perm_0, x = v_9_cast_fp16)[name = tensor("transpose_202")]; + tensor x_117_cast_fp16 = matmul(transpose_x = x_117_transpose_x_0, transpose_y = x_117_transpose_y_0, x = input_249_cast_fp16, y = value_11_cast_fp16)[name = tensor("x_117_cast_fp16")]; + tensor var_1220_perm_0 = const()[name = tensor("op_1220_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1221 = const()[name = tensor("op_1221"), val = tensor([1, -1, 512])]; + tensor var_1220_cast_fp16 = transpose(perm = var_1220_perm_0, x = x_117_cast_fp16)[name = tensor("transpose_198")]; + tensor input_251_cast_fp16 = reshape(shape = var_1221, x = var_1220_cast_fp16)[name = tensor("input_251_cast_fp16")]; + tensor encoder_layers_4_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59594176)))]; + tensor linear_43_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_self_attn_linear_out_weight_to_fp16, x = input_251_cast_fp16)[name = tensor("linear_43_cast_fp16")]; + tensor input_255_cast_fp16 = add(x = input_245_cast_fp16, y = linear_43_cast_fp16)[name = tensor("input_255_cast_fp16")]; + tensor x_121_axes_0 = const()[name = tensor("x_121_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60118528)))]; + tensor encoder_layers_4_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60119616)))]; + tensor x_121_cast_fp16 = layer_norm(axes = x_121_axes_0, beta = encoder_layers_4_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_norm_conv_weight_to_fp16, x = input_255_cast_fp16)[name = tensor("x_121_cast_fp16")]; + tensor input_257_perm_0 = const()[name = tensor("input_257_perm_0"), val = tensor([0, 2, 1])]; + tensor input_259_pad_type_0 = const()[name = tensor("input_259_pad_type_0"), val = tensor("valid")]; + tensor input_259_strides_0 = const()[name = tensor("input_259_strides_0"), val = tensor([1])]; + tensor input_259_pad_0 = const()[name = tensor("input_259_pad_0"), val = tensor([0, 0])]; + tensor input_259_dilations_0 = const()[name = tensor("input_259_dilations_0"), val = tensor([1])]; + tensor input_259_groups_0 = const()[name = tensor("input_259_groups_0"), val = tensor(1)]; + tensor encoder_layers_4_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_4_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60120704)))]; + tensor input_257_cast_fp16 = transpose(perm = input_257_perm_0, x = x_121_cast_fp16)[name = tensor("transpose_197")]; + tensor input_259_cast_fp16 = conv(dilations = input_259_dilations_0, groups = input_259_groups_0, pad = input_259_pad_0, pad_type = input_259_pad_type_0, strides = input_259_strides_0, weight = encoder_layers_4_conv_pointwise_conv1_weight_to_fp16, x = input_257_cast_fp16)[name = tensor("input_259_cast_fp16")]; + tensor x_123_split_num_splits_0 = const()[name = tensor("x_123_split_num_splits_0"), val = tensor(2)]; + tensor x_123_split_axis_0 = const()[name = tensor("x_123_split_axis_0"), val = tensor(1)]; + tensor x_123_split_cast_fp16_0, tensor x_123_split_cast_fp16_1 = split(axis = x_123_split_axis_0, num_splits = x_123_split_num_splits_0, x = input_259_cast_fp16)[name = tensor("x_123_split_cast_fp16")]; + tensor x_123_split_1_sigmoid_cast_fp16 = sigmoid(x = x_123_split_cast_fp16_1)[name = tensor("x_123_split_1_sigmoid_cast_fp16")]; + tensor x_123_cast_fp16 = mul(x = x_123_split_cast_fp16_0, y = x_123_split_1_sigmoid_cast_fp16)[name = tensor("x_123_cast_fp16")]; + tensor input_261_cast_fp16 = select(a = var_40_to_fp16, b = x_123_cast_fp16, cond = var_418)[name = tensor("input_261_cast_fp16")]; + tensor new_x_19_interleave_0 = const()[name = tensor("new_x_19_interleave_0"), val = tensor(false)]; + tensor new_x_19_cast_fp16 = concat(axis = var_62, interleave = new_x_19_interleave_0, values = (cache_19_cast_fp16, input_261_cast_fp16))[name = tensor("new_x_19_cast_fp16")]; + tensor next_cache_9_begin_0 = const()[name = tensor("next_cache_9_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_9_end_0 = const()[name = tensor("next_cache_9_end_0"), val = tensor([1, 512, 9])]; + tensor next_cache_9_end_mask_0 = const()[name = tensor("next_cache_9_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_9_cast_fp16 = slice_by_index(begin = next_cache_9_begin_0, end = next_cache_9_end_0, end_mask = next_cache_9_end_mask_0, x = new_x_19_cast_fp16)[name = tensor("next_cache_9_cast_fp16")]; + tensor var_1262_begin_0 = const()[name = tensor("op_1262_begin_0"), val = tensor([0, 0, 1])]; + tensor var_1262_end_0 = const()[name = tensor("op_1262_end_0"), val = tensor([1, 512, 9])]; + tensor var_1262_end_mask_0 = const()[name = tensor("op_1262_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1262_cast_fp16 = slice_by_index(begin = var_1262_begin_0, end = var_1262_end_0, end_mask = var_1262_end_mask_0, x = next_cache_9_cast_fp16)[name = tensor("op_1262_cast_fp16")]; + tensor x_125_pad_type_0 = const()[name = tensor("x_125_pad_type_0"), val = tensor("valid")]; + tensor x_125_groups_0 = const()[name = tensor("x_125_groups_0"), val = tensor(512)]; + tensor x_125_strides_0 = const()[name = tensor("x_125_strides_0"), val = tensor([1])]; + tensor x_125_pad_0 = const()[name = tensor("x_125_pad_0"), val = tensor([0, 0])]; + tensor x_125_dilations_0 = const()[name = tensor("x_125_dilations_0"), val = tensor([1])]; + tensor encoder_layers_4_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_4_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61169344)))]; + tensor x_125_cast_fp16 = conv(dilations = x_125_dilations_0, groups = x_125_groups_0, pad = x_125_pad_0, pad_type = x_125_pad_type_0, strides = x_125_strides_0, weight = encoder_layers_4_conv_depthwise_conv_weight_to_fp16, x = new_x_19_cast_fp16)[name = tensor("x_125_cast_fp16")]; + tensor input_263_perm_0 = const()[name = tensor("input_263_perm_0"), val = tensor([0, 2, 1])]; + tensor x_127_axes_0 = const()[name = tensor("x_127_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_4_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61178624)))]; + tensor encoder_layers_4_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_4_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61179712)))]; + tensor input_263_cast_fp16 = transpose(perm = input_263_perm_0, x = x_125_cast_fp16)[name = tensor("transpose_196")]; + tensor x_127_cast_fp16 = layer_norm(axes = x_127_axes_0, beta = encoder_layers_4_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_conv_batch_norm_weight_to_fp16, x = input_263_cast_fp16)[name = tensor("x_127_cast_fp16")]; + tensor input_265_perm_0 = const()[name = tensor("input_265_perm_0"), val = tensor([0, 2, 1])]; + tensor input_265_cast_fp16 = transpose(perm = input_265_perm_0, x = x_127_cast_fp16)[name = tensor("transpose_195")]; + tensor input_267_cast_fp16 = silu(x = input_265_cast_fp16)[name = tensor("input_267_cast_fp16")]; + tensor x_129_pad_type_0 = const()[name = tensor("x_129_pad_type_0"), val = tensor("valid")]; + tensor x_129_strides_0 = const()[name = tensor("x_129_strides_0"), val = tensor([1])]; + tensor x_129_pad_0 = const()[name = tensor("x_129_pad_0"), val = tensor([0, 0])]; + tensor x_129_dilations_0 = const()[name = tensor("x_129_dilations_0"), val = tensor([1])]; + tensor x_129_groups_0 = const()[name = tensor("x_129_groups_0"), val = tensor(1)]; + tensor encoder_layers_4_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_4_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61180800)))]; + tensor x_129_cast_fp16 = conv(dilations = x_129_dilations_0, groups = x_129_groups_0, pad = x_129_pad_0, pad_type = x_129_pad_type_0, strides = x_129_strides_0, weight = encoder_layers_4_conv_pointwise_conv2_weight_to_fp16, x = input_267_cast_fp16)[name = tensor("x_129_cast_fp16")]; + tensor input_269_perm_0 = const()[name = tensor("input_269_perm_0"), val = tensor([0, 2, 1])]; + tensor input_269_cast_fp16 = transpose(perm = input_269_perm_0, x = x_129_cast_fp16)[name = tensor("transpose_194")]; + tensor input_271_cast_fp16 = add(x = input_255_cast_fp16, y = input_269_cast_fp16)[name = tensor("input_271_cast_fp16")]; + tensor input_273_axes_0 = const()[name = tensor("input_273_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61705152)))]; + tensor encoder_layers_4_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61706240)))]; + tensor input_273_cast_fp16 = layer_norm(axes = input_273_axes_0, beta = encoder_layers_4_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_norm_feed_forward2_weight_to_fp16, x = input_271_cast_fp16)[name = tensor("input_273_cast_fp16")]; + tensor encoder_layers_4_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_4_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61707328)))]; + tensor linear_44_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_4_feed_forward2_linear1_weight_to_fp16, x = input_273_cast_fp16)[name = tensor("linear_44_cast_fp16")]; + tensor input_277_cast_fp16 = silu(x = linear_44_cast_fp16)[name = tensor("input_277_cast_fp16")]; + tensor encoder_layers_4_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_4_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63804544)))]; + tensor linear_45_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_feed_forward2_linear2_weight_to_fp16, x = input_277_cast_fp16)[name = tensor("linear_45_cast_fp16")]; + tensor var_1303_to_fp16 = const()[name = tensor("op_1303_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1304_cast_fp16 = mul(x = linear_45_cast_fp16, y = var_1303_to_fp16)[name = tensor("op_1304_cast_fp16")]; + tensor input_283_cast_fp16 = add(x = input_271_cast_fp16, y = var_1304_cast_fp16)[name = tensor("input_283_cast_fp16")]; + tensor input_285_axes_0 = const()[name = tensor("input_285_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65901760)))]; + tensor encoder_layers_4_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65902848)))]; + tensor input_285_cast_fp16 = layer_norm(axes = input_285_axes_0, beta = encoder_layers_4_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_norm_out_weight_to_fp16, x = input_283_cast_fp16)[name = tensor("input_285_cast_fp16")]; + tensor cache_21_begin_0 = const()[name = tensor("cache_21_begin_0"), val = tensor([5, 0, 0, 0])]; + tensor cache_21_end_0 = const()[name = tensor("cache_21_end_0"), val = tensor([6, 1, 70, 512])]; + tensor cache_21_end_mask_0 = const()[name = tensor("cache_21_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_21_squeeze_mask_0 = const()[name = tensor("cache_21_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_21_cast_fp16 = slice_by_index(begin = cache_21_begin_0, end = cache_21_end_0, end_mask = cache_21_end_mask_0, squeeze_mask = cache_21_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_21_cast_fp16")]; + tensor cache_23_begin_0 = const()[name = tensor("cache_23_begin_0"), val = tensor([5, 0, 0, 0])]; + tensor cache_23_end_0 = const()[name = tensor("cache_23_end_0"), val = tensor([6, 1, 512, 8])]; + tensor cache_23_end_mask_0 = const()[name = tensor("cache_23_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_23_squeeze_mask_0 = const()[name = tensor("cache_23_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_23_cast_fp16 = slice_by_index(begin = cache_23_begin_0, end = cache_23_end_0, end_mask = cache_23_end_mask_0, squeeze_mask = cache_23_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_23_cast_fp16")]; + tensor input_287_axes_0 = const()[name = tensor("input_287_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65903936)))]; + tensor encoder_layers_5_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65905024)))]; + tensor input_287_cast_fp16 = layer_norm(axes = input_287_axes_0, beta = encoder_layers_5_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_norm_feed_forward1_weight_to_fp16, x = input_285_cast_fp16)[name = tensor("input_287_cast_fp16")]; + tensor encoder_layers_5_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_5_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65906112)))]; + tensor linear_46_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_5_feed_forward1_linear1_weight_to_fp16, x = input_287_cast_fp16)[name = tensor("linear_46_cast_fp16")]; + tensor input_291_cast_fp16 = silu(x = linear_46_cast_fp16)[name = tensor("input_291_cast_fp16")]; + tensor encoder_layers_5_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_5_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68003328)))]; + tensor linear_47_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_feed_forward1_linear2_weight_to_fp16, x = input_291_cast_fp16)[name = tensor("linear_47_cast_fp16")]; + tensor var_1338_to_fp16 = const()[name = tensor("op_1338_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1339_cast_fp16 = mul(x = linear_47_cast_fp16, y = var_1338_to_fp16)[name = tensor("op_1339_cast_fp16")]; + tensor input_297_cast_fp16 = add(x = input_285_cast_fp16, y = var_1339_cast_fp16)[name = tensor("input_297_cast_fp16")]; + tensor key_11_axes_0 = const()[name = tensor("key_11_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70100544)))]; + tensor encoder_layers_5_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70101632)))]; + tensor key_11_cast_fp16 = layer_norm(axes = key_11_axes_0, beta = encoder_layers_5_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_norm_self_att_weight_to_fp16, x = input_297_cast_fp16)[name = tensor("key_11_cast_fp16")]; + tensor input_299_interleave_0 = const()[name = tensor("input_299_interleave_0"), val = tensor(false)]; + tensor input_299_cast_fp16 = concat(axis = var_64, interleave = input_299_interleave_0, values = (cache_21_cast_fp16, key_11_cast_fp16))[name = tensor("input_299_cast_fp16")]; + tensor var_1361_begin_0 = const()[name = tensor("op_1361_begin_0"), val = tensor([0, 1, 0])]; + tensor var_1361_end_0 = const()[name = tensor("op_1361_end_0"), val = tensor([1, 70, 512])]; + tensor var_1361_end_mask_0 = const()[name = tensor("op_1361_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1361_cast_fp16 = slice_by_index(begin = var_1361_begin_0, end = var_1361_end_0, end_mask = var_1361_end_mask_0, x = cache_21_cast_fp16)[name = tensor("op_1361_cast_fp16")]; + tensor var_1364_begin_0 = const()[name = tensor("op_1364_begin_0"), val = tensor([0, 0, 0])]; + tensor var_1364_end_0 = const()[name = tensor("op_1364_end_0"), val = tensor([1, 1, 512])]; + tensor var_1364_end_mask_0 = const()[name = tensor("op_1364_end_mask_0"), val = tensor([true, false, true])]; + tensor var_1364_cast_fp16 = slice_by_index(begin = var_1364_begin_0, end = var_1364_end_0, end_mask = var_1364_end_mask_0, x = key_11_cast_fp16)[name = tensor("op_1364_cast_fp16")]; + tensor var_1367_interleave_0 = const()[name = tensor("op_1367_interleave_0"), val = tensor(false)]; + tensor var_1367_cast_fp16 = concat(axis = var_64, interleave = var_1367_interleave_0, values = (var_1361_cast_fp16, var_1364_cast_fp16))[name = tensor("op_1367_cast_fp16")]; + tensor encoder_layers_5_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70102720)))]; + tensor linear_48_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_self_attn_linear_q_weight_to_fp16, x = key_11_cast_fp16)[name = tensor("linear_48_cast_fp16")]; + tensor var_1371 = const()[name = tensor("op_1371"), val = tensor([1, -1, 8, 64])]; + tensor q_31_cast_fp16 = reshape(shape = var_1371, x = linear_48_cast_fp16)[name = tensor("q_31_cast_fp16")]; + tensor encoder_layers_5_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70627072)))]; + tensor linear_49_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_self_attn_linear_k_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("linear_49_cast_fp16")]; + tensor var_1375 = const()[name = tensor("op_1375"), val = tensor([1, -1, 8, 64])]; + tensor k_21_cast_fp16 = reshape(shape = var_1375, x = linear_49_cast_fp16)[name = tensor("k_21_cast_fp16")]; + tensor encoder_layers_5_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71151424)))]; + tensor linear_50_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_self_attn_linear_v_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("linear_50_cast_fp16")]; + tensor var_1379 = const()[name = tensor("op_1379"), val = tensor([1, -1, 8, 64])]; + tensor v_11_cast_fp16 = reshape(shape = var_1379, x = linear_50_cast_fp16)[name = tensor("v_11_cast_fp16")]; + tensor value_13_perm_0 = const()[name = tensor("value_13_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_5_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71675776)))]; + tensor var_1391_cast_fp16 = add(x = q_31_cast_fp16, y = encoder_layers_5_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1391_cast_fp16")]; + tensor encoder_layers_5_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71676864)))]; + tensor var_1393_cast_fp16 = add(x = q_31_cast_fp16, y = encoder_layers_5_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1393_cast_fp16")]; + tensor q_with_bias_v_11_perm_0 = const()[name = tensor("q_with_bias_v_11_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_137_transpose_x_0 = const()[name = tensor("x_137_transpose_x_0"), val = tensor(false)]; + tensor x_137_transpose_y_0 = const()[name = tensor("x_137_transpose_y_0"), val = tensor(false)]; + tensor var_1395_to_fp16 = const()[name = tensor("op_1395_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71677952)))]; + tensor q_with_bias_v_11_cast_fp16 = transpose(perm = q_with_bias_v_11_perm_0, x = var_1393_cast_fp16)[name = tensor("transpose_192")]; + tensor x_137_cast_fp16 = matmul(transpose_x = x_137_transpose_x_0, transpose_y = x_137_transpose_y_0, x = q_with_bias_v_11_cast_fp16, y = var_1395_to_fp16)[name = tensor("x_137_cast_fp16")]; + tensor x_139_pad_0 = const()[name = tensor("x_139_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_139_mode_0 = const()[name = tensor("x_139_mode_0"), val = tensor("constant")]; + tensor const_88_to_fp16 = const()[name = tensor("const_88_to_fp16"), val = tensor(0x0p+0)]; + tensor x_139_cast_fp16 = pad(constant_val = const_88_to_fp16, mode = x_139_mode_0, pad = x_139_pad_0, x = x_137_cast_fp16)[name = tensor("x_139_cast_fp16")]; + tensor var_1403 = const()[name = tensor("op_1403"), val = tensor([1, 8, -1, 3])]; + tensor x_141_cast_fp16 = reshape(shape = var_1403, x = x_139_cast_fp16)[name = tensor("x_141_cast_fp16")]; + tensor var_1407_begin_0 = const()[name = tensor("op_1407_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1407_end_0 = const()[name = tensor("op_1407_end_0"), val = tensor([1, 8, 146, 3])]; + tensor var_1407_end_mask_0 = const()[name = tensor("op_1407_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1407_cast_fp16 = slice_by_index(begin = var_1407_begin_0, end = var_1407_end_0, end_mask = var_1407_end_mask_0, x = x_141_cast_fp16)[name = tensor("op_1407_cast_fp16")]; + tensor var_1408 = const()[name = tensor("op_1408"), val = tensor([1, 8, 3, 145])]; + tensor matrix_bd_21_cast_fp16 = reshape(shape = var_1408, x = var_1407_cast_fp16)[name = tensor("matrix_bd_21_cast_fp16")]; + tensor matrix_ac_11_transpose_x_0 = const()[name = tensor("matrix_ac_11_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_11_transpose_y_0 = const()[name = tensor("matrix_ac_11_transpose_y_0"), val = tensor(false)]; + tensor transpose_61_perm_0 = const()[name = tensor("transpose_61_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_62_perm_0 = const()[name = tensor("transpose_62_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_62 = transpose(perm = transpose_62_perm_0, x = k_21_cast_fp16)[name = tensor("transpose_190")]; + tensor transpose_61 = transpose(perm = transpose_61_perm_0, x = var_1391_cast_fp16)[name = tensor("transpose_191")]; + tensor matrix_ac_11_cast_fp16 = matmul(transpose_x = matrix_ac_11_transpose_x_0, transpose_y = matrix_ac_11_transpose_y_0, x = transpose_61, y = transpose_62)[name = tensor("matrix_ac_11_cast_fp16")]; + tensor matrix_bd_23_begin_0 = const()[name = tensor("matrix_bd_23_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_23_end_0 = const()[name = tensor("matrix_bd_23_end_0"), val = tensor([1, 8, 3, 73])]; + tensor matrix_bd_23_end_mask_0 = const()[name = tensor("matrix_bd_23_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_23_cast_fp16 = slice_by_index(begin = matrix_bd_23_begin_0, end = matrix_bd_23_end_0, end_mask = matrix_bd_23_end_mask_0, x = matrix_bd_21_cast_fp16)[name = tensor("matrix_bd_23_cast_fp16")]; + tensor var_1417_cast_fp16 = add(x = matrix_ac_11_cast_fp16, y = matrix_bd_23_cast_fp16)[name = tensor("op_1417_cast_fp16")]; + tensor _inversed_scores_21_y_0_to_fp16 = const()[name = tensor("_inversed_scores_21_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_21_cast_fp16 = mul(x = var_1417_cast_fp16, y = _inversed_scores_21_y_0_to_fp16)[name = tensor("_inversed_scores_21_cast_fp16")]; + tensor scores_23_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_21_cast_fp16, cond = mask_3)[name = tensor("scores_23_cast_fp16")]; + tensor var_1423_cast_fp16 = softmax(axis = var_62, x = scores_23_cast_fp16)[name = tensor("op_1423_cast_fp16")]; + tensor input_301_cast_fp16 = select(a = var_40_to_fp16, b = var_1423_cast_fp16, cond = mask_3)[name = tensor("input_301_cast_fp16")]; + tensor x_143_transpose_x_0 = const()[name = tensor("x_143_transpose_x_0"), val = tensor(false)]; + tensor x_143_transpose_y_0 = const()[name = tensor("x_143_transpose_y_0"), val = tensor(false)]; + tensor value_13_cast_fp16 = transpose(perm = value_13_perm_0, x = v_11_cast_fp16)[name = tensor("transpose_193")]; + tensor x_143_cast_fp16 = matmul(transpose_x = x_143_transpose_x_0, transpose_y = x_143_transpose_y_0, x = input_301_cast_fp16, y = value_13_cast_fp16)[name = tensor("x_143_cast_fp16")]; + tensor var_1427_perm_0 = const()[name = tensor("op_1427_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1428 = const()[name = tensor("op_1428"), val = tensor([1, -1, 512])]; + tensor var_1427_cast_fp16 = transpose(perm = var_1427_perm_0, x = x_143_cast_fp16)[name = tensor("transpose_189")]; + tensor input_303_cast_fp16 = reshape(shape = var_1428, x = var_1427_cast_fp16)[name = tensor("input_303_cast_fp16")]; + tensor encoder_layers_5_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71826496)))]; + tensor linear_52_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_self_attn_linear_out_weight_to_fp16, x = input_303_cast_fp16)[name = tensor("linear_52_cast_fp16")]; + tensor input_307_cast_fp16 = add(x = input_297_cast_fp16, y = linear_52_cast_fp16)[name = tensor("input_307_cast_fp16")]; + tensor x_147_axes_0 = const()[name = tensor("x_147_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72350848)))]; + tensor encoder_layers_5_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72351936)))]; + tensor x_147_cast_fp16 = layer_norm(axes = x_147_axes_0, beta = encoder_layers_5_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_norm_conv_weight_to_fp16, x = input_307_cast_fp16)[name = tensor("x_147_cast_fp16")]; + tensor input_309_perm_0 = const()[name = tensor("input_309_perm_0"), val = tensor([0, 2, 1])]; + tensor input_311_pad_type_0 = const()[name = tensor("input_311_pad_type_0"), val = tensor("valid")]; + tensor input_311_strides_0 = const()[name = tensor("input_311_strides_0"), val = tensor([1])]; + tensor input_311_pad_0 = const()[name = tensor("input_311_pad_0"), val = tensor([0, 0])]; + tensor input_311_dilations_0 = const()[name = tensor("input_311_dilations_0"), val = tensor([1])]; + tensor input_311_groups_0 = const()[name = tensor("input_311_groups_0"), val = tensor(1)]; + tensor encoder_layers_5_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_5_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72353024)))]; + tensor input_309_cast_fp16 = transpose(perm = input_309_perm_0, x = x_147_cast_fp16)[name = tensor("transpose_188")]; + tensor input_311_cast_fp16 = conv(dilations = input_311_dilations_0, groups = input_311_groups_0, pad = input_311_pad_0, pad_type = input_311_pad_type_0, strides = input_311_strides_0, weight = encoder_layers_5_conv_pointwise_conv1_weight_to_fp16, x = input_309_cast_fp16)[name = tensor("input_311_cast_fp16")]; + tensor x_149_split_num_splits_0 = const()[name = tensor("x_149_split_num_splits_0"), val = tensor(2)]; + tensor x_149_split_axis_0 = const()[name = tensor("x_149_split_axis_0"), val = tensor(1)]; + tensor x_149_split_cast_fp16_0, tensor x_149_split_cast_fp16_1 = split(axis = x_149_split_axis_0, num_splits = x_149_split_num_splits_0, x = input_311_cast_fp16)[name = tensor("x_149_split_cast_fp16")]; + tensor x_149_split_1_sigmoid_cast_fp16 = sigmoid(x = x_149_split_cast_fp16_1)[name = tensor("x_149_split_1_sigmoid_cast_fp16")]; + tensor x_149_cast_fp16 = mul(x = x_149_split_cast_fp16_0, y = x_149_split_1_sigmoid_cast_fp16)[name = tensor("x_149_cast_fp16")]; + tensor input_313_cast_fp16 = select(a = var_40_to_fp16, b = x_149_cast_fp16, cond = var_418)[name = tensor("input_313_cast_fp16")]; + tensor new_x_23_interleave_0 = const()[name = tensor("new_x_23_interleave_0"), val = tensor(false)]; + tensor new_x_23_cast_fp16 = concat(axis = var_62, interleave = new_x_23_interleave_0, values = (cache_23_cast_fp16, input_313_cast_fp16))[name = tensor("new_x_23_cast_fp16")]; + tensor next_cache_11_begin_0 = const()[name = tensor("next_cache_11_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_11_end_0 = const()[name = tensor("next_cache_11_end_0"), val = tensor([1, 512, 9])]; + tensor next_cache_11_end_mask_0 = const()[name = tensor("next_cache_11_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_11_cast_fp16 = slice_by_index(begin = next_cache_11_begin_0, end = next_cache_11_end_0, end_mask = next_cache_11_end_mask_0, x = new_x_23_cast_fp16)[name = tensor("next_cache_11_cast_fp16")]; + tensor var_1469_begin_0 = const()[name = tensor("op_1469_begin_0"), val = tensor([0, 0, 1])]; + tensor var_1469_end_0 = const()[name = tensor("op_1469_end_0"), val = tensor([1, 512, 9])]; + tensor var_1469_end_mask_0 = const()[name = tensor("op_1469_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1469_cast_fp16 = slice_by_index(begin = var_1469_begin_0, end = var_1469_end_0, end_mask = var_1469_end_mask_0, x = next_cache_11_cast_fp16)[name = tensor("op_1469_cast_fp16")]; + tensor x_151_pad_type_0 = const()[name = tensor("x_151_pad_type_0"), val = tensor("valid")]; + tensor x_151_groups_0 = const()[name = tensor("x_151_groups_0"), val = tensor(512)]; + tensor x_151_strides_0 = const()[name = tensor("x_151_strides_0"), val = tensor([1])]; + tensor x_151_pad_0 = const()[name = tensor("x_151_pad_0"), val = tensor([0, 0])]; + tensor x_151_dilations_0 = const()[name = tensor("x_151_dilations_0"), val = tensor([1])]; + tensor encoder_layers_5_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_5_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73401664)))]; + tensor x_151_cast_fp16 = conv(dilations = x_151_dilations_0, groups = x_151_groups_0, pad = x_151_pad_0, pad_type = x_151_pad_type_0, strides = x_151_strides_0, weight = encoder_layers_5_conv_depthwise_conv_weight_to_fp16, x = new_x_23_cast_fp16)[name = tensor("x_151_cast_fp16")]; + tensor input_315_perm_0 = const()[name = tensor("input_315_perm_0"), val = tensor([0, 2, 1])]; + tensor x_153_axes_0 = const()[name = tensor("x_153_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_5_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73410944)))]; + tensor encoder_layers_5_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_5_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73412032)))]; + tensor input_315_cast_fp16 = transpose(perm = input_315_perm_0, x = x_151_cast_fp16)[name = tensor("transpose_187")]; + tensor x_153_cast_fp16 = layer_norm(axes = x_153_axes_0, beta = encoder_layers_5_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_conv_batch_norm_weight_to_fp16, x = input_315_cast_fp16)[name = tensor("x_153_cast_fp16")]; + tensor input_317_perm_0 = const()[name = tensor("input_317_perm_0"), val = tensor([0, 2, 1])]; + tensor input_317_cast_fp16 = transpose(perm = input_317_perm_0, x = x_153_cast_fp16)[name = tensor("transpose_186")]; + tensor input_319_cast_fp16 = silu(x = input_317_cast_fp16)[name = tensor("input_319_cast_fp16")]; + tensor x_155_pad_type_0 = const()[name = tensor("x_155_pad_type_0"), val = tensor("valid")]; + tensor x_155_strides_0 = const()[name = tensor("x_155_strides_0"), val = tensor([1])]; + tensor x_155_pad_0 = const()[name = tensor("x_155_pad_0"), val = tensor([0, 0])]; + tensor x_155_dilations_0 = const()[name = tensor("x_155_dilations_0"), val = tensor([1])]; + tensor x_155_groups_0 = const()[name = tensor("x_155_groups_0"), val = tensor(1)]; + tensor encoder_layers_5_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_5_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73413120)))]; + tensor x_155_cast_fp16 = conv(dilations = x_155_dilations_0, groups = x_155_groups_0, pad = x_155_pad_0, pad_type = x_155_pad_type_0, strides = x_155_strides_0, weight = encoder_layers_5_conv_pointwise_conv2_weight_to_fp16, x = input_319_cast_fp16)[name = tensor("x_155_cast_fp16")]; + tensor input_321_perm_0 = const()[name = tensor("input_321_perm_0"), val = tensor([0, 2, 1])]; + tensor input_321_cast_fp16 = transpose(perm = input_321_perm_0, x = x_155_cast_fp16)[name = tensor("transpose_185")]; + tensor input_323_cast_fp16 = add(x = input_307_cast_fp16, y = input_321_cast_fp16)[name = tensor("input_323_cast_fp16")]; + tensor input_325_axes_0 = const()[name = tensor("input_325_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73937472)))]; + tensor encoder_layers_5_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73938560)))]; + tensor input_325_cast_fp16 = layer_norm(axes = input_325_axes_0, beta = encoder_layers_5_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_norm_feed_forward2_weight_to_fp16, x = input_323_cast_fp16)[name = tensor("input_325_cast_fp16")]; + tensor encoder_layers_5_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_5_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73939648)))]; + tensor linear_53_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_5_feed_forward2_linear1_weight_to_fp16, x = input_325_cast_fp16)[name = tensor("linear_53_cast_fp16")]; + tensor input_329_cast_fp16 = silu(x = linear_53_cast_fp16)[name = tensor("input_329_cast_fp16")]; + tensor encoder_layers_5_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_5_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76036864)))]; + tensor linear_54_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_feed_forward2_linear2_weight_to_fp16, x = input_329_cast_fp16)[name = tensor("linear_54_cast_fp16")]; + tensor var_1510_to_fp16 = const()[name = tensor("op_1510_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1511_cast_fp16 = mul(x = linear_54_cast_fp16, y = var_1510_to_fp16)[name = tensor("op_1511_cast_fp16")]; + tensor input_335_cast_fp16 = add(x = input_323_cast_fp16, y = var_1511_cast_fp16)[name = tensor("input_335_cast_fp16")]; + tensor input_337_axes_0 = const()[name = tensor("input_337_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78134080)))]; + tensor encoder_layers_5_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78135168)))]; + tensor input_337_cast_fp16 = layer_norm(axes = input_337_axes_0, beta = encoder_layers_5_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_norm_out_weight_to_fp16, x = input_335_cast_fp16)[name = tensor("input_337_cast_fp16")]; + tensor cache_25_begin_0 = const()[name = tensor("cache_25_begin_0"), val = tensor([6, 0, 0, 0])]; + tensor cache_25_end_0 = const()[name = tensor("cache_25_end_0"), val = tensor([7, 1, 70, 512])]; + tensor cache_25_end_mask_0 = const()[name = tensor("cache_25_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_25_squeeze_mask_0 = const()[name = tensor("cache_25_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_25_cast_fp16 = slice_by_index(begin = cache_25_begin_0, end = cache_25_end_0, end_mask = cache_25_end_mask_0, squeeze_mask = cache_25_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_25_cast_fp16")]; + tensor cache_27_begin_0 = const()[name = tensor("cache_27_begin_0"), val = tensor([6, 0, 0, 0])]; + tensor cache_27_end_0 = const()[name = tensor("cache_27_end_0"), val = tensor([7, 1, 512, 8])]; + tensor cache_27_end_mask_0 = const()[name = tensor("cache_27_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_27_squeeze_mask_0 = const()[name = tensor("cache_27_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_27_cast_fp16 = slice_by_index(begin = cache_27_begin_0, end = cache_27_end_0, end_mask = cache_27_end_mask_0, squeeze_mask = cache_27_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_27_cast_fp16")]; + tensor input_339_axes_0 = const()[name = tensor("input_339_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78136256)))]; + tensor encoder_layers_6_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78137344)))]; + tensor input_339_cast_fp16 = layer_norm(axes = input_339_axes_0, beta = encoder_layers_6_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_norm_feed_forward1_weight_to_fp16, x = input_337_cast_fp16)[name = tensor("input_339_cast_fp16")]; + tensor encoder_layers_6_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_6_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78138432)))]; + tensor linear_55_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_6_feed_forward1_linear1_weight_to_fp16, x = input_339_cast_fp16)[name = tensor("linear_55_cast_fp16")]; + tensor input_343_cast_fp16 = silu(x = linear_55_cast_fp16)[name = tensor("input_343_cast_fp16")]; + tensor encoder_layers_6_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_6_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80235648)))]; + tensor linear_56_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_feed_forward1_linear2_weight_to_fp16, x = input_343_cast_fp16)[name = tensor("linear_56_cast_fp16")]; + tensor var_1545_to_fp16 = const()[name = tensor("op_1545_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1546_cast_fp16 = mul(x = linear_56_cast_fp16, y = var_1545_to_fp16)[name = tensor("op_1546_cast_fp16")]; + tensor input_349_cast_fp16 = add(x = input_337_cast_fp16, y = var_1546_cast_fp16)[name = tensor("input_349_cast_fp16")]; + tensor key_13_axes_0 = const()[name = tensor("key_13_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82332864)))]; + tensor encoder_layers_6_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82333952)))]; + tensor key_13_cast_fp16 = layer_norm(axes = key_13_axes_0, beta = encoder_layers_6_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_norm_self_att_weight_to_fp16, x = input_349_cast_fp16)[name = tensor("key_13_cast_fp16")]; + tensor input_351_interleave_0 = const()[name = tensor("input_351_interleave_0"), val = tensor(false)]; + tensor input_351_cast_fp16 = concat(axis = var_64, interleave = input_351_interleave_0, values = (cache_25_cast_fp16, key_13_cast_fp16))[name = tensor("input_351_cast_fp16")]; + tensor var_1568_begin_0 = const()[name = tensor("op_1568_begin_0"), val = tensor([0, 1, 0])]; + tensor var_1568_end_0 = const()[name = tensor("op_1568_end_0"), val = tensor([1, 70, 512])]; + tensor var_1568_end_mask_0 = const()[name = tensor("op_1568_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1568_cast_fp16 = slice_by_index(begin = var_1568_begin_0, end = var_1568_end_0, end_mask = var_1568_end_mask_0, x = cache_25_cast_fp16)[name = tensor("op_1568_cast_fp16")]; + tensor var_1571_begin_0 = const()[name = tensor("op_1571_begin_0"), val = tensor([0, 0, 0])]; + tensor var_1571_end_0 = const()[name = tensor("op_1571_end_0"), val = tensor([1, 1, 512])]; + tensor var_1571_end_mask_0 = const()[name = tensor("op_1571_end_mask_0"), val = tensor([true, false, true])]; + tensor var_1571_cast_fp16 = slice_by_index(begin = var_1571_begin_0, end = var_1571_end_0, end_mask = var_1571_end_mask_0, x = key_13_cast_fp16)[name = tensor("op_1571_cast_fp16")]; + tensor var_1574_interleave_0 = const()[name = tensor("op_1574_interleave_0"), val = tensor(false)]; + tensor var_1574_cast_fp16 = concat(axis = var_64, interleave = var_1574_interleave_0, values = (var_1568_cast_fp16, var_1571_cast_fp16))[name = tensor("op_1574_cast_fp16")]; + tensor encoder_layers_6_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82335040)))]; + tensor linear_57_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_self_attn_linear_q_weight_to_fp16, x = key_13_cast_fp16)[name = tensor("linear_57_cast_fp16")]; + tensor var_1578 = const()[name = tensor("op_1578"), val = tensor([1, -1, 8, 64])]; + tensor q_37_cast_fp16 = reshape(shape = var_1578, x = linear_57_cast_fp16)[name = tensor("q_37_cast_fp16")]; + tensor encoder_layers_6_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82859392)))]; + tensor linear_58_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_self_attn_linear_k_weight_to_fp16, x = input_351_cast_fp16)[name = tensor("linear_58_cast_fp16")]; + tensor var_1582 = const()[name = tensor("op_1582"), val = tensor([1, -1, 8, 64])]; + tensor k_25_cast_fp16 = reshape(shape = var_1582, x = linear_58_cast_fp16)[name = tensor("k_25_cast_fp16")]; + tensor encoder_layers_6_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83383744)))]; + tensor linear_59_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_self_attn_linear_v_weight_to_fp16, x = input_351_cast_fp16)[name = tensor("linear_59_cast_fp16")]; + tensor var_1586 = const()[name = tensor("op_1586"), val = tensor([1, -1, 8, 64])]; + tensor v_13_cast_fp16 = reshape(shape = var_1586, x = linear_59_cast_fp16)[name = tensor("v_13_cast_fp16")]; + tensor value_15_perm_0 = const()[name = tensor("value_15_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_6_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83908096)))]; + tensor var_1598_cast_fp16 = add(x = q_37_cast_fp16, y = encoder_layers_6_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1598_cast_fp16")]; + tensor encoder_layers_6_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83909184)))]; + tensor var_1600_cast_fp16 = add(x = q_37_cast_fp16, y = encoder_layers_6_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1600_cast_fp16")]; + tensor q_with_bias_v_13_perm_0 = const()[name = tensor("q_with_bias_v_13_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_163_transpose_x_0 = const()[name = tensor("x_163_transpose_x_0"), val = tensor(false)]; + tensor x_163_transpose_y_0 = const()[name = tensor("x_163_transpose_y_0"), val = tensor(false)]; + tensor var_1602_to_fp16 = const()[name = tensor("op_1602_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83910272)))]; + tensor q_with_bias_v_13_cast_fp16 = transpose(perm = q_with_bias_v_13_perm_0, x = var_1600_cast_fp16)[name = tensor("transpose_183")]; + tensor x_163_cast_fp16 = matmul(transpose_x = x_163_transpose_x_0, transpose_y = x_163_transpose_y_0, x = q_with_bias_v_13_cast_fp16, y = var_1602_to_fp16)[name = tensor("x_163_cast_fp16")]; + tensor x_165_pad_0 = const()[name = tensor("x_165_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_165_mode_0 = const()[name = tensor("x_165_mode_0"), val = tensor("constant")]; + tensor const_101_to_fp16 = const()[name = tensor("const_101_to_fp16"), val = tensor(0x0p+0)]; + tensor x_165_cast_fp16 = pad(constant_val = const_101_to_fp16, mode = x_165_mode_0, pad = x_165_pad_0, x = x_163_cast_fp16)[name = tensor("x_165_cast_fp16")]; + tensor var_1610 = const()[name = tensor("op_1610"), val = tensor([1, 8, -1, 3])]; + tensor x_167_cast_fp16 = reshape(shape = var_1610, x = x_165_cast_fp16)[name = tensor("x_167_cast_fp16")]; + tensor var_1614_begin_0 = const()[name = tensor("op_1614_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1614_end_0 = const()[name = tensor("op_1614_end_0"), val = tensor([1, 8, 146, 3])]; + tensor var_1614_end_mask_0 = const()[name = tensor("op_1614_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1614_cast_fp16 = slice_by_index(begin = var_1614_begin_0, end = var_1614_end_0, end_mask = var_1614_end_mask_0, x = x_167_cast_fp16)[name = tensor("op_1614_cast_fp16")]; + tensor var_1615 = const()[name = tensor("op_1615"), val = tensor([1, 8, 3, 145])]; + tensor matrix_bd_25_cast_fp16 = reshape(shape = var_1615, x = var_1614_cast_fp16)[name = tensor("matrix_bd_25_cast_fp16")]; + tensor matrix_ac_13_transpose_x_0 = const()[name = tensor("matrix_ac_13_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_13_transpose_y_0 = const()[name = tensor("matrix_ac_13_transpose_y_0"), val = tensor(false)]; + tensor transpose_63_perm_0 = const()[name = tensor("transpose_63_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_64_perm_0 = const()[name = tensor("transpose_64_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_64 = transpose(perm = transpose_64_perm_0, x = k_25_cast_fp16)[name = tensor("transpose_181")]; + tensor transpose_63 = transpose(perm = transpose_63_perm_0, x = var_1598_cast_fp16)[name = tensor("transpose_182")]; + tensor matrix_ac_13_cast_fp16 = matmul(transpose_x = matrix_ac_13_transpose_x_0, transpose_y = matrix_ac_13_transpose_y_0, x = transpose_63, y = transpose_64)[name = tensor("matrix_ac_13_cast_fp16")]; + tensor matrix_bd_27_begin_0 = const()[name = tensor("matrix_bd_27_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_27_end_0 = const()[name = tensor("matrix_bd_27_end_0"), val = tensor([1, 8, 3, 73])]; + tensor matrix_bd_27_end_mask_0 = const()[name = tensor("matrix_bd_27_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_27_cast_fp16 = slice_by_index(begin = matrix_bd_27_begin_0, end = matrix_bd_27_end_0, end_mask = matrix_bd_27_end_mask_0, x = matrix_bd_25_cast_fp16)[name = tensor("matrix_bd_27_cast_fp16")]; + tensor var_1624_cast_fp16 = add(x = matrix_ac_13_cast_fp16, y = matrix_bd_27_cast_fp16)[name = tensor("op_1624_cast_fp16")]; + tensor _inversed_scores_25_y_0_to_fp16 = const()[name = tensor("_inversed_scores_25_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_25_cast_fp16 = mul(x = var_1624_cast_fp16, y = _inversed_scores_25_y_0_to_fp16)[name = tensor("_inversed_scores_25_cast_fp16")]; + tensor scores_27_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_25_cast_fp16, cond = mask_3)[name = tensor("scores_27_cast_fp16")]; + tensor var_1630_cast_fp16 = softmax(axis = var_62, x = scores_27_cast_fp16)[name = tensor("op_1630_cast_fp16")]; + tensor input_353_cast_fp16 = select(a = var_40_to_fp16, b = var_1630_cast_fp16, cond = mask_3)[name = tensor("input_353_cast_fp16")]; + tensor x_169_transpose_x_0 = const()[name = tensor("x_169_transpose_x_0"), val = tensor(false)]; + tensor x_169_transpose_y_0 = const()[name = tensor("x_169_transpose_y_0"), val = tensor(false)]; + tensor value_15_cast_fp16 = transpose(perm = value_15_perm_0, x = v_13_cast_fp16)[name = tensor("transpose_184")]; + tensor x_169_cast_fp16 = matmul(transpose_x = x_169_transpose_x_0, transpose_y = x_169_transpose_y_0, x = input_353_cast_fp16, y = value_15_cast_fp16)[name = tensor("x_169_cast_fp16")]; + tensor var_1634_perm_0 = const()[name = tensor("op_1634_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1635 = const()[name = tensor("op_1635"), val = tensor([1, -1, 512])]; + tensor var_1634_cast_fp16 = transpose(perm = var_1634_perm_0, x = x_169_cast_fp16)[name = tensor("transpose_180")]; + tensor input_355_cast_fp16 = reshape(shape = var_1635, x = var_1634_cast_fp16)[name = tensor("input_355_cast_fp16")]; + tensor encoder_layers_6_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84058816)))]; + tensor linear_61_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_self_attn_linear_out_weight_to_fp16, x = input_355_cast_fp16)[name = tensor("linear_61_cast_fp16")]; + tensor input_359_cast_fp16 = add(x = input_349_cast_fp16, y = linear_61_cast_fp16)[name = tensor("input_359_cast_fp16")]; + tensor x_173_axes_0 = const()[name = tensor("x_173_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84583168)))]; + tensor encoder_layers_6_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84584256)))]; + tensor x_173_cast_fp16 = layer_norm(axes = x_173_axes_0, beta = encoder_layers_6_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_norm_conv_weight_to_fp16, x = input_359_cast_fp16)[name = tensor("x_173_cast_fp16")]; + tensor input_361_perm_0 = const()[name = tensor("input_361_perm_0"), val = tensor([0, 2, 1])]; + tensor input_363_pad_type_0 = const()[name = tensor("input_363_pad_type_0"), val = tensor("valid")]; + tensor input_363_strides_0 = const()[name = tensor("input_363_strides_0"), val = tensor([1])]; + tensor input_363_pad_0 = const()[name = tensor("input_363_pad_0"), val = tensor([0, 0])]; + tensor input_363_dilations_0 = const()[name = tensor("input_363_dilations_0"), val = tensor([1])]; + tensor input_363_groups_0 = const()[name = tensor("input_363_groups_0"), val = tensor(1)]; + tensor encoder_layers_6_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_6_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84585344)))]; + tensor input_361_cast_fp16 = transpose(perm = input_361_perm_0, x = x_173_cast_fp16)[name = tensor("transpose_179")]; + tensor input_363_cast_fp16 = conv(dilations = input_363_dilations_0, groups = input_363_groups_0, pad = input_363_pad_0, pad_type = input_363_pad_type_0, strides = input_363_strides_0, weight = encoder_layers_6_conv_pointwise_conv1_weight_to_fp16, x = input_361_cast_fp16)[name = tensor("input_363_cast_fp16")]; + tensor x_175_split_num_splits_0 = const()[name = tensor("x_175_split_num_splits_0"), val = tensor(2)]; + tensor x_175_split_axis_0 = const()[name = tensor("x_175_split_axis_0"), val = tensor(1)]; + tensor x_175_split_cast_fp16_0, tensor x_175_split_cast_fp16_1 = split(axis = x_175_split_axis_0, num_splits = x_175_split_num_splits_0, x = input_363_cast_fp16)[name = tensor("x_175_split_cast_fp16")]; + tensor x_175_split_1_sigmoid_cast_fp16 = sigmoid(x = x_175_split_cast_fp16_1)[name = tensor("x_175_split_1_sigmoid_cast_fp16")]; + tensor x_175_cast_fp16 = mul(x = x_175_split_cast_fp16_0, y = x_175_split_1_sigmoid_cast_fp16)[name = tensor("x_175_cast_fp16")]; + tensor input_365_cast_fp16 = select(a = var_40_to_fp16, b = x_175_cast_fp16, cond = var_418)[name = tensor("input_365_cast_fp16")]; + tensor new_x_27_interleave_0 = const()[name = tensor("new_x_27_interleave_0"), val = tensor(false)]; + tensor new_x_27_cast_fp16 = concat(axis = var_62, interleave = new_x_27_interleave_0, values = (cache_27_cast_fp16, input_365_cast_fp16))[name = tensor("new_x_27_cast_fp16")]; + tensor next_cache_13_begin_0 = const()[name = tensor("next_cache_13_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_13_end_0 = const()[name = tensor("next_cache_13_end_0"), val = tensor([1, 512, 9])]; + tensor next_cache_13_end_mask_0 = const()[name = tensor("next_cache_13_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_13_cast_fp16 = slice_by_index(begin = next_cache_13_begin_0, end = next_cache_13_end_0, end_mask = next_cache_13_end_mask_0, x = new_x_27_cast_fp16)[name = tensor("next_cache_13_cast_fp16")]; + tensor var_1676_begin_0 = const()[name = tensor("op_1676_begin_0"), val = tensor([0, 0, 1])]; + tensor var_1676_end_0 = const()[name = tensor("op_1676_end_0"), val = tensor([1, 512, 9])]; + tensor var_1676_end_mask_0 = const()[name = tensor("op_1676_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1676_cast_fp16 = slice_by_index(begin = var_1676_begin_0, end = var_1676_end_0, end_mask = var_1676_end_mask_0, x = next_cache_13_cast_fp16)[name = tensor("op_1676_cast_fp16")]; + tensor x_177_pad_type_0 = const()[name = tensor("x_177_pad_type_0"), val = tensor("valid")]; + tensor x_177_groups_0 = const()[name = tensor("x_177_groups_0"), val = tensor(512)]; + tensor x_177_strides_0 = const()[name = tensor("x_177_strides_0"), val = tensor([1])]; + tensor x_177_pad_0 = const()[name = tensor("x_177_pad_0"), val = tensor([0, 0])]; + tensor x_177_dilations_0 = const()[name = tensor("x_177_dilations_0"), val = tensor([1])]; + tensor encoder_layers_6_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_6_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85633984)))]; + tensor x_177_cast_fp16 = conv(dilations = x_177_dilations_0, groups = x_177_groups_0, pad = x_177_pad_0, pad_type = x_177_pad_type_0, strides = x_177_strides_0, weight = encoder_layers_6_conv_depthwise_conv_weight_to_fp16, x = new_x_27_cast_fp16)[name = tensor("x_177_cast_fp16")]; + tensor input_367_perm_0 = const()[name = tensor("input_367_perm_0"), val = tensor([0, 2, 1])]; + tensor x_179_axes_0 = const()[name = tensor("x_179_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_6_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85643264)))]; + tensor encoder_layers_6_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_6_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85644352)))]; + tensor input_367_cast_fp16 = transpose(perm = input_367_perm_0, x = x_177_cast_fp16)[name = tensor("transpose_178")]; + tensor x_179_cast_fp16 = layer_norm(axes = x_179_axes_0, beta = encoder_layers_6_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_conv_batch_norm_weight_to_fp16, x = input_367_cast_fp16)[name = tensor("x_179_cast_fp16")]; + tensor input_369_perm_0 = const()[name = tensor("input_369_perm_0"), val = tensor([0, 2, 1])]; + tensor input_369_cast_fp16 = transpose(perm = input_369_perm_0, x = x_179_cast_fp16)[name = tensor("transpose_177")]; + tensor input_371_cast_fp16 = silu(x = input_369_cast_fp16)[name = tensor("input_371_cast_fp16")]; + tensor x_181_pad_type_0 = const()[name = tensor("x_181_pad_type_0"), val = tensor("valid")]; + tensor x_181_strides_0 = const()[name = tensor("x_181_strides_0"), val = tensor([1])]; + tensor x_181_pad_0 = const()[name = tensor("x_181_pad_0"), val = tensor([0, 0])]; + tensor x_181_dilations_0 = const()[name = tensor("x_181_dilations_0"), val = tensor([1])]; + tensor x_181_groups_0 = const()[name = tensor("x_181_groups_0"), val = tensor(1)]; + tensor encoder_layers_6_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_6_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85645440)))]; + tensor x_181_cast_fp16 = conv(dilations = x_181_dilations_0, groups = x_181_groups_0, pad = x_181_pad_0, pad_type = x_181_pad_type_0, strides = x_181_strides_0, weight = encoder_layers_6_conv_pointwise_conv2_weight_to_fp16, x = input_371_cast_fp16)[name = tensor("x_181_cast_fp16")]; + tensor input_373_perm_0 = const()[name = tensor("input_373_perm_0"), val = tensor([0, 2, 1])]; + tensor input_373_cast_fp16 = transpose(perm = input_373_perm_0, x = x_181_cast_fp16)[name = tensor("transpose_176")]; + tensor input_375_cast_fp16 = add(x = input_359_cast_fp16, y = input_373_cast_fp16)[name = tensor("input_375_cast_fp16")]; + tensor input_377_axes_0 = const()[name = tensor("input_377_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86169792)))]; + tensor encoder_layers_6_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86170880)))]; + tensor input_377_cast_fp16 = layer_norm(axes = input_377_axes_0, beta = encoder_layers_6_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_norm_feed_forward2_weight_to_fp16, x = input_375_cast_fp16)[name = tensor("input_377_cast_fp16")]; + tensor encoder_layers_6_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_6_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86171968)))]; + tensor linear_62_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_6_feed_forward2_linear1_weight_to_fp16, x = input_377_cast_fp16)[name = tensor("linear_62_cast_fp16")]; + tensor input_381_cast_fp16 = silu(x = linear_62_cast_fp16)[name = tensor("input_381_cast_fp16")]; + tensor encoder_layers_6_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_6_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88269184)))]; + tensor linear_63_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_feed_forward2_linear2_weight_to_fp16, x = input_381_cast_fp16)[name = tensor("linear_63_cast_fp16")]; + tensor var_1717_to_fp16 = const()[name = tensor("op_1717_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1718_cast_fp16 = mul(x = linear_63_cast_fp16, y = var_1717_to_fp16)[name = tensor("op_1718_cast_fp16")]; + tensor input_387_cast_fp16 = add(x = input_375_cast_fp16, y = var_1718_cast_fp16)[name = tensor("input_387_cast_fp16")]; + tensor input_389_axes_0 = const()[name = tensor("input_389_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90366400)))]; + tensor encoder_layers_6_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90367488)))]; + tensor input_389_cast_fp16 = layer_norm(axes = input_389_axes_0, beta = encoder_layers_6_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_norm_out_weight_to_fp16, x = input_387_cast_fp16)[name = tensor("input_389_cast_fp16")]; + tensor cache_29_begin_0 = const()[name = tensor("cache_29_begin_0"), val = tensor([7, 0, 0, 0])]; + tensor cache_29_end_0 = const()[name = tensor("cache_29_end_0"), val = tensor([8, 1, 70, 512])]; + tensor cache_29_end_mask_0 = const()[name = tensor("cache_29_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_29_squeeze_mask_0 = const()[name = tensor("cache_29_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_29_cast_fp16 = slice_by_index(begin = cache_29_begin_0, end = cache_29_end_0, end_mask = cache_29_end_mask_0, squeeze_mask = cache_29_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_29_cast_fp16")]; + tensor cache_31_begin_0 = const()[name = tensor("cache_31_begin_0"), val = tensor([7, 0, 0, 0])]; + tensor cache_31_end_0 = const()[name = tensor("cache_31_end_0"), val = tensor([8, 1, 512, 8])]; + tensor cache_31_end_mask_0 = const()[name = tensor("cache_31_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_31_squeeze_mask_0 = const()[name = tensor("cache_31_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_31_cast_fp16 = slice_by_index(begin = cache_31_begin_0, end = cache_31_end_0, end_mask = cache_31_end_mask_0, squeeze_mask = cache_31_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_31_cast_fp16")]; + tensor input_391_axes_0 = const()[name = tensor("input_391_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90368576)))]; + tensor encoder_layers_7_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90369664)))]; + tensor input_391_cast_fp16 = layer_norm(axes = input_391_axes_0, beta = encoder_layers_7_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_norm_feed_forward1_weight_to_fp16, x = input_389_cast_fp16)[name = tensor("input_391_cast_fp16")]; + tensor encoder_layers_7_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_7_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90370752)))]; + tensor linear_64_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_7_feed_forward1_linear1_weight_to_fp16, x = input_391_cast_fp16)[name = tensor("linear_64_cast_fp16")]; + tensor input_395_cast_fp16 = silu(x = linear_64_cast_fp16)[name = tensor("input_395_cast_fp16")]; + tensor encoder_layers_7_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_7_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92467968)))]; + tensor linear_65_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_feed_forward1_linear2_weight_to_fp16, x = input_395_cast_fp16)[name = tensor("linear_65_cast_fp16")]; + tensor var_1752_to_fp16 = const()[name = tensor("op_1752_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1753_cast_fp16 = mul(x = linear_65_cast_fp16, y = var_1752_to_fp16)[name = tensor("op_1753_cast_fp16")]; + tensor input_401_cast_fp16 = add(x = input_389_cast_fp16, y = var_1753_cast_fp16)[name = tensor("input_401_cast_fp16")]; + tensor key_15_axes_0 = const()[name = tensor("key_15_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94565184)))]; + tensor encoder_layers_7_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94566272)))]; + tensor key_15_cast_fp16 = layer_norm(axes = key_15_axes_0, beta = encoder_layers_7_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_norm_self_att_weight_to_fp16, x = input_401_cast_fp16)[name = tensor("key_15_cast_fp16")]; + tensor input_403_interleave_0 = const()[name = tensor("input_403_interleave_0"), val = tensor(false)]; + tensor input_403_cast_fp16 = concat(axis = var_64, interleave = input_403_interleave_0, values = (cache_29_cast_fp16, key_15_cast_fp16))[name = tensor("input_403_cast_fp16")]; + tensor var_1775_begin_0 = const()[name = tensor("op_1775_begin_0"), val = tensor([0, 1, 0])]; + tensor var_1775_end_0 = const()[name = tensor("op_1775_end_0"), val = tensor([1, 70, 512])]; + tensor var_1775_end_mask_0 = const()[name = tensor("op_1775_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1775_cast_fp16 = slice_by_index(begin = var_1775_begin_0, end = var_1775_end_0, end_mask = var_1775_end_mask_0, x = cache_29_cast_fp16)[name = tensor("op_1775_cast_fp16")]; + tensor var_1778_begin_0 = const()[name = tensor("op_1778_begin_0"), val = tensor([0, 0, 0])]; + tensor var_1778_end_0 = const()[name = tensor("op_1778_end_0"), val = tensor([1, 1, 512])]; + tensor var_1778_end_mask_0 = const()[name = tensor("op_1778_end_mask_0"), val = tensor([true, false, true])]; + tensor var_1778_cast_fp16 = slice_by_index(begin = var_1778_begin_0, end = var_1778_end_0, end_mask = var_1778_end_mask_0, x = key_15_cast_fp16)[name = tensor("op_1778_cast_fp16")]; + tensor var_1781_interleave_0 = const()[name = tensor("op_1781_interleave_0"), val = tensor(false)]; + tensor var_1781_cast_fp16 = concat(axis = var_64, interleave = var_1781_interleave_0, values = (var_1775_cast_fp16, var_1778_cast_fp16))[name = tensor("op_1781_cast_fp16")]; + tensor encoder_layers_7_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94567360)))]; + tensor linear_66_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_self_attn_linear_q_weight_to_fp16, x = key_15_cast_fp16)[name = tensor("linear_66_cast_fp16")]; + tensor var_1785 = const()[name = tensor("op_1785"), val = tensor([1, -1, 8, 64])]; + tensor q_43_cast_fp16 = reshape(shape = var_1785, x = linear_66_cast_fp16)[name = tensor("q_43_cast_fp16")]; + tensor encoder_layers_7_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95091712)))]; + tensor linear_67_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_self_attn_linear_k_weight_to_fp16, x = input_403_cast_fp16)[name = tensor("linear_67_cast_fp16")]; + tensor var_1789 = const()[name = tensor("op_1789"), val = tensor([1, -1, 8, 64])]; + tensor k_29_cast_fp16 = reshape(shape = var_1789, x = linear_67_cast_fp16)[name = tensor("k_29_cast_fp16")]; + tensor encoder_layers_7_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95616064)))]; + tensor linear_68_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_self_attn_linear_v_weight_to_fp16, x = input_403_cast_fp16)[name = tensor("linear_68_cast_fp16")]; + tensor var_1793 = const()[name = tensor("op_1793"), val = tensor([1, -1, 8, 64])]; + tensor v_15_cast_fp16 = reshape(shape = var_1793, x = linear_68_cast_fp16)[name = tensor("v_15_cast_fp16")]; + tensor value_17_perm_0 = const()[name = tensor("value_17_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_7_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96140416)))]; + tensor var_1805_cast_fp16 = add(x = q_43_cast_fp16, y = encoder_layers_7_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1805_cast_fp16")]; + tensor encoder_layers_7_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96141504)))]; + tensor var_1807_cast_fp16 = add(x = q_43_cast_fp16, y = encoder_layers_7_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1807_cast_fp16")]; + tensor q_with_bias_v_15_perm_0 = const()[name = tensor("q_with_bias_v_15_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_189_transpose_x_0 = const()[name = tensor("x_189_transpose_x_0"), val = tensor(false)]; + tensor x_189_transpose_y_0 = const()[name = tensor("x_189_transpose_y_0"), val = tensor(false)]; + tensor var_1809_to_fp16 = const()[name = tensor("op_1809_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96142592)))]; + tensor q_with_bias_v_15_cast_fp16 = transpose(perm = q_with_bias_v_15_perm_0, x = var_1807_cast_fp16)[name = tensor("transpose_174")]; + tensor x_189_cast_fp16 = matmul(transpose_x = x_189_transpose_x_0, transpose_y = x_189_transpose_y_0, x = q_with_bias_v_15_cast_fp16, y = var_1809_to_fp16)[name = tensor("x_189_cast_fp16")]; + tensor x_191_pad_0 = const()[name = tensor("x_191_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_191_mode_0 = const()[name = tensor("x_191_mode_0"), val = tensor("constant")]; + tensor const_114_to_fp16 = const()[name = tensor("const_114_to_fp16"), val = tensor(0x0p+0)]; + tensor x_191_cast_fp16 = pad(constant_val = const_114_to_fp16, mode = x_191_mode_0, pad = x_191_pad_0, x = x_189_cast_fp16)[name = tensor("x_191_cast_fp16")]; + tensor var_1817 = const()[name = tensor("op_1817"), val = tensor([1, 8, -1, 3])]; + tensor x_193_cast_fp16 = reshape(shape = var_1817, x = x_191_cast_fp16)[name = tensor("x_193_cast_fp16")]; + tensor var_1821_begin_0 = const()[name = tensor("op_1821_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1821_end_0 = const()[name = tensor("op_1821_end_0"), val = tensor([1, 8, 146, 3])]; + tensor var_1821_end_mask_0 = const()[name = tensor("op_1821_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1821_cast_fp16 = slice_by_index(begin = var_1821_begin_0, end = var_1821_end_0, end_mask = var_1821_end_mask_0, x = x_193_cast_fp16)[name = tensor("op_1821_cast_fp16")]; + tensor var_1822 = const()[name = tensor("op_1822"), val = tensor([1, 8, 3, 145])]; + tensor matrix_bd_29_cast_fp16 = reshape(shape = var_1822, x = var_1821_cast_fp16)[name = tensor("matrix_bd_29_cast_fp16")]; + tensor matrix_ac_15_transpose_x_0 = const()[name = tensor("matrix_ac_15_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_15_transpose_y_0 = const()[name = tensor("matrix_ac_15_transpose_y_0"), val = tensor(false)]; + tensor transpose_65_perm_0 = const()[name = tensor("transpose_65_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_66_perm_0 = const()[name = tensor("transpose_66_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_66 = transpose(perm = transpose_66_perm_0, x = k_29_cast_fp16)[name = tensor("transpose_172")]; + tensor transpose_65 = transpose(perm = transpose_65_perm_0, x = var_1805_cast_fp16)[name = tensor("transpose_173")]; + tensor matrix_ac_15_cast_fp16 = matmul(transpose_x = matrix_ac_15_transpose_x_0, transpose_y = matrix_ac_15_transpose_y_0, x = transpose_65, y = transpose_66)[name = tensor("matrix_ac_15_cast_fp16")]; + tensor matrix_bd_31_begin_0 = const()[name = tensor("matrix_bd_31_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_31_end_0 = const()[name = tensor("matrix_bd_31_end_0"), val = tensor([1, 8, 3, 73])]; + tensor matrix_bd_31_end_mask_0 = const()[name = tensor("matrix_bd_31_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_31_cast_fp16 = slice_by_index(begin = matrix_bd_31_begin_0, end = matrix_bd_31_end_0, end_mask = matrix_bd_31_end_mask_0, x = matrix_bd_29_cast_fp16)[name = tensor("matrix_bd_31_cast_fp16")]; + tensor var_1831_cast_fp16 = add(x = matrix_ac_15_cast_fp16, y = matrix_bd_31_cast_fp16)[name = tensor("op_1831_cast_fp16")]; + tensor _inversed_scores_29_y_0_to_fp16 = const()[name = tensor("_inversed_scores_29_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_29_cast_fp16 = mul(x = var_1831_cast_fp16, y = _inversed_scores_29_y_0_to_fp16)[name = tensor("_inversed_scores_29_cast_fp16")]; + tensor scores_31_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_29_cast_fp16, cond = mask_3)[name = tensor("scores_31_cast_fp16")]; + tensor var_1837_cast_fp16 = softmax(axis = var_62, x = scores_31_cast_fp16)[name = tensor("op_1837_cast_fp16")]; + tensor input_405_cast_fp16 = select(a = var_40_to_fp16, b = var_1837_cast_fp16, cond = mask_3)[name = tensor("input_405_cast_fp16")]; + tensor x_195_transpose_x_0 = const()[name = tensor("x_195_transpose_x_0"), val = tensor(false)]; + tensor x_195_transpose_y_0 = const()[name = tensor("x_195_transpose_y_0"), val = tensor(false)]; + tensor value_17_cast_fp16 = transpose(perm = value_17_perm_0, x = v_15_cast_fp16)[name = tensor("transpose_175")]; + tensor x_195_cast_fp16 = matmul(transpose_x = x_195_transpose_x_0, transpose_y = x_195_transpose_y_0, x = input_405_cast_fp16, y = value_17_cast_fp16)[name = tensor("x_195_cast_fp16")]; + tensor var_1841_perm_0 = const()[name = tensor("op_1841_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1842 = const()[name = tensor("op_1842"), val = tensor([1, -1, 512])]; + tensor var_1841_cast_fp16 = transpose(perm = var_1841_perm_0, x = x_195_cast_fp16)[name = tensor("transpose_171")]; + tensor input_407_cast_fp16 = reshape(shape = var_1842, x = var_1841_cast_fp16)[name = tensor("input_407_cast_fp16")]; + tensor encoder_layers_7_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96291136)))]; + tensor linear_70_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_self_attn_linear_out_weight_to_fp16, x = input_407_cast_fp16)[name = tensor("linear_70_cast_fp16")]; + tensor input_411_cast_fp16 = add(x = input_401_cast_fp16, y = linear_70_cast_fp16)[name = tensor("input_411_cast_fp16")]; + tensor x_199_axes_0 = const()[name = tensor("x_199_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96815488)))]; + tensor encoder_layers_7_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96816576)))]; + tensor x_199_cast_fp16 = layer_norm(axes = x_199_axes_0, beta = encoder_layers_7_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_norm_conv_weight_to_fp16, x = input_411_cast_fp16)[name = tensor("x_199_cast_fp16")]; + tensor input_413_perm_0 = const()[name = tensor("input_413_perm_0"), val = tensor([0, 2, 1])]; + tensor input_415_pad_type_0 = const()[name = tensor("input_415_pad_type_0"), val = tensor("valid")]; + tensor input_415_strides_0 = const()[name = tensor("input_415_strides_0"), val = tensor([1])]; + tensor input_415_pad_0 = const()[name = tensor("input_415_pad_0"), val = tensor([0, 0])]; + tensor input_415_dilations_0 = const()[name = tensor("input_415_dilations_0"), val = tensor([1])]; + tensor input_415_groups_0 = const()[name = tensor("input_415_groups_0"), val = tensor(1)]; + tensor encoder_layers_7_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_7_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96817664)))]; + tensor input_413_cast_fp16 = transpose(perm = input_413_perm_0, x = x_199_cast_fp16)[name = tensor("transpose_170")]; + tensor input_415_cast_fp16 = conv(dilations = input_415_dilations_0, groups = input_415_groups_0, pad = input_415_pad_0, pad_type = input_415_pad_type_0, strides = input_415_strides_0, weight = encoder_layers_7_conv_pointwise_conv1_weight_to_fp16, x = input_413_cast_fp16)[name = tensor("input_415_cast_fp16")]; + tensor x_201_split_num_splits_0 = const()[name = tensor("x_201_split_num_splits_0"), val = tensor(2)]; + tensor x_201_split_axis_0 = const()[name = tensor("x_201_split_axis_0"), val = tensor(1)]; + tensor x_201_split_cast_fp16_0, tensor x_201_split_cast_fp16_1 = split(axis = x_201_split_axis_0, num_splits = x_201_split_num_splits_0, x = input_415_cast_fp16)[name = tensor("x_201_split_cast_fp16")]; + tensor x_201_split_1_sigmoid_cast_fp16 = sigmoid(x = x_201_split_cast_fp16_1)[name = tensor("x_201_split_1_sigmoid_cast_fp16")]; + tensor x_201_cast_fp16 = mul(x = x_201_split_cast_fp16_0, y = x_201_split_1_sigmoid_cast_fp16)[name = tensor("x_201_cast_fp16")]; + tensor input_417_cast_fp16 = select(a = var_40_to_fp16, b = x_201_cast_fp16, cond = var_418)[name = tensor("input_417_cast_fp16")]; + tensor new_x_31_interleave_0 = const()[name = tensor("new_x_31_interleave_0"), val = tensor(false)]; + tensor new_x_31_cast_fp16 = concat(axis = var_62, interleave = new_x_31_interleave_0, values = (cache_31_cast_fp16, input_417_cast_fp16))[name = tensor("new_x_31_cast_fp16")]; + tensor next_cache_15_begin_0 = const()[name = tensor("next_cache_15_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_15_end_0 = const()[name = tensor("next_cache_15_end_0"), val = tensor([1, 512, 9])]; + tensor next_cache_15_end_mask_0 = const()[name = tensor("next_cache_15_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_15_cast_fp16 = slice_by_index(begin = next_cache_15_begin_0, end = next_cache_15_end_0, end_mask = next_cache_15_end_mask_0, x = new_x_31_cast_fp16)[name = tensor("next_cache_15_cast_fp16")]; + tensor var_1883_begin_0 = const()[name = tensor("op_1883_begin_0"), val = tensor([0, 0, 1])]; + tensor var_1883_end_0 = const()[name = tensor("op_1883_end_0"), val = tensor([1, 512, 9])]; + tensor var_1883_end_mask_0 = const()[name = tensor("op_1883_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1883_cast_fp16 = slice_by_index(begin = var_1883_begin_0, end = var_1883_end_0, end_mask = var_1883_end_mask_0, x = next_cache_15_cast_fp16)[name = tensor("op_1883_cast_fp16")]; + tensor x_203_pad_type_0 = const()[name = tensor("x_203_pad_type_0"), val = tensor("valid")]; + tensor x_203_groups_0 = const()[name = tensor("x_203_groups_0"), val = tensor(512)]; + tensor x_203_strides_0 = const()[name = tensor("x_203_strides_0"), val = tensor([1])]; + tensor x_203_pad_0 = const()[name = tensor("x_203_pad_0"), val = tensor([0, 0])]; + tensor x_203_dilations_0 = const()[name = tensor("x_203_dilations_0"), val = tensor([1])]; + tensor encoder_layers_7_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_7_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97866304)))]; + tensor x_203_cast_fp16 = conv(dilations = x_203_dilations_0, groups = x_203_groups_0, pad = x_203_pad_0, pad_type = x_203_pad_type_0, strides = x_203_strides_0, weight = encoder_layers_7_conv_depthwise_conv_weight_to_fp16, x = new_x_31_cast_fp16)[name = tensor("x_203_cast_fp16")]; + tensor input_419_perm_0 = const()[name = tensor("input_419_perm_0"), val = tensor([0, 2, 1])]; + tensor x_205_axes_0 = const()[name = tensor("x_205_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_7_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97875584)))]; + tensor encoder_layers_7_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_7_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97876672)))]; + tensor input_419_cast_fp16 = transpose(perm = input_419_perm_0, x = x_203_cast_fp16)[name = tensor("transpose_169")]; + tensor x_205_cast_fp16 = layer_norm(axes = x_205_axes_0, beta = encoder_layers_7_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_conv_batch_norm_weight_to_fp16, x = input_419_cast_fp16)[name = tensor("x_205_cast_fp16")]; + tensor input_421_perm_0 = const()[name = tensor("input_421_perm_0"), val = tensor([0, 2, 1])]; + tensor input_421_cast_fp16 = transpose(perm = input_421_perm_0, x = x_205_cast_fp16)[name = tensor("transpose_168")]; + tensor input_423_cast_fp16 = silu(x = input_421_cast_fp16)[name = tensor("input_423_cast_fp16")]; + tensor x_207_pad_type_0 = const()[name = tensor("x_207_pad_type_0"), val = tensor("valid")]; + tensor x_207_strides_0 = const()[name = tensor("x_207_strides_0"), val = tensor([1])]; + tensor x_207_pad_0 = const()[name = tensor("x_207_pad_0"), val = tensor([0, 0])]; + tensor x_207_dilations_0 = const()[name = tensor("x_207_dilations_0"), val = tensor([1])]; + tensor x_207_groups_0 = const()[name = tensor("x_207_groups_0"), val = tensor(1)]; + tensor encoder_layers_7_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_7_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97877760)))]; + tensor x_207_cast_fp16 = conv(dilations = x_207_dilations_0, groups = x_207_groups_0, pad = x_207_pad_0, pad_type = x_207_pad_type_0, strides = x_207_strides_0, weight = encoder_layers_7_conv_pointwise_conv2_weight_to_fp16, x = input_423_cast_fp16)[name = tensor("x_207_cast_fp16")]; + tensor input_425_perm_0 = const()[name = tensor("input_425_perm_0"), val = tensor([0, 2, 1])]; + tensor input_425_cast_fp16 = transpose(perm = input_425_perm_0, x = x_207_cast_fp16)[name = tensor("transpose_167")]; + tensor input_427_cast_fp16 = add(x = input_411_cast_fp16, y = input_425_cast_fp16)[name = tensor("input_427_cast_fp16")]; + tensor input_429_axes_0 = const()[name = tensor("input_429_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98402112)))]; + tensor encoder_layers_7_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98403200)))]; + tensor input_429_cast_fp16 = layer_norm(axes = input_429_axes_0, beta = encoder_layers_7_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_norm_feed_forward2_weight_to_fp16, x = input_427_cast_fp16)[name = tensor("input_429_cast_fp16")]; + tensor encoder_layers_7_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_7_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98404288)))]; + tensor linear_71_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_7_feed_forward2_linear1_weight_to_fp16, x = input_429_cast_fp16)[name = tensor("linear_71_cast_fp16")]; + tensor input_433_cast_fp16 = silu(x = linear_71_cast_fp16)[name = tensor("input_433_cast_fp16")]; + tensor encoder_layers_7_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_7_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100501504)))]; + tensor linear_72_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_feed_forward2_linear2_weight_to_fp16, x = input_433_cast_fp16)[name = tensor("linear_72_cast_fp16")]; + tensor var_1924_to_fp16 = const()[name = tensor("op_1924_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1925_cast_fp16 = mul(x = linear_72_cast_fp16, y = var_1924_to_fp16)[name = tensor("op_1925_cast_fp16")]; + tensor input_439_cast_fp16 = add(x = input_427_cast_fp16, y = var_1925_cast_fp16)[name = tensor("input_439_cast_fp16")]; + tensor input_441_axes_0 = const()[name = tensor("input_441_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102598720)))]; + tensor encoder_layers_7_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102599808)))]; + tensor input_441_cast_fp16 = layer_norm(axes = input_441_axes_0, beta = encoder_layers_7_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_norm_out_weight_to_fp16, x = input_439_cast_fp16)[name = tensor("input_441_cast_fp16")]; + tensor cache_33_begin_0 = const()[name = tensor("cache_33_begin_0"), val = tensor([8, 0, 0, 0])]; + tensor cache_33_end_0 = const()[name = tensor("cache_33_end_0"), val = tensor([9, 1, 70, 512])]; + tensor cache_33_end_mask_0 = const()[name = tensor("cache_33_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_33_squeeze_mask_0 = const()[name = tensor("cache_33_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_33_cast_fp16 = slice_by_index(begin = cache_33_begin_0, end = cache_33_end_0, end_mask = cache_33_end_mask_0, squeeze_mask = cache_33_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_33_cast_fp16")]; + tensor cache_35_begin_0 = const()[name = tensor("cache_35_begin_0"), val = tensor([8, 0, 0, 0])]; + tensor cache_35_end_0 = const()[name = tensor("cache_35_end_0"), val = tensor([9, 1, 512, 8])]; + tensor cache_35_end_mask_0 = const()[name = tensor("cache_35_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_35_squeeze_mask_0 = const()[name = tensor("cache_35_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_35_cast_fp16 = slice_by_index(begin = cache_35_begin_0, end = cache_35_end_0, end_mask = cache_35_end_mask_0, squeeze_mask = cache_35_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_35_cast_fp16")]; + tensor input_443_axes_0 = const()[name = tensor("input_443_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102600896)))]; + tensor encoder_layers_8_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102601984)))]; + tensor input_443_cast_fp16 = layer_norm(axes = input_443_axes_0, beta = encoder_layers_8_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_norm_feed_forward1_weight_to_fp16, x = input_441_cast_fp16)[name = tensor("input_443_cast_fp16")]; + tensor encoder_layers_8_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_8_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102603072)))]; + tensor linear_73_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_8_feed_forward1_linear1_weight_to_fp16, x = input_443_cast_fp16)[name = tensor("linear_73_cast_fp16")]; + tensor input_447_cast_fp16 = silu(x = linear_73_cast_fp16)[name = tensor("input_447_cast_fp16")]; + tensor encoder_layers_8_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_8_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(104700288)))]; + tensor linear_74_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_feed_forward1_linear2_weight_to_fp16, x = input_447_cast_fp16)[name = tensor("linear_74_cast_fp16")]; + tensor var_1959_to_fp16 = const()[name = tensor("op_1959_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1960_cast_fp16 = mul(x = linear_74_cast_fp16, y = var_1959_to_fp16)[name = tensor("op_1960_cast_fp16")]; + tensor input_453_cast_fp16 = add(x = input_441_cast_fp16, y = var_1960_cast_fp16)[name = tensor("input_453_cast_fp16")]; + tensor key_17_axes_0 = const()[name = tensor("key_17_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106797504)))]; + tensor encoder_layers_8_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106798592)))]; + tensor key_17_cast_fp16 = layer_norm(axes = key_17_axes_0, beta = encoder_layers_8_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_norm_self_att_weight_to_fp16, x = input_453_cast_fp16)[name = tensor("key_17_cast_fp16")]; + tensor input_455_interleave_0 = const()[name = tensor("input_455_interleave_0"), val = tensor(false)]; + tensor input_455_cast_fp16 = concat(axis = var_64, interleave = input_455_interleave_0, values = (cache_33_cast_fp16, key_17_cast_fp16))[name = tensor("input_455_cast_fp16")]; + tensor var_1982_begin_0 = const()[name = tensor("op_1982_begin_0"), val = tensor([0, 1, 0])]; + tensor var_1982_end_0 = const()[name = tensor("op_1982_end_0"), val = tensor([1, 70, 512])]; + tensor var_1982_end_mask_0 = const()[name = tensor("op_1982_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1982_cast_fp16 = slice_by_index(begin = var_1982_begin_0, end = var_1982_end_0, end_mask = var_1982_end_mask_0, x = cache_33_cast_fp16)[name = tensor("op_1982_cast_fp16")]; + tensor var_1985_begin_0 = const()[name = tensor("op_1985_begin_0"), val = tensor([0, 0, 0])]; + tensor var_1985_end_0 = const()[name = tensor("op_1985_end_0"), val = tensor([1, 1, 512])]; + tensor var_1985_end_mask_0 = const()[name = tensor("op_1985_end_mask_0"), val = tensor([true, false, true])]; + tensor var_1985_cast_fp16 = slice_by_index(begin = var_1985_begin_0, end = var_1985_end_0, end_mask = var_1985_end_mask_0, x = key_17_cast_fp16)[name = tensor("op_1985_cast_fp16")]; + tensor var_1988_interleave_0 = const()[name = tensor("op_1988_interleave_0"), val = tensor(false)]; + tensor var_1988_cast_fp16 = concat(axis = var_64, interleave = var_1988_interleave_0, values = (var_1982_cast_fp16, var_1985_cast_fp16))[name = tensor("op_1988_cast_fp16")]; + tensor encoder_layers_8_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106799680)))]; + tensor linear_75_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_self_attn_linear_q_weight_to_fp16, x = key_17_cast_fp16)[name = tensor("linear_75_cast_fp16")]; + tensor var_1992 = const()[name = tensor("op_1992"), val = tensor([1, -1, 8, 64])]; + tensor q_49_cast_fp16 = reshape(shape = var_1992, x = linear_75_cast_fp16)[name = tensor("q_49_cast_fp16")]; + tensor encoder_layers_8_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107324032)))]; + tensor linear_76_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_self_attn_linear_k_weight_to_fp16, x = input_455_cast_fp16)[name = tensor("linear_76_cast_fp16")]; + tensor var_1996 = const()[name = tensor("op_1996"), val = tensor([1, -1, 8, 64])]; + tensor k_33_cast_fp16 = reshape(shape = var_1996, x = linear_76_cast_fp16)[name = tensor("k_33_cast_fp16")]; + tensor encoder_layers_8_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107848384)))]; + tensor linear_77_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_self_attn_linear_v_weight_to_fp16, x = input_455_cast_fp16)[name = tensor("linear_77_cast_fp16")]; + tensor var_2000 = const()[name = tensor("op_2000"), val = tensor([1, -1, 8, 64])]; + tensor v_17_cast_fp16 = reshape(shape = var_2000, x = linear_77_cast_fp16)[name = tensor("v_17_cast_fp16")]; + tensor value_19_perm_0 = const()[name = tensor("value_19_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_8_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108372736)))]; + tensor var_2012_cast_fp16 = add(x = q_49_cast_fp16, y = encoder_layers_8_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2012_cast_fp16")]; + tensor encoder_layers_8_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108373824)))]; + tensor var_2014_cast_fp16 = add(x = q_49_cast_fp16, y = encoder_layers_8_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2014_cast_fp16")]; + tensor q_with_bias_v_17_perm_0 = const()[name = tensor("q_with_bias_v_17_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_215_transpose_x_0 = const()[name = tensor("x_215_transpose_x_0"), val = tensor(false)]; + tensor x_215_transpose_y_0 = const()[name = tensor("x_215_transpose_y_0"), val = tensor(false)]; + tensor var_2016_to_fp16 = const()[name = tensor("op_2016_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108374912)))]; + tensor q_with_bias_v_17_cast_fp16 = transpose(perm = q_with_bias_v_17_perm_0, x = var_2014_cast_fp16)[name = tensor("transpose_165")]; + tensor x_215_cast_fp16 = matmul(transpose_x = x_215_transpose_x_0, transpose_y = x_215_transpose_y_0, x = q_with_bias_v_17_cast_fp16, y = var_2016_to_fp16)[name = tensor("x_215_cast_fp16")]; + tensor x_217_pad_0 = const()[name = tensor("x_217_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_217_mode_0 = const()[name = tensor("x_217_mode_0"), val = tensor("constant")]; + tensor const_127_to_fp16 = const()[name = tensor("const_127_to_fp16"), val = tensor(0x0p+0)]; + tensor x_217_cast_fp16 = pad(constant_val = const_127_to_fp16, mode = x_217_mode_0, pad = x_217_pad_0, x = x_215_cast_fp16)[name = tensor("x_217_cast_fp16")]; + tensor var_2024 = const()[name = tensor("op_2024"), val = tensor([1, 8, -1, 3])]; + tensor x_219_cast_fp16 = reshape(shape = var_2024, x = x_217_cast_fp16)[name = tensor("x_219_cast_fp16")]; + tensor var_2028_begin_0 = const()[name = tensor("op_2028_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2028_end_0 = const()[name = tensor("op_2028_end_0"), val = tensor([1, 8, 146, 3])]; + tensor var_2028_end_mask_0 = const()[name = tensor("op_2028_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2028_cast_fp16 = slice_by_index(begin = var_2028_begin_0, end = var_2028_end_0, end_mask = var_2028_end_mask_0, x = x_219_cast_fp16)[name = tensor("op_2028_cast_fp16")]; + tensor var_2029 = const()[name = tensor("op_2029"), val = tensor([1, 8, 3, 145])]; + tensor matrix_bd_33_cast_fp16 = reshape(shape = var_2029, x = var_2028_cast_fp16)[name = tensor("matrix_bd_33_cast_fp16")]; + tensor matrix_ac_17_transpose_x_0 = const()[name = tensor("matrix_ac_17_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_17_transpose_y_0 = const()[name = tensor("matrix_ac_17_transpose_y_0"), val = tensor(false)]; + tensor transpose_67_perm_0 = const()[name = tensor("transpose_67_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_68_perm_0 = const()[name = tensor("transpose_68_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_68 = transpose(perm = transpose_68_perm_0, x = k_33_cast_fp16)[name = tensor("transpose_163")]; + tensor transpose_67 = transpose(perm = transpose_67_perm_0, x = var_2012_cast_fp16)[name = tensor("transpose_164")]; + tensor matrix_ac_17_cast_fp16 = matmul(transpose_x = matrix_ac_17_transpose_x_0, transpose_y = matrix_ac_17_transpose_y_0, x = transpose_67, y = transpose_68)[name = tensor("matrix_ac_17_cast_fp16")]; + tensor matrix_bd_35_begin_0 = const()[name = tensor("matrix_bd_35_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_35_end_0 = const()[name = tensor("matrix_bd_35_end_0"), val = tensor([1, 8, 3, 73])]; + tensor matrix_bd_35_end_mask_0 = const()[name = tensor("matrix_bd_35_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_35_cast_fp16 = slice_by_index(begin = matrix_bd_35_begin_0, end = matrix_bd_35_end_0, end_mask = matrix_bd_35_end_mask_0, x = matrix_bd_33_cast_fp16)[name = tensor("matrix_bd_35_cast_fp16")]; + tensor var_2038_cast_fp16 = add(x = matrix_ac_17_cast_fp16, y = matrix_bd_35_cast_fp16)[name = tensor("op_2038_cast_fp16")]; + tensor _inversed_scores_33_y_0_to_fp16 = const()[name = tensor("_inversed_scores_33_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_33_cast_fp16 = mul(x = var_2038_cast_fp16, y = _inversed_scores_33_y_0_to_fp16)[name = tensor("_inversed_scores_33_cast_fp16")]; + tensor scores_35_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_33_cast_fp16, cond = mask_3)[name = tensor("scores_35_cast_fp16")]; + tensor var_2044_cast_fp16 = softmax(axis = var_62, x = scores_35_cast_fp16)[name = tensor("op_2044_cast_fp16")]; + tensor input_457_cast_fp16 = select(a = var_40_to_fp16, b = var_2044_cast_fp16, cond = mask_3)[name = tensor("input_457_cast_fp16")]; + tensor x_221_transpose_x_0 = const()[name = tensor("x_221_transpose_x_0"), val = tensor(false)]; + tensor x_221_transpose_y_0 = const()[name = tensor("x_221_transpose_y_0"), val = tensor(false)]; + tensor value_19_cast_fp16 = transpose(perm = value_19_perm_0, x = v_17_cast_fp16)[name = tensor("transpose_166")]; + tensor x_221_cast_fp16 = matmul(transpose_x = x_221_transpose_x_0, transpose_y = x_221_transpose_y_0, x = input_457_cast_fp16, y = value_19_cast_fp16)[name = tensor("x_221_cast_fp16")]; + tensor var_2048_perm_0 = const()[name = tensor("op_2048_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2049 = const()[name = tensor("op_2049"), val = tensor([1, -1, 512])]; + tensor var_2048_cast_fp16 = transpose(perm = var_2048_perm_0, x = x_221_cast_fp16)[name = tensor("transpose_162")]; + tensor input_459_cast_fp16 = reshape(shape = var_2049, x = var_2048_cast_fp16)[name = tensor("input_459_cast_fp16")]; + tensor encoder_layers_8_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108523456)))]; + tensor linear_79_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_self_attn_linear_out_weight_to_fp16, x = input_459_cast_fp16)[name = tensor("linear_79_cast_fp16")]; + tensor input_463_cast_fp16 = add(x = input_453_cast_fp16, y = linear_79_cast_fp16)[name = tensor("input_463_cast_fp16")]; + tensor x_225_axes_0 = const()[name = tensor("x_225_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109047808)))]; + tensor encoder_layers_8_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109048896)))]; + tensor x_225_cast_fp16 = layer_norm(axes = x_225_axes_0, beta = encoder_layers_8_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_norm_conv_weight_to_fp16, x = input_463_cast_fp16)[name = tensor("x_225_cast_fp16")]; + tensor input_465_perm_0 = const()[name = tensor("input_465_perm_0"), val = tensor([0, 2, 1])]; + tensor input_467_pad_type_0 = const()[name = tensor("input_467_pad_type_0"), val = tensor("valid")]; + tensor input_467_strides_0 = const()[name = tensor("input_467_strides_0"), val = tensor([1])]; + tensor input_467_pad_0 = const()[name = tensor("input_467_pad_0"), val = tensor([0, 0])]; + tensor input_467_dilations_0 = const()[name = tensor("input_467_dilations_0"), val = tensor([1])]; + tensor input_467_groups_0 = const()[name = tensor("input_467_groups_0"), val = tensor(1)]; + tensor encoder_layers_8_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_8_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109049984)))]; + tensor input_465_cast_fp16 = transpose(perm = input_465_perm_0, x = x_225_cast_fp16)[name = tensor("transpose_161")]; + tensor input_467_cast_fp16 = conv(dilations = input_467_dilations_0, groups = input_467_groups_0, pad = input_467_pad_0, pad_type = input_467_pad_type_0, strides = input_467_strides_0, weight = encoder_layers_8_conv_pointwise_conv1_weight_to_fp16, x = input_465_cast_fp16)[name = tensor("input_467_cast_fp16")]; + tensor x_227_split_num_splits_0 = const()[name = tensor("x_227_split_num_splits_0"), val = tensor(2)]; + tensor x_227_split_axis_0 = const()[name = tensor("x_227_split_axis_0"), val = tensor(1)]; + tensor x_227_split_cast_fp16_0, tensor x_227_split_cast_fp16_1 = split(axis = x_227_split_axis_0, num_splits = x_227_split_num_splits_0, x = input_467_cast_fp16)[name = tensor("x_227_split_cast_fp16")]; + tensor x_227_split_1_sigmoid_cast_fp16 = sigmoid(x = x_227_split_cast_fp16_1)[name = tensor("x_227_split_1_sigmoid_cast_fp16")]; + tensor x_227_cast_fp16 = mul(x = x_227_split_cast_fp16_0, y = x_227_split_1_sigmoid_cast_fp16)[name = tensor("x_227_cast_fp16")]; + tensor input_469_cast_fp16 = select(a = var_40_to_fp16, b = x_227_cast_fp16, cond = var_418)[name = tensor("input_469_cast_fp16")]; + tensor new_x_35_interleave_0 = const()[name = tensor("new_x_35_interleave_0"), val = tensor(false)]; + tensor new_x_35_cast_fp16 = concat(axis = var_62, interleave = new_x_35_interleave_0, values = (cache_35_cast_fp16, input_469_cast_fp16))[name = tensor("new_x_35_cast_fp16")]; + tensor next_cache_17_begin_0 = const()[name = tensor("next_cache_17_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_17_end_0 = const()[name = tensor("next_cache_17_end_0"), val = tensor([1, 512, 9])]; + tensor next_cache_17_end_mask_0 = const()[name = tensor("next_cache_17_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_17_cast_fp16 = slice_by_index(begin = next_cache_17_begin_0, end = next_cache_17_end_0, end_mask = next_cache_17_end_mask_0, x = new_x_35_cast_fp16)[name = tensor("next_cache_17_cast_fp16")]; + tensor var_2090_begin_0 = const()[name = tensor("op_2090_begin_0"), val = tensor([0, 0, 1])]; + tensor var_2090_end_0 = const()[name = tensor("op_2090_end_0"), val = tensor([1, 512, 9])]; + tensor var_2090_end_mask_0 = const()[name = tensor("op_2090_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2090_cast_fp16 = slice_by_index(begin = var_2090_begin_0, end = var_2090_end_0, end_mask = var_2090_end_mask_0, x = next_cache_17_cast_fp16)[name = tensor("op_2090_cast_fp16")]; + tensor x_229_pad_type_0 = const()[name = tensor("x_229_pad_type_0"), val = tensor("valid")]; + tensor x_229_groups_0 = const()[name = tensor("x_229_groups_0"), val = tensor(512)]; + tensor x_229_strides_0 = const()[name = tensor("x_229_strides_0"), val = tensor([1])]; + tensor x_229_pad_0 = const()[name = tensor("x_229_pad_0"), val = tensor([0, 0])]; + tensor x_229_dilations_0 = const()[name = tensor("x_229_dilations_0"), val = tensor([1])]; + tensor encoder_layers_8_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_8_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110098624)))]; + tensor x_229_cast_fp16 = conv(dilations = x_229_dilations_0, groups = x_229_groups_0, pad = x_229_pad_0, pad_type = x_229_pad_type_0, strides = x_229_strides_0, weight = encoder_layers_8_conv_depthwise_conv_weight_to_fp16, x = new_x_35_cast_fp16)[name = tensor("x_229_cast_fp16")]; + tensor input_471_perm_0 = const()[name = tensor("input_471_perm_0"), val = tensor([0, 2, 1])]; + tensor x_231_axes_0 = const()[name = tensor("x_231_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_8_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110107904)))]; + tensor encoder_layers_8_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_8_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110108992)))]; + tensor input_471_cast_fp16 = transpose(perm = input_471_perm_0, x = x_229_cast_fp16)[name = tensor("transpose_160")]; + tensor x_231_cast_fp16 = layer_norm(axes = x_231_axes_0, beta = encoder_layers_8_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_conv_batch_norm_weight_to_fp16, x = input_471_cast_fp16)[name = tensor("x_231_cast_fp16")]; + tensor input_473_perm_0 = const()[name = tensor("input_473_perm_0"), val = tensor([0, 2, 1])]; + tensor input_473_cast_fp16 = transpose(perm = input_473_perm_0, x = x_231_cast_fp16)[name = tensor("transpose_159")]; + tensor input_475_cast_fp16 = silu(x = input_473_cast_fp16)[name = tensor("input_475_cast_fp16")]; + tensor x_233_pad_type_0 = const()[name = tensor("x_233_pad_type_0"), val = tensor("valid")]; + tensor x_233_strides_0 = const()[name = tensor("x_233_strides_0"), val = tensor([1])]; + tensor x_233_pad_0 = const()[name = tensor("x_233_pad_0"), val = tensor([0, 0])]; + tensor x_233_dilations_0 = const()[name = tensor("x_233_dilations_0"), val = tensor([1])]; + tensor x_233_groups_0 = const()[name = tensor("x_233_groups_0"), val = tensor(1)]; + tensor encoder_layers_8_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_8_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110110080)))]; + tensor x_233_cast_fp16 = conv(dilations = x_233_dilations_0, groups = x_233_groups_0, pad = x_233_pad_0, pad_type = x_233_pad_type_0, strides = x_233_strides_0, weight = encoder_layers_8_conv_pointwise_conv2_weight_to_fp16, x = input_475_cast_fp16)[name = tensor("x_233_cast_fp16")]; + tensor input_477_perm_0 = const()[name = tensor("input_477_perm_0"), val = tensor([0, 2, 1])]; + tensor input_477_cast_fp16 = transpose(perm = input_477_perm_0, x = x_233_cast_fp16)[name = tensor("transpose_158")]; + tensor input_479_cast_fp16 = add(x = input_463_cast_fp16, y = input_477_cast_fp16)[name = tensor("input_479_cast_fp16")]; + tensor input_481_axes_0 = const()[name = tensor("input_481_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110634432)))]; + tensor encoder_layers_8_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110635520)))]; + tensor input_481_cast_fp16 = layer_norm(axes = input_481_axes_0, beta = encoder_layers_8_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_norm_feed_forward2_weight_to_fp16, x = input_479_cast_fp16)[name = tensor("input_481_cast_fp16")]; + tensor encoder_layers_8_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_8_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110636608)))]; + tensor linear_80_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_8_feed_forward2_linear1_weight_to_fp16, x = input_481_cast_fp16)[name = tensor("linear_80_cast_fp16")]; + tensor input_485_cast_fp16 = silu(x = linear_80_cast_fp16)[name = tensor("input_485_cast_fp16")]; + tensor encoder_layers_8_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_8_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(112733824)))]; + tensor linear_81_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_feed_forward2_linear2_weight_to_fp16, x = input_485_cast_fp16)[name = tensor("linear_81_cast_fp16")]; + tensor var_2131_to_fp16 = const()[name = tensor("op_2131_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2132_cast_fp16 = mul(x = linear_81_cast_fp16, y = var_2131_to_fp16)[name = tensor("op_2132_cast_fp16")]; + tensor input_491_cast_fp16 = add(x = input_479_cast_fp16, y = var_2132_cast_fp16)[name = tensor("input_491_cast_fp16")]; + tensor input_493_axes_0 = const()[name = tensor("input_493_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114831040)))]; + tensor encoder_layers_8_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114832128)))]; + tensor input_493_cast_fp16 = layer_norm(axes = input_493_axes_0, beta = encoder_layers_8_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_norm_out_weight_to_fp16, x = input_491_cast_fp16)[name = tensor("input_493_cast_fp16")]; + tensor cache_37_begin_0 = const()[name = tensor("cache_37_begin_0"), val = tensor([9, 0, 0, 0])]; + tensor cache_37_end_0 = const()[name = tensor("cache_37_end_0"), val = tensor([10, 1, 70, 512])]; + tensor cache_37_end_mask_0 = const()[name = tensor("cache_37_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_37_squeeze_mask_0 = const()[name = tensor("cache_37_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_37_cast_fp16 = slice_by_index(begin = cache_37_begin_0, end = cache_37_end_0, end_mask = cache_37_end_mask_0, squeeze_mask = cache_37_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_37_cast_fp16")]; + tensor cache_39_begin_0 = const()[name = tensor("cache_39_begin_0"), val = tensor([9, 0, 0, 0])]; + tensor cache_39_end_0 = const()[name = tensor("cache_39_end_0"), val = tensor([10, 1, 512, 8])]; + tensor cache_39_end_mask_0 = const()[name = tensor("cache_39_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_39_squeeze_mask_0 = const()[name = tensor("cache_39_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_39_cast_fp16 = slice_by_index(begin = cache_39_begin_0, end = cache_39_end_0, end_mask = cache_39_end_mask_0, squeeze_mask = cache_39_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_39_cast_fp16")]; + tensor input_495_axes_0 = const()[name = tensor("input_495_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114833216)))]; + tensor encoder_layers_9_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114834304)))]; + tensor input_495_cast_fp16 = layer_norm(axes = input_495_axes_0, beta = encoder_layers_9_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_norm_feed_forward1_weight_to_fp16, x = input_493_cast_fp16)[name = tensor("input_495_cast_fp16")]; + tensor encoder_layers_9_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_9_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114835392)))]; + tensor linear_82_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_9_feed_forward1_linear1_weight_to_fp16, x = input_495_cast_fp16)[name = tensor("linear_82_cast_fp16")]; + tensor input_499_cast_fp16 = silu(x = linear_82_cast_fp16)[name = tensor("input_499_cast_fp16")]; + tensor encoder_layers_9_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_9_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116932608)))]; + tensor linear_83_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_feed_forward1_linear2_weight_to_fp16, x = input_499_cast_fp16)[name = tensor("linear_83_cast_fp16")]; + tensor var_2166_to_fp16 = const()[name = tensor("op_2166_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2167_cast_fp16 = mul(x = linear_83_cast_fp16, y = var_2166_to_fp16)[name = tensor("op_2167_cast_fp16")]; + tensor input_505_cast_fp16 = add(x = input_493_cast_fp16, y = var_2167_cast_fp16)[name = tensor("input_505_cast_fp16")]; + tensor key_19_axes_0 = const()[name = tensor("key_19_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119029824)))]; + tensor encoder_layers_9_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119030912)))]; + tensor key_19_cast_fp16 = layer_norm(axes = key_19_axes_0, beta = encoder_layers_9_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_norm_self_att_weight_to_fp16, x = input_505_cast_fp16)[name = tensor("key_19_cast_fp16")]; + tensor input_507_interleave_0 = const()[name = tensor("input_507_interleave_0"), val = tensor(false)]; + tensor input_507_cast_fp16 = concat(axis = var_64, interleave = input_507_interleave_0, values = (cache_37_cast_fp16, key_19_cast_fp16))[name = tensor("input_507_cast_fp16")]; + tensor var_2189_begin_0 = const()[name = tensor("op_2189_begin_0"), val = tensor([0, 1, 0])]; + tensor var_2189_end_0 = const()[name = tensor("op_2189_end_0"), val = tensor([1, 70, 512])]; + tensor var_2189_end_mask_0 = const()[name = tensor("op_2189_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2189_cast_fp16 = slice_by_index(begin = var_2189_begin_0, end = var_2189_end_0, end_mask = var_2189_end_mask_0, x = cache_37_cast_fp16)[name = tensor("op_2189_cast_fp16")]; + tensor var_2192_begin_0 = const()[name = tensor("op_2192_begin_0"), val = tensor([0, 0, 0])]; + tensor var_2192_end_0 = const()[name = tensor("op_2192_end_0"), val = tensor([1, 1, 512])]; + tensor var_2192_end_mask_0 = const()[name = tensor("op_2192_end_mask_0"), val = tensor([true, false, true])]; + tensor var_2192_cast_fp16 = slice_by_index(begin = var_2192_begin_0, end = var_2192_end_0, end_mask = var_2192_end_mask_0, x = key_19_cast_fp16)[name = tensor("op_2192_cast_fp16")]; + tensor var_2195_interleave_0 = const()[name = tensor("op_2195_interleave_0"), val = tensor(false)]; + tensor var_2195_cast_fp16 = concat(axis = var_64, interleave = var_2195_interleave_0, values = (var_2189_cast_fp16, var_2192_cast_fp16))[name = tensor("op_2195_cast_fp16")]; + tensor encoder_layers_9_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119032000)))]; + tensor linear_84_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_self_attn_linear_q_weight_to_fp16, x = key_19_cast_fp16)[name = tensor("linear_84_cast_fp16")]; + tensor var_2199 = const()[name = tensor("op_2199"), val = tensor([1, -1, 8, 64])]; + tensor q_55_cast_fp16 = reshape(shape = var_2199, x = linear_84_cast_fp16)[name = tensor("q_55_cast_fp16")]; + tensor encoder_layers_9_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119556352)))]; + tensor linear_85_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_self_attn_linear_k_weight_to_fp16, x = input_507_cast_fp16)[name = tensor("linear_85_cast_fp16")]; + tensor var_2203 = const()[name = tensor("op_2203"), val = tensor([1, -1, 8, 64])]; + tensor k_37_cast_fp16 = reshape(shape = var_2203, x = linear_85_cast_fp16)[name = tensor("k_37_cast_fp16")]; + tensor encoder_layers_9_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120080704)))]; + tensor linear_86_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_self_attn_linear_v_weight_to_fp16, x = input_507_cast_fp16)[name = tensor("linear_86_cast_fp16")]; + tensor var_2207 = const()[name = tensor("op_2207"), val = tensor([1, -1, 8, 64])]; + tensor v_19_cast_fp16 = reshape(shape = var_2207, x = linear_86_cast_fp16)[name = tensor("v_19_cast_fp16")]; + tensor value_21_perm_0 = const()[name = tensor("value_21_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_9_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120605056)))]; + tensor var_2219_cast_fp16 = add(x = q_55_cast_fp16, y = encoder_layers_9_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2219_cast_fp16")]; + tensor encoder_layers_9_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120606144)))]; + tensor var_2221_cast_fp16 = add(x = q_55_cast_fp16, y = encoder_layers_9_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2221_cast_fp16")]; + tensor q_with_bias_v_19_perm_0 = const()[name = tensor("q_with_bias_v_19_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_241_transpose_x_0 = const()[name = tensor("x_241_transpose_x_0"), val = tensor(false)]; + tensor x_241_transpose_y_0 = const()[name = tensor("x_241_transpose_y_0"), val = tensor(false)]; + tensor var_2223_to_fp16 = const()[name = tensor("op_2223_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120607232)))]; + tensor q_with_bias_v_19_cast_fp16 = transpose(perm = q_with_bias_v_19_perm_0, x = var_2221_cast_fp16)[name = tensor("transpose_156")]; + tensor x_241_cast_fp16 = matmul(transpose_x = x_241_transpose_x_0, transpose_y = x_241_transpose_y_0, x = q_with_bias_v_19_cast_fp16, y = var_2223_to_fp16)[name = tensor("x_241_cast_fp16")]; + tensor x_243_pad_0 = const()[name = tensor("x_243_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_243_mode_0 = const()[name = tensor("x_243_mode_0"), val = tensor("constant")]; + tensor const_140_to_fp16 = const()[name = tensor("const_140_to_fp16"), val = tensor(0x0p+0)]; + tensor x_243_cast_fp16 = pad(constant_val = const_140_to_fp16, mode = x_243_mode_0, pad = x_243_pad_0, x = x_241_cast_fp16)[name = tensor("x_243_cast_fp16")]; + tensor var_2231 = const()[name = tensor("op_2231"), val = tensor([1, 8, -1, 3])]; + tensor x_245_cast_fp16 = reshape(shape = var_2231, x = x_243_cast_fp16)[name = tensor("x_245_cast_fp16")]; + tensor var_2235_begin_0 = const()[name = tensor("op_2235_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2235_end_0 = const()[name = tensor("op_2235_end_0"), val = tensor([1, 8, 146, 3])]; + tensor var_2235_end_mask_0 = const()[name = tensor("op_2235_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2235_cast_fp16 = slice_by_index(begin = var_2235_begin_0, end = var_2235_end_0, end_mask = var_2235_end_mask_0, x = x_245_cast_fp16)[name = tensor("op_2235_cast_fp16")]; + tensor var_2236 = const()[name = tensor("op_2236"), val = tensor([1, 8, 3, 145])]; + tensor matrix_bd_37_cast_fp16 = reshape(shape = var_2236, x = var_2235_cast_fp16)[name = tensor("matrix_bd_37_cast_fp16")]; + tensor matrix_ac_19_transpose_x_0 = const()[name = tensor("matrix_ac_19_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_19_transpose_y_0 = const()[name = tensor("matrix_ac_19_transpose_y_0"), val = tensor(false)]; + tensor transpose_69_perm_0 = const()[name = tensor("transpose_69_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_70_perm_0 = const()[name = tensor("transpose_70_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_70 = transpose(perm = transpose_70_perm_0, x = k_37_cast_fp16)[name = tensor("transpose_154")]; + tensor transpose_69 = transpose(perm = transpose_69_perm_0, x = var_2219_cast_fp16)[name = tensor("transpose_155")]; + tensor matrix_ac_19_cast_fp16 = matmul(transpose_x = matrix_ac_19_transpose_x_0, transpose_y = matrix_ac_19_transpose_y_0, x = transpose_69, y = transpose_70)[name = tensor("matrix_ac_19_cast_fp16")]; + tensor matrix_bd_39_begin_0 = const()[name = tensor("matrix_bd_39_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_39_end_0 = const()[name = tensor("matrix_bd_39_end_0"), val = tensor([1, 8, 3, 73])]; + tensor matrix_bd_39_end_mask_0 = const()[name = tensor("matrix_bd_39_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_39_cast_fp16 = slice_by_index(begin = matrix_bd_39_begin_0, end = matrix_bd_39_end_0, end_mask = matrix_bd_39_end_mask_0, x = matrix_bd_37_cast_fp16)[name = tensor("matrix_bd_39_cast_fp16")]; + tensor var_2245_cast_fp16 = add(x = matrix_ac_19_cast_fp16, y = matrix_bd_39_cast_fp16)[name = tensor("op_2245_cast_fp16")]; + tensor _inversed_scores_37_y_0_to_fp16 = const()[name = tensor("_inversed_scores_37_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_37_cast_fp16 = mul(x = var_2245_cast_fp16, y = _inversed_scores_37_y_0_to_fp16)[name = tensor("_inversed_scores_37_cast_fp16")]; + tensor scores_39_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_37_cast_fp16, cond = mask_3)[name = tensor("scores_39_cast_fp16")]; + tensor var_2251_cast_fp16 = softmax(axis = var_62, x = scores_39_cast_fp16)[name = tensor("op_2251_cast_fp16")]; + tensor input_509_cast_fp16 = select(a = var_40_to_fp16, b = var_2251_cast_fp16, cond = mask_3)[name = tensor("input_509_cast_fp16")]; + tensor x_247_transpose_x_0 = const()[name = tensor("x_247_transpose_x_0"), val = tensor(false)]; + tensor x_247_transpose_y_0 = const()[name = tensor("x_247_transpose_y_0"), val = tensor(false)]; + tensor value_21_cast_fp16 = transpose(perm = value_21_perm_0, x = v_19_cast_fp16)[name = tensor("transpose_157")]; + tensor x_247_cast_fp16 = matmul(transpose_x = x_247_transpose_x_0, transpose_y = x_247_transpose_y_0, x = input_509_cast_fp16, y = value_21_cast_fp16)[name = tensor("x_247_cast_fp16")]; + tensor var_2255_perm_0 = const()[name = tensor("op_2255_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2256 = const()[name = tensor("op_2256"), val = tensor([1, -1, 512])]; + tensor var_2255_cast_fp16 = transpose(perm = var_2255_perm_0, x = x_247_cast_fp16)[name = tensor("transpose_153")]; + tensor input_511_cast_fp16 = reshape(shape = var_2256, x = var_2255_cast_fp16)[name = tensor("input_511_cast_fp16")]; + tensor encoder_layers_9_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120755776)))]; + tensor linear_88_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_self_attn_linear_out_weight_to_fp16, x = input_511_cast_fp16)[name = tensor("linear_88_cast_fp16")]; + tensor input_515_cast_fp16 = add(x = input_505_cast_fp16, y = linear_88_cast_fp16)[name = tensor("input_515_cast_fp16")]; + tensor x_251_axes_0 = const()[name = tensor("x_251_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121280128)))]; + tensor encoder_layers_9_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121281216)))]; + tensor x_251_cast_fp16 = layer_norm(axes = x_251_axes_0, beta = encoder_layers_9_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_norm_conv_weight_to_fp16, x = input_515_cast_fp16)[name = tensor("x_251_cast_fp16")]; + tensor input_517_perm_0 = const()[name = tensor("input_517_perm_0"), val = tensor([0, 2, 1])]; + tensor input_519_pad_type_0 = const()[name = tensor("input_519_pad_type_0"), val = tensor("valid")]; + tensor input_519_strides_0 = const()[name = tensor("input_519_strides_0"), val = tensor([1])]; + tensor input_519_pad_0 = const()[name = tensor("input_519_pad_0"), val = tensor([0, 0])]; + tensor input_519_dilations_0 = const()[name = tensor("input_519_dilations_0"), val = tensor([1])]; + tensor input_519_groups_0 = const()[name = tensor("input_519_groups_0"), val = tensor(1)]; + tensor encoder_layers_9_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_9_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121282304)))]; + tensor input_517_cast_fp16 = transpose(perm = input_517_perm_0, x = x_251_cast_fp16)[name = tensor("transpose_152")]; + tensor input_519_cast_fp16 = conv(dilations = input_519_dilations_0, groups = input_519_groups_0, pad = input_519_pad_0, pad_type = input_519_pad_type_0, strides = input_519_strides_0, weight = encoder_layers_9_conv_pointwise_conv1_weight_to_fp16, x = input_517_cast_fp16)[name = tensor("input_519_cast_fp16")]; + tensor x_253_split_num_splits_0 = const()[name = tensor("x_253_split_num_splits_0"), val = tensor(2)]; + tensor x_253_split_axis_0 = const()[name = tensor("x_253_split_axis_0"), val = tensor(1)]; + tensor x_253_split_cast_fp16_0, tensor x_253_split_cast_fp16_1 = split(axis = x_253_split_axis_0, num_splits = x_253_split_num_splits_0, x = input_519_cast_fp16)[name = tensor("x_253_split_cast_fp16")]; + tensor x_253_split_1_sigmoid_cast_fp16 = sigmoid(x = x_253_split_cast_fp16_1)[name = tensor("x_253_split_1_sigmoid_cast_fp16")]; + tensor x_253_cast_fp16 = mul(x = x_253_split_cast_fp16_0, y = x_253_split_1_sigmoid_cast_fp16)[name = tensor("x_253_cast_fp16")]; + tensor input_521_cast_fp16 = select(a = var_40_to_fp16, b = x_253_cast_fp16, cond = var_418)[name = tensor("input_521_cast_fp16")]; + tensor new_x_39_interleave_0 = const()[name = tensor("new_x_39_interleave_0"), val = tensor(false)]; + tensor new_x_39_cast_fp16 = concat(axis = var_62, interleave = new_x_39_interleave_0, values = (cache_39_cast_fp16, input_521_cast_fp16))[name = tensor("new_x_39_cast_fp16")]; + tensor next_cache_19_begin_0 = const()[name = tensor("next_cache_19_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_19_end_0 = const()[name = tensor("next_cache_19_end_0"), val = tensor([1, 512, 9])]; + tensor next_cache_19_end_mask_0 = const()[name = tensor("next_cache_19_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_19_cast_fp16 = slice_by_index(begin = next_cache_19_begin_0, end = next_cache_19_end_0, end_mask = next_cache_19_end_mask_0, x = new_x_39_cast_fp16)[name = tensor("next_cache_19_cast_fp16")]; + tensor var_2297_begin_0 = const()[name = tensor("op_2297_begin_0"), val = tensor([0, 0, 1])]; + tensor var_2297_end_0 = const()[name = tensor("op_2297_end_0"), val = tensor([1, 512, 9])]; + tensor var_2297_end_mask_0 = const()[name = tensor("op_2297_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2297_cast_fp16 = slice_by_index(begin = var_2297_begin_0, end = var_2297_end_0, end_mask = var_2297_end_mask_0, x = next_cache_19_cast_fp16)[name = tensor("op_2297_cast_fp16")]; + tensor x_255_pad_type_0 = const()[name = tensor("x_255_pad_type_0"), val = tensor("valid")]; + tensor x_255_groups_0 = const()[name = tensor("x_255_groups_0"), val = tensor(512)]; + tensor x_255_strides_0 = const()[name = tensor("x_255_strides_0"), val = tensor([1])]; + tensor x_255_pad_0 = const()[name = tensor("x_255_pad_0"), val = tensor([0, 0])]; + tensor x_255_dilations_0 = const()[name = tensor("x_255_dilations_0"), val = tensor([1])]; + tensor encoder_layers_9_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_9_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122330944)))]; + tensor x_255_cast_fp16 = conv(dilations = x_255_dilations_0, groups = x_255_groups_0, pad = x_255_pad_0, pad_type = x_255_pad_type_0, strides = x_255_strides_0, weight = encoder_layers_9_conv_depthwise_conv_weight_to_fp16, x = new_x_39_cast_fp16)[name = tensor("x_255_cast_fp16")]; + tensor input_523_perm_0 = const()[name = tensor("input_523_perm_0"), val = tensor([0, 2, 1])]; + tensor x_257_axes_0 = const()[name = tensor("x_257_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_9_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122340224)))]; + tensor encoder_layers_9_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_9_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122341312)))]; + tensor input_523_cast_fp16 = transpose(perm = input_523_perm_0, x = x_255_cast_fp16)[name = tensor("transpose_151")]; + tensor x_257_cast_fp16 = layer_norm(axes = x_257_axes_0, beta = encoder_layers_9_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_conv_batch_norm_weight_to_fp16, x = input_523_cast_fp16)[name = tensor("x_257_cast_fp16")]; + tensor input_525_perm_0 = const()[name = tensor("input_525_perm_0"), val = tensor([0, 2, 1])]; + tensor input_525_cast_fp16 = transpose(perm = input_525_perm_0, x = x_257_cast_fp16)[name = tensor("transpose_150")]; + tensor input_527_cast_fp16 = silu(x = input_525_cast_fp16)[name = tensor("input_527_cast_fp16")]; + tensor x_259_pad_type_0 = const()[name = tensor("x_259_pad_type_0"), val = tensor("valid")]; + tensor x_259_strides_0 = const()[name = tensor("x_259_strides_0"), val = tensor([1])]; + tensor x_259_pad_0 = const()[name = tensor("x_259_pad_0"), val = tensor([0, 0])]; + tensor x_259_dilations_0 = const()[name = tensor("x_259_dilations_0"), val = tensor([1])]; + tensor x_259_groups_0 = const()[name = tensor("x_259_groups_0"), val = tensor(1)]; + tensor encoder_layers_9_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_9_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122342400)))]; + tensor x_259_cast_fp16 = conv(dilations = x_259_dilations_0, groups = x_259_groups_0, pad = x_259_pad_0, pad_type = x_259_pad_type_0, strides = x_259_strides_0, weight = encoder_layers_9_conv_pointwise_conv2_weight_to_fp16, x = input_527_cast_fp16)[name = tensor("x_259_cast_fp16")]; + tensor input_529_perm_0 = const()[name = tensor("input_529_perm_0"), val = tensor([0, 2, 1])]; + tensor input_529_cast_fp16 = transpose(perm = input_529_perm_0, x = x_259_cast_fp16)[name = tensor("transpose_149")]; + tensor input_531_cast_fp16 = add(x = input_515_cast_fp16, y = input_529_cast_fp16)[name = tensor("input_531_cast_fp16")]; + tensor input_533_axes_0 = const()[name = tensor("input_533_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122866752)))]; + tensor encoder_layers_9_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122867840)))]; + tensor input_533_cast_fp16 = layer_norm(axes = input_533_axes_0, beta = encoder_layers_9_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_norm_feed_forward2_weight_to_fp16, x = input_531_cast_fp16)[name = tensor("input_533_cast_fp16")]; + tensor encoder_layers_9_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_9_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122868928)))]; + tensor linear_89_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_9_feed_forward2_linear1_weight_to_fp16, x = input_533_cast_fp16)[name = tensor("linear_89_cast_fp16")]; + tensor input_537_cast_fp16 = silu(x = linear_89_cast_fp16)[name = tensor("input_537_cast_fp16")]; + tensor encoder_layers_9_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_9_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(124966144)))]; + tensor linear_90_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_feed_forward2_linear2_weight_to_fp16, x = input_537_cast_fp16)[name = tensor("linear_90_cast_fp16")]; + tensor var_2338_to_fp16 = const()[name = tensor("op_2338_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2339_cast_fp16 = mul(x = linear_90_cast_fp16, y = var_2338_to_fp16)[name = tensor("op_2339_cast_fp16")]; + tensor input_543_cast_fp16 = add(x = input_531_cast_fp16, y = var_2339_cast_fp16)[name = tensor("input_543_cast_fp16")]; + tensor input_545_axes_0 = const()[name = tensor("input_545_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127063360)))]; + tensor encoder_layers_9_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127064448)))]; + tensor input_545_cast_fp16 = layer_norm(axes = input_545_axes_0, beta = encoder_layers_9_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_norm_out_weight_to_fp16, x = input_543_cast_fp16)[name = tensor("input_545_cast_fp16")]; + tensor cache_41_begin_0 = const()[name = tensor("cache_41_begin_0"), val = tensor([10, 0, 0, 0])]; + tensor cache_41_end_0 = const()[name = tensor("cache_41_end_0"), val = tensor([11, 1, 70, 512])]; + tensor cache_41_end_mask_0 = const()[name = tensor("cache_41_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_41_squeeze_mask_0 = const()[name = tensor("cache_41_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_41_cast_fp16 = slice_by_index(begin = cache_41_begin_0, end = cache_41_end_0, end_mask = cache_41_end_mask_0, squeeze_mask = cache_41_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_41_cast_fp16")]; + tensor cache_43_begin_0 = const()[name = tensor("cache_43_begin_0"), val = tensor([10, 0, 0, 0])]; + tensor cache_43_end_0 = const()[name = tensor("cache_43_end_0"), val = tensor([11, 1, 512, 8])]; + tensor cache_43_end_mask_0 = const()[name = tensor("cache_43_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_43_squeeze_mask_0 = const()[name = tensor("cache_43_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_43_cast_fp16 = slice_by_index(begin = cache_43_begin_0, end = cache_43_end_0, end_mask = cache_43_end_mask_0, squeeze_mask = cache_43_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_43_cast_fp16")]; + tensor input_547_axes_0 = const()[name = tensor("input_547_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127065536)))]; + tensor encoder_layers_10_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127066624)))]; + tensor input_547_cast_fp16 = layer_norm(axes = input_547_axes_0, beta = encoder_layers_10_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_norm_feed_forward1_weight_to_fp16, x = input_545_cast_fp16)[name = tensor("input_547_cast_fp16")]; + tensor encoder_layers_10_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_10_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127067712)))]; + tensor linear_91_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_10_feed_forward1_linear1_weight_to_fp16, x = input_547_cast_fp16)[name = tensor("linear_91_cast_fp16")]; + tensor input_551_cast_fp16 = silu(x = linear_91_cast_fp16)[name = tensor("input_551_cast_fp16")]; + tensor encoder_layers_10_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_10_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129164928)))]; + tensor linear_92_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_feed_forward1_linear2_weight_to_fp16, x = input_551_cast_fp16)[name = tensor("linear_92_cast_fp16")]; + tensor var_2373_to_fp16 = const()[name = tensor("op_2373_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2374_cast_fp16 = mul(x = linear_92_cast_fp16, y = var_2373_to_fp16)[name = tensor("op_2374_cast_fp16")]; + tensor input_557_cast_fp16 = add(x = input_545_cast_fp16, y = var_2374_cast_fp16)[name = tensor("input_557_cast_fp16")]; + tensor key_21_axes_0 = const()[name = tensor("key_21_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131262144)))]; + tensor encoder_layers_10_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131263232)))]; + tensor key_21_cast_fp16 = layer_norm(axes = key_21_axes_0, beta = encoder_layers_10_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_norm_self_att_weight_to_fp16, x = input_557_cast_fp16)[name = tensor("key_21_cast_fp16")]; + tensor input_559_interleave_0 = const()[name = tensor("input_559_interleave_0"), val = tensor(false)]; + tensor input_559_cast_fp16 = concat(axis = var_64, interleave = input_559_interleave_0, values = (cache_41_cast_fp16, key_21_cast_fp16))[name = tensor("input_559_cast_fp16")]; + tensor var_2396_begin_0 = const()[name = tensor("op_2396_begin_0"), val = tensor([0, 1, 0])]; + tensor var_2396_end_0 = const()[name = tensor("op_2396_end_0"), val = tensor([1, 70, 512])]; + tensor var_2396_end_mask_0 = const()[name = tensor("op_2396_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2396_cast_fp16 = slice_by_index(begin = var_2396_begin_0, end = var_2396_end_0, end_mask = var_2396_end_mask_0, x = cache_41_cast_fp16)[name = tensor("op_2396_cast_fp16")]; + tensor var_2399_begin_0 = const()[name = tensor("op_2399_begin_0"), val = tensor([0, 0, 0])]; + tensor var_2399_end_0 = const()[name = tensor("op_2399_end_0"), val = tensor([1, 1, 512])]; + tensor var_2399_end_mask_0 = const()[name = tensor("op_2399_end_mask_0"), val = tensor([true, false, true])]; + tensor var_2399_cast_fp16 = slice_by_index(begin = var_2399_begin_0, end = var_2399_end_0, end_mask = var_2399_end_mask_0, x = key_21_cast_fp16)[name = tensor("op_2399_cast_fp16")]; + tensor var_2402_interleave_0 = const()[name = tensor("op_2402_interleave_0"), val = tensor(false)]; + tensor var_2402_cast_fp16 = concat(axis = var_64, interleave = var_2402_interleave_0, values = (var_2396_cast_fp16, var_2399_cast_fp16))[name = tensor("op_2402_cast_fp16")]; + tensor encoder_layers_10_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131264320)))]; + tensor linear_93_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_self_attn_linear_q_weight_to_fp16, x = key_21_cast_fp16)[name = tensor("linear_93_cast_fp16")]; + tensor var_2406 = const()[name = tensor("op_2406"), val = tensor([1, -1, 8, 64])]; + tensor q_61_cast_fp16 = reshape(shape = var_2406, x = linear_93_cast_fp16)[name = tensor("q_61_cast_fp16")]; + tensor encoder_layers_10_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131788672)))]; + tensor linear_94_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_self_attn_linear_k_weight_to_fp16, x = input_559_cast_fp16)[name = tensor("linear_94_cast_fp16")]; + tensor var_2410 = const()[name = tensor("op_2410"), val = tensor([1, -1, 8, 64])]; + tensor k_41_cast_fp16 = reshape(shape = var_2410, x = linear_94_cast_fp16)[name = tensor("k_41_cast_fp16")]; + tensor encoder_layers_10_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132313024)))]; + tensor linear_95_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_self_attn_linear_v_weight_to_fp16, x = input_559_cast_fp16)[name = tensor("linear_95_cast_fp16")]; + tensor var_2414 = const()[name = tensor("op_2414"), val = tensor([1, -1, 8, 64])]; + tensor v_21_cast_fp16 = reshape(shape = var_2414, x = linear_95_cast_fp16)[name = tensor("v_21_cast_fp16")]; + tensor value_23_perm_0 = const()[name = tensor("value_23_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_10_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132837376)))]; + tensor var_2426_cast_fp16 = add(x = q_61_cast_fp16, y = encoder_layers_10_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2426_cast_fp16")]; + tensor encoder_layers_10_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132838464)))]; + tensor var_2428_cast_fp16 = add(x = q_61_cast_fp16, y = encoder_layers_10_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2428_cast_fp16")]; + tensor q_with_bias_v_21_perm_0 = const()[name = tensor("q_with_bias_v_21_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_267_transpose_x_0 = const()[name = tensor("x_267_transpose_x_0"), val = tensor(false)]; + tensor x_267_transpose_y_0 = const()[name = tensor("x_267_transpose_y_0"), val = tensor(false)]; + tensor var_2430_to_fp16 = const()[name = tensor("op_2430_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132839552)))]; + tensor q_with_bias_v_21_cast_fp16 = transpose(perm = q_with_bias_v_21_perm_0, x = var_2428_cast_fp16)[name = tensor("transpose_147")]; + tensor x_267_cast_fp16 = matmul(transpose_x = x_267_transpose_x_0, transpose_y = x_267_transpose_y_0, x = q_with_bias_v_21_cast_fp16, y = var_2430_to_fp16)[name = tensor("x_267_cast_fp16")]; + tensor x_269_pad_0 = const()[name = tensor("x_269_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_269_mode_0 = const()[name = tensor("x_269_mode_0"), val = tensor("constant")]; + tensor const_153_to_fp16 = const()[name = tensor("const_153_to_fp16"), val = tensor(0x0p+0)]; + tensor x_269_cast_fp16 = pad(constant_val = const_153_to_fp16, mode = x_269_mode_0, pad = x_269_pad_0, x = x_267_cast_fp16)[name = tensor("x_269_cast_fp16")]; + tensor var_2438 = const()[name = tensor("op_2438"), val = tensor([1, 8, -1, 3])]; + tensor x_271_cast_fp16 = reshape(shape = var_2438, x = x_269_cast_fp16)[name = tensor("x_271_cast_fp16")]; + tensor var_2442_begin_0 = const()[name = tensor("op_2442_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2442_end_0 = const()[name = tensor("op_2442_end_0"), val = tensor([1, 8, 146, 3])]; + tensor var_2442_end_mask_0 = const()[name = tensor("op_2442_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2442_cast_fp16 = slice_by_index(begin = var_2442_begin_0, end = var_2442_end_0, end_mask = var_2442_end_mask_0, x = x_271_cast_fp16)[name = tensor("op_2442_cast_fp16")]; + tensor var_2443 = const()[name = tensor("op_2443"), val = tensor([1, 8, 3, 145])]; + tensor matrix_bd_41_cast_fp16 = reshape(shape = var_2443, x = var_2442_cast_fp16)[name = tensor("matrix_bd_41_cast_fp16")]; + tensor matrix_ac_21_transpose_x_0 = const()[name = tensor("matrix_ac_21_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_21_transpose_y_0 = const()[name = tensor("matrix_ac_21_transpose_y_0"), val = tensor(false)]; + tensor transpose_71_perm_0 = const()[name = tensor("transpose_71_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_72_perm_0 = const()[name = tensor("transpose_72_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_72 = transpose(perm = transpose_72_perm_0, x = k_41_cast_fp16)[name = tensor("transpose_145")]; + tensor transpose_71 = transpose(perm = transpose_71_perm_0, x = var_2426_cast_fp16)[name = tensor("transpose_146")]; + tensor matrix_ac_21_cast_fp16 = matmul(transpose_x = matrix_ac_21_transpose_x_0, transpose_y = matrix_ac_21_transpose_y_0, x = transpose_71, y = transpose_72)[name = tensor("matrix_ac_21_cast_fp16")]; + tensor matrix_bd_43_begin_0 = const()[name = tensor("matrix_bd_43_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_43_end_0 = const()[name = tensor("matrix_bd_43_end_0"), val = tensor([1, 8, 3, 73])]; + tensor matrix_bd_43_end_mask_0 = const()[name = tensor("matrix_bd_43_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_43_cast_fp16 = slice_by_index(begin = matrix_bd_43_begin_0, end = matrix_bd_43_end_0, end_mask = matrix_bd_43_end_mask_0, x = matrix_bd_41_cast_fp16)[name = tensor("matrix_bd_43_cast_fp16")]; + tensor var_2452_cast_fp16 = add(x = matrix_ac_21_cast_fp16, y = matrix_bd_43_cast_fp16)[name = tensor("op_2452_cast_fp16")]; + tensor _inversed_scores_41_y_0_to_fp16 = const()[name = tensor("_inversed_scores_41_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_41_cast_fp16 = mul(x = var_2452_cast_fp16, y = _inversed_scores_41_y_0_to_fp16)[name = tensor("_inversed_scores_41_cast_fp16")]; + tensor scores_43_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_41_cast_fp16, cond = mask_3)[name = tensor("scores_43_cast_fp16")]; + tensor var_2458_cast_fp16 = softmax(axis = var_62, x = scores_43_cast_fp16)[name = tensor("op_2458_cast_fp16")]; + tensor input_561_cast_fp16 = select(a = var_40_to_fp16, b = var_2458_cast_fp16, cond = mask_3)[name = tensor("input_561_cast_fp16")]; + tensor x_273_transpose_x_0 = const()[name = tensor("x_273_transpose_x_0"), val = tensor(false)]; + tensor x_273_transpose_y_0 = const()[name = tensor("x_273_transpose_y_0"), val = tensor(false)]; + tensor value_23_cast_fp16 = transpose(perm = value_23_perm_0, x = v_21_cast_fp16)[name = tensor("transpose_148")]; + tensor x_273_cast_fp16 = matmul(transpose_x = x_273_transpose_x_0, transpose_y = x_273_transpose_y_0, x = input_561_cast_fp16, y = value_23_cast_fp16)[name = tensor("x_273_cast_fp16")]; + tensor var_2462_perm_0 = const()[name = tensor("op_2462_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2463 = const()[name = tensor("op_2463"), val = tensor([1, -1, 512])]; + tensor var_2462_cast_fp16 = transpose(perm = var_2462_perm_0, x = x_273_cast_fp16)[name = tensor("transpose_144")]; + tensor input_563_cast_fp16 = reshape(shape = var_2463, x = var_2462_cast_fp16)[name = tensor("input_563_cast_fp16")]; + tensor encoder_layers_10_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132988096)))]; + tensor linear_97_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_self_attn_linear_out_weight_to_fp16, x = input_563_cast_fp16)[name = tensor("linear_97_cast_fp16")]; + tensor input_567_cast_fp16 = add(x = input_557_cast_fp16, y = linear_97_cast_fp16)[name = tensor("input_567_cast_fp16")]; + tensor x_277_axes_0 = const()[name = tensor("x_277_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133512448)))]; + tensor encoder_layers_10_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133513536)))]; + tensor x_277_cast_fp16 = layer_norm(axes = x_277_axes_0, beta = encoder_layers_10_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_norm_conv_weight_to_fp16, x = input_567_cast_fp16)[name = tensor("x_277_cast_fp16")]; + tensor input_569_perm_0 = const()[name = tensor("input_569_perm_0"), val = tensor([0, 2, 1])]; + tensor input_571_pad_type_0 = const()[name = tensor("input_571_pad_type_0"), val = tensor("valid")]; + tensor input_571_strides_0 = const()[name = tensor("input_571_strides_0"), val = tensor([1])]; + tensor input_571_pad_0 = const()[name = tensor("input_571_pad_0"), val = tensor([0, 0])]; + tensor input_571_dilations_0 = const()[name = tensor("input_571_dilations_0"), val = tensor([1])]; + tensor input_571_groups_0 = const()[name = tensor("input_571_groups_0"), val = tensor(1)]; + tensor encoder_layers_10_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_10_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133514624)))]; + tensor input_569_cast_fp16 = transpose(perm = input_569_perm_0, x = x_277_cast_fp16)[name = tensor("transpose_143")]; + tensor input_571_cast_fp16 = conv(dilations = input_571_dilations_0, groups = input_571_groups_0, pad = input_571_pad_0, pad_type = input_571_pad_type_0, strides = input_571_strides_0, weight = encoder_layers_10_conv_pointwise_conv1_weight_to_fp16, x = input_569_cast_fp16)[name = tensor("input_571_cast_fp16")]; + tensor x_279_split_num_splits_0 = const()[name = tensor("x_279_split_num_splits_0"), val = tensor(2)]; + tensor x_279_split_axis_0 = const()[name = tensor("x_279_split_axis_0"), val = tensor(1)]; + tensor x_279_split_cast_fp16_0, tensor x_279_split_cast_fp16_1 = split(axis = x_279_split_axis_0, num_splits = x_279_split_num_splits_0, x = input_571_cast_fp16)[name = tensor("x_279_split_cast_fp16")]; + tensor x_279_split_1_sigmoid_cast_fp16 = sigmoid(x = x_279_split_cast_fp16_1)[name = tensor("x_279_split_1_sigmoid_cast_fp16")]; + tensor x_279_cast_fp16 = mul(x = x_279_split_cast_fp16_0, y = x_279_split_1_sigmoid_cast_fp16)[name = tensor("x_279_cast_fp16")]; + tensor input_573_cast_fp16 = select(a = var_40_to_fp16, b = x_279_cast_fp16, cond = var_418)[name = tensor("input_573_cast_fp16")]; + tensor new_x_43_interleave_0 = const()[name = tensor("new_x_43_interleave_0"), val = tensor(false)]; + tensor new_x_43_cast_fp16 = concat(axis = var_62, interleave = new_x_43_interleave_0, values = (cache_43_cast_fp16, input_573_cast_fp16))[name = tensor("new_x_43_cast_fp16")]; + tensor next_cache_21_begin_0 = const()[name = tensor("next_cache_21_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_21_end_0 = const()[name = tensor("next_cache_21_end_0"), val = tensor([1, 512, 9])]; + tensor next_cache_21_end_mask_0 = const()[name = tensor("next_cache_21_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_21_cast_fp16 = slice_by_index(begin = next_cache_21_begin_0, end = next_cache_21_end_0, end_mask = next_cache_21_end_mask_0, x = new_x_43_cast_fp16)[name = tensor("next_cache_21_cast_fp16")]; + tensor var_2504_begin_0 = const()[name = tensor("op_2504_begin_0"), val = tensor([0, 0, 1])]; + tensor var_2504_end_0 = const()[name = tensor("op_2504_end_0"), val = tensor([1, 512, 9])]; + tensor var_2504_end_mask_0 = const()[name = tensor("op_2504_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2504_cast_fp16 = slice_by_index(begin = var_2504_begin_0, end = var_2504_end_0, end_mask = var_2504_end_mask_0, x = next_cache_21_cast_fp16)[name = tensor("op_2504_cast_fp16")]; + tensor x_281_pad_type_0 = const()[name = tensor("x_281_pad_type_0"), val = tensor("valid")]; + tensor x_281_groups_0 = const()[name = tensor("x_281_groups_0"), val = tensor(512)]; + tensor x_281_strides_0 = const()[name = tensor("x_281_strides_0"), val = tensor([1])]; + tensor x_281_pad_0 = const()[name = tensor("x_281_pad_0"), val = tensor([0, 0])]; + tensor x_281_dilations_0 = const()[name = tensor("x_281_dilations_0"), val = tensor([1])]; + tensor encoder_layers_10_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_10_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134563264)))]; + tensor x_281_cast_fp16 = conv(dilations = x_281_dilations_0, groups = x_281_groups_0, pad = x_281_pad_0, pad_type = x_281_pad_type_0, strides = x_281_strides_0, weight = encoder_layers_10_conv_depthwise_conv_weight_to_fp16, x = new_x_43_cast_fp16)[name = tensor("x_281_cast_fp16")]; + tensor input_575_perm_0 = const()[name = tensor("input_575_perm_0"), val = tensor([0, 2, 1])]; + tensor x_283_axes_0 = const()[name = tensor("x_283_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_10_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134572544)))]; + tensor encoder_layers_10_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_10_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134573632)))]; + tensor input_575_cast_fp16 = transpose(perm = input_575_perm_0, x = x_281_cast_fp16)[name = tensor("transpose_142")]; + tensor x_283_cast_fp16 = layer_norm(axes = x_283_axes_0, beta = encoder_layers_10_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_conv_batch_norm_weight_to_fp16, x = input_575_cast_fp16)[name = tensor("x_283_cast_fp16")]; + tensor input_577_perm_0 = const()[name = tensor("input_577_perm_0"), val = tensor([0, 2, 1])]; + tensor input_577_cast_fp16 = transpose(perm = input_577_perm_0, x = x_283_cast_fp16)[name = tensor("transpose_141")]; + tensor input_579_cast_fp16 = silu(x = input_577_cast_fp16)[name = tensor("input_579_cast_fp16")]; + tensor x_285_pad_type_0 = const()[name = tensor("x_285_pad_type_0"), val = tensor("valid")]; + tensor x_285_strides_0 = const()[name = tensor("x_285_strides_0"), val = tensor([1])]; + tensor x_285_pad_0 = const()[name = tensor("x_285_pad_0"), val = tensor([0, 0])]; + tensor x_285_dilations_0 = const()[name = tensor("x_285_dilations_0"), val = tensor([1])]; + tensor x_285_groups_0 = const()[name = tensor("x_285_groups_0"), val = tensor(1)]; + tensor encoder_layers_10_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_10_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134574720)))]; + tensor x_285_cast_fp16 = conv(dilations = x_285_dilations_0, groups = x_285_groups_0, pad = x_285_pad_0, pad_type = x_285_pad_type_0, strides = x_285_strides_0, weight = encoder_layers_10_conv_pointwise_conv2_weight_to_fp16, x = input_579_cast_fp16)[name = tensor("x_285_cast_fp16")]; + tensor input_581_perm_0 = const()[name = tensor("input_581_perm_0"), val = tensor([0, 2, 1])]; + tensor input_581_cast_fp16 = transpose(perm = input_581_perm_0, x = x_285_cast_fp16)[name = tensor("transpose_140")]; + tensor input_583_cast_fp16 = add(x = input_567_cast_fp16, y = input_581_cast_fp16)[name = tensor("input_583_cast_fp16")]; + tensor input_585_axes_0 = const()[name = tensor("input_585_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135099072)))]; + tensor encoder_layers_10_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135100160)))]; + tensor input_585_cast_fp16 = layer_norm(axes = input_585_axes_0, beta = encoder_layers_10_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_norm_feed_forward2_weight_to_fp16, x = input_583_cast_fp16)[name = tensor("input_585_cast_fp16")]; + tensor encoder_layers_10_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_10_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135101248)))]; + tensor linear_98_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_10_feed_forward2_linear1_weight_to_fp16, x = input_585_cast_fp16)[name = tensor("linear_98_cast_fp16")]; + tensor input_589_cast_fp16 = silu(x = linear_98_cast_fp16)[name = tensor("input_589_cast_fp16")]; + tensor encoder_layers_10_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_10_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(137198464)))]; + tensor linear_99_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_feed_forward2_linear2_weight_to_fp16, x = input_589_cast_fp16)[name = tensor("linear_99_cast_fp16")]; + tensor var_2545_to_fp16 = const()[name = tensor("op_2545_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2546_cast_fp16 = mul(x = linear_99_cast_fp16, y = var_2545_to_fp16)[name = tensor("op_2546_cast_fp16")]; + tensor input_595_cast_fp16 = add(x = input_583_cast_fp16, y = var_2546_cast_fp16)[name = tensor("input_595_cast_fp16")]; + tensor input_597_axes_0 = const()[name = tensor("input_597_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139295680)))]; + tensor encoder_layers_10_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139296768)))]; + tensor input_597_cast_fp16 = layer_norm(axes = input_597_axes_0, beta = encoder_layers_10_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_norm_out_weight_to_fp16, x = input_595_cast_fp16)[name = tensor("input_597_cast_fp16")]; + tensor cache_45_begin_0 = const()[name = tensor("cache_45_begin_0"), val = tensor([11, 0, 0, 0])]; + tensor cache_45_end_0 = const()[name = tensor("cache_45_end_0"), val = tensor([12, 1, 70, 512])]; + tensor cache_45_end_mask_0 = const()[name = tensor("cache_45_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_45_squeeze_mask_0 = const()[name = tensor("cache_45_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_45_cast_fp16 = slice_by_index(begin = cache_45_begin_0, end = cache_45_end_0, end_mask = cache_45_end_mask_0, squeeze_mask = cache_45_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_45_cast_fp16")]; + tensor cache_47_begin_0 = const()[name = tensor("cache_47_begin_0"), val = tensor([11, 0, 0, 0])]; + tensor cache_47_end_0 = const()[name = tensor("cache_47_end_0"), val = tensor([12, 1, 512, 8])]; + tensor cache_47_end_mask_0 = const()[name = tensor("cache_47_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_47_squeeze_mask_0 = const()[name = tensor("cache_47_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_47_cast_fp16 = slice_by_index(begin = cache_47_begin_0, end = cache_47_end_0, end_mask = cache_47_end_mask_0, squeeze_mask = cache_47_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_47_cast_fp16")]; + tensor input_599_axes_0 = const()[name = tensor("input_599_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139297856)))]; + tensor encoder_layers_11_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139298944)))]; + tensor input_599_cast_fp16 = layer_norm(axes = input_599_axes_0, beta = encoder_layers_11_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_norm_feed_forward1_weight_to_fp16, x = input_597_cast_fp16)[name = tensor("input_599_cast_fp16")]; + tensor encoder_layers_11_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_11_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139300032)))]; + tensor linear_100_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_11_feed_forward1_linear1_weight_to_fp16, x = input_599_cast_fp16)[name = tensor("linear_100_cast_fp16")]; + tensor input_603_cast_fp16 = silu(x = linear_100_cast_fp16)[name = tensor("input_603_cast_fp16")]; + tensor encoder_layers_11_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_11_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141397248)))]; + tensor linear_101_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_feed_forward1_linear2_weight_to_fp16, x = input_603_cast_fp16)[name = tensor("linear_101_cast_fp16")]; + tensor var_2580_to_fp16 = const()[name = tensor("op_2580_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2581_cast_fp16 = mul(x = linear_101_cast_fp16, y = var_2580_to_fp16)[name = tensor("op_2581_cast_fp16")]; + tensor input_609_cast_fp16 = add(x = input_597_cast_fp16, y = var_2581_cast_fp16)[name = tensor("input_609_cast_fp16")]; + tensor key_23_axes_0 = const()[name = tensor("key_23_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143494464)))]; + tensor encoder_layers_11_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143495552)))]; + tensor key_23_cast_fp16 = layer_norm(axes = key_23_axes_0, beta = encoder_layers_11_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_norm_self_att_weight_to_fp16, x = input_609_cast_fp16)[name = tensor("key_23_cast_fp16")]; + tensor input_611_interleave_0 = const()[name = tensor("input_611_interleave_0"), val = tensor(false)]; + tensor input_611_cast_fp16 = concat(axis = var_64, interleave = input_611_interleave_0, values = (cache_45_cast_fp16, key_23_cast_fp16))[name = tensor("input_611_cast_fp16")]; + tensor var_2603_begin_0 = const()[name = tensor("op_2603_begin_0"), val = tensor([0, 1, 0])]; + tensor var_2603_end_0 = const()[name = tensor("op_2603_end_0"), val = tensor([1, 70, 512])]; + tensor var_2603_end_mask_0 = const()[name = tensor("op_2603_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2603_cast_fp16 = slice_by_index(begin = var_2603_begin_0, end = var_2603_end_0, end_mask = var_2603_end_mask_0, x = cache_45_cast_fp16)[name = tensor("op_2603_cast_fp16")]; + tensor var_2606_begin_0 = const()[name = tensor("op_2606_begin_0"), val = tensor([0, 0, 0])]; + tensor var_2606_end_0 = const()[name = tensor("op_2606_end_0"), val = tensor([1, 1, 512])]; + tensor var_2606_end_mask_0 = const()[name = tensor("op_2606_end_mask_0"), val = tensor([true, false, true])]; + tensor var_2606_cast_fp16 = slice_by_index(begin = var_2606_begin_0, end = var_2606_end_0, end_mask = var_2606_end_mask_0, x = key_23_cast_fp16)[name = tensor("op_2606_cast_fp16")]; + tensor var_2609_interleave_0 = const()[name = tensor("op_2609_interleave_0"), val = tensor(false)]; + tensor var_2609_cast_fp16 = concat(axis = var_64, interleave = var_2609_interleave_0, values = (var_2603_cast_fp16, var_2606_cast_fp16))[name = tensor("op_2609_cast_fp16")]; + tensor encoder_layers_11_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143496640)))]; + tensor linear_102_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_self_attn_linear_q_weight_to_fp16, x = key_23_cast_fp16)[name = tensor("linear_102_cast_fp16")]; + tensor var_2613 = const()[name = tensor("op_2613"), val = tensor([1, -1, 8, 64])]; + tensor q_67_cast_fp16 = reshape(shape = var_2613, x = linear_102_cast_fp16)[name = tensor("q_67_cast_fp16")]; + tensor encoder_layers_11_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144020992)))]; + tensor linear_103_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_self_attn_linear_k_weight_to_fp16, x = input_611_cast_fp16)[name = tensor("linear_103_cast_fp16")]; + tensor var_2617 = const()[name = tensor("op_2617"), val = tensor([1, -1, 8, 64])]; + tensor k_45_cast_fp16 = reshape(shape = var_2617, x = linear_103_cast_fp16)[name = tensor("k_45_cast_fp16")]; + tensor encoder_layers_11_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144545344)))]; + tensor linear_104_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_self_attn_linear_v_weight_to_fp16, x = input_611_cast_fp16)[name = tensor("linear_104_cast_fp16")]; + tensor var_2621 = const()[name = tensor("op_2621"), val = tensor([1, -1, 8, 64])]; + tensor v_23_cast_fp16 = reshape(shape = var_2621, x = linear_104_cast_fp16)[name = tensor("v_23_cast_fp16")]; + tensor value_25_perm_0 = const()[name = tensor("value_25_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_11_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145069696)))]; + tensor var_2633_cast_fp16 = add(x = q_67_cast_fp16, y = encoder_layers_11_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2633_cast_fp16")]; + tensor encoder_layers_11_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145070784)))]; + tensor var_2635_cast_fp16 = add(x = q_67_cast_fp16, y = encoder_layers_11_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2635_cast_fp16")]; + tensor q_with_bias_v_23_perm_0 = const()[name = tensor("q_with_bias_v_23_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_293_transpose_x_0 = const()[name = tensor("x_293_transpose_x_0"), val = tensor(false)]; + tensor x_293_transpose_y_0 = const()[name = tensor("x_293_transpose_y_0"), val = tensor(false)]; + tensor var_2637_to_fp16 = const()[name = tensor("op_2637_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145071872)))]; + tensor q_with_bias_v_23_cast_fp16 = transpose(perm = q_with_bias_v_23_perm_0, x = var_2635_cast_fp16)[name = tensor("transpose_138")]; + tensor x_293_cast_fp16 = matmul(transpose_x = x_293_transpose_x_0, transpose_y = x_293_transpose_y_0, x = q_with_bias_v_23_cast_fp16, y = var_2637_to_fp16)[name = tensor("x_293_cast_fp16")]; + tensor x_295_pad_0 = const()[name = tensor("x_295_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_295_mode_0 = const()[name = tensor("x_295_mode_0"), val = tensor("constant")]; + tensor const_166_to_fp16 = const()[name = tensor("const_166_to_fp16"), val = tensor(0x0p+0)]; + tensor x_295_cast_fp16 = pad(constant_val = const_166_to_fp16, mode = x_295_mode_0, pad = x_295_pad_0, x = x_293_cast_fp16)[name = tensor("x_295_cast_fp16")]; + tensor var_2645 = const()[name = tensor("op_2645"), val = tensor([1, 8, -1, 3])]; + tensor x_297_cast_fp16 = reshape(shape = var_2645, x = x_295_cast_fp16)[name = tensor("x_297_cast_fp16")]; + tensor var_2649_begin_0 = const()[name = tensor("op_2649_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2649_end_0 = const()[name = tensor("op_2649_end_0"), val = tensor([1, 8, 146, 3])]; + tensor var_2649_end_mask_0 = const()[name = tensor("op_2649_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2649_cast_fp16 = slice_by_index(begin = var_2649_begin_0, end = var_2649_end_0, end_mask = var_2649_end_mask_0, x = x_297_cast_fp16)[name = tensor("op_2649_cast_fp16")]; + tensor var_2650 = const()[name = tensor("op_2650"), val = tensor([1, 8, 3, 145])]; + tensor matrix_bd_45_cast_fp16 = reshape(shape = var_2650, x = var_2649_cast_fp16)[name = tensor("matrix_bd_45_cast_fp16")]; + tensor matrix_ac_23_transpose_x_0 = const()[name = tensor("matrix_ac_23_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_23_transpose_y_0 = const()[name = tensor("matrix_ac_23_transpose_y_0"), val = tensor(false)]; + tensor transpose_73_perm_0 = const()[name = tensor("transpose_73_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_74_perm_0 = const()[name = tensor("transpose_74_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_74 = transpose(perm = transpose_74_perm_0, x = k_45_cast_fp16)[name = tensor("transpose_136")]; + tensor transpose_73 = transpose(perm = transpose_73_perm_0, x = var_2633_cast_fp16)[name = tensor("transpose_137")]; + tensor matrix_ac_23_cast_fp16 = matmul(transpose_x = matrix_ac_23_transpose_x_0, transpose_y = matrix_ac_23_transpose_y_0, x = transpose_73, y = transpose_74)[name = tensor("matrix_ac_23_cast_fp16")]; + tensor matrix_bd_47_begin_0 = const()[name = tensor("matrix_bd_47_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_47_end_0 = const()[name = tensor("matrix_bd_47_end_0"), val = tensor([1, 8, 3, 73])]; + tensor matrix_bd_47_end_mask_0 = const()[name = tensor("matrix_bd_47_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_47_cast_fp16 = slice_by_index(begin = matrix_bd_47_begin_0, end = matrix_bd_47_end_0, end_mask = matrix_bd_47_end_mask_0, x = matrix_bd_45_cast_fp16)[name = tensor("matrix_bd_47_cast_fp16")]; + tensor var_2659_cast_fp16 = add(x = matrix_ac_23_cast_fp16, y = matrix_bd_47_cast_fp16)[name = tensor("op_2659_cast_fp16")]; + tensor _inversed_scores_45_y_0_to_fp16 = const()[name = tensor("_inversed_scores_45_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_45_cast_fp16 = mul(x = var_2659_cast_fp16, y = _inversed_scores_45_y_0_to_fp16)[name = tensor("_inversed_scores_45_cast_fp16")]; + tensor scores_47_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_45_cast_fp16, cond = mask_3)[name = tensor("scores_47_cast_fp16")]; + tensor var_2665_cast_fp16 = softmax(axis = var_62, x = scores_47_cast_fp16)[name = tensor("op_2665_cast_fp16")]; + tensor input_613_cast_fp16 = select(a = var_40_to_fp16, b = var_2665_cast_fp16, cond = mask_3)[name = tensor("input_613_cast_fp16")]; + tensor x_299_transpose_x_0 = const()[name = tensor("x_299_transpose_x_0"), val = tensor(false)]; + tensor x_299_transpose_y_0 = const()[name = tensor("x_299_transpose_y_0"), val = tensor(false)]; + tensor value_25_cast_fp16 = transpose(perm = value_25_perm_0, x = v_23_cast_fp16)[name = tensor("transpose_139")]; + tensor x_299_cast_fp16 = matmul(transpose_x = x_299_transpose_x_0, transpose_y = x_299_transpose_y_0, x = input_613_cast_fp16, y = value_25_cast_fp16)[name = tensor("x_299_cast_fp16")]; + tensor var_2669_perm_0 = const()[name = tensor("op_2669_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2670 = const()[name = tensor("op_2670"), val = tensor([1, -1, 512])]; + tensor var_2669_cast_fp16 = transpose(perm = var_2669_perm_0, x = x_299_cast_fp16)[name = tensor("transpose_135")]; + tensor input_615_cast_fp16 = reshape(shape = var_2670, x = var_2669_cast_fp16)[name = tensor("input_615_cast_fp16")]; + tensor encoder_layers_11_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145220416)))]; + tensor linear_106_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_self_attn_linear_out_weight_to_fp16, x = input_615_cast_fp16)[name = tensor("linear_106_cast_fp16")]; + tensor input_619_cast_fp16 = add(x = input_609_cast_fp16, y = linear_106_cast_fp16)[name = tensor("input_619_cast_fp16")]; + tensor x_303_axes_0 = const()[name = tensor("x_303_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145744768)))]; + tensor encoder_layers_11_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145745856)))]; + tensor x_303_cast_fp16 = layer_norm(axes = x_303_axes_0, beta = encoder_layers_11_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_norm_conv_weight_to_fp16, x = input_619_cast_fp16)[name = tensor("x_303_cast_fp16")]; + tensor input_621_perm_0 = const()[name = tensor("input_621_perm_0"), val = tensor([0, 2, 1])]; + tensor input_623_pad_type_0 = const()[name = tensor("input_623_pad_type_0"), val = tensor("valid")]; + tensor input_623_strides_0 = const()[name = tensor("input_623_strides_0"), val = tensor([1])]; + tensor input_623_pad_0 = const()[name = tensor("input_623_pad_0"), val = tensor([0, 0])]; + tensor input_623_dilations_0 = const()[name = tensor("input_623_dilations_0"), val = tensor([1])]; + tensor input_623_groups_0 = const()[name = tensor("input_623_groups_0"), val = tensor(1)]; + tensor encoder_layers_11_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_11_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145746944)))]; + tensor input_621_cast_fp16 = transpose(perm = input_621_perm_0, x = x_303_cast_fp16)[name = tensor("transpose_134")]; + tensor input_623_cast_fp16 = conv(dilations = input_623_dilations_0, groups = input_623_groups_0, pad = input_623_pad_0, pad_type = input_623_pad_type_0, strides = input_623_strides_0, weight = encoder_layers_11_conv_pointwise_conv1_weight_to_fp16, x = input_621_cast_fp16)[name = tensor("input_623_cast_fp16")]; + tensor x_305_split_num_splits_0 = const()[name = tensor("x_305_split_num_splits_0"), val = tensor(2)]; + tensor x_305_split_axis_0 = const()[name = tensor("x_305_split_axis_0"), val = tensor(1)]; + tensor x_305_split_cast_fp16_0, tensor x_305_split_cast_fp16_1 = split(axis = x_305_split_axis_0, num_splits = x_305_split_num_splits_0, x = input_623_cast_fp16)[name = tensor("x_305_split_cast_fp16")]; + tensor x_305_split_1_sigmoid_cast_fp16 = sigmoid(x = x_305_split_cast_fp16_1)[name = tensor("x_305_split_1_sigmoid_cast_fp16")]; + tensor x_305_cast_fp16 = mul(x = x_305_split_cast_fp16_0, y = x_305_split_1_sigmoid_cast_fp16)[name = tensor("x_305_cast_fp16")]; + tensor input_625_cast_fp16 = select(a = var_40_to_fp16, b = x_305_cast_fp16, cond = var_418)[name = tensor("input_625_cast_fp16")]; + tensor new_x_47_interleave_0 = const()[name = tensor("new_x_47_interleave_0"), val = tensor(false)]; + tensor new_x_47_cast_fp16 = concat(axis = var_62, interleave = new_x_47_interleave_0, values = (cache_47_cast_fp16, input_625_cast_fp16))[name = tensor("new_x_47_cast_fp16")]; + tensor next_cache_23_begin_0 = const()[name = tensor("next_cache_23_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_23_end_0 = const()[name = tensor("next_cache_23_end_0"), val = tensor([1, 512, 9])]; + tensor next_cache_23_end_mask_0 = const()[name = tensor("next_cache_23_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_23_cast_fp16 = slice_by_index(begin = next_cache_23_begin_0, end = next_cache_23_end_0, end_mask = next_cache_23_end_mask_0, x = new_x_47_cast_fp16)[name = tensor("next_cache_23_cast_fp16")]; + tensor var_2711_begin_0 = const()[name = tensor("op_2711_begin_0"), val = tensor([0, 0, 1])]; + tensor var_2711_end_0 = const()[name = tensor("op_2711_end_0"), val = tensor([1, 512, 9])]; + tensor var_2711_end_mask_0 = const()[name = tensor("op_2711_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2711_cast_fp16 = slice_by_index(begin = var_2711_begin_0, end = var_2711_end_0, end_mask = var_2711_end_mask_0, x = next_cache_23_cast_fp16)[name = tensor("op_2711_cast_fp16")]; + tensor x_307_pad_type_0 = const()[name = tensor("x_307_pad_type_0"), val = tensor("valid")]; + tensor x_307_groups_0 = const()[name = tensor("x_307_groups_0"), val = tensor(512)]; + tensor x_307_strides_0 = const()[name = tensor("x_307_strides_0"), val = tensor([1])]; + tensor x_307_pad_0 = const()[name = tensor("x_307_pad_0"), val = tensor([0, 0])]; + tensor x_307_dilations_0 = const()[name = tensor("x_307_dilations_0"), val = tensor([1])]; + tensor encoder_layers_11_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_11_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146795584)))]; + tensor x_307_cast_fp16 = conv(dilations = x_307_dilations_0, groups = x_307_groups_0, pad = x_307_pad_0, pad_type = x_307_pad_type_0, strides = x_307_strides_0, weight = encoder_layers_11_conv_depthwise_conv_weight_to_fp16, x = new_x_47_cast_fp16)[name = tensor("x_307_cast_fp16")]; + tensor input_627_perm_0 = const()[name = tensor("input_627_perm_0"), val = tensor([0, 2, 1])]; + tensor x_309_axes_0 = const()[name = tensor("x_309_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_11_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146804864)))]; + tensor encoder_layers_11_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_11_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146805952)))]; + tensor input_627_cast_fp16 = transpose(perm = input_627_perm_0, x = x_307_cast_fp16)[name = tensor("transpose_133")]; + tensor x_309_cast_fp16 = layer_norm(axes = x_309_axes_0, beta = encoder_layers_11_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_conv_batch_norm_weight_to_fp16, x = input_627_cast_fp16)[name = tensor("x_309_cast_fp16")]; + tensor input_629_perm_0 = const()[name = tensor("input_629_perm_0"), val = tensor([0, 2, 1])]; + tensor input_629_cast_fp16 = transpose(perm = input_629_perm_0, x = x_309_cast_fp16)[name = tensor("transpose_132")]; + tensor input_631_cast_fp16 = silu(x = input_629_cast_fp16)[name = tensor("input_631_cast_fp16")]; + tensor x_311_pad_type_0 = const()[name = tensor("x_311_pad_type_0"), val = tensor("valid")]; + tensor x_311_strides_0 = const()[name = tensor("x_311_strides_0"), val = tensor([1])]; + tensor x_311_pad_0 = const()[name = tensor("x_311_pad_0"), val = tensor([0, 0])]; + tensor x_311_dilations_0 = const()[name = tensor("x_311_dilations_0"), val = tensor([1])]; + tensor x_311_groups_0 = const()[name = tensor("x_311_groups_0"), val = tensor(1)]; + tensor encoder_layers_11_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_11_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146807040)))]; + tensor x_311_cast_fp16 = conv(dilations = x_311_dilations_0, groups = x_311_groups_0, pad = x_311_pad_0, pad_type = x_311_pad_type_0, strides = x_311_strides_0, weight = encoder_layers_11_conv_pointwise_conv2_weight_to_fp16, x = input_631_cast_fp16)[name = tensor("x_311_cast_fp16")]; + tensor input_633_perm_0 = const()[name = tensor("input_633_perm_0"), val = tensor([0, 2, 1])]; + tensor input_633_cast_fp16 = transpose(perm = input_633_perm_0, x = x_311_cast_fp16)[name = tensor("transpose_131")]; + tensor input_635_cast_fp16 = add(x = input_619_cast_fp16, y = input_633_cast_fp16)[name = tensor("input_635_cast_fp16")]; + tensor input_637_axes_0 = const()[name = tensor("input_637_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147331392)))]; + tensor encoder_layers_11_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147332480)))]; + tensor input_637_cast_fp16 = layer_norm(axes = input_637_axes_0, beta = encoder_layers_11_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_norm_feed_forward2_weight_to_fp16, x = input_635_cast_fp16)[name = tensor("input_637_cast_fp16")]; + tensor encoder_layers_11_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_11_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147333568)))]; + tensor linear_107_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_11_feed_forward2_linear1_weight_to_fp16, x = input_637_cast_fp16)[name = tensor("linear_107_cast_fp16")]; + tensor input_641_cast_fp16 = silu(x = linear_107_cast_fp16)[name = tensor("input_641_cast_fp16")]; + tensor encoder_layers_11_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_11_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(149430784)))]; + tensor linear_108_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_feed_forward2_linear2_weight_to_fp16, x = input_641_cast_fp16)[name = tensor("linear_108_cast_fp16")]; + tensor var_2752_to_fp16 = const()[name = tensor("op_2752_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2753_cast_fp16 = mul(x = linear_108_cast_fp16, y = var_2752_to_fp16)[name = tensor("op_2753_cast_fp16")]; + tensor input_647_cast_fp16 = add(x = input_635_cast_fp16, y = var_2753_cast_fp16)[name = tensor("input_647_cast_fp16")]; + tensor input_649_axes_0 = const()[name = tensor("input_649_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151528000)))]; + tensor encoder_layers_11_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151529088)))]; + tensor input_649_cast_fp16 = layer_norm(axes = input_649_axes_0, beta = encoder_layers_11_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_norm_out_weight_to_fp16, x = input_647_cast_fp16)[name = tensor("input_649_cast_fp16")]; + tensor cache_49_begin_0 = const()[name = tensor("cache_49_begin_0"), val = tensor([12, 0, 0, 0])]; + tensor cache_49_end_0 = const()[name = tensor("cache_49_end_0"), val = tensor([13, 1, 70, 512])]; + tensor cache_49_end_mask_0 = const()[name = tensor("cache_49_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_49_squeeze_mask_0 = const()[name = tensor("cache_49_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_49_cast_fp16 = slice_by_index(begin = cache_49_begin_0, end = cache_49_end_0, end_mask = cache_49_end_mask_0, squeeze_mask = cache_49_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_49_cast_fp16")]; + tensor cache_51_begin_0 = const()[name = tensor("cache_51_begin_0"), val = tensor([12, 0, 0, 0])]; + tensor cache_51_end_0 = const()[name = tensor("cache_51_end_0"), val = tensor([13, 1, 512, 8])]; + tensor cache_51_end_mask_0 = const()[name = tensor("cache_51_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_51_squeeze_mask_0 = const()[name = tensor("cache_51_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_51_cast_fp16 = slice_by_index(begin = cache_51_begin_0, end = cache_51_end_0, end_mask = cache_51_end_mask_0, squeeze_mask = cache_51_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_51_cast_fp16")]; + tensor input_651_axes_0 = const()[name = tensor("input_651_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151530176)))]; + tensor encoder_layers_12_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151531264)))]; + tensor input_651_cast_fp16 = layer_norm(axes = input_651_axes_0, beta = encoder_layers_12_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_norm_feed_forward1_weight_to_fp16, x = input_649_cast_fp16)[name = tensor("input_651_cast_fp16")]; + tensor encoder_layers_12_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_12_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151532352)))]; + tensor linear_109_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_12_feed_forward1_linear1_weight_to_fp16, x = input_651_cast_fp16)[name = tensor("linear_109_cast_fp16")]; + tensor input_655_cast_fp16 = silu(x = linear_109_cast_fp16)[name = tensor("input_655_cast_fp16")]; + tensor encoder_layers_12_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_12_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(153629568)))]; + tensor linear_110_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_feed_forward1_linear2_weight_to_fp16, x = input_655_cast_fp16)[name = tensor("linear_110_cast_fp16")]; + tensor var_2787_to_fp16 = const()[name = tensor("op_2787_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2788_cast_fp16 = mul(x = linear_110_cast_fp16, y = var_2787_to_fp16)[name = tensor("op_2788_cast_fp16")]; + tensor input_661_cast_fp16 = add(x = input_649_cast_fp16, y = var_2788_cast_fp16)[name = tensor("input_661_cast_fp16")]; + tensor key_25_axes_0 = const()[name = tensor("key_25_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(155726784)))]; + tensor encoder_layers_12_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(155727872)))]; + tensor key_25_cast_fp16 = layer_norm(axes = key_25_axes_0, beta = encoder_layers_12_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_norm_self_att_weight_to_fp16, x = input_661_cast_fp16)[name = tensor("key_25_cast_fp16")]; + tensor input_663_interleave_0 = const()[name = tensor("input_663_interleave_0"), val = tensor(false)]; + tensor input_663_cast_fp16 = concat(axis = var_64, interleave = input_663_interleave_0, values = (cache_49_cast_fp16, key_25_cast_fp16))[name = tensor("input_663_cast_fp16")]; + tensor var_2810_begin_0 = const()[name = tensor("op_2810_begin_0"), val = tensor([0, 1, 0])]; + tensor var_2810_end_0 = const()[name = tensor("op_2810_end_0"), val = tensor([1, 70, 512])]; + tensor var_2810_end_mask_0 = const()[name = tensor("op_2810_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2810_cast_fp16 = slice_by_index(begin = var_2810_begin_0, end = var_2810_end_0, end_mask = var_2810_end_mask_0, x = cache_49_cast_fp16)[name = tensor("op_2810_cast_fp16")]; + tensor var_2813_begin_0 = const()[name = tensor("op_2813_begin_0"), val = tensor([0, 0, 0])]; + tensor var_2813_end_0 = const()[name = tensor("op_2813_end_0"), val = tensor([1, 1, 512])]; + tensor var_2813_end_mask_0 = const()[name = tensor("op_2813_end_mask_0"), val = tensor([true, false, true])]; + tensor var_2813_cast_fp16 = slice_by_index(begin = var_2813_begin_0, end = var_2813_end_0, end_mask = var_2813_end_mask_0, x = key_25_cast_fp16)[name = tensor("op_2813_cast_fp16")]; + tensor var_2816_interleave_0 = const()[name = tensor("op_2816_interleave_0"), val = tensor(false)]; + tensor var_2816_cast_fp16 = concat(axis = var_64, interleave = var_2816_interleave_0, values = (var_2810_cast_fp16, var_2813_cast_fp16))[name = tensor("op_2816_cast_fp16")]; + tensor encoder_layers_12_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(155728960)))]; + tensor linear_111_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_self_attn_linear_q_weight_to_fp16, x = key_25_cast_fp16)[name = tensor("linear_111_cast_fp16")]; + tensor var_2820 = const()[name = tensor("op_2820"), val = tensor([1, -1, 8, 64])]; + tensor q_73_cast_fp16 = reshape(shape = var_2820, x = linear_111_cast_fp16)[name = tensor("q_73_cast_fp16")]; + tensor encoder_layers_12_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156253312)))]; + tensor linear_112_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_self_attn_linear_k_weight_to_fp16, x = input_663_cast_fp16)[name = tensor("linear_112_cast_fp16")]; + tensor var_2824 = const()[name = tensor("op_2824"), val = tensor([1, -1, 8, 64])]; + tensor k_49_cast_fp16 = reshape(shape = var_2824, x = linear_112_cast_fp16)[name = tensor("k_49_cast_fp16")]; + tensor encoder_layers_12_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156777664)))]; + tensor linear_113_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_self_attn_linear_v_weight_to_fp16, x = input_663_cast_fp16)[name = tensor("linear_113_cast_fp16")]; + tensor var_2828 = const()[name = tensor("op_2828"), val = tensor([1, -1, 8, 64])]; + tensor v_25_cast_fp16 = reshape(shape = var_2828, x = linear_113_cast_fp16)[name = tensor("v_25_cast_fp16")]; + tensor value_27_perm_0 = const()[name = tensor("value_27_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_12_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157302016)))]; + tensor var_2840_cast_fp16 = add(x = q_73_cast_fp16, y = encoder_layers_12_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2840_cast_fp16")]; + tensor encoder_layers_12_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157303104)))]; + tensor var_2842_cast_fp16 = add(x = q_73_cast_fp16, y = encoder_layers_12_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2842_cast_fp16")]; + tensor q_with_bias_v_25_perm_0 = const()[name = tensor("q_with_bias_v_25_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_319_transpose_x_0 = const()[name = tensor("x_319_transpose_x_0"), val = tensor(false)]; + tensor x_319_transpose_y_0 = const()[name = tensor("x_319_transpose_y_0"), val = tensor(false)]; + tensor var_2844_to_fp16 = const()[name = tensor("op_2844_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157304192)))]; + tensor q_with_bias_v_25_cast_fp16 = transpose(perm = q_with_bias_v_25_perm_0, x = var_2842_cast_fp16)[name = tensor("transpose_129")]; + tensor x_319_cast_fp16 = matmul(transpose_x = x_319_transpose_x_0, transpose_y = x_319_transpose_y_0, x = q_with_bias_v_25_cast_fp16, y = var_2844_to_fp16)[name = tensor("x_319_cast_fp16")]; + tensor x_321_pad_0 = const()[name = tensor("x_321_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_321_mode_0 = const()[name = tensor("x_321_mode_0"), val = tensor("constant")]; + tensor const_179_to_fp16 = const()[name = tensor("const_179_to_fp16"), val = tensor(0x0p+0)]; + tensor x_321_cast_fp16 = pad(constant_val = const_179_to_fp16, mode = x_321_mode_0, pad = x_321_pad_0, x = x_319_cast_fp16)[name = tensor("x_321_cast_fp16")]; + tensor var_2852 = const()[name = tensor("op_2852"), val = tensor([1, 8, -1, 3])]; + tensor x_323_cast_fp16 = reshape(shape = var_2852, x = x_321_cast_fp16)[name = tensor("x_323_cast_fp16")]; + tensor var_2856_begin_0 = const()[name = tensor("op_2856_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2856_end_0 = const()[name = tensor("op_2856_end_0"), val = tensor([1, 8, 146, 3])]; + tensor var_2856_end_mask_0 = const()[name = tensor("op_2856_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2856_cast_fp16 = slice_by_index(begin = var_2856_begin_0, end = var_2856_end_0, end_mask = var_2856_end_mask_0, x = x_323_cast_fp16)[name = tensor("op_2856_cast_fp16")]; + tensor var_2857 = const()[name = tensor("op_2857"), val = tensor([1, 8, 3, 145])]; + tensor matrix_bd_49_cast_fp16 = reshape(shape = var_2857, x = var_2856_cast_fp16)[name = tensor("matrix_bd_49_cast_fp16")]; + tensor matrix_ac_25_transpose_x_0 = const()[name = tensor("matrix_ac_25_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_25_transpose_y_0 = const()[name = tensor("matrix_ac_25_transpose_y_0"), val = tensor(false)]; + tensor transpose_75_perm_0 = const()[name = tensor("transpose_75_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_76_perm_0 = const()[name = tensor("transpose_76_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_76 = transpose(perm = transpose_76_perm_0, x = k_49_cast_fp16)[name = tensor("transpose_127")]; + tensor transpose_75 = transpose(perm = transpose_75_perm_0, x = var_2840_cast_fp16)[name = tensor("transpose_128")]; + tensor matrix_ac_25_cast_fp16 = matmul(transpose_x = matrix_ac_25_transpose_x_0, transpose_y = matrix_ac_25_transpose_y_0, x = transpose_75, y = transpose_76)[name = tensor("matrix_ac_25_cast_fp16")]; + tensor matrix_bd_51_begin_0 = const()[name = tensor("matrix_bd_51_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_51_end_0 = const()[name = tensor("matrix_bd_51_end_0"), val = tensor([1, 8, 3, 73])]; + tensor matrix_bd_51_end_mask_0 = const()[name = tensor("matrix_bd_51_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_51_cast_fp16 = slice_by_index(begin = matrix_bd_51_begin_0, end = matrix_bd_51_end_0, end_mask = matrix_bd_51_end_mask_0, x = matrix_bd_49_cast_fp16)[name = tensor("matrix_bd_51_cast_fp16")]; + tensor var_2866_cast_fp16 = add(x = matrix_ac_25_cast_fp16, y = matrix_bd_51_cast_fp16)[name = tensor("op_2866_cast_fp16")]; + tensor _inversed_scores_49_y_0_to_fp16 = const()[name = tensor("_inversed_scores_49_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_49_cast_fp16 = mul(x = var_2866_cast_fp16, y = _inversed_scores_49_y_0_to_fp16)[name = tensor("_inversed_scores_49_cast_fp16")]; + tensor scores_51_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_49_cast_fp16, cond = mask_3)[name = tensor("scores_51_cast_fp16")]; + tensor var_2872_cast_fp16 = softmax(axis = var_62, x = scores_51_cast_fp16)[name = tensor("op_2872_cast_fp16")]; + tensor input_665_cast_fp16 = select(a = var_40_to_fp16, b = var_2872_cast_fp16, cond = mask_3)[name = tensor("input_665_cast_fp16")]; + tensor x_325_transpose_x_0 = const()[name = tensor("x_325_transpose_x_0"), val = tensor(false)]; + tensor x_325_transpose_y_0 = const()[name = tensor("x_325_transpose_y_0"), val = tensor(false)]; + tensor value_27_cast_fp16 = transpose(perm = value_27_perm_0, x = v_25_cast_fp16)[name = tensor("transpose_130")]; + tensor x_325_cast_fp16 = matmul(transpose_x = x_325_transpose_x_0, transpose_y = x_325_transpose_y_0, x = input_665_cast_fp16, y = value_27_cast_fp16)[name = tensor("x_325_cast_fp16")]; + tensor var_2876_perm_0 = const()[name = tensor("op_2876_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2877 = const()[name = tensor("op_2877"), val = tensor([1, -1, 512])]; + tensor var_2876_cast_fp16 = transpose(perm = var_2876_perm_0, x = x_325_cast_fp16)[name = tensor("transpose_126")]; + tensor input_667_cast_fp16 = reshape(shape = var_2877, x = var_2876_cast_fp16)[name = tensor("input_667_cast_fp16")]; + tensor encoder_layers_12_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157452736)))]; + tensor linear_115_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_self_attn_linear_out_weight_to_fp16, x = input_667_cast_fp16)[name = tensor("linear_115_cast_fp16")]; + tensor input_671_cast_fp16 = add(x = input_661_cast_fp16, y = linear_115_cast_fp16)[name = tensor("input_671_cast_fp16")]; + tensor x_329_axes_0 = const()[name = tensor("x_329_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157977088)))]; + tensor encoder_layers_12_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157978176)))]; + tensor x_329_cast_fp16 = layer_norm(axes = x_329_axes_0, beta = encoder_layers_12_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_norm_conv_weight_to_fp16, x = input_671_cast_fp16)[name = tensor("x_329_cast_fp16")]; + tensor input_673_perm_0 = const()[name = tensor("input_673_perm_0"), val = tensor([0, 2, 1])]; + tensor input_675_pad_type_0 = const()[name = tensor("input_675_pad_type_0"), val = tensor("valid")]; + tensor input_675_strides_0 = const()[name = tensor("input_675_strides_0"), val = tensor([1])]; + tensor input_675_pad_0 = const()[name = tensor("input_675_pad_0"), val = tensor([0, 0])]; + tensor input_675_dilations_0 = const()[name = tensor("input_675_dilations_0"), val = tensor([1])]; + tensor input_675_groups_0 = const()[name = tensor("input_675_groups_0"), val = tensor(1)]; + tensor encoder_layers_12_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_12_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157979264)))]; + tensor input_673_cast_fp16 = transpose(perm = input_673_perm_0, x = x_329_cast_fp16)[name = tensor("transpose_125")]; + tensor input_675_cast_fp16 = conv(dilations = input_675_dilations_0, groups = input_675_groups_0, pad = input_675_pad_0, pad_type = input_675_pad_type_0, strides = input_675_strides_0, weight = encoder_layers_12_conv_pointwise_conv1_weight_to_fp16, x = input_673_cast_fp16)[name = tensor("input_675_cast_fp16")]; + tensor x_331_split_num_splits_0 = const()[name = tensor("x_331_split_num_splits_0"), val = tensor(2)]; + tensor x_331_split_axis_0 = const()[name = tensor("x_331_split_axis_0"), val = tensor(1)]; + tensor x_331_split_cast_fp16_0, tensor x_331_split_cast_fp16_1 = split(axis = x_331_split_axis_0, num_splits = x_331_split_num_splits_0, x = input_675_cast_fp16)[name = tensor("x_331_split_cast_fp16")]; + tensor x_331_split_1_sigmoid_cast_fp16 = sigmoid(x = x_331_split_cast_fp16_1)[name = tensor("x_331_split_1_sigmoid_cast_fp16")]; + tensor x_331_cast_fp16 = mul(x = x_331_split_cast_fp16_0, y = x_331_split_1_sigmoid_cast_fp16)[name = tensor("x_331_cast_fp16")]; + tensor input_677_cast_fp16 = select(a = var_40_to_fp16, b = x_331_cast_fp16, cond = var_418)[name = tensor("input_677_cast_fp16")]; + tensor new_x_51_interleave_0 = const()[name = tensor("new_x_51_interleave_0"), val = tensor(false)]; + tensor new_x_51_cast_fp16 = concat(axis = var_62, interleave = new_x_51_interleave_0, values = (cache_51_cast_fp16, input_677_cast_fp16))[name = tensor("new_x_51_cast_fp16")]; + tensor next_cache_25_begin_0 = const()[name = tensor("next_cache_25_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_25_end_0 = const()[name = tensor("next_cache_25_end_0"), val = tensor([1, 512, 9])]; + tensor next_cache_25_end_mask_0 = const()[name = tensor("next_cache_25_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_25_cast_fp16 = slice_by_index(begin = next_cache_25_begin_0, end = next_cache_25_end_0, end_mask = next_cache_25_end_mask_0, x = new_x_51_cast_fp16)[name = tensor("next_cache_25_cast_fp16")]; + tensor var_2918_begin_0 = const()[name = tensor("op_2918_begin_0"), val = tensor([0, 0, 1])]; + tensor var_2918_end_0 = const()[name = tensor("op_2918_end_0"), val = tensor([1, 512, 9])]; + tensor var_2918_end_mask_0 = const()[name = tensor("op_2918_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2918_cast_fp16 = slice_by_index(begin = var_2918_begin_0, end = var_2918_end_0, end_mask = var_2918_end_mask_0, x = next_cache_25_cast_fp16)[name = tensor("op_2918_cast_fp16")]; + tensor x_333_pad_type_0 = const()[name = tensor("x_333_pad_type_0"), val = tensor("valid")]; + tensor x_333_groups_0 = const()[name = tensor("x_333_groups_0"), val = tensor(512)]; + tensor x_333_strides_0 = const()[name = tensor("x_333_strides_0"), val = tensor([1])]; + tensor x_333_pad_0 = const()[name = tensor("x_333_pad_0"), val = tensor([0, 0])]; + tensor x_333_dilations_0 = const()[name = tensor("x_333_dilations_0"), val = tensor([1])]; + tensor encoder_layers_12_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_12_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159027904)))]; + tensor x_333_cast_fp16 = conv(dilations = x_333_dilations_0, groups = x_333_groups_0, pad = x_333_pad_0, pad_type = x_333_pad_type_0, strides = x_333_strides_0, weight = encoder_layers_12_conv_depthwise_conv_weight_to_fp16, x = new_x_51_cast_fp16)[name = tensor("x_333_cast_fp16")]; + tensor input_679_perm_0 = const()[name = tensor("input_679_perm_0"), val = tensor([0, 2, 1])]; + tensor x_335_axes_0 = const()[name = tensor("x_335_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_12_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159037184)))]; + tensor encoder_layers_12_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_12_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159038272)))]; + tensor input_679_cast_fp16 = transpose(perm = input_679_perm_0, x = x_333_cast_fp16)[name = tensor("transpose_124")]; + tensor x_335_cast_fp16 = layer_norm(axes = x_335_axes_0, beta = encoder_layers_12_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_conv_batch_norm_weight_to_fp16, x = input_679_cast_fp16)[name = tensor("x_335_cast_fp16")]; + tensor input_681_perm_0 = const()[name = tensor("input_681_perm_0"), val = tensor([0, 2, 1])]; + tensor input_681_cast_fp16 = transpose(perm = input_681_perm_0, x = x_335_cast_fp16)[name = tensor("transpose_123")]; + tensor input_683_cast_fp16 = silu(x = input_681_cast_fp16)[name = tensor("input_683_cast_fp16")]; + tensor x_337_pad_type_0 = const()[name = tensor("x_337_pad_type_0"), val = tensor("valid")]; + tensor x_337_strides_0 = const()[name = tensor("x_337_strides_0"), val = tensor([1])]; + tensor x_337_pad_0 = const()[name = tensor("x_337_pad_0"), val = tensor([0, 0])]; + tensor x_337_dilations_0 = const()[name = tensor("x_337_dilations_0"), val = tensor([1])]; + tensor x_337_groups_0 = const()[name = tensor("x_337_groups_0"), val = tensor(1)]; + tensor encoder_layers_12_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_12_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159039360)))]; + tensor x_337_cast_fp16 = conv(dilations = x_337_dilations_0, groups = x_337_groups_0, pad = x_337_pad_0, pad_type = x_337_pad_type_0, strides = x_337_strides_0, weight = encoder_layers_12_conv_pointwise_conv2_weight_to_fp16, x = input_683_cast_fp16)[name = tensor("x_337_cast_fp16")]; + tensor input_685_perm_0 = const()[name = tensor("input_685_perm_0"), val = tensor([0, 2, 1])]; + tensor input_685_cast_fp16 = transpose(perm = input_685_perm_0, x = x_337_cast_fp16)[name = tensor("transpose_122")]; + tensor input_687_cast_fp16 = add(x = input_671_cast_fp16, y = input_685_cast_fp16)[name = tensor("input_687_cast_fp16")]; + tensor input_689_axes_0 = const()[name = tensor("input_689_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159563712)))]; + tensor encoder_layers_12_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159564800)))]; + tensor input_689_cast_fp16 = layer_norm(axes = input_689_axes_0, beta = encoder_layers_12_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_norm_feed_forward2_weight_to_fp16, x = input_687_cast_fp16)[name = tensor("input_689_cast_fp16")]; + tensor encoder_layers_12_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_12_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159565888)))]; + tensor linear_116_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_12_feed_forward2_linear1_weight_to_fp16, x = input_689_cast_fp16)[name = tensor("linear_116_cast_fp16")]; + tensor input_693_cast_fp16 = silu(x = linear_116_cast_fp16)[name = tensor("input_693_cast_fp16")]; + tensor encoder_layers_12_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_12_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161663104)))]; + tensor linear_117_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_feed_forward2_linear2_weight_to_fp16, x = input_693_cast_fp16)[name = tensor("linear_117_cast_fp16")]; + tensor var_2959_to_fp16 = const()[name = tensor("op_2959_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2960_cast_fp16 = mul(x = linear_117_cast_fp16, y = var_2959_to_fp16)[name = tensor("op_2960_cast_fp16")]; + tensor input_699_cast_fp16 = add(x = input_687_cast_fp16, y = var_2960_cast_fp16)[name = tensor("input_699_cast_fp16")]; + tensor input_701_axes_0 = const()[name = tensor("input_701_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163760320)))]; + tensor encoder_layers_12_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163761408)))]; + tensor input_701_cast_fp16 = layer_norm(axes = input_701_axes_0, beta = encoder_layers_12_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_norm_out_weight_to_fp16, x = input_699_cast_fp16)[name = tensor("input_701_cast_fp16")]; + tensor cache_53_begin_0 = const()[name = tensor("cache_53_begin_0"), val = tensor([13, 0, 0, 0])]; + tensor cache_53_end_0 = const()[name = tensor("cache_53_end_0"), val = tensor([14, 1, 70, 512])]; + tensor cache_53_end_mask_0 = const()[name = tensor("cache_53_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_53_squeeze_mask_0 = const()[name = tensor("cache_53_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_53_cast_fp16 = slice_by_index(begin = cache_53_begin_0, end = cache_53_end_0, end_mask = cache_53_end_mask_0, squeeze_mask = cache_53_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_53_cast_fp16")]; + tensor cache_55_begin_0 = const()[name = tensor("cache_55_begin_0"), val = tensor([13, 0, 0, 0])]; + tensor cache_55_end_0 = const()[name = tensor("cache_55_end_0"), val = tensor([14, 1, 512, 8])]; + tensor cache_55_end_mask_0 = const()[name = tensor("cache_55_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_55_squeeze_mask_0 = const()[name = tensor("cache_55_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_55_cast_fp16 = slice_by_index(begin = cache_55_begin_0, end = cache_55_end_0, end_mask = cache_55_end_mask_0, squeeze_mask = cache_55_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_55_cast_fp16")]; + tensor input_703_axes_0 = const()[name = tensor("input_703_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163762496)))]; + tensor encoder_layers_13_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163763584)))]; + tensor input_703_cast_fp16 = layer_norm(axes = input_703_axes_0, beta = encoder_layers_13_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_norm_feed_forward1_weight_to_fp16, x = input_701_cast_fp16)[name = tensor("input_703_cast_fp16")]; + tensor encoder_layers_13_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_13_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163764672)))]; + tensor linear_118_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_13_feed_forward1_linear1_weight_to_fp16, x = input_703_cast_fp16)[name = tensor("linear_118_cast_fp16")]; + tensor input_707_cast_fp16 = silu(x = linear_118_cast_fp16)[name = tensor("input_707_cast_fp16")]; + tensor encoder_layers_13_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_13_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165861888)))]; + tensor linear_119_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_feed_forward1_linear2_weight_to_fp16, x = input_707_cast_fp16)[name = tensor("linear_119_cast_fp16")]; + tensor var_2994_to_fp16 = const()[name = tensor("op_2994_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2995_cast_fp16 = mul(x = linear_119_cast_fp16, y = var_2994_to_fp16)[name = tensor("op_2995_cast_fp16")]; + tensor input_713_cast_fp16 = add(x = input_701_cast_fp16, y = var_2995_cast_fp16)[name = tensor("input_713_cast_fp16")]; + tensor key_27_axes_0 = const()[name = tensor("key_27_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(167959104)))]; + tensor encoder_layers_13_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(167960192)))]; + tensor key_27_cast_fp16 = layer_norm(axes = key_27_axes_0, beta = encoder_layers_13_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_norm_self_att_weight_to_fp16, x = input_713_cast_fp16)[name = tensor("key_27_cast_fp16")]; + tensor input_715_interleave_0 = const()[name = tensor("input_715_interleave_0"), val = tensor(false)]; + tensor input_715_cast_fp16 = concat(axis = var_64, interleave = input_715_interleave_0, values = (cache_53_cast_fp16, key_27_cast_fp16))[name = tensor("input_715_cast_fp16")]; + tensor var_3017_begin_0 = const()[name = tensor("op_3017_begin_0"), val = tensor([0, 1, 0])]; + tensor var_3017_end_0 = const()[name = tensor("op_3017_end_0"), val = tensor([1, 70, 512])]; + tensor var_3017_end_mask_0 = const()[name = tensor("op_3017_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3017_cast_fp16 = slice_by_index(begin = var_3017_begin_0, end = var_3017_end_0, end_mask = var_3017_end_mask_0, x = cache_53_cast_fp16)[name = tensor("op_3017_cast_fp16")]; + tensor var_3020_begin_0 = const()[name = tensor("op_3020_begin_0"), val = tensor([0, 0, 0])]; + tensor var_3020_end_0 = const()[name = tensor("op_3020_end_0"), val = tensor([1, 1, 512])]; + tensor var_3020_end_mask_0 = const()[name = tensor("op_3020_end_mask_0"), val = tensor([true, false, true])]; + tensor var_3020_cast_fp16 = slice_by_index(begin = var_3020_begin_0, end = var_3020_end_0, end_mask = var_3020_end_mask_0, x = key_27_cast_fp16)[name = tensor("op_3020_cast_fp16")]; + tensor var_3023_interleave_0 = const()[name = tensor("op_3023_interleave_0"), val = tensor(false)]; + tensor var_3023_cast_fp16 = concat(axis = var_64, interleave = var_3023_interleave_0, values = (var_3017_cast_fp16, var_3020_cast_fp16))[name = tensor("op_3023_cast_fp16")]; + tensor encoder_layers_13_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(167961280)))]; + tensor linear_120_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_self_attn_linear_q_weight_to_fp16, x = key_27_cast_fp16)[name = tensor("linear_120_cast_fp16")]; + tensor var_3027 = const()[name = tensor("op_3027"), val = tensor([1, -1, 8, 64])]; + tensor q_79_cast_fp16 = reshape(shape = var_3027, x = linear_120_cast_fp16)[name = tensor("q_79_cast_fp16")]; + tensor encoder_layers_13_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168485632)))]; + tensor linear_121_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_self_attn_linear_k_weight_to_fp16, x = input_715_cast_fp16)[name = tensor("linear_121_cast_fp16")]; + tensor var_3031 = const()[name = tensor("op_3031"), val = tensor([1, -1, 8, 64])]; + tensor k_53_cast_fp16 = reshape(shape = var_3031, x = linear_121_cast_fp16)[name = tensor("k_53_cast_fp16")]; + tensor encoder_layers_13_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169009984)))]; + tensor linear_122_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_self_attn_linear_v_weight_to_fp16, x = input_715_cast_fp16)[name = tensor("linear_122_cast_fp16")]; + tensor var_3035 = const()[name = tensor("op_3035"), val = tensor([1, -1, 8, 64])]; + tensor v_27_cast_fp16 = reshape(shape = var_3035, x = linear_122_cast_fp16)[name = tensor("v_27_cast_fp16")]; + tensor value_29_perm_0 = const()[name = tensor("value_29_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_13_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169534336)))]; + tensor var_3047_cast_fp16 = add(x = q_79_cast_fp16, y = encoder_layers_13_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3047_cast_fp16")]; + tensor encoder_layers_13_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169535424)))]; + tensor var_3049_cast_fp16 = add(x = q_79_cast_fp16, y = encoder_layers_13_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3049_cast_fp16")]; + tensor q_with_bias_v_27_perm_0 = const()[name = tensor("q_with_bias_v_27_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_345_transpose_x_0 = const()[name = tensor("x_345_transpose_x_0"), val = tensor(false)]; + tensor x_345_transpose_y_0 = const()[name = tensor("x_345_transpose_y_0"), val = tensor(false)]; + tensor var_3051_to_fp16 = const()[name = tensor("op_3051_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169536512)))]; + tensor q_with_bias_v_27_cast_fp16 = transpose(perm = q_with_bias_v_27_perm_0, x = var_3049_cast_fp16)[name = tensor("transpose_120")]; + tensor x_345_cast_fp16 = matmul(transpose_x = x_345_transpose_x_0, transpose_y = x_345_transpose_y_0, x = q_with_bias_v_27_cast_fp16, y = var_3051_to_fp16)[name = tensor("x_345_cast_fp16")]; + tensor x_347_pad_0 = const()[name = tensor("x_347_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_347_mode_0 = const()[name = tensor("x_347_mode_0"), val = tensor("constant")]; + tensor const_192_to_fp16 = const()[name = tensor("const_192_to_fp16"), val = tensor(0x0p+0)]; + tensor x_347_cast_fp16 = pad(constant_val = const_192_to_fp16, mode = x_347_mode_0, pad = x_347_pad_0, x = x_345_cast_fp16)[name = tensor("x_347_cast_fp16")]; + tensor var_3059 = const()[name = tensor("op_3059"), val = tensor([1, 8, -1, 3])]; + tensor x_349_cast_fp16 = reshape(shape = var_3059, x = x_347_cast_fp16)[name = tensor("x_349_cast_fp16")]; + tensor var_3063_begin_0 = const()[name = tensor("op_3063_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_3063_end_0 = const()[name = tensor("op_3063_end_0"), val = tensor([1, 8, 146, 3])]; + tensor var_3063_end_mask_0 = const()[name = tensor("op_3063_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_3063_cast_fp16 = slice_by_index(begin = var_3063_begin_0, end = var_3063_end_0, end_mask = var_3063_end_mask_0, x = x_349_cast_fp16)[name = tensor("op_3063_cast_fp16")]; + tensor var_3064 = const()[name = tensor("op_3064"), val = tensor([1, 8, 3, 145])]; + tensor matrix_bd_53_cast_fp16 = reshape(shape = var_3064, x = var_3063_cast_fp16)[name = tensor("matrix_bd_53_cast_fp16")]; + tensor matrix_ac_27_transpose_x_0 = const()[name = tensor("matrix_ac_27_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_27_transpose_y_0 = const()[name = tensor("matrix_ac_27_transpose_y_0"), val = tensor(false)]; + tensor transpose_77_perm_0 = const()[name = tensor("transpose_77_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_78_perm_0 = const()[name = tensor("transpose_78_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_78 = transpose(perm = transpose_78_perm_0, x = k_53_cast_fp16)[name = tensor("transpose_118")]; + tensor transpose_77 = transpose(perm = transpose_77_perm_0, x = var_3047_cast_fp16)[name = tensor("transpose_119")]; + tensor matrix_ac_27_cast_fp16 = matmul(transpose_x = matrix_ac_27_transpose_x_0, transpose_y = matrix_ac_27_transpose_y_0, x = transpose_77, y = transpose_78)[name = tensor("matrix_ac_27_cast_fp16")]; + tensor matrix_bd_55_begin_0 = const()[name = tensor("matrix_bd_55_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_55_end_0 = const()[name = tensor("matrix_bd_55_end_0"), val = tensor([1, 8, 3, 73])]; + tensor matrix_bd_55_end_mask_0 = const()[name = tensor("matrix_bd_55_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_55_cast_fp16 = slice_by_index(begin = matrix_bd_55_begin_0, end = matrix_bd_55_end_0, end_mask = matrix_bd_55_end_mask_0, x = matrix_bd_53_cast_fp16)[name = tensor("matrix_bd_55_cast_fp16")]; + tensor var_3073_cast_fp16 = add(x = matrix_ac_27_cast_fp16, y = matrix_bd_55_cast_fp16)[name = tensor("op_3073_cast_fp16")]; + tensor _inversed_scores_53_y_0_to_fp16 = const()[name = tensor("_inversed_scores_53_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_53_cast_fp16 = mul(x = var_3073_cast_fp16, y = _inversed_scores_53_y_0_to_fp16)[name = tensor("_inversed_scores_53_cast_fp16")]; + tensor scores_55_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_53_cast_fp16, cond = mask_3)[name = tensor("scores_55_cast_fp16")]; + tensor var_3079_cast_fp16 = softmax(axis = var_62, x = scores_55_cast_fp16)[name = tensor("op_3079_cast_fp16")]; + tensor input_717_cast_fp16 = select(a = var_40_to_fp16, b = var_3079_cast_fp16, cond = mask_3)[name = tensor("input_717_cast_fp16")]; + tensor x_351_transpose_x_0 = const()[name = tensor("x_351_transpose_x_0"), val = tensor(false)]; + tensor x_351_transpose_y_0 = const()[name = tensor("x_351_transpose_y_0"), val = tensor(false)]; + tensor value_29_cast_fp16 = transpose(perm = value_29_perm_0, x = v_27_cast_fp16)[name = tensor("transpose_121")]; + tensor x_351_cast_fp16 = matmul(transpose_x = x_351_transpose_x_0, transpose_y = x_351_transpose_y_0, x = input_717_cast_fp16, y = value_29_cast_fp16)[name = tensor("x_351_cast_fp16")]; + tensor var_3083_perm_0 = const()[name = tensor("op_3083_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_3084 = const()[name = tensor("op_3084"), val = tensor([1, -1, 512])]; + tensor var_3083_cast_fp16 = transpose(perm = var_3083_perm_0, x = x_351_cast_fp16)[name = tensor("transpose_117")]; + tensor input_719_cast_fp16 = reshape(shape = var_3084, x = var_3083_cast_fp16)[name = tensor("input_719_cast_fp16")]; + tensor encoder_layers_13_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169685056)))]; + tensor linear_124_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_self_attn_linear_out_weight_to_fp16, x = input_719_cast_fp16)[name = tensor("linear_124_cast_fp16")]; + tensor input_723_cast_fp16 = add(x = input_713_cast_fp16, y = linear_124_cast_fp16)[name = tensor("input_723_cast_fp16")]; + tensor x_355_axes_0 = const()[name = tensor("x_355_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170209408)))]; + tensor encoder_layers_13_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170210496)))]; + tensor x_355_cast_fp16 = layer_norm(axes = x_355_axes_0, beta = encoder_layers_13_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_norm_conv_weight_to_fp16, x = input_723_cast_fp16)[name = tensor("x_355_cast_fp16")]; + tensor input_725_perm_0 = const()[name = tensor("input_725_perm_0"), val = tensor([0, 2, 1])]; + tensor input_727_pad_type_0 = const()[name = tensor("input_727_pad_type_0"), val = tensor("valid")]; + tensor input_727_strides_0 = const()[name = tensor("input_727_strides_0"), val = tensor([1])]; + tensor input_727_pad_0 = const()[name = tensor("input_727_pad_0"), val = tensor([0, 0])]; + tensor input_727_dilations_0 = const()[name = tensor("input_727_dilations_0"), val = tensor([1])]; + tensor input_727_groups_0 = const()[name = tensor("input_727_groups_0"), val = tensor(1)]; + tensor encoder_layers_13_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_13_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170211584)))]; + tensor input_725_cast_fp16 = transpose(perm = input_725_perm_0, x = x_355_cast_fp16)[name = tensor("transpose_116")]; + tensor input_727_cast_fp16 = conv(dilations = input_727_dilations_0, groups = input_727_groups_0, pad = input_727_pad_0, pad_type = input_727_pad_type_0, strides = input_727_strides_0, weight = encoder_layers_13_conv_pointwise_conv1_weight_to_fp16, x = input_725_cast_fp16)[name = tensor("input_727_cast_fp16")]; + tensor x_357_split_num_splits_0 = const()[name = tensor("x_357_split_num_splits_0"), val = tensor(2)]; + tensor x_357_split_axis_0 = const()[name = tensor("x_357_split_axis_0"), val = tensor(1)]; + tensor x_357_split_cast_fp16_0, tensor x_357_split_cast_fp16_1 = split(axis = x_357_split_axis_0, num_splits = x_357_split_num_splits_0, x = input_727_cast_fp16)[name = tensor("x_357_split_cast_fp16")]; + tensor x_357_split_1_sigmoid_cast_fp16 = sigmoid(x = x_357_split_cast_fp16_1)[name = tensor("x_357_split_1_sigmoid_cast_fp16")]; + tensor x_357_cast_fp16 = mul(x = x_357_split_cast_fp16_0, y = x_357_split_1_sigmoid_cast_fp16)[name = tensor("x_357_cast_fp16")]; + tensor input_729_cast_fp16 = select(a = var_40_to_fp16, b = x_357_cast_fp16, cond = var_418)[name = tensor("input_729_cast_fp16")]; + tensor new_x_55_interleave_0 = const()[name = tensor("new_x_55_interleave_0"), val = tensor(false)]; + tensor new_x_55_cast_fp16 = concat(axis = var_62, interleave = new_x_55_interleave_0, values = (cache_55_cast_fp16, input_729_cast_fp16))[name = tensor("new_x_55_cast_fp16")]; + tensor next_cache_27_begin_0 = const()[name = tensor("next_cache_27_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_27_end_0 = const()[name = tensor("next_cache_27_end_0"), val = tensor([1, 512, 9])]; + tensor next_cache_27_end_mask_0 = const()[name = tensor("next_cache_27_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_27_cast_fp16 = slice_by_index(begin = next_cache_27_begin_0, end = next_cache_27_end_0, end_mask = next_cache_27_end_mask_0, x = new_x_55_cast_fp16)[name = tensor("next_cache_27_cast_fp16")]; + tensor var_3125_begin_0 = const()[name = tensor("op_3125_begin_0"), val = tensor([0, 0, 1])]; + tensor var_3125_end_0 = const()[name = tensor("op_3125_end_0"), val = tensor([1, 512, 9])]; + tensor var_3125_end_mask_0 = const()[name = tensor("op_3125_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3125_cast_fp16 = slice_by_index(begin = var_3125_begin_0, end = var_3125_end_0, end_mask = var_3125_end_mask_0, x = next_cache_27_cast_fp16)[name = tensor("op_3125_cast_fp16")]; + tensor x_359_pad_type_0 = const()[name = tensor("x_359_pad_type_0"), val = tensor("valid")]; + tensor x_359_groups_0 = const()[name = tensor("x_359_groups_0"), val = tensor(512)]; + tensor x_359_strides_0 = const()[name = tensor("x_359_strides_0"), val = tensor([1])]; + tensor x_359_pad_0 = const()[name = tensor("x_359_pad_0"), val = tensor([0, 0])]; + tensor x_359_dilations_0 = const()[name = tensor("x_359_dilations_0"), val = tensor([1])]; + tensor encoder_layers_13_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_13_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171260224)))]; + tensor x_359_cast_fp16 = conv(dilations = x_359_dilations_0, groups = x_359_groups_0, pad = x_359_pad_0, pad_type = x_359_pad_type_0, strides = x_359_strides_0, weight = encoder_layers_13_conv_depthwise_conv_weight_to_fp16, x = new_x_55_cast_fp16)[name = tensor("x_359_cast_fp16")]; + tensor input_731_perm_0 = const()[name = tensor("input_731_perm_0"), val = tensor([0, 2, 1])]; + tensor x_361_axes_0 = const()[name = tensor("x_361_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_13_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171269504)))]; + tensor encoder_layers_13_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_13_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171270592)))]; + tensor input_731_cast_fp16 = transpose(perm = input_731_perm_0, x = x_359_cast_fp16)[name = tensor("transpose_115")]; + tensor x_361_cast_fp16 = layer_norm(axes = x_361_axes_0, beta = encoder_layers_13_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_conv_batch_norm_weight_to_fp16, x = input_731_cast_fp16)[name = tensor("x_361_cast_fp16")]; + tensor input_733_perm_0 = const()[name = tensor("input_733_perm_0"), val = tensor([0, 2, 1])]; + tensor input_733_cast_fp16 = transpose(perm = input_733_perm_0, x = x_361_cast_fp16)[name = tensor("transpose_114")]; + tensor input_735_cast_fp16 = silu(x = input_733_cast_fp16)[name = tensor("input_735_cast_fp16")]; + tensor x_363_pad_type_0 = const()[name = tensor("x_363_pad_type_0"), val = tensor("valid")]; + tensor x_363_strides_0 = const()[name = tensor("x_363_strides_0"), val = tensor([1])]; + tensor x_363_pad_0 = const()[name = tensor("x_363_pad_0"), val = tensor([0, 0])]; + tensor x_363_dilations_0 = const()[name = tensor("x_363_dilations_0"), val = tensor([1])]; + tensor x_363_groups_0 = const()[name = tensor("x_363_groups_0"), val = tensor(1)]; + tensor encoder_layers_13_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_13_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171271680)))]; + tensor x_363_cast_fp16 = conv(dilations = x_363_dilations_0, groups = x_363_groups_0, pad = x_363_pad_0, pad_type = x_363_pad_type_0, strides = x_363_strides_0, weight = encoder_layers_13_conv_pointwise_conv2_weight_to_fp16, x = input_735_cast_fp16)[name = tensor("x_363_cast_fp16")]; + tensor input_737_perm_0 = const()[name = tensor("input_737_perm_0"), val = tensor([0, 2, 1])]; + tensor input_737_cast_fp16 = transpose(perm = input_737_perm_0, x = x_363_cast_fp16)[name = tensor("transpose_113")]; + tensor input_739_cast_fp16 = add(x = input_723_cast_fp16, y = input_737_cast_fp16)[name = tensor("input_739_cast_fp16")]; + tensor input_741_axes_0 = const()[name = tensor("input_741_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171796032)))]; + tensor encoder_layers_13_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171797120)))]; + tensor input_741_cast_fp16 = layer_norm(axes = input_741_axes_0, beta = encoder_layers_13_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_norm_feed_forward2_weight_to_fp16, x = input_739_cast_fp16)[name = tensor("input_741_cast_fp16")]; + tensor encoder_layers_13_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_13_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171798208)))]; + tensor linear_125_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_13_feed_forward2_linear1_weight_to_fp16, x = input_741_cast_fp16)[name = tensor("linear_125_cast_fp16")]; + tensor input_745_cast_fp16 = silu(x = linear_125_cast_fp16)[name = tensor("input_745_cast_fp16")]; + tensor encoder_layers_13_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_13_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173895424)))]; + tensor linear_126_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_feed_forward2_linear2_weight_to_fp16, x = input_745_cast_fp16)[name = tensor("linear_126_cast_fp16")]; + tensor var_3166_to_fp16 = const()[name = tensor("op_3166_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3167_cast_fp16 = mul(x = linear_126_cast_fp16, y = var_3166_to_fp16)[name = tensor("op_3167_cast_fp16")]; + tensor input_751_cast_fp16 = add(x = input_739_cast_fp16, y = var_3167_cast_fp16)[name = tensor("input_751_cast_fp16")]; + tensor input_753_axes_0 = const()[name = tensor("input_753_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175992640)))]; + tensor encoder_layers_13_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175993728)))]; + tensor input_753_cast_fp16 = layer_norm(axes = input_753_axes_0, beta = encoder_layers_13_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_norm_out_weight_to_fp16, x = input_751_cast_fp16)[name = tensor("input_753_cast_fp16")]; + tensor cache_57_begin_0 = const()[name = tensor("cache_57_begin_0"), val = tensor([14, 0, 0, 0])]; + tensor cache_57_end_0 = const()[name = tensor("cache_57_end_0"), val = tensor([15, 1, 70, 512])]; + tensor cache_57_end_mask_0 = const()[name = tensor("cache_57_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_57_squeeze_mask_0 = const()[name = tensor("cache_57_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_57_cast_fp16 = slice_by_index(begin = cache_57_begin_0, end = cache_57_end_0, end_mask = cache_57_end_mask_0, squeeze_mask = cache_57_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_57_cast_fp16")]; + tensor cache_59_begin_0 = const()[name = tensor("cache_59_begin_0"), val = tensor([14, 0, 0, 0])]; + tensor cache_59_end_0 = const()[name = tensor("cache_59_end_0"), val = tensor([15, 1, 512, 8])]; + tensor cache_59_end_mask_0 = const()[name = tensor("cache_59_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_59_squeeze_mask_0 = const()[name = tensor("cache_59_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_59_cast_fp16 = slice_by_index(begin = cache_59_begin_0, end = cache_59_end_0, end_mask = cache_59_end_mask_0, squeeze_mask = cache_59_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_59_cast_fp16")]; + tensor input_755_axes_0 = const()[name = tensor("input_755_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175994816)))]; + tensor encoder_layers_14_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175995904)))]; + tensor input_755_cast_fp16 = layer_norm(axes = input_755_axes_0, beta = encoder_layers_14_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_norm_feed_forward1_weight_to_fp16, x = input_753_cast_fp16)[name = tensor("input_755_cast_fp16")]; + tensor encoder_layers_14_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_14_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175996992)))]; + tensor linear_127_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_14_feed_forward1_linear1_weight_to_fp16, x = input_755_cast_fp16)[name = tensor("linear_127_cast_fp16")]; + tensor input_759_cast_fp16 = silu(x = linear_127_cast_fp16)[name = tensor("input_759_cast_fp16")]; + tensor encoder_layers_14_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_14_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178094208)))]; + tensor linear_128_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_feed_forward1_linear2_weight_to_fp16, x = input_759_cast_fp16)[name = tensor("linear_128_cast_fp16")]; + tensor var_3201_to_fp16 = const()[name = tensor("op_3201_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3202_cast_fp16 = mul(x = linear_128_cast_fp16, y = var_3201_to_fp16)[name = tensor("op_3202_cast_fp16")]; + tensor input_765_cast_fp16 = add(x = input_753_cast_fp16, y = var_3202_cast_fp16)[name = tensor("input_765_cast_fp16")]; + tensor key_29_axes_0 = const()[name = tensor("key_29_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180191424)))]; + tensor encoder_layers_14_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180192512)))]; + tensor key_29_cast_fp16 = layer_norm(axes = key_29_axes_0, beta = encoder_layers_14_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_norm_self_att_weight_to_fp16, x = input_765_cast_fp16)[name = tensor("key_29_cast_fp16")]; + tensor input_767_interleave_0 = const()[name = tensor("input_767_interleave_0"), val = tensor(false)]; + tensor input_767_cast_fp16 = concat(axis = var_64, interleave = input_767_interleave_0, values = (cache_57_cast_fp16, key_29_cast_fp16))[name = tensor("input_767_cast_fp16")]; + tensor var_3224_begin_0 = const()[name = tensor("op_3224_begin_0"), val = tensor([0, 1, 0])]; + tensor var_3224_end_0 = const()[name = tensor("op_3224_end_0"), val = tensor([1, 70, 512])]; + tensor var_3224_end_mask_0 = const()[name = tensor("op_3224_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3224_cast_fp16 = slice_by_index(begin = var_3224_begin_0, end = var_3224_end_0, end_mask = var_3224_end_mask_0, x = cache_57_cast_fp16)[name = tensor("op_3224_cast_fp16")]; + tensor var_3227_begin_0 = const()[name = tensor("op_3227_begin_0"), val = tensor([0, 0, 0])]; + tensor var_3227_end_0 = const()[name = tensor("op_3227_end_0"), val = tensor([1, 1, 512])]; + tensor var_3227_end_mask_0 = const()[name = tensor("op_3227_end_mask_0"), val = tensor([true, false, true])]; + tensor var_3227_cast_fp16 = slice_by_index(begin = var_3227_begin_0, end = var_3227_end_0, end_mask = var_3227_end_mask_0, x = key_29_cast_fp16)[name = tensor("op_3227_cast_fp16")]; + tensor var_3230_interleave_0 = const()[name = tensor("op_3230_interleave_0"), val = tensor(false)]; + tensor var_3230_cast_fp16 = concat(axis = var_64, interleave = var_3230_interleave_0, values = (var_3224_cast_fp16, var_3227_cast_fp16))[name = tensor("op_3230_cast_fp16")]; + tensor encoder_layers_14_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180193600)))]; + tensor linear_129_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_self_attn_linear_q_weight_to_fp16, x = key_29_cast_fp16)[name = tensor("linear_129_cast_fp16")]; + tensor var_3234 = const()[name = tensor("op_3234"), val = tensor([1, -1, 8, 64])]; + tensor q_85_cast_fp16 = reshape(shape = var_3234, x = linear_129_cast_fp16)[name = tensor("q_85_cast_fp16")]; + tensor encoder_layers_14_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180717952)))]; + tensor linear_130_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_self_attn_linear_k_weight_to_fp16, x = input_767_cast_fp16)[name = tensor("linear_130_cast_fp16")]; + tensor var_3238 = const()[name = tensor("op_3238"), val = tensor([1, -1, 8, 64])]; + tensor k_57_cast_fp16 = reshape(shape = var_3238, x = linear_130_cast_fp16)[name = tensor("k_57_cast_fp16")]; + tensor encoder_layers_14_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181242304)))]; + tensor linear_131_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_self_attn_linear_v_weight_to_fp16, x = input_767_cast_fp16)[name = tensor("linear_131_cast_fp16")]; + tensor var_3242 = const()[name = tensor("op_3242"), val = tensor([1, -1, 8, 64])]; + tensor v_29_cast_fp16 = reshape(shape = var_3242, x = linear_131_cast_fp16)[name = tensor("v_29_cast_fp16")]; + tensor value_31_perm_0 = const()[name = tensor("value_31_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_14_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181766656)))]; + tensor var_3254_cast_fp16 = add(x = q_85_cast_fp16, y = encoder_layers_14_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3254_cast_fp16")]; + tensor encoder_layers_14_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181767744)))]; + tensor var_3256_cast_fp16 = add(x = q_85_cast_fp16, y = encoder_layers_14_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3256_cast_fp16")]; + tensor q_with_bias_v_29_perm_0 = const()[name = tensor("q_with_bias_v_29_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_371_transpose_x_0 = const()[name = tensor("x_371_transpose_x_0"), val = tensor(false)]; + tensor x_371_transpose_y_0 = const()[name = tensor("x_371_transpose_y_0"), val = tensor(false)]; + tensor var_3258_to_fp16 = const()[name = tensor("op_3258_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181768832)))]; + tensor q_with_bias_v_29_cast_fp16 = transpose(perm = q_with_bias_v_29_perm_0, x = var_3256_cast_fp16)[name = tensor("transpose_111")]; + tensor x_371_cast_fp16 = matmul(transpose_x = x_371_transpose_x_0, transpose_y = x_371_transpose_y_0, x = q_with_bias_v_29_cast_fp16, y = var_3258_to_fp16)[name = tensor("x_371_cast_fp16")]; + tensor x_373_pad_0 = const()[name = tensor("x_373_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_373_mode_0 = const()[name = tensor("x_373_mode_0"), val = tensor("constant")]; + tensor const_205_to_fp16 = const()[name = tensor("const_205_to_fp16"), val = tensor(0x0p+0)]; + tensor x_373_cast_fp16 = pad(constant_val = const_205_to_fp16, mode = x_373_mode_0, pad = x_373_pad_0, x = x_371_cast_fp16)[name = tensor("x_373_cast_fp16")]; + tensor var_3266 = const()[name = tensor("op_3266"), val = tensor([1, 8, -1, 3])]; + tensor x_375_cast_fp16 = reshape(shape = var_3266, x = x_373_cast_fp16)[name = tensor("x_375_cast_fp16")]; + tensor var_3270_begin_0 = const()[name = tensor("op_3270_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_3270_end_0 = const()[name = tensor("op_3270_end_0"), val = tensor([1, 8, 146, 3])]; + tensor var_3270_end_mask_0 = const()[name = tensor("op_3270_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_3270_cast_fp16 = slice_by_index(begin = var_3270_begin_0, end = var_3270_end_0, end_mask = var_3270_end_mask_0, x = x_375_cast_fp16)[name = tensor("op_3270_cast_fp16")]; + tensor var_3271 = const()[name = tensor("op_3271"), val = tensor([1, 8, 3, 145])]; + tensor matrix_bd_57_cast_fp16 = reshape(shape = var_3271, x = var_3270_cast_fp16)[name = tensor("matrix_bd_57_cast_fp16")]; + tensor matrix_ac_29_transpose_x_0 = const()[name = tensor("matrix_ac_29_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_29_transpose_y_0 = const()[name = tensor("matrix_ac_29_transpose_y_0"), val = tensor(false)]; + tensor transpose_79_perm_0 = const()[name = tensor("transpose_79_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_80_perm_0 = const()[name = tensor("transpose_80_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_80 = transpose(perm = transpose_80_perm_0, x = k_57_cast_fp16)[name = tensor("transpose_109")]; + tensor transpose_79 = transpose(perm = transpose_79_perm_0, x = var_3254_cast_fp16)[name = tensor("transpose_110")]; + tensor matrix_ac_29_cast_fp16 = matmul(transpose_x = matrix_ac_29_transpose_x_0, transpose_y = matrix_ac_29_transpose_y_0, x = transpose_79, y = transpose_80)[name = tensor("matrix_ac_29_cast_fp16")]; + tensor matrix_bd_59_begin_0 = const()[name = tensor("matrix_bd_59_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_59_end_0 = const()[name = tensor("matrix_bd_59_end_0"), val = tensor([1, 8, 3, 73])]; + tensor matrix_bd_59_end_mask_0 = const()[name = tensor("matrix_bd_59_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_59_cast_fp16 = slice_by_index(begin = matrix_bd_59_begin_0, end = matrix_bd_59_end_0, end_mask = matrix_bd_59_end_mask_0, x = matrix_bd_57_cast_fp16)[name = tensor("matrix_bd_59_cast_fp16")]; + tensor var_3280_cast_fp16 = add(x = matrix_ac_29_cast_fp16, y = matrix_bd_59_cast_fp16)[name = tensor("op_3280_cast_fp16")]; + tensor _inversed_scores_57_y_0_to_fp16 = const()[name = tensor("_inversed_scores_57_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_57_cast_fp16 = mul(x = var_3280_cast_fp16, y = _inversed_scores_57_y_0_to_fp16)[name = tensor("_inversed_scores_57_cast_fp16")]; + tensor scores_59_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_57_cast_fp16, cond = mask_3)[name = tensor("scores_59_cast_fp16")]; + tensor var_3286_cast_fp16 = softmax(axis = var_62, x = scores_59_cast_fp16)[name = tensor("op_3286_cast_fp16")]; + tensor input_769_cast_fp16 = select(a = var_40_to_fp16, b = var_3286_cast_fp16, cond = mask_3)[name = tensor("input_769_cast_fp16")]; + tensor x_377_transpose_x_0 = const()[name = tensor("x_377_transpose_x_0"), val = tensor(false)]; + tensor x_377_transpose_y_0 = const()[name = tensor("x_377_transpose_y_0"), val = tensor(false)]; + tensor value_31_cast_fp16 = transpose(perm = value_31_perm_0, x = v_29_cast_fp16)[name = tensor("transpose_112")]; + tensor x_377_cast_fp16 = matmul(transpose_x = x_377_transpose_x_0, transpose_y = x_377_transpose_y_0, x = input_769_cast_fp16, y = value_31_cast_fp16)[name = tensor("x_377_cast_fp16")]; + tensor var_3290_perm_0 = const()[name = tensor("op_3290_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_3291 = const()[name = tensor("op_3291"), val = tensor([1, -1, 512])]; + tensor var_3290_cast_fp16 = transpose(perm = var_3290_perm_0, x = x_377_cast_fp16)[name = tensor("transpose_108")]; + tensor input_771_cast_fp16 = reshape(shape = var_3291, x = var_3290_cast_fp16)[name = tensor("input_771_cast_fp16")]; + tensor encoder_layers_14_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181917376)))]; + tensor linear_133_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_self_attn_linear_out_weight_to_fp16, x = input_771_cast_fp16)[name = tensor("linear_133_cast_fp16")]; + tensor input_775_cast_fp16 = add(x = input_765_cast_fp16, y = linear_133_cast_fp16)[name = tensor("input_775_cast_fp16")]; + tensor x_381_axes_0 = const()[name = tensor("x_381_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182441728)))]; + tensor encoder_layers_14_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182442816)))]; + tensor x_381_cast_fp16 = layer_norm(axes = x_381_axes_0, beta = encoder_layers_14_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_norm_conv_weight_to_fp16, x = input_775_cast_fp16)[name = tensor("x_381_cast_fp16")]; + tensor input_777_perm_0 = const()[name = tensor("input_777_perm_0"), val = tensor([0, 2, 1])]; + tensor input_779_pad_type_0 = const()[name = tensor("input_779_pad_type_0"), val = tensor("valid")]; + tensor input_779_strides_0 = const()[name = tensor("input_779_strides_0"), val = tensor([1])]; + tensor input_779_pad_0 = const()[name = tensor("input_779_pad_0"), val = tensor([0, 0])]; + tensor input_779_dilations_0 = const()[name = tensor("input_779_dilations_0"), val = tensor([1])]; + tensor input_779_groups_0 = const()[name = tensor("input_779_groups_0"), val = tensor(1)]; + tensor encoder_layers_14_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_14_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182443904)))]; + tensor input_777_cast_fp16 = transpose(perm = input_777_perm_0, x = x_381_cast_fp16)[name = tensor("transpose_107")]; + tensor input_779_cast_fp16 = conv(dilations = input_779_dilations_0, groups = input_779_groups_0, pad = input_779_pad_0, pad_type = input_779_pad_type_0, strides = input_779_strides_0, weight = encoder_layers_14_conv_pointwise_conv1_weight_to_fp16, x = input_777_cast_fp16)[name = tensor("input_779_cast_fp16")]; + tensor x_383_split_num_splits_0 = const()[name = tensor("x_383_split_num_splits_0"), val = tensor(2)]; + tensor x_383_split_axis_0 = const()[name = tensor("x_383_split_axis_0"), val = tensor(1)]; + tensor x_383_split_cast_fp16_0, tensor x_383_split_cast_fp16_1 = split(axis = x_383_split_axis_0, num_splits = x_383_split_num_splits_0, x = input_779_cast_fp16)[name = tensor("x_383_split_cast_fp16")]; + tensor x_383_split_1_sigmoid_cast_fp16 = sigmoid(x = x_383_split_cast_fp16_1)[name = tensor("x_383_split_1_sigmoid_cast_fp16")]; + tensor x_383_cast_fp16 = mul(x = x_383_split_cast_fp16_0, y = x_383_split_1_sigmoid_cast_fp16)[name = tensor("x_383_cast_fp16")]; + tensor input_781_cast_fp16 = select(a = var_40_to_fp16, b = x_383_cast_fp16, cond = var_418)[name = tensor("input_781_cast_fp16")]; + tensor new_x_59_interleave_0 = const()[name = tensor("new_x_59_interleave_0"), val = tensor(false)]; + tensor new_x_59_cast_fp16 = concat(axis = var_62, interleave = new_x_59_interleave_0, values = (cache_59_cast_fp16, input_781_cast_fp16))[name = tensor("new_x_59_cast_fp16")]; + tensor next_cache_29_begin_0 = const()[name = tensor("next_cache_29_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_29_end_0 = const()[name = tensor("next_cache_29_end_0"), val = tensor([1, 512, 9])]; + tensor next_cache_29_end_mask_0 = const()[name = tensor("next_cache_29_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_29_cast_fp16 = slice_by_index(begin = next_cache_29_begin_0, end = next_cache_29_end_0, end_mask = next_cache_29_end_mask_0, x = new_x_59_cast_fp16)[name = tensor("next_cache_29_cast_fp16")]; + tensor var_3332_begin_0 = const()[name = tensor("op_3332_begin_0"), val = tensor([0, 0, 1])]; + tensor var_3332_end_0 = const()[name = tensor("op_3332_end_0"), val = tensor([1, 512, 9])]; + tensor var_3332_end_mask_0 = const()[name = tensor("op_3332_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3332_cast_fp16 = slice_by_index(begin = var_3332_begin_0, end = var_3332_end_0, end_mask = var_3332_end_mask_0, x = next_cache_29_cast_fp16)[name = tensor("op_3332_cast_fp16")]; + tensor x_385_pad_type_0 = const()[name = tensor("x_385_pad_type_0"), val = tensor("valid")]; + tensor x_385_groups_0 = const()[name = tensor("x_385_groups_0"), val = tensor(512)]; + tensor x_385_strides_0 = const()[name = tensor("x_385_strides_0"), val = tensor([1])]; + tensor x_385_pad_0 = const()[name = tensor("x_385_pad_0"), val = tensor([0, 0])]; + tensor x_385_dilations_0 = const()[name = tensor("x_385_dilations_0"), val = tensor([1])]; + tensor encoder_layers_14_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_14_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183492544)))]; + tensor x_385_cast_fp16 = conv(dilations = x_385_dilations_0, groups = x_385_groups_0, pad = x_385_pad_0, pad_type = x_385_pad_type_0, strides = x_385_strides_0, weight = encoder_layers_14_conv_depthwise_conv_weight_to_fp16, x = new_x_59_cast_fp16)[name = tensor("x_385_cast_fp16")]; + tensor input_783_perm_0 = const()[name = tensor("input_783_perm_0"), val = tensor([0, 2, 1])]; + tensor x_387_axes_0 = const()[name = tensor("x_387_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_14_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183501824)))]; + tensor encoder_layers_14_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_14_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183502912)))]; + tensor input_783_cast_fp16 = transpose(perm = input_783_perm_0, x = x_385_cast_fp16)[name = tensor("transpose_106")]; + tensor x_387_cast_fp16 = layer_norm(axes = x_387_axes_0, beta = encoder_layers_14_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_conv_batch_norm_weight_to_fp16, x = input_783_cast_fp16)[name = tensor("x_387_cast_fp16")]; + tensor input_785_perm_0 = const()[name = tensor("input_785_perm_0"), val = tensor([0, 2, 1])]; + tensor input_785_cast_fp16 = transpose(perm = input_785_perm_0, x = x_387_cast_fp16)[name = tensor("transpose_105")]; + tensor input_787_cast_fp16 = silu(x = input_785_cast_fp16)[name = tensor("input_787_cast_fp16")]; + tensor x_389_pad_type_0 = const()[name = tensor("x_389_pad_type_0"), val = tensor("valid")]; + tensor x_389_strides_0 = const()[name = tensor("x_389_strides_0"), val = tensor([1])]; + tensor x_389_pad_0 = const()[name = tensor("x_389_pad_0"), val = tensor([0, 0])]; + tensor x_389_dilations_0 = const()[name = tensor("x_389_dilations_0"), val = tensor([1])]; + tensor x_389_groups_0 = const()[name = tensor("x_389_groups_0"), val = tensor(1)]; + tensor encoder_layers_14_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_14_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183504000)))]; + tensor x_389_cast_fp16 = conv(dilations = x_389_dilations_0, groups = x_389_groups_0, pad = x_389_pad_0, pad_type = x_389_pad_type_0, strides = x_389_strides_0, weight = encoder_layers_14_conv_pointwise_conv2_weight_to_fp16, x = input_787_cast_fp16)[name = tensor("x_389_cast_fp16")]; + tensor input_789_perm_0 = const()[name = tensor("input_789_perm_0"), val = tensor([0, 2, 1])]; + tensor input_789_cast_fp16 = transpose(perm = input_789_perm_0, x = x_389_cast_fp16)[name = tensor("transpose_104")]; + tensor input_791_cast_fp16 = add(x = input_775_cast_fp16, y = input_789_cast_fp16)[name = tensor("input_791_cast_fp16")]; + tensor input_793_axes_0 = const()[name = tensor("input_793_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184028352)))]; + tensor encoder_layers_14_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184029440)))]; + tensor input_793_cast_fp16 = layer_norm(axes = input_793_axes_0, beta = encoder_layers_14_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_norm_feed_forward2_weight_to_fp16, x = input_791_cast_fp16)[name = tensor("input_793_cast_fp16")]; + tensor encoder_layers_14_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_14_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184030528)))]; + tensor linear_134_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_14_feed_forward2_linear1_weight_to_fp16, x = input_793_cast_fp16)[name = tensor("linear_134_cast_fp16")]; + tensor input_797_cast_fp16 = silu(x = linear_134_cast_fp16)[name = tensor("input_797_cast_fp16")]; + tensor encoder_layers_14_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_14_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186127744)))]; + tensor linear_135_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_feed_forward2_linear2_weight_to_fp16, x = input_797_cast_fp16)[name = tensor("linear_135_cast_fp16")]; + tensor var_3373_to_fp16 = const()[name = tensor("op_3373_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3374_cast_fp16 = mul(x = linear_135_cast_fp16, y = var_3373_to_fp16)[name = tensor("op_3374_cast_fp16")]; + tensor input_803_cast_fp16 = add(x = input_791_cast_fp16, y = var_3374_cast_fp16)[name = tensor("input_803_cast_fp16")]; + tensor input_805_axes_0 = const()[name = tensor("input_805_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188224960)))]; + tensor encoder_layers_14_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188226048)))]; + tensor input_805_cast_fp16 = layer_norm(axes = input_805_axes_0, beta = encoder_layers_14_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_norm_out_weight_to_fp16, x = input_803_cast_fp16)[name = tensor("input_805_cast_fp16")]; + tensor cache_61_begin_0 = const()[name = tensor("cache_61_begin_0"), val = tensor([15, 0, 0, 0])]; + tensor cache_61_end_0 = const()[name = tensor("cache_61_end_0"), val = tensor([16, 1, 70, 512])]; + tensor cache_61_end_mask_0 = const()[name = tensor("cache_61_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_61_squeeze_mask_0 = const()[name = tensor("cache_61_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_61_cast_fp16 = slice_by_index(begin = cache_61_begin_0, end = cache_61_end_0, end_mask = cache_61_end_mask_0, squeeze_mask = cache_61_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_61_cast_fp16")]; + tensor cache_63_begin_0 = const()[name = tensor("cache_63_begin_0"), val = tensor([15, 0, 0, 0])]; + tensor cache_63_end_0 = const()[name = tensor("cache_63_end_0"), val = tensor([16, 1, 512, 8])]; + tensor cache_63_end_mask_0 = const()[name = tensor("cache_63_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_63_squeeze_mask_0 = const()[name = tensor("cache_63_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_63_cast_fp16 = slice_by_index(begin = cache_63_begin_0, end = cache_63_end_0, end_mask = cache_63_end_mask_0, squeeze_mask = cache_63_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_63_cast_fp16")]; + tensor input_807_axes_0 = const()[name = tensor("input_807_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188227136)))]; + tensor encoder_layers_15_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188228224)))]; + tensor input_807_cast_fp16 = layer_norm(axes = input_807_axes_0, beta = encoder_layers_15_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_norm_feed_forward1_weight_to_fp16, x = input_805_cast_fp16)[name = tensor("input_807_cast_fp16")]; + tensor encoder_layers_15_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_15_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188229312)))]; + tensor linear_136_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_15_feed_forward1_linear1_weight_to_fp16, x = input_807_cast_fp16)[name = tensor("linear_136_cast_fp16")]; + tensor input_811_cast_fp16 = silu(x = linear_136_cast_fp16)[name = tensor("input_811_cast_fp16")]; + tensor encoder_layers_15_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_15_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190326528)))]; + tensor linear_137_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_feed_forward1_linear2_weight_to_fp16, x = input_811_cast_fp16)[name = tensor("linear_137_cast_fp16")]; + tensor var_3408_to_fp16 = const()[name = tensor("op_3408_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3409_cast_fp16 = mul(x = linear_137_cast_fp16, y = var_3408_to_fp16)[name = tensor("op_3409_cast_fp16")]; + tensor input_817_cast_fp16 = add(x = input_805_cast_fp16, y = var_3409_cast_fp16)[name = tensor("input_817_cast_fp16")]; + tensor key_31_axes_0 = const()[name = tensor("key_31_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192423744)))]; + tensor encoder_layers_15_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192424832)))]; + tensor key_31_cast_fp16 = layer_norm(axes = key_31_axes_0, beta = encoder_layers_15_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_norm_self_att_weight_to_fp16, x = input_817_cast_fp16)[name = tensor("key_31_cast_fp16")]; + tensor input_819_interleave_0 = const()[name = tensor("input_819_interleave_0"), val = tensor(false)]; + tensor input_819_cast_fp16 = concat(axis = var_64, interleave = input_819_interleave_0, values = (cache_61_cast_fp16, key_31_cast_fp16))[name = tensor("input_819_cast_fp16")]; + tensor var_3431_begin_0 = const()[name = tensor("op_3431_begin_0"), val = tensor([0, 1, 0])]; + tensor var_3431_end_0 = const()[name = tensor("op_3431_end_0"), val = tensor([1, 70, 512])]; + tensor var_3431_end_mask_0 = const()[name = tensor("op_3431_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3431_cast_fp16 = slice_by_index(begin = var_3431_begin_0, end = var_3431_end_0, end_mask = var_3431_end_mask_0, x = cache_61_cast_fp16)[name = tensor("op_3431_cast_fp16")]; + tensor var_3434_begin_0 = const()[name = tensor("op_3434_begin_0"), val = tensor([0, 0, 0])]; + tensor var_3434_end_0 = const()[name = tensor("op_3434_end_0"), val = tensor([1, 1, 512])]; + tensor var_3434_end_mask_0 = const()[name = tensor("op_3434_end_mask_0"), val = tensor([true, false, true])]; + tensor var_3434_cast_fp16 = slice_by_index(begin = var_3434_begin_0, end = var_3434_end_0, end_mask = var_3434_end_mask_0, x = key_31_cast_fp16)[name = tensor("op_3434_cast_fp16")]; + tensor var_3437_interleave_0 = const()[name = tensor("op_3437_interleave_0"), val = tensor(false)]; + tensor var_3437_cast_fp16 = concat(axis = var_64, interleave = var_3437_interleave_0, values = (var_3431_cast_fp16, var_3434_cast_fp16))[name = tensor("op_3437_cast_fp16")]; + tensor encoder_layers_15_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192425920)))]; + tensor linear_138_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_self_attn_linear_q_weight_to_fp16, x = key_31_cast_fp16)[name = tensor("linear_138_cast_fp16")]; + tensor var_3441 = const()[name = tensor("op_3441"), val = tensor([1, -1, 8, 64])]; + tensor q_91_cast_fp16 = reshape(shape = var_3441, x = linear_138_cast_fp16)[name = tensor("q_91_cast_fp16")]; + tensor encoder_layers_15_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192950272)))]; + tensor linear_139_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_self_attn_linear_k_weight_to_fp16, x = input_819_cast_fp16)[name = tensor("linear_139_cast_fp16")]; + tensor var_3445 = const()[name = tensor("op_3445"), val = tensor([1, -1, 8, 64])]; + tensor k_61_cast_fp16 = reshape(shape = var_3445, x = linear_139_cast_fp16)[name = tensor("k_61_cast_fp16")]; + tensor encoder_layers_15_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(193474624)))]; + tensor linear_140_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_self_attn_linear_v_weight_to_fp16, x = input_819_cast_fp16)[name = tensor("linear_140_cast_fp16")]; + tensor var_3449 = const()[name = tensor("op_3449"), val = tensor([1, -1, 8, 64])]; + tensor v_31_cast_fp16 = reshape(shape = var_3449, x = linear_140_cast_fp16)[name = tensor("v_31_cast_fp16")]; + tensor value_33_perm_0 = const()[name = tensor("value_33_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_15_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(193998976)))]; + tensor var_3461_cast_fp16 = add(x = q_91_cast_fp16, y = encoder_layers_15_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3461_cast_fp16")]; + tensor encoder_layers_15_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194000064)))]; + tensor var_3463_cast_fp16 = add(x = q_91_cast_fp16, y = encoder_layers_15_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3463_cast_fp16")]; + tensor q_with_bias_v_31_perm_0 = const()[name = tensor("q_with_bias_v_31_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_397_transpose_x_0 = const()[name = tensor("x_397_transpose_x_0"), val = tensor(false)]; + tensor x_397_transpose_y_0 = const()[name = tensor("x_397_transpose_y_0"), val = tensor(false)]; + tensor var_3465_to_fp16 = const()[name = tensor("op_3465_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194001152)))]; + tensor q_with_bias_v_31_cast_fp16 = transpose(perm = q_with_bias_v_31_perm_0, x = var_3463_cast_fp16)[name = tensor("transpose_102")]; + tensor x_397_cast_fp16 = matmul(transpose_x = x_397_transpose_x_0, transpose_y = x_397_transpose_y_0, x = q_with_bias_v_31_cast_fp16, y = var_3465_to_fp16)[name = tensor("x_397_cast_fp16")]; + tensor x_399_pad_0 = const()[name = tensor("x_399_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_399_mode_0 = const()[name = tensor("x_399_mode_0"), val = tensor("constant")]; + tensor const_218_to_fp16 = const()[name = tensor("const_218_to_fp16"), val = tensor(0x0p+0)]; + tensor x_399_cast_fp16 = pad(constant_val = const_218_to_fp16, mode = x_399_mode_0, pad = x_399_pad_0, x = x_397_cast_fp16)[name = tensor("x_399_cast_fp16")]; + tensor var_3473 = const()[name = tensor("op_3473"), val = tensor([1, 8, -1, 3])]; + tensor x_401_cast_fp16 = reshape(shape = var_3473, x = x_399_cast_fp16)[name = tensor("x_401_cast_fp16")]; + tensor var_3477_begin_0 = const()[name = tensor("op_3477_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_3477_end_0 = const()[name = tensor("op_3477_end_0"), val = tensor([1, 8, 146, 3])]; + tensor var_3477_end_mask_0 = const()[name = tensor("op_3477_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_3477_cast_fp16 = slice_by_index(begin = var_3477_begin_0, end = var_3477_end_0, end_mask = var_3477_end_mask_0, x = x_401_cast_fp16)[name = tensor("op_3477_cast_fp16")]; + tensor var_3478 = const()[name = tensor("op_3478"), val = tensor([1, 8, 3, 145])]; + tensor matrix_bd_61_cast_fp16 = reshape(shape = var_3478, x = var_3477_cast_fp16)[name = tensor("matrix_bd_61_cast_fp16")]; + tensor matrix_ac_31_transpose_x_0 = const()[name = tensor("matrix_ac_31_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_31_transpose_y_0 = const()[name = tensor("matrix_ac_31_transpose_y_0"), val = tensor(false)]; + tensor transpose_81_perm_0 = const()[name = tensor("transpose_81_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_82_perm_0 = const()[name = tensor("transpose_82_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_82 = transpose(perm = transpose_82_perm_0, x = k_61_cast_fp16)[name = tensor("transpose_100")]; + tensor transpose_81 = transpose(perm = transpose_81_perm_0, x = var_3461_cast_fp16)[name = tensor("transpose_101")]; + tensor matrix_ac_31_cast_fp16 = matmul(transpose_x = matrix_ac_31_transpose_x_0, transpose_y = matrix_ac_31_transpose_y_0, x = transpose_81, y = transpose_82)[name = tensor("matrix_ac_31_cast_fp16")]; + tensor matrix_bd_63_begin_0 = const()[name = tensor("matrix_bd_63_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_63_end_0 = const()[name = tensor("matrix_bd_63_end_0"), val = tensor([1, 8, 3, 73])]; + tensor matrix_bd_63_end_mask_0 = const()[name = tensor("matrix_bd_63_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_63_cast_fp16 = slice_by_index(begin = matrix_bd_63_begin_0, end = matrix_bd_63_end_0, end_mask = matrix_bd_63_end_mask_0, x = matrix_bd_61_cast_fp16)[name = tensor("matrix_bd_63_cast_fp16")]; + tensor var_3487_cast_fp16 = add(x = matrix_ac_31_cast_fp16, y = matrix_bd_63_cast_fp16)[name = tensor("op_3487_cast_fp16")]; + tensor _inversed_scores_61_y_0_to_fp16 = const()[name = tensor("_inversed_scores_61_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_61_cast_fp16 = mul(x = var_3487_cast_fp16, y = _inversed_scores_61_y_0_to_fp16)[name = tensor("_inversed_scores_61_cast_fp16")]; + tensor scores_63_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_61_cast_fp16, cond = mask_3)[name = tensor("scores_63_cast_fp16")]; + tensor var_3493_cast_fp16 = softmax(axis = var_62, x = scores_63_cast_fp16)[name = tensor("op_3493_cast_fp16")]; + tensor input_821_cast_fp16 = select(a = var_40_to_fp16, b = var_3493_cast_fp16, cond = mask_3)[name = tensor("input_821_cast_fp16")]; + tensor x_403_transpose_x_0 = const()[name = tensor("x_403_transpose_x_0"), val = tensor(false)]; + tensor x_403_transpose_y_0 = const()[name = tensor("x_403_transpose_y_0"), val = tensor(false)]; + tensor value_33_cast_fp16 = transpose(perm = value_33_perm_0, x = v_31_cast_fp16)[name = tensor("transpose_103")]; + tensor x_403_cast_fp16 = matmul(transpose_x = x_403_transpose_x_0, transpose_y = x_403_transpose_y_0, x = input_821_cast_fp16, y = value_33_cast_fp16)[name = tensor("x_403_cast_fp16")]; + tensor var_3497_perm_0 = const()[name = tensor("op_3497_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_3498 = const()[name = tensor("op_3498"), val = tensor([1, -1, 512])]; + tensor var_3497_cast_fp16 = transpose(perm = var_3497_perm_0, x = x_403_cast_fp16)[name = tensor("transpose_99")]; + tensor input_823_cast_fp16 = reshape(shape = var_3498, x = var_3497_cast_fp16)[name = tensor("input_823_cast_fp16")]; + tensor encoder_layers_15_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194149696)))]; + tensor linear_142_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_self_attn_linear_out_weight_to_fp16, x = input_823_cast_fp16)[name = tensor("linear_142_cast_fp16")]; + tensor input_827_cast_fp16 = add(x = input_817_cast_fp16, y = linear_142_cast_fp16)[name = tensor("input_827_cast_fp16")]; + tensor x_407_axes_0 = const()[name = tensor("x_407_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194674048)))]; + tensor encoder_layers_15_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194675136)))]; + tensor x_407_cast_fp16 = layer_norm(axes = x_407_axes_0, beta = encoder_layers_15_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_norm_conv_weight_to_fp16, x = input_827_cast_fp16)[name = tensor("x_407_cast_fp16")]; + tensor input_829_perm_0 = const()[name = tensor("input_829_perm_0"), val = tensor([0, 2, 1])]; + tensor input_831_pad_type_0 = const()[name = tensor("input_831_pad_type_0"), val = tensor("valid")]; + tensor input_831_strides_0 = const()[name = tensor("input_831_strides_0"), val = tensor([1])]; + tensor input_831_pad_0 = const()[name = tensor("input_831_pad_0"), val = tensor([0, 0])]; + tensor input_831_dilations_0 = const()[name = tensor("input_831_dilations_0"), val = tensor([1])]; + tensor input_831_groups_0 = const()[name = tensor("input_831_groups_0"), val = tensor(1)]; + tensor encoder_layers_15_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_15_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194676224)))]; + tensor input_829_cast_fp16 = transpose(perm = input_829_perm_0, x = x_407_cast_fp16)[name = tensor("transpose_98")]; + tensor input_831_cast_fp16 = conv(dilations = input_831_dilations_0, groups = input_831_groups_0, pad = input_831_pad_0, pad_type = input_831_pad_type_0, strides = input_831_strides_0, weight = encoder_layers_15_conv_pointwise_conv1_weight_to_fp16, x = input_829_cast_fp16)[name = tensor("input_831_cast_fp16")]; + tensor x_409_split_num_splits_0 = const()[name = tensor("x_409_split_num_splits_0"), val = tensor(2)]; + tensor x_409_split_axis_0 = const()[name = tensor("x_409_split_axis_0"), val = tensor(1)]; + tensor x_409_split_cast_fp16_0, tensor x_409_split_cast_fp16_1 = split(axis = x_409_split_axis_0, num_splits = x_409_split_num_splits_0, x = input_831_cast_fp16)[name = tensor("x_409_split_cast_fp16")]; + tensor x_409_split_1_sigmoid_cast_fp16 = sigmoid(x = x_409_split_cast_fp16_1)[name = tensor("x_409_split_1_sigmoid_cast_fp16")]; + tensor x_409_cast_fp16 = mul(x = x_409_split_cast_fp16_0, y = x_409_split_1_sigmoid_cast_fp16)[name = tensor("x_409_cast_fp16")]; + tensor input_833_cast_fp16 = select(a = var_40_to_fp16, b = x_409_cast_fp16, cond = var_418)[name = tensor("input_833_cast_fp16")]; + tensor new_x_63_interleave_0 = const()[name = tensor("new_x_63_interleave_0"), val = tensor(false)]; + tensor new_x_63_cast_fp16 = concat(axis = var_62, interleave = new_x_63_interleave_0, values = (cache_63_cast_fp16, input_833_cast_fp16))[name = tensor("new_x_63_cast_fp16")]; + tensor next_cache_31_begin_0 = const()[name = tensor("next_cache_31_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_31_end_0 = const()[name = tensor("next_cache_31_end_0"), val = tensor([1, 512, 9])]; + tensor next_cache_31_end_mask_0 = const()[name = tensor("next_cache_31_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_31_cast_fp16 = slice_by_index(begin = next_cache_31_begin_0, end = next_cache_31_end_0, end_mask = next_cache_31_end_mask_0, x = new_x_63_cast_fp16)[name = tensor("next_cache_31_cast_fp16")]; + tensor var_3539_begin_0 = const()[name = tensor("op_3539_begin_0"), val = tensor([0, 0, 1])]; + tensor var_3539_end_0 = const()[name = tensor("op_3539_end_0"), val = tensor([1, 512, 9])]; + tensor var_3539_end_mask_0 = const()[name = tensor("op_3539_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3539_cast_fp16 = slice_by_index(begin = var_3539_begin_0, end = var_3539_end_0, end_mask = var_3539_end_mask_0, x = next_cache_31_cast_fp16)[name = tensor("op_3539_cast_fp16")]; + tensor x_411_pad_type_0 = const()[name = tensor("x_411_pad_type_0"), val = tensor("valid")]; + tensor x_411_groups_0 = const()[name = tensor("x_411_groups_0"), val = tensor(512)]; + tensor x_411_strides_0 = const()[name = tensor("x_411_strides_0"), val = tensor([1])]; + tensor x_411_pad_0 = const()[name = tensor("x_411_pad_0"), val = tensor([0, 0])]; + tensor x_411_dilations_0 = const()[name = tensor("x_411_dilations_0"), val = tensor([1])]; + tensor encoder_layers_15_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_15_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195724864)))]; + tensor x_411_cast_fp16 = conv(dilations = x_411_dilations_0, groups = x_411_groups_0, pad = x_411_pad_0, pad_type = x_411_pad_type_0, strides = x_411_strides_0, weight = encoder_layers_15_conv_depthwise_conv_weight_to_fp16, x = new_x_63_cast_fp16)[name = tensor("x_411_cast_fp16")]; + tensor input_835_perm_0 = const()[name = tensor("input_835_perm_0"), val = tensor([0, 2, 1])]; + tensor x_413_axes_0 = const()[name = tensor("x_413_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_15_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195734144)))]; + tensor encoder_layers_15_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_15_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195735232)))]; + tensor input_835_cast_fp16 = transpose(perm = input_835_perm_0, x = x_411_cast_fp16)[name = tensor("transpose_97")]; + tensor x_413_cast_fp16 = layer_norm(axes = x_413_axes_0, beta = encoder_layers_15_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_conv_batch_norm_weight_to_fp16, x = input_835_cast_fp16)[name = tensor("x_413_cast_fp16")]; + tensor input_837_perm_0 = const()[name = tensor("input_837_perm_0"), val = tensor([0, 2, 1])]; + tensor input_837_cast_fp16 = transpose(perm = input_837_perm_0, x = x_413_cast_fp16)[name = tensor("transpose_96")]; + tensor input_839_cast_fp16 = silu(x = input_837_cast_fp16)[name = tensor("input_839_cast_fp16")]; + tensor x_415_pad_type_0 = const()[name = tensor("x_415_pad_type_0"), val = tensor("valid")]; + tensor x_415_strides_0 = const()[name = tensor("x_415_strides_0"), val = tensor([1])]; + tensor x_415_pad_0 = const()[name = tensor("x_415_pad_0"), val = tensor([0, 0])]; + tensor x_415_dilations_0 = const()[name = tensor("x_415_dilations_0"), val = tensor([1])]; + tensor x_415_groups_0 = const()[name = tensor("x_415_groups_0"), val = tensor(1)]; + tensor encoder_layers_15_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_15_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195736320)))]; + tensor x_415_cast_fp16 = conv(dilations = x_415_dilations_0, groups = x_415_groups_0, pad = x_415_pad_0, pad_type = x_415_pad_type_0, strides = x_415_strides_0, weight = encoder_layers_15_conv_pointwise_conv2_weight_to_fp16, x = input_839_cast_fp16)[name = tensor("x_415_cast_fp16")]; + tensor input_841_perm_0 = const()[name = tensor("input_841_perm_0"), val = tensor([0, 2, 1])]; + tensor input_841_cast_fp16 = transpose(perm = input_841_perm_0, x = x_415_cast_fp16)[name = tensor("transpose_95")]; + tensor input_843_cast_fp16 = add(x = input_827_cast_fp16, y = input_841_cast_fp16)[name = tensor("input_843_cast_fp16")]; + tensor input_845_axes_0 = const()[name = tensor("input_845_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196260672)))]; + tensor encoder_layers_15_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196261760)))]; + tensor input_845_cast_fp16 = layer_norm(axes = input_845_axes_0, beta = encoder_layers_15_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_norm_feed_forward2_weight_to_fp16, x = input_843_cast_fp16)[name = tensor("input_845_cast_fp16")]; + tensor encoder_layers_15_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_15_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196262848)))]; + tensor linear_143_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_15_feed_forward2_linear1_weight_to_fp16, x = input_845_cast_fp16)[name = tensor("linear_143_cast_fp16")]; + tensor input_849_cast_fp16 = silu(x = linear_143_cast_fp16)[name = tensor("input_849_cast_fp16")]; + tensor encoder_layers_15_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_15_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198360064)))]; + tensor linear_144_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_feed_forward2_linear2_weight_to_fp16, x = input_849_cast_fp16)[name = tensor("linear_144_cast_fp16")]; + tensor var_3580_to_fp16 = const()[name = tensor("op_3580_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3581_cast_fp16 = mul(x = linear_144_cast_fp16, y = var_3580_to_fp16)[name = tensor("op_3581_cast_fp16")]; + tensor input_855_cast_fp16 = add(x = input_843_cast_fp16, y = var_3581_cast_fp16)[name = tensor("input_855_cast_fp16")]; + tensor input_857_axes_0 = const()[name = tensor("input_857_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200457280)))]; + tensor encoder_layers_15_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200458368)))]; + tensor input_857_cast_fp16 = layer_norm(axes = input_857_axes_0, beta = encoder_layers_15_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_norm_out_weight_to_fp16, x = input_855_cast_fp16)[name = tensor("input_857_cast_fp16")]; + tensor cache_65_begin_0 = const()[name = tensor("cache_65_begin_0"), val = tensor([16, 0, 0, 0])]; + tensor cache_65_end_0 = const()[name = tensor("cache_65_end_0"), val = tensor([17, 1, 70, 512])]; + tensor cache_65_end_mask_0 = const()[name = tensor("cache_65_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_65_squeeze_mask_0 = const()[name = tensor("cache_65_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_65_cast_fp16 = slice_by_index(begin = cache_65_begin_0, end = cache_65_end_0, end_mask = cache_65_end_mask_0, squeeze_mask = cache_65_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_65_cast_fp16")]; + tensor cache_begin_0 = const()[name = tensor("cache_begin_0"), val = tensor([16, 0, 0, 0])]; + tensor cache_end_0 = const()[name = tensor("cache_end_0"), val = tensor([17, 1, 512, 8])]; + tensor cache_end_mask_0 = const()[name = tensor("cache_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_squeeze_mask_0 = const()[name = tensor("cache_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_cast_fp16 = slice_by_index(begin = cache_begin_0, end = cache_end_0, end_mask = cache_end_mask_0, squeeze_mask = cache_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_cast_fp16")]; + tensor input_859_axes_0 = const()[name = tensor("input_859_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200459456)))]; + tensor encoder_layers_16_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200460544)))]; + tensor input_859_cast_fp16 = layer_norm(axes = input_859_axes_0, beta = encoder_layers_16_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_norm_feed_forward1_weight_to_fp16, x = input_857_cast_fp16)[name = tensor("input_859_cast_fp16")]; + tensor encoder_layers_16_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_16_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200461632)))]; + tensor linear_145_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_16_feed_forward1_linear1_weight_to_fp16, x = input_859_cast_fp16)[name = tensor("linear_145_cast_fp16")]; + tensor input_863_cast_fp16 = silu(x = linear_145_cast_fp16)[name = tensor("input_863_cast_fp16")]; + tensor encoder_layers_16_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_16_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202558848)))]; + tensor linear_146_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_feed_forward1_linear2_weight_to_fp16, x = input_863_cast_fp16)[name = tensor("linear_146_cast_fp16")]; + tensor var_3615_to_fp16 = const()[name = tensor("op_3615_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3616_cast_fp16 = mul(x = linear_146_cast_fp16, y = var_3615_to_fp16)[name = tensor("op_3616_cast_fp16")]; + tensor input_869_cast_fp16 = add(x = input_857_cast_fp16, y = var_3616_cast_fp16)[name = tensor("input_869_cast_fp16")]; + tensor key_axes_0 = const()[name = tensor("key_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204656064)))]; + tensor encoder_layers_16_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204657152)))]; + tensor key_cast_fp16 = layer_norm(axes = key_axes_0, beta = encoder_layers_16_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_norm_self_att_weight_to_fp16, x = input_869_cast_fp16)[name = tensor("key_cast_fp16")]; + tensor input_871_interleave_0 = const()[name = tensor("input_871_interleave_0"), val = tensor(false)]; + tensor input_871_cast_fp16 = concat(axis = var_64, interleave = input_871_interleave_0, values = (cache_65_cast_fp16, key_cast_fp16))[name = tensor("input_871_cast_fp16")]; + tensor var_3638_begin_0 = const()[name = tensor("op_3638_begin_0"), val = tensor([0, 1, 0])]; + tensor var_3638_end_0 = const()[name = tensor("op_3638_end_0"), val = tensor([1, 70, 512])]; + tensor var_3638_end_mask_0 = const()[name = tensor("op_3638_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3638_cast_fp16 = slice_by_index(begin = var_3638_begin_0, end = var_3638_end_0, end_mask = var_3638_end_mask_0, x = cache_65_cast_fp16)[name = tensor("op_3638_cast_fp16")]; + tensor var_3641_begin_0 = const()[name = tensor("op_3641_begin_0"), val = tensor([0, 0, 0])]; + tensor var_3641_end_0 = const()[name = tensor("op_3641_end_0"), val = tensor([1, 1, 512])]; + tensor var_3641_end_mask_0 = const()[name = tensor("op_3641_end_mask_0"), val = tensor([true, false, true])]; + tensor var_3641_cast_fp16 = slice_by_index(begin = var_3641_begin_0, end = var_3641_end_0, end_mask = var_3641_end_mask_0, x = key_cast_fp16)[name = tensor("op_3641_cast_fp16")]; + tensor cache_last_channel_cur_interleave_0 = const()[name = tensor("cache_last_channel_cur_interleave_0"), val = tensor(false)]; + tensor cache_last_channel_cur_cast_fp16 = concat(axis = var_64, interleave = cache_last_channel_cur_interleave_0, values = (var_3638_cast_fp16, var_3641_cast_fp16))[name = tensor("cache_last_channel_cur_cast_fp16")]; + tensor encoder_layers_16_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204658240)))]; + tensor linear_147_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_self_attn_linear_q_weight_to_fp16, x = key_cast_fp16)[name = tensor("linear_147_cast_fp16")]; + tensor var_3648 = const()[name = tensor("op_3648"), val = tensor([1, -1, 8, 64])]; + tensor q_97_cast_fp16 = reshape(shape = var_3648, x = linear_147_cast_fp16)[name = tensor("q_97_cast_fp16")]; + tensor encoder_layers_16_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205182592)))]; + tensor linear_148_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_self_attn_linear_k_weight_to_fp16, x = input_871_cast_fp16)[name = tensor("linear_148_cast_fp16")]; + tensor var_3652 = const()[name = tensor("op_3652"), val = tensor([1, -1, 8, 64])]; + tensor k_65_cast_fp16 = reshape(shape = var_3652, x = linear_148_cast_fp16)[name = tensor("k_65_cast_fp16")]; + tensor encoder_layers_16_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205706944)))]; + tensor linear_149_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_self_attn_linear_v_weight_to_fp16, x = input_871_cast_fp16)[name = tensor("linear_149_cast_fp16")]; + tensor var_3656 = const()[name = tensor("op_3656"), val = tensor([1, -1, 8, 64])]; + tensor v_cast_fp16 = reshape(shape = var_3656, x = linear_149_cast_fp16)[name = tensor("v_cast_fp16")]; + tensor value_perm_0 = const()[name = tensor("value_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_16_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206231296)))]; + tensor var_3668_cast_fp16 = add(x = q_97_cast_fp16, y = encoder_layers_16_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3668_cast_fp16")]; + tensor encoder_layers_16_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206232384)))]; + tensor var_3670_cast_fp16 = add(x = q_97_cast_fp16, y = encoder_layers_16_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3670_cast_fp16")]; + tensor q_with_bias_v_perm_0 = const()[name = tensor("q_with_bias_v_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_423_transpose_x_0 = const()[name = tensor("x_423_transpose_x_0"), val = tensor(false)]; + tensor x_423_transpose_y_0 = const()[name = tensor("x_423_transpose_y_0"), val = tensor(false)]; + tensor var_3672_to_fp16 = const()[name = tensor("op_3672_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206233472)))]; + tensor q_with_bias_v_cast_fp16 = transpose(perm = q_with_bias_v_perm_0, x = var_3670_cast_fp16)[name = tensor("transpose_93")]; + tensor x_423_cast_fp16 = matmul(transpose_x = x_423_transpose_x_0, transpose_y = x_423_transpose_y_0, x = q_with_bias_v_cast_fp16, y = var_3672_to_fp16)[name = tensor("x_423_cast_fp16")]; + tensor x_425_pad_0 = const()[name = tensor("x_425_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_425_mode_0 = const()[name = tensor("x_425_mode_0"), val = tensor("constant")]; + tensor const_231_to_fp16 = const()[name = tensor("const_231_to_fp16"), val = tensor(0x0p+0)]; + tensor x_425_cast_fp16 = pad(constant_val = const_231_to_fp16, mode = x_425_mode_0, pad = x_425_pad_0, x = x_423_cast_fp16)[name = tensor("x_425_cast_fp16")]; + tensor var_3680 = const()[name = tensor("op_3680"), val = tensor([1, 8, -1, 3])]; + tensor x_427_cast_fp16 = reshape(shape = var_3680, x = x_425_cast_fp16)[name = tensor("x_427_cast_fp16")]; + tensor var_3684_begin_0 = const()[name = tensor("op_3684_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_3684_end_0 = const()[name = tensor("op_3684_end_0"), val = tensor([1, 8, 146, 3])]; + tensor var_3684_end_mask_0 = const()[name = tensor("op_3684_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_3684_cast_fp16 = slice_by_index(begin = var_3684_begin_0, end = var_3684_end_0, end_mask = var_3684_end_mask_0, x = x_427_cast_fp16)[name = tensor("op_3684_cast_fp16")]; + tensor var_3685 = const()[name = tensor("op_3685"), val = tensor([1, 8, 3, 145])]; + tensor matrix_bd_65_cast_fp16 = reshape(shape = var_3685, x = var_3684_cast_fp16)[name = tensor("matrix_bd_65_cast_fp16")]; + tensor matrix_ac_transpose_x_0 = const()[name = tensor("matrix_ac_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_transpose_y_0 = const()[name = tensor("matrix_ac_transpose_y_0"), val = tensor(false)]; + tensor transpose_83_perm_0 = const()[name = tensor("transpose_83_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_84_perm_0 = const()[name = tensor("transpose_84_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_84 = transpose(perm = transpose_84_perm_0, x = k_65_cast_fp16)[name = tensor("transpose_91")]; + tensor transpose_83 = transpose(perm = transpose_83_perm_0, x = var_3668_cast_fp16)[name = tensor("transpose_92")]; + tensor matrix_ac_cast_fp16 = matmul(transpose_x = matrix_ac_transpose_x_0, transpose_y = matrix_ac_transpose_y_0, x = transpose_83, y = transpose_84)[name = tensor("matrix_ac_cast_fp16")]; + tensor matrix_bd_begin_0 = const()[name = tensor("matrix_bd_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_end_0 = const()[name = tensor("matrix_bd_end_0"), val = tensor([1, 8, 3, 73])]; + tensor matrix_bd_end_mask_0 = const()[name = tensor("matrix_bd_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_cast_fp16 = slice_by_index(begin = matrix_bd_begin_0, end = matrix_bd_end_0, end_mask = matrix_bd_end_mask_0, x = matrix_bd_65_cast_fp16)[name = tensor("matrix_bd_cast_fp16")]; + tensor var_3694_cast_fp16 = add(x = matrix_ac_cast_fp16, y = matrix_bd_cast_fp16)[name = tensor("op_3694_cast_fp16")]; + tensor _inversed_scores_65_y_0_to_fp16 = const()[name = tensor("_inversed_scores_65_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_65_cast_fp16 = mul(x = var_3694_cast_fp16, y = _inversed_scores_65_y_0_to_fp16)[name = tensor("_inversed_scores_65_cast_fp16")]; + tensor scores_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_65_cast_fp16, cond = mask_3)[name = tensor("scores_cast_fp16")]; + tensor var_3700_cast_fp16 = softmax(axis = var_62, x = scores_cast_fp16)[name = tensor("op_3700_cast_fp16")]; + tensor input_873_cast_fp16 = select(a = var_40_to_fp16, b = var_3700_cast_fp16, cond = mask_3)[name = tensor("input_873_cast_fp16")]; + tensor x_429_transpose_x_0 = const()[name = tensor("x_429_transpose_x_0"), val = tensor(false)]; + tensor x_429_transpose_y_0 = const()[name = tensor("x_429_transpose_y_0"), val = tensor(false)]; + tensor value_cast_fp16 = transpose(perm = value_perm_0, x = v_cast_fp16)[name = tensor("transpose_94")]; + tensor x_429_cast_fp16 = matmul(transpose_x = x_429_transpose_x_0, transpose_y = x_429_transpose_y_0, x = input_873_cast_fp16, y = value_cast_fp16)[name = tensor("x_429_cast_fp16")]; + tensor var_3704_perm_0 = const()[name = tensor("op_3704_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_3705 = const()[name = tensor("op_3705"), val = tensor([1, -1, 512])]; + tensor var_3704_cast_fp16 = transpose(perm = var_3704_perm_0, x = x_429_cast_fp16)[name = tensor("transpose_90")]; + tensor input_875_cast_fp16 = reshape(shape = var_3705, x = var_3704_cast_fp16)[name = tensor("input_875_cast_fp16")]; + tensor encoder_layers_16_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206382016)))]; + tensor linear_151_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_self_attn_linear_out_weight_to_fp16, x = input_875_cast_fp16)[name = tensor("linear_151_cast_fp16")]; + tensor input_879_cast_fp16 = add(x = input_869_cast_fp16, y = linear_151_cast_fp16)[name = tensor("input_879_cast_fp16")]; + tensor x_433_axes_0 = const()[name = tensor("x_433_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206906368)))]; + tensor encoder_layers_16_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206907456)))]; + tensor x_433_cast_fp16 = layer_norm(axes = x_433_axes_0, beta = encoder_layers_16_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_norm_conv_weight_to_fp16, x = input_879_cast_fp16)[name = tensor("x_433_cast_fp16")]; + tensor input_881_perm_0 = const()[name = tensor("input_881_perm_0"), val = tensor([0, 2, 1])]; + tensor input_883_pad_type_0 = const()[name = tensor("input_883_pad_type_0"), val = tensor("valid")]; + tensor input_883_strides_0 = const()[name = tensor("input_883_strides_0"), val = tensor([1])]; + tensor input_883_pad_0 = const()[name = tensor("input_883_pad_0"), val = tensor([0, 0])]; + tensor input_883_dilations_0 = const()[name = tensor("input_883_dilations_0"), val = tensor([1])]; + tensor input_883_groups_0 = const()[name = tensor("input_883_groups_0"), val = tensor(1)]; + tensor encoder_layers_16_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_16_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206908544)))]; + tensor input_881_cast_fp16 = transpose(perm = input_881_perm_0, x = x_433_cast_fp16)[name = tensor("transpose_89")]; + tensor input_883_cast_fp16 = conv(dilations = input_883_dilations_0, groups = input_883_groups_0, pad = input_883_pad_0, pad_type = input_883_pad_type_0, strides = input_883_strides_0, weight = encoder_layers_16_conv_pointwise_conv1_weight_to_fp16, x = input_881_cast_fp16)[name = tensor("input_883_cast_fp16")]; + tensor x_435_split_num_splits_0 = const()[name = tensor("x_435_split_num_splits_0"), val = tensor(2)]; + tensor x_435_split_axis_0 = const()[name = tensor("x_435_split_axis_0"), val = tensor(1)]; + tensor x_435_split_cast_fp16_0, tensor x_435_split_cast_fp16_1 = split(axis = x_435_split_axis_0, num_splits = x_435_split_num_splits_0, x = input_883_cast_fp16)[name = tensor("x_435_split_cast_fp16")]; + tensor x_435_split_1_sigmoid_cast_fp16 = sigmoid(x = x_435_split_cast_fp16_1)[name = tensor("x_435_split_1_sigmoid_cast_fp16")]; + tensor x_435_cast_fp16 = mul(x = x_435_split_cast_fp16_0, y = x_435_split_1_sigmoid_cast_fp16)[name = tensor("x_435_cast_fp16")]; + tensor input_885_cast_fp16 = select(a = var_40_to_fp16, b = x_435_cast_fp16, cond = var_418)[name = tensor("input_885_cast_fp16")]; + tensor new_x_interleave_0 = const()[name = tensor("new_x_interleave_0"), val = tensor(false)]; + tensor new_x_cast_fp16 = concat(axis = var_62, interleave = new_x_interleave_0, values = (cache_cast_fp16, input_885_cast_fp16))[name = tensor("new_x_cast_fp16")]; + tensor next_cache_begin_0 = const()[name = tensor("next_cache_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_end_0 = const()[name = tensor("next_cache_end_0"), val = tensor([1, 512, 9])]; + tensor next_cache_end_mask_0 = const()[name = tensor("next_cache_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_cast_fp16 = slice_by_index(begin = next_cache_begin_0, end = next_cache_end_0, end_mask = next_cache_end_mask_0, x = new_x_cast_fp16)[name = tensor("next_cache_cast_fp16")]; + tensor cache_last_time_cur_begin_0 = const()[name = tensor("cache_last_time_cur_begin_0"), val = tensor([0, 0, 1])]; + tensor cache_last_time_cur_end_0 = const()[name = tensor("cache_last_time_cur_end_0"), val = tensor([1, 512, 9])]; + tensor cache_last_time_cur_end_mask_0 = const()[name = tensor("cache_last_time_cur_end_mask_0"), val = tensor([true, true, true])]; + tensor cache_last_time_cur_cast_fp16 = slice_by_index(begin = cache_last_time_cur_begin_0, end = cache_last_time_cur_end_0, end_mask = cache_last_time_cur_end_mask_0, x = next_cache_cast_fp16)[name = tensor("cache_last_time_cur_cast_fp16")]; + tensor x_437_pad_type_0 = const()[name = tensor("x_437_pad_type_0"), val = tensor("valid")]; + tensor x_437_groups_0 = const()[name = tensor("x_437_groups_0"), val = tensor(512)]; + tensor x_437_strides_0 = const()[name = tensor("x_437_strides_0"), val = tensor([1])]; + tensor x_437_pad_0 = const()[name = tensor("x_437_pad_0"), val = tensor([0, 0])]; + tensor x_437_dilations_0 = const()[name = tensor("x_437_dilations_0"), val = tensor([1])]; + tensor encoder_layers_16_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_16_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207957184)))]; + tensor x_437_cast_fp16 = conv(dilations = x_437_dilations_0, groups = x_437_groups_0, pad = x_437_pad_0, pad_type = x_437_pad_type_0, strides = x_437_strides_0, weight = encoder_layers_16_conv_depthwise_conv_weight_to_fp16, x = new_x_cast_fp16)[name = tensor("x_437_cast_fp16")]; + tensor input_887_perm_0 = const()[name = tensor("input_887_perm_0"), val = tensor([0, 2, 1])]; + tensor x_439_axes_0 = const()[name = tensor("x_439_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_16_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207966464)))]; + tensor encoder_layers_16_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_16_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207967552)))]; + tensor input_887_cast_fp16 = transpose(perm = input_887_perm_0, x = x_437_cast_fp16)[name = tensor("transpose_88")]; + tensor x_439_cast_fp16 = layer_norm(axes = x_439_axes_0, beta = encoder_layers_16_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_conv_batch_norm_weight_to_fp16, x = input_887_cast_fp16)[name = tensor("x_439_cast_fp16")]; + tensor input_889_perm_0 = const()[name = tensor("input_889_perm_0"), val = tensor([0, 2, 1])]; + tensor input_889_cast_fp16 = transpose(perm = input_889_perm_0, x = x_439_cast_fp16)[name = tensor("transpose_87")]; + tensor input_891_cast_fp16 = silu(x = input_889_cast_fp16)[name = tensor("input_891_cast_fp16")]; + tensor x_441_pad_type_0 = const()[name = tensor("x_441_pad_type_0"), val = tensor("valid")]; + tensor x_441_strides_0 = const()[name = tensor("x_441_strides_0"), val = tensor([1])]; + tensor x_441_pad_0 = const()[name = tensor("x_441_pad_0"), val = tensor([0, 0])]; + tensor x_441_dilations_0 = const()[name = tensor("x_441_dilations_0"), val = tensor([1])]; + tensor x_441_groups_0 = const()[name = tensor("x_441_groups_0"), val = tensor(1)]; + tensor encoder_layers_16_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_16_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207968640)))]; + tensor x_441_cast_fp16 = conv(dilations = x_441_dilations_0, groups = x_441_groups_0, pad = x_441_pad_0, pad_type = x_441_pad_type_0, strides = x_441_strides_0, weight = encoder_layers_16_conv_pointwise_conv2_weight_to_fp16, x = input_891_cast_fp16)[name = tensor("x_441_cast_fp16")]; + tensor input_893_perm_0 = const()[name = tensor("input_893_perm_0"), val = tensor([0, 2, 1])]; + tensor input_893_cast_fp16 = transpose(perm = input_893_perm_0, x = x_441_cast_fp16)[name = tensor("transpose_86")]; + tensor input_895_cast_fp16 = add(x = input_879_cast_fp16, y = input_893_cast_fp16)[name = tensor("input_895_cast_fp16")]; + tensor input_897_axes_0 = const()[name = tensor("input_897_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208492992)))]; + tensor encoder_layers_16_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208494080)))]; + tensor input_897_cast_fp16 = layer_norm(axes = input_897_axes_0, beta = encoder_layers_16_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_norm_feed_forward2_weight_to_fp16, x = input_895_cast_fp16)[name = tensor("input_897_cast_fp16")]; + tensor encoder_layers_16_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_16_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208495168)))]; + tensor linear_152_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_16_feed_forward2_linear1_weight_to_fp16, x = input_897_cast_fp16)[name = tensor("linear_152_cast_fp16")]; + tensor input_901_cast_fp16 = silu(x = linear_152_cast_fp16)[name = tensor("input_901_cast_fp16")]; + tensor encoder_layers_16_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_16_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210592384)))]; + tensor linear_153_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_feed_forward2_linear2_weight_to_fp16, x = input_901_cast_fp16)[name = tensor("linear_153_cast_fp16")]; + tensor var_3787_to_fp16 = const()[name = tensor("op_3787_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3788_cast_fp16 = mul(x = linear_153_cast_fp16, y = var_3787_to_fp16)[name = tensor("op_3788_cast_fp16")]; + tensor input_cast_fp16 = add(x = input_895_cast_fp16, y = var_3788_cast_fp16)[name = tensor("input_cast_fp16")]; + tensor audio_signal_axes_0 = const()[name = tensor("audio_signal_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212689600)))]; + tensor encoder_layers_16_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212690688)))]; + tensor audio_signal_cast_fp16 = layer_norm(axes = audio_signal_axes_0, beta = encoder_layers_16_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_norm_out_weight_to_fp16, x = input_cast_fp16)[name = tensor("audio_signal_cast_fp16")]; + tensor obj_1_perm_0 = const()[name = tensor("obj_1_perm_0"), val = tensor([0, 2, 1])]; + tensor cast_178_dtype_0 = const()[name = tensor("cast_178_dtype_0"), val = tensor("int32")]; + tensor obj_5_axis_0 = const()[name = tensor("obj_5_axis_0"), val = tensor(0)]; + tensor obj_5_cast_fp16 = stack(axis = obj_5_axis_0, values = (var_332_cast_fp16, var_539_cast_fp16, var_746_cast_fp16, var_953_cast_fp16, var_1160_cast_fp16, var_1367_cast_fp16, var_1574_cast_fp16, var_1781_cast_fp16, var_1988_cast_fp16, var_2195_cast_fp16, var_2402_cast_fp16, var_2609_cast_fp16, var_2816_cast_fp16, var_3023_cast_fp16, var_3230_cast_fp16, var_3437_cast_fp16, cache_last_channel_cur_cast_fp16))[name = tensor("obj_5_cast_fp16")]; + tensor obj_7_axis_0 = const()[name = tensor("obj_7_axis_0"), val = tensor(0)]; + tensor obj_7_cast_fp16 = stack(axis = obj_7_axis_0, values = (var_434_cast_fp16, var_641_cast_fp16, var_848_cast_fp16, var_1055_cast_fp16, var_1262_cast_fp16, var_1469_cast_fp16, var_1676_cast_fp16, var_1883_cast_fp16, var_2090_cast_fp16, var_2297_cast_fp16, var_2504_cast_fp16, var_2711_cast_fp16, var_2918_cast_fp16, var_3125_cast_fp16, var_3332_cast_fp16, var_3539_cast_fp16, cache_last_time_cur_cast_fp16))[name = tensor("obj_7_cast_fp16")]; + tensor obj_7_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_7_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor var_3804 = add(x = cache_last_channel_len, y = cache_keep_size)[name = tensor("op_3804")]; + tensor var_3804_promoted_to_fp16_dtype_0 = const()[name = tensor("op_3804_promoted_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor const_237_to_fp16 = const()[name = tensor("const_237_to_fp16"), val = tensor(-inf)]; + tensor var_45_promoted_to_fp16 = const()[name = tensor("op_45_promoted_to_fp16"), val = tensor(0x1.18p+6)]; + tensor var_3804_to_fp16 = cast(dtype = var_3804_promoted_to_fp16_dtype_0, x = var_3804)[name = tensor("cast_186")]; + tensor clip_1_cast_fp16 = clip(alpha = const_237_to_fp16, beta = var_45_promoted_to_fp16, x = var_3804_to_fp16)[name = tensor("clip_1_cast_fp16")]; + tensor var_3831_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_3831_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor var_3846_begin_0 = const()[name = tensor("op_3846_begin_0"), val = tensor([0, 0, 0])]; + tensor var_3846_end_0 = const()[name = tensor("op_3846_end_0"), val = tensor([1, 512, 2])]; + tensor var_3846_end_mask_0 = const()[name = tensor("op_3846_end_mask_0"), val = tensor([true, true, false])]; + tensor obj_1_cast_fp16 = transpose(perm = obj_1_perm_0, x = audio_signal_cast_fp16)[name = tensor("transpose_85")]; + tensor var_3846_cast_fp16 = slice_by_index(begin = var_3846_begin_0, end = var_3846_end_0, end_mask = var_3846_end_mask_0, x = obj_1_cast_fp16)[name = tensor("op_3846_cast_fp16")]; + tensor var_3846_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_3846_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor cast_178_promoted_to_fp16_dtype_0 = const()[name = tensor("cast_178_promoted_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor const_238_to_fp16 = const()[name = tensor("const_238_to_fp16"), val = tensor(-inf)]; + tensor var_3848_promoted_to_fp16 = const()[name = tensor("op_3848_promoted_to_fp16"), val = tensor(0x1p+1)]; + tensor clip_0_cast_fp16_to_int32 = cast(dtype = cast_178_dtype_0, x = clip_0_cast_fp16)[name = tensor("cast_188")]; + tensor clip_0_cast_fp16_to_int32_to_fp16 = cast(dtype = cast_178_promoted_to_fp16_dtype_0, x = clip_0_cast_fp16_to_int32)[name = tensor("cast_183")]; + tensor clip_2_cast_fp16 = clip(alpha = const_238_to_fp16, beta = var_3848_promoted_to_fp16, x = clip_0_cast_fp16_to_int32_to_fp16)[name = tensor("clip_2_cast_fp16")]; + tensor cast_179_dtype_0 = const()[name = tensor("cast_179_dtype_0"), val = tensor("int32")]; + tensor cast_180_dtype_0 = const()[name = tensor("cast_180_dtype_0"), val = tensor("int32")]; + tensor new_cache_last_channel_len = cast(dtype = cast_180_dtype_0, x = clip_1_cast_fp16)[name = tensor("cast_181")]; + tensor encoded_length = cast(dtype = cast_179_dtype_0, x = clip_2_cast_fp16)[name = tensor("cast_182")]; + tensor encoded_output = cast(dtype = var_3846_cast_fp16_to_fp32_dtype_0, x = var_3846_cast_fp16)[name = tensor("cast_184")]; + tensor new_cache_last_channel = cast(dtype = var_3831_cast_fp16_to_fp32_dtype_0, x = obj_5_cast_fp16)[name = tensor("cast_185")]; + tensor new_cache_last_time = cast(dtype = obj_7_cast_fp16_to_fp32_dtype_0, x = obj_7_cast_fp16)[name = tensor("cast_187")]; + tensor new_pre_cache = cast(dtype = var_28_cast_fp16_to_fp32_dtype_0, x = var_28_cast_fp16)[name = tensor("cast_194")]; + } -> (encoded_output, encoded_length, new_pre_cache, new_cache_last_channel, new_cache_last_time, new_cache_last_channel_len); +} \ No newline at end of file diff --git a/160ms/streaming_encoder.mlmodelc/weights/weight.bin b/160ms/streaming_encoder.mlmodelc/weights/weight.bin new file mode 100644 index 0000000000000000000000000000000000000000..36f875726932df5278655f7d4ef4e024c7ef9370 --- /dev/null +++ b/160ms/streaming_encoder.mlmodelc/weights/weight.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:12cd781a4300b52b6687587b7d8e37e0ce5c8ccb1dbea036008275e6abf5070c +size 212691776 diff --git a/160ms/streaming_encoder.mlpackage/Data/com.apple.CoreML/model.mlmodel b/160ms/streaming_encoder.mlpackage/Data/com.apple.CoreML/model.mlmodel new file mode 100644 index 0000000000000000000000000000000000000000..70022ef0bc4f8becf4754afe70d501e12951987a --- /dev/null +++ b/160ms/streaming_encoder.mlpackage/Data/com.apple.CoreML/model.mlmodel @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ca562061f63b3e908e835f45da5ea7668b560678c57c5279d7abd7d1e150846 +size 539800 diff --git a/160ms/streaming_encoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/160ms/streaming_encoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin new file mode 100644 index 0000000000000000000000000000000000000000..36f875726932df5278655f7d4ef4e024c7ef9370 --- /dev/null +++ b/160ms/streaming_encoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:12cd781a4300b52b6687587b7d8e37e0ce5c8ccb1dbea036008275e6abf5070c +size 212691776 diff --git a/160ms/streaming_encoder.mlpackage/Manifest.json b/160ms/streaming_encoder.mlpackage/Manifest.json new file mode 100644 index 0000000000000000000000000000000000000000..98d489310c78257a8108a9ffc3cefb329029e499 --- /dev/null +++ b/160ms/streaming_encoder.mlpackage/Manifest.json @@ -0,0 +1,18 @@ +{ + "fileFormatVersion": "1.0.0", + "itemInfoEntries": { + "58E1A989-930B-49FA-AF42-1F7D3F023CA4": { + "author": "com.apple.CoreML", + "description": "CoreML Model Weights", + "name": "weights", + "path": "com.apple.CoreML/weights" + }, + "8E215896-5930-4E70-B597-BEF2276F4D32": { + "author": "com.apple.CoreML", + "description": "CoreML Model Specification", + "name": "model.mlmodel", + "path": "com.apple.CoreML/model.mlmodel" + } + }, + "rootModelIdentifier": "8E215896-5930-4E70-B597-BEF2276F4D32" +} diff --git a/160ms/vocab.json b/160ms/vocab.json new file mode 100644 index 0000000000000000000000000000000000000000..462081190dbb331a0dde13e99d5cfa99b302f0a7 --- /dev/null +++ b/160ms/vocab.json @@ -0,0 +1,1028 @@ +{ + "0": "", + "1": "▁t", + "2": "▁th", + "3": "▁a", + "4": "▁i", + "5": "▁the", + "6": "▁s", + "7": "re", + "8": "▁w", + "9": "▁o", + "10": "in", + "11": "at", + "12": "er", + "13": "nd", + "14": "ou", + "15": "▁c", + "16": "▁b", + "17": "▁h", + "18": "en", + "19": "on", + "20": "▁m", + "21": "▁f", + "22": "ing", + "23": "▁p", + "24": "▁to", + "25": "▁and", + "26": "▁d", + "27": "an", + "28": "or", + "29": "es", + "30": "▁y", + "31": "▁l", + "32": "▁of", + "33": "ll", + "34": "▁in", + "35": "ed", + "36": "it", + "37": "▁g", + "38": "is", + "39": "▁you", + "40": "▁n", + "41": "ar", + "42": "om", + "43": "as", + "44": "ve", + "45": "▁e", + "46": "ic", + "47": "▁it", + "48": "al", + "49": "us", + "50": "▁wh", + "51": "▁we", + "52": "▁be", + "53": "ion", + "54": "ow", + "55": "le", + "56": "▁is", + "57": "et", + "58": "ent", + "59": "ot", + "60": "ut", + "61": "▁re", + "62": "▁on", + "63": "ay", + "64": "▁ha", + "65": "ig", + "66": "▁so", + "67": "ct", + "68": "▁he", + "69": "▁for", + "70": "ver", + "71": "ke", + "72": "ro", + "73": "▁st", + "74": "id", + "75": "▁go", + "76": "all", + "77": "se", + "78": "ly", + "79": "▁u", + "80": "ch", + "81": "st", + "82": "ld", + "83": "▁k", + "84": "ce", + "85": "ur", + "86": "▁li", + "87": "am", + "88": "▁r", + "89": "ht", + "90": "▁j", + "91": "ith", + "92": "▁se", + "93": "ir", + "94": "▁as", + "95": "▁an", + "96": "im", + "97": "▁do", + "98": "ad", + "99": "▁was", + "100": "ight", + "101": "th", + "102": "▁are", + "103": "▁but", + "104": "▁sh", + "105": "ust", + "106": "ally", + "107": "▁not", + "108": "▁or", + "109": "▁com", + "110": "▁can", + "111": "▁me", + "112": "op", + "113": "▁mo", + "114": "▁at", + "115": "ill", + "116": "▁ch", + "117": "▁ne", + "118": "ant", + "119": "▁de", + "120": "▁kn", + "121": "▁one", + "122": "il", + "123": "ol", + "124": "▁con", + "125": "ter", + "126": "▁ab", + "127": "▁fr", + "128": "ere", + "129": "ck", + "130": "▁al", + "131": "▁all", + "132": "qu", + "133": "▁pro", + "134": "▁som", + "135": "ould", + "136": "▁tw", + "137": "ul", + "138": "ra", + "139": "od", + "140": "ers", + "141": "▁su", + "142": "ive", + "143": "▁v", + "144": "use", + "145": "ate", + "146": "ge", + "147": "if", + "148": "▁ex", + "149": "ess", + "150": "pp", + "151": "▁lo", + "152": "out", + "153": "▁if", + "154": "est", + "155": "ain", + "156": "ist", + "157": "and", + "158": "ea", + "159": "very", + "160": "art", + "161": "▁wor", + "162": "▁my", + "163": "ab", + "164": "ment", + "165": "▁bec", + "166": "un", + "167": "ity", + "168": "ri", + "169": "pe", + "170": "ions", + "171": "▁by", + "172": "ok", + "173": "our", + "174": "ort", + "175": "ind", + "176": "ink", + "177": "nt", + "178": "▁up", + "179": "um", + "180": "▁don", + "181": "▁get", + "182": "red", + "183": "▁out", + "184": "el", + "185": "ause", + "186": "res", + "187": "▁ma", + "188": "ich", + "189": "▁us", + "190": "rou", + "191": "▁int", + "192": "em", + "193": "os", + "194": "ies", + "195": "ie", + "196": "▁pl", + "197": "▁tr", + "198": "ven", + "199": "ous", + "200": "▁le", + "201": "▁two", + "202": "ard", + "203": "ine", + "204": "▁co", + "205": "een", + "206": "▁now", + "207": "ty", + "208": "her", + "209": "ack", + "210": "▁pe", + "211": "ame", + "212": "▁how", + "213": "▁who", + "214": "▁see", + "215": "▁tim", + "216": "ect", + "217": "ast", + "218": "▁our", + "219": "ci", + "220": "ree", + "221": "ople", + "222": "gh", + "223": "▁no", + "224": "▁had", + "225": "▁man", + "226": "▁qu", + "227": "▁en", + "228": "ide", + "229": "ure", + "230": "ud", + "231": "so", + "232": "▁his", + "233": "▁sa", + "234": "▁sp", + "235": "▁say", + "236": "ose", + "237": "ther", + "238": "▁act", + "239": "▁ta", + "240": "▁cl", + "241": "ings", + "242": "pt", + "243": "king", + "244": "▁any", + "245": "▁has", + "246": "▁un", + "247": "iv", + "248": "▁im", + "249": "▁ag", + "250": "▁te", + "251": "▁fe", + "252": "one", + "253": "per", + "254": "ong", + "255": "▁po", + "256": "▁ad", + "257": "ff", + "258": "ore", + "259": "itt", + "260": "ans", + "261": "iz", + "262": "eah", + "263": "reat", + "264": "act", + "265": "own", + "266": "hing", + "267": "enty", + "268": "age", + "269": "ber", + "270": "ice", + "271": "▁am", + "272": "ple", + "273": "are", + "274": "▁per", + "275": "und", + "276": "ite", + "277": "ix", + "278": "pl", + "279": "▁way", + "280": "▁did", + "281": "▁pr", + "282": "▁got", + "283": "ars", + "284": "▁she", + "285": "▁let", + "286": "ag", + "287": "▁ac", + "288": "int", + "289": "▁ar", + "290": "ry", + "291": "ign", + "292": "ish", + "293": "▁fir", + "294": "ace", + "295": "ble", + "296": "og", + "297": "ue", + "298": "▁ye", + "299": "ap", + "300": "iff", + "301": "▁ro", + "302": "▁her", + "303": "nder", + "304": "▁ok", + "305": "▁res", + "306": "▁gu", + "307": "ence", + "308": "▁may", + "309": "ated", + "310": "ip", + "311": "▁bo", + "312": "▁him", + "313": "way", + "314": "ac", + "315": "ical", + "316": "ass", + "317": "ase", + "318": "▁dis", + "319": "able", + "320": "ick", + "321": "▁app", + "322": "ance", + "323": "▁pre", + "324": "▁six", + "325": "▁off", + "326": "▁new", + "327": "ia", + "328": "orm", + "329": "ank", + "330": "▁lot", + "331": "ach", + "332": "▁fo", + "333": "inet", + "334": "ire", + "335": "ary", + "336": "ult", + "337": "▁tal", + "338": "▁mu", + "339": "▁bl", + "340": "ount", + "341": "sel", + "342": "vel", + "343": "▁br", + "344": "▁imp", + "345": "ep", + "346": "cess", + "347": "ord", + "348": "▁sc", + "349": "▁inc", + "350": "ound", + "351": "ang", + "352": "be", + "353": "ress", + "354": "uct", + "355": "▁ind", + "356": "▁af", + "357": "ving", + "358": "▁oh", + "359": "▁bet", + "360": "▁use", + "361": "ome", + "362": "ens", + "363": "ys", + "364": "▁bu", + "365": "co", + "366": "ory", + "367": "ater", + "368": "ild", + "369": "ght", + "370": "ial", + "371": "▁day", + "372": "ning", + "373": "na", + "374": "ile", + "375": "▁spe", + "376": "▁mar", + "377": "ody", + "378": "ough", + "379": "ade", + "380": "vers", + "381": "xt", + "382": "▁fl", + "383": "▁ke", + "384": "ian", + "385": "▁sy", + "386": "▁put", + "387": "fore", + "388": "ub", + "389": "▁ph", + "390": "fe", + "391": "▁em", + "392": "▁ser", + "393": "form", + "394": "ting", + "395": "te", + "396": "av", + "397": "ious", + "398": "▁rec", + "399": "ks", + "400": "▁gr", + "401": "ces", + "402": "wn", + "403": "ors", + "404": "▁jo", + "405": "ents", + "406": "▁des", + "407": "▁try", + "408": "▁equ", + "409": "▁z", + "410": "▁rem", + "411": "▁str", + "412": "self", + "413": "▁bit", + "414": "ph", + "415": "ved", + "416": "▁why", + "417": "▁bas", + "418": "▁hel", + "419": "▁rel", + "420": "ath", + "421": "ject", + "422": "ail", + "423": "▁la", + "424": "ual", + "425": "▁god", + "426": "▁nat", + "427": "erm", + "428": "day", + "429": "▁id", + "430": "ft", + "431": "▁wr", + "432": "▁min", + "433": "ates", + "434": "▁gen", + "435": "tain", + "436": "▁ob", + "437": "ull", + "438": "ict", + "439": "▁tra", + "440": "▁end", + "441": "▁hig", + "442": "▁fif", + "443": "oth", + "444": "tern", + "445": "▁its", + "446": "vent", + "447": "▁sm", + "448": "ons", + "449": "▁add", + "450": "iss", + "451": "▁bel", + "452": "ful", + "453": "get", + "454": "▁ele", + "455": "▁rep", + "456": "ak", + "457": "▁ho", + "458": "▁pos", + "459": "▁num", + "460": "ange", + "461": "ves", + "462": "ific", + "463": "urn", + "464": "ise", + "465": "▁cr", + "466": "▁um", + "467": "ward", + "468": "▁reg", + "469": "ady", + "470": "ower", + "471": "uc", + "472": "▁dec", + "473": "lic", + "474": "▁set", + "475": "▁gon", + "476": "▁op", + "477": "▁ear", + "478": "▁sub", + "479": "▁sl", + "480": "les", + "481": "stem", + "482": "cial", + "483": "olog", + "484": "atch", + "485": "ily", + "486": "body", + "487": "nds", + "488": "ular", + "489": "ren", + "490": "▁own", + "491": "▁too", + "492": "cent", + "493": "ible", + "494": "pect", + "495": "ered", + "496": "ways", + "497": "teen", + "498": "▁uh", + "499": "▁big", + "500": "▁mod", + "501": "▁att", + "502": "▁car", + "503": "gr", + "504": "▁acc", + "505": "ied", + "506": "mun", + "507": "ib", + "508": "▁mon", + "509": "▁sch", + "510": "▁pol", + "511": "▁dat", + "512": "▁fin", + "513": "▁sim", + "514": "▁inv", + "515": "▁def", + "516": "ked", + "517": "▁ent", + "518": "▁yes", + "519": "ows", + "520": "ics", + "521": "ited", + "522": "ute", + "523": "ism", + "524": "ps", + "525": "▁ed", + "526": "▁el", + "527": "ably", + "528": "ppen", + "529": "als", + "530": "▁ten", + "531": "ract", + "532": "ss", + "533": "▁ass", + "534": "▁met", + "535": "gan", + "536": "▁eng", + "537": "▁stu", + "538": "ween", + "539": "arch", + "540": "▁gl", + "541": "▁cor", + "542": "▁dr", + "543": "vern", + "544": "▁ty", + "545": "▁run", + "546": "hip", + "547": "cus", + "548": "cond", + "549": "▁ins", + "550": "irty", + "551": "▁pub", + "552": "lud", + "553": "llow", + "554": "▁cou", + "555": "ew", + "556": "iew", + "557": "▁sur", + "558": "ero", + "559": "ood", + "560": "ness", + "561": "▁fun", + "562": "▁eff", + "563": "cept", + "564": "▁ca", + "565": "▁exp", + "566": "duct", + "567": "▁sw", + "568": "ize", + "569": "ope", + "570": "▁par", + "571": "kes", + "572": "cy", + "573": "▁ev", + "574": "▁ref", + "575": "ell", + "576": "▁bus", + "577": "ug", + "578": "rib", + "579": "▁cur", + "580": "mo", + "581": "ock", + "582": "ures", + "583": "air", + "584": "▁war", + "585": "str", + "586": "▁med", + "587": "▁wa", + "588": "▁val", + "589": "▁sin", + "590": "blem", + "591": "▁fam", + "592": "li", + "593": "▁far", + "594": "▁cle", + "595": "▁col", + "596": "mon", + "597": "▁gra", + "598": "led", + "599": "ense", + "600": "tin", + "601": "ues", + "602": "its", + "603": "▁mem", + "604": "▁inf", + "605": "▁eas", + "606": "ideo", + "607": "▁top", + "608": "io", + "609": "pan", + "610": "▁hum", + "611": "▁old", + "612": "ead", + "613": "▁ord", + "614": "ric", + "615": "ants", + "616": "oy", + "617": "esn", + "618": "uck", + "619": "ason", + "620": "ced", + "621": "ool", + "622": "rat", + "623": "ouse", + "624": "▁lar", + "625": "▁art", + "626": "▁wee", + "627": "▁cer", + "628": "ized", + "629": "▁mat", + "630": "con", + "631": "erg", + "632": "land", + "633": "ines", + "634": "▁chr", + "635": "▁aut", + "636": "▁lea", + "637": "▁sou", + "638": "oney", + "639": "tty", + "640": "▁ple", + "641": "ulat", + "642": "oks", + "643": "▁few", + "644": "▁sol", + "645": "▁che", + "646": "chn", + "647": "ird", + "648": "▁bre", + "649": "▁dur", + "650": "▁wom", + "651": "me", + "652": "izat", + "653": "eric", + "654": "ote", + "655": "▁uni", + "656": "eren", + "657": "arn", + "658": "ross", + "659": "ices", + "660": "ten", + "661": "eral", + "662": "ever", + "663": "ieve", + "664": "lish", + "665": "ash", + "666": "▁opp", + "667": "alth", + "668": "ger", + "669": "▁sk", + "670": "▁red", + "671": "peri", + "672": "▁det", + "673": "▁ext", + "674": "ner", + "675": "ah", + "676": "▁var", + "677": "▁loc", + "678": "gram", + "679": "ists", + "680": "ives", + "681": "▁es", + "682": "▁nor", + "683": "tro", + "684": "ale", + "685": "▁iss", + "686": "▁pri", + "687": "gin", + "688": "az", + "689": "oc", + "690": "▁pop", + "691": "ern", + "692": "▁sit", + "693": "ket", + "694": "▁pa", + "695": "▁law", + "696": "ages", + "697": "br", + "698": "▁cam", + "699": "▁mom", + "700": "osed", + "701": "▁bro", + "702": "ne", + "703": "bs", + "704": "▁cre", + "705": "erat", + "706": "▁sec", + "707": "▁cap", + "708": "▁vis", + "709": "▁pat", + "710": "ield", + "711": "iet", + "712": "▁tri", + "713": "up", + "714": "▁bra", + "715": "ts", + "716": "▁mot", + "717": "▁unt", + "718": "put", + "719": "bo", + "720": "ork", + "721": "mer", + "722": "ital", + "723": "▁air", + "724": "ined", + "725": "▁beh", + "726": "▁adv", + "727": "▁ret", + "728": "imes", + "729": "▁tea", + "730": "ural", + "731": "sid", + "732": "ters", + "733": "▁pur", + "734": "▁sci", + "735": "bers", + "736": "ient", + "737": "ier", + "738": "cc", + "739": "sw", + "740": "▁av", + "741": "reen", + "742": "ode", + "743": "ont", + "744": "▁dra", + "745": "ann", + "746": "nect", + "747": "▁x", + "748": "▁eu", + "749": "ton", + "750": "inat", + "751": "ene", + "752": "ared", + "753": "els", + "754": "▁mor", + "755": "▁rat", + "756": "cri", + "757": "▁men", + "758": "▁ah", + "759": "ames", + "760": "▁arm", + "761": "eak", + "762": "▁pay", + "763": "▁hal", + "764": "ins", + "765": "ilit", + "766": "stit", + "767": "▁ra", + "768": "▁leg", + "769": "cl", + "770": "pr", + "771": "▁wal", + "772": "▁bad", + "773": "▁ge", + "774": "roup", + "775": "▁mus", + "776": "man", + "777": "▁gi", + "778": "eds", + "779": "▁aw", + "780": "po", + "781": "ark", + "782": "row", + "783": "▁dep", + "784": "ully", + "785": "ral", + "786": "lect", + "787": "pend", + "788": "▁sev", + "789": "ime", + "790": "gest", + "791": "here", + "792": "▁yet", + "793": "ted", + "794": "▁rev", + "795": "ds", + "796": "▁ask", + "797": "less", + "798": "▁di", + "799": "ets", + "800": "line", + "801": "▁aff", + "802": "ired", + "803": "▁est", + "804": "ken", + "805": "vid", + "806": "most", + "807": "ivid", + "808": "unch", + "809": "par", + "810": "med", + "811": "rop", + "812": "ased", + "813": "eone", + "814": "▁ve", + "815": "▁abs", + "816": "ergy", + "817": "ret", + "818": "▁saw", + "819": "▁ey", + "820": "▁cal", + "821": "uat", + "822": "▁mid", + "823": "vat", + "824": "ream", + "825": "vice", + "826": "ians", + "827": "rent", + "828": "ctor", + "829": "err", + "830": "ush", + "831": "ases", + "832": "▁suc", + "833": "erms", + "834": "ave", + "835": "angu", + "836": "ries", + "837": "▁wo", + "838": "arts", + "839": "▁fil", + "840": "▁fat", + "841": "▁cho", + "842": "orts", + "843": "▁fre", + "844": "ee", + "845": "ught", + "846": "eng", + "847": "ump", + "848": "▁bar", + "849": "ying", + "850": "ane", + "851": "▁tem", + "852": "anks", + "853": "ury", + "854": "iat", + "855": "mit", + "856": "trol", + "857": "▁net", + "858": "▁maj", + "859": "▁cra", + "860": "ling", + "861": "▁fig", + "862": "orn", + "863": "icat", + "864": "pany", + "865": "▁occ", + "866": "ott", + "867": "ands", + "868": "▁exc", + "869": "▁mr", + "870": "ency", + "871": "rope", + "872": "itch", + "873": "▁lit", + "874": "abil", + "875": "not", + "876": "ma", + "877": "▁typ", + "878": "▁opt", + "879": "ob", + "880": "ser", + "881": "ety", + "882": "ms", + "883": "peci", + "884": "aces", + "885": "aut", + "886": "▁hon", + "887": "cuss", + "888": "▁sal", + "889": "▁sor", + "890": "att", + "891": "▁lab", + "892": "▁har", + "893": "urch", + "894": "nded", + "895": "uce", + "896": "ids", + "897": "▁hy", + "898": "▁fut", + "899": "▁ste", + "900": "ours", + "901": "ems", + "902": "utes", + "903": "ng", + "904": "ta", + "905": "▁won", + "906": "▁fa", + "907": "▁env", + "908": "ards", + "909": "▁job", + "910": "ium", + "911": "▁dot", + "912": "▁obv", + "913": "ina", + "914": "side", + "915": "elve", + "916": "cu", + "917": "▁jes", + "918": "▁pot", + "919": "▁pie", + "920": "▁tre", + "921": "▁hey", + "922": "▁mag", + "923": "ron", + "924": "▁key", + "925": "swer", + "926": "▁win", + "927": "ucat", + "928": "work", + "929": "ides", + "930": "▁low", + "931": "▁vol", + "932": "▁oth", + "933": "atic", + "934": "lf", + "935": "ads", + "936": "inds", + "937": "com", + "938": "ths", + "939": "▁ver", + "940": "ised", + "941": "lo", + "942": "▁squ", + "943": "▁cut", + "944": "oked", + "945": "irit", + "946": "ateg", + "947": "ppy", + "948": "mitt", + "949": "come", + "950": "hn", + "951": "igin", + "952": "mand", + "953": "▁dam", + "954": "ho", + "955": "▁da", + "956": "▁fur", + "957": "iron", + "958": "ilar", + "959": "▁fac", + "960": "▁neg", + "961": "▁ago", + "962": "ged", + "963": "miss", + "964": "enth", + "965": "▁dou", + "966": "▁hit", + "967": "▁guy", + "968": "▁bi", + "969": "ove", + "970": "fess", + "971": "ples", + "972": "owed", + "973": "ured", + "974": "▁ris", + "975": "ints", + "976": "rew", + "977": "▁sum", + "978": "▁hu", + "979": "ploy", + "980": "ude", + "981": "ried", + "982": "▁cir", + "983": "▁dev", + "984": "ear", + "985": "▁tot", + "986": "▁ann", + "987": "duc", + "988": "ik", + "989": "pon", + "990": "sted", + "991": "▁ide", + "992": "▁'", + "993": "ipp", + "994": "▁eat", + "995": "▁dom", + "996": "▁", + "997": "e", + "998": "t", + "999": "o", + "1000": "a", + "1001": "i", + "1002": "n", + "1003": "s", + "1004": "r", + "1005": "h", + "1006": "l", + "1007": "d", + "1008": "u", + "1009": "c", + "1010": "m", + "1011": "y", + "1012": "g", + "1013": "w", + "1014": "f", + "1015": "p", + "1016": "b", + "1017": "v", + "1018": "k", + "1019": "'", + "1020": "j", + "1021": "x", + "1022": "q", + "1023": "z", + "1024": "", + "1025": "" +} \ No newline at end of file diff --git a/320ms/.DS_Store b/320ms/.DS_Store new file mode 100644 index 0000000000000000000000000000000000000000..8e1aae8afb05f089f83507db6d67efd90fb51604 Binary files /dev/null and b/320ms/.DS_Store differ diff --git a/320ms/convert_parakeet_eou.py b/320ms/convert_parakeet_eou.py new file mode 100644 index 0000000000000000000000000000000000000000..9c865f90c8132840b5755dcd62fcfac15e3f0352 --- /dev/null +++ b/320ms/convert_parakeet_eou.py @@ -0,0 +1,740 @@ +#!/usr/bin/env python3 +"""CLI for exporting Parakeet Realtime EOU 120M components to CoreML. + +This model is a cache-aware streaming FastConformer-RNNT model optimized for +low-latency speech recognition with end-of-utterance detection. + +Key differences from Parakeet TDT v3: +- Smaller model (120M vs 600M params) +- No duration outputs (standard RNNT, not TDT) +- Cache-aware streaming encoder (17 layers, attention context [70,1]) +- Special token for end-of-utterance detection +- Optimized for 80-160ms latency + +Reference: https://huggingface.co/nvidia/parakeet_realtime_eou_120m-v1 +""" +from __future__ import annotations + +import json +from dataclasses import asdict +from pathlib import Path +from typing import Dict, Optional, Tuple + +import coremltools as ct +import numpy as np +import soundfile as sf +import torch +import typer + +import nemo.collections.asr as nemo_asr + +from individual_components import ( + DecoderWrapper, + EncoderWrapper, + ExportSettings, + JointWrapper, + JointDecisionWrapper, + JointDecisionSingleStep, + PreprocessorWrapper, + MelEncoderWrapper, + _coreml_convert, +) + +def apply_stft_patch(): + # Monkey patch coremltools.stft to handle extra arguments from newer torch versions + try: + import coremltools.converters.mil.frontend.torch.ops as torch_ops + _original_stft = torch_ops.stft + + def patched_stft(context, node): + if len(node.inputs) > 8: + node.inputs = node.inputs[:8] + return _original_stft(context, node) + + torch_ops.stft = patched_stft + if "stft" in torch_ops._TORCH_OPS_REGISTRY: + torch_ops._TORCH_OPS_REGISTRY["stft"] = patched_stft + print("Monkey patched coremltools.stft for compatibility.") + except Exception as e: + print(f"Warning: Could not monkey patch stft: {e}") + +DEFAULT_MODEL_ID = "nvidia/parakeet_realtime_eou_120m-v1" +AUTHOR = "Fluid Inference" + + +def _compute_length(seconds: float, sample_rate: int) -> int: + return int(round(seconds * sample_rate)) + + +def _prepare_audio( + validation_audio: Optional[Path], + sample_rate: int, + max_samples: int, + seed: Optional[int], +) -> torch.Tensor: + if validation_audio is None: + if seed is not None: + torch.manual_seed(seed) + audio = torch.randn(1, max_samples, dtype=torch.float32) + return audio + + data, sr = sf.read(str(validation_audio), dtype="float32") + if sr != sample_rate: + raise typer.BadParameter( + f"Validation audio sample rate {sr} does not match model rate {sample_rate}" + ) + + if data.ndim > 1: + data = data[:, 0] + + if data.size == 0: + raise typer.BadParameter("Validation audio is empty") + + if data.size < max_samples: + pad_width = max_samples - data.size + data = np.pad(data, (0, pad_width)) + elif data.size > max_samples: + data = data[:max_samples] + + audio = torch.from_numpy(data).unsqueeze(0).to(dtype=torch.float32) + return audio + + +def _save_mlpackage(model: ct.models.MLModel, path: Path, description: str) -> None: + try: + model.minimum_deployment_target = ct.target.iOS17 + except Exception: + pass + model.short_description = description + model.author = AUTHOR + path.parent.mkdir(parents=True, exist_ok=True) + model.save(str(path)) + + +def _tensor_shape(tensor: torch.Tensor) -> Tuple[int, ...]: + return tuple(int(dim) for dim in tensor.shape) + + +def _parse_compute_units(name: str) -> ct.ComputeUnit: + """Parse a human-friendly compute units string into ct.ComputeUnit.""" + normalized = str(name).strip().upper() + mapping = { + "ALL": ct.ComputeUnit.ALL, + "CPU_ONLY": ct.ComputeUnit.CPU_ONLY, + "CPU_AND_GPU": ct.ComputeUnit.CPU_AND_GPU, + "CPU_AND_NE": ct.ComputeUnit.CPU_AND_NE, + "CPU_AND_NEURALENGINE": ct.ComputeUnit.CPU_AND_NE, + } + if normalized not in mapping: + raise typer.BadParameter( + f"Unknown compute units '{name}'. Choose from: " + ", ".join(mapping.keys()) + ) + return mapping[normalized] + + +def _parse_compute_precision(name: Optional[str]) -> Optional[ct.precision]: + """Parse compute precision string into ct.precision or None.""" + if name is None: + return None + normalized = str(name).strip().upper() + if normalized == "": + return None + mapping = { + "FLOAT32": ct.precision.FLOAT32, + "FLOAT16": ct.precision.FLOAT16, + } + if normalized not in mapping: + raise typer.BadParameter( + f"Unknown compute precision '{name}'. Choose from: " + + ", ".join(mapping.keys()) + ) + return mapping[normalized] + + +app = typer.Typer(add_completion=False, pretty_exceptions_show_locals=False) + + +@app.command() +def convert( + nemo_path: Optional[Path] = typer.Option( + None, + "--nemo-path", + exists=True, + resolve_path=True, + help="Path to parakeet_realtime_eou_120m-v1.nemo checkpoint (skip to auto-download)", + ), + model_id: str = typer.Option( + DEFAULT_MODEL_ID, + "--model-id", + help="Model identifier to download when --nemo-path is omitted", + ), + output_dir: Path = typer.Option( + Path("parakeet_eou_coreml"), + help="Directory where mlpackages and metadata will be written", + ), + preprocessor_cu: str = typer.Option( + "CPU_ONLY", + "--preprocessor-cu", + help="Compute units for preprocessor (default CPU_ONLY)", + ), + mel_encoder_cu: str = typer.Option( + "CPU_ONLY", + "--mel-encoder-cu", + help="Compute units for fused mel+encoder (default CPU_ONLY)", + ), + compute_precision: Optional[str] = typer.Option( + None, + "--compute-precision", + help="Export precision: FLOAT32 (default) or FLOAT16 to shrink non-quantized weights.", + ), + max_audio_seconds: float = typer.Option( + 15.0, + "--max-audio-seconds", + help="Maximum audio duration in seconds for the fixed window export", + ), + validation_audio: Optional[Path] = typer.Option( + None, + "--validation-audio", + exists=True, + resolve_path=True, + help="Path to a 16kHz WAV file for tracing (uses random if not provided)", + ), +) -> None: + """Export all Parakeet Realtime EOU sub-modules to CoreML. + + This exports the cache-aware streaming FastConformer-RNNT model for + low-latency speech recognition with end-of-utterance detection. + """ + export_settings = ExportSettings( + output_dir=output_dir, + compute_units=ct.ComputeUnit.CPU_ONLY, + deployment_target=ct.target.iOS17, + compute_precision=_parse_compute_precision(compute_precision), + max_audio_seconds=max_audio_seconds, + max_symbol_steps=1, + ) + + typer.echo("Export configuration:") + typer.echo(asdict(export_settings)) + + output_dir.mkdir(parents=True, exist_ok=True) + pre_cu = _parse_compute_units(preprocessor_cu) + melenc_cu = _parse_compute_units(mel_encoder_cu) + + if nemo_path is not None: + typer.echo(f"Loading NeMo model from {nemo_path}…") + # Try loading as generic ASRModel first, then specific class + try: + asr_model = nemo_asr.models.ASRModel.restore_from( + str(nemo_path), map_location="cpu" + ) + except Exception: + # Fallback to EncDecRNNTBPEModel + asr_model = nemo_asr.models.EncDecRNNTBPEModel.restore_from( + str(nemo_path), map_location="cpu" + ) + checkpoint_meta = { + "type": "file", + "path": str(nemo_path), + } + else: + typer.echo(f"Downloading NeMo model via {model_id}…") + # Use ASRModel.from_pretrained as recommended for this model + try: + asr_model = nemo_asr.models.ASRModel.from_pretrained( + model_id, map_location="cpu" + ) + except Exception: + # Fallback to EncDecRNNTBPEModel + asr_model = nemo_asr.models.EncDecRNNTBPEModel.from_pretrained( + model_id, map_location="cpu" + ) + checkpoint_meta = { + "type": "pretrained", + "model_id": model_id, + } + asr_model.eval() + + # Print model info + typer.echo(f"Model class: {type(asr_model).__name__}") + typer.echo(f"Encoder class: {type(asr_model.encoder).__name__}") + + sample_rate = int(asr_model.cfg.preprocessor.sample_rate) + max_samples = _compute_length(export_settings.max_audio_seconds, sample_rate) + + # Prepare audio for tracing + if validation_audio is not None: + typer.echo(f"Using validation audio: {validation_audio}") + audio_tensor = _prepare_audio(validation_audio, sample_rate, max_samples, seed=None) + else: + typer.echo("Using random audio for tracing (seed=42)") + audio_tensor = _prepare_audio(None, sample_rate, max_samples, seed=42) + + audio_length = torch.tensor([max_samples], dtype=torch.int32) + + preprocessor = PreprocessorWrapper(asr_model.preprocessor.eval()) + encoder = EncoderWrapper(asr_model.encoder.eval()) + decoder = DecoderWrapper(asr_model.decoder.eval()) + joint = JointWrapper(asr_model.joint.eval()) + + decoder_export_flag = getattr(asr_model.decoder, "_rnnt_export", False) + asr_model.decoder._rnnt_export = True + + try: + with torch.no_grad(): + mel_ref, mel_length_ref = preprocessor(audio_tensor, audio_length) + mel_length_ref = mel_length_ref.to(dtype=torch.int32) + encoder_ref, encoder_length_ref, frame_times_ref = encoder( + mel_ref, mel_length_ref + ) + encoder_length_ref = encoder_length_ref.to(dtype=torch.int32) + + # Clone tensors to drop inference flags + mel_ref = mel_ref.clone().detach() + mel_length_ref = mel_length_ref.clone().detach() + encoder_ref = encoder_ref.clone().detach() + encoder_length_ref = encoder_length_ref.clone().detach() + frame_times_ref = frame_times_ref.clone().detach() + + vocab_size = int(asr_model.tokenizer.vocab_size) + decoder_hidden = int(asr_model.decoder.pred_hidden) + decoder_layers = int(asr_model.decoder.pred_rnn_layers) + + # Check if model has extra outputs (TDT-style duration) + num_extra = getattr(asr_model.joint, "num_extra_outputs", 0) + typer.echo(f"Vocab size: {vocab_size}, num_extra_outputs: {num_extra}") + + targets = torch.full( + (1, export_settings.max_symbol_steps), + fill_value=asr_model.decoder.blank_idx, + dtype=torch.int32, + ) + target_lengths = torch.tensor( + [export_settings.max_symbol_steps], dtype=torch.int32 + ) + zero_state = torch.zeros( + decoder_layers, + 1, + decoder_hidden, + dtype=torch.float32, + ) + + with torch.no_grad(): + decoder_ref, h_ref, c_ref = decoder( + targets, target_lengths, zero_state, zero_state + ) + joint_ref = joint(encoder_ref, decoder_ref) + + decoder_ref = decoder_ref.clone() + h_ref = h_ref.clone() + c_ref = c_ref.clone() + joint_ref = joint_ref.clone() + + typer.echo(f"Encoder output shape: {encoder_ref.shape}") + typer.echo(f"Decoder output shape: {decoder_ref.shape}") + typer.echo(f"Joint output shape: {joint_ref.shape}") + + # === Export Preprocessor === + typer.echo("Tracing and converting preprocessor…") + preprocessor = preprocessor.cpu() + audio_tensor = audio_tensor.cpu() + audio_length = audio_length.cpu() + traced_preprocessor = torch.jit.trace( + preprocessor, (audio_tensor, audio_length), strict=False + ) + traced_preprocessor.eval() + preprocessor_inputs = [ + ct.TensorType( + name="audio_signal", + shape=(1, ct.RangeDim(1, max_samples)), + dtype=np.float32, + ), + ct.TensorType(name="audio_length", shape=(1,), dtype=np.int32), + ] + preprocessor_outputs = [ + ct.TensorType(name="mel", dtype=np.float32), + ct.TensorType(name="mel_length", dtype=np.int32), + ] + preprocessor_model = _coreml_convert( + traced_preprocessor, + preprocessor_inputs, + preprocessor_outputs, + export_settings, + compute_units_override=pre_cu, + ) + preprocessor_path = output_dir / "parakeet_eou_preprocessor.mlpackage" + _save_mlpackage( + preprocessor_model, + preprocessor_path, + f"Parakeet EOU preprocessor ({max_audio_seconds}s window)", + ) + + # === Export Encoder === + typer.echo("Tracing and converting encoder…") + traced_encoder = torch.jit.trace( + encoder, (mel_ref, mel_length_ref), strict=False + ) + traced_encoder.eval() + encoder_inputs = [ + ct.TensorType( + name="mel", shape=_tensor_shape(mel_ref), dtype=np.float32 + ), + ct.TensorType(name="mel_length", shape=(1,), dtype=np.int32), + ] + encoder_outputs = [ + ct.TensorType(name="encoder", dtype=np.float32), + ct.TensorType(name="encoder_length", dtype=np.int32), + ct.TensorType(name="frame_times", dtype=np.float32), + ] + encoder_model = _coreml_convert( + traced_encoder, + encoder_inputs, + encoder_outputs, + export_settings, + compute_units_override=ct.ComputeUnit.CPU_ONLY, + ) + encoder_path = output_dir / "parakeet_eou_encoder.mlpackage" + _save_mlpackage( + encoder_model, + encoder_path, + f"Parakeet EOU encoder ({max_audio_seconds}s window)", + ) + + # === Export Fused Mel+Encoder === + typer.echo("Tracing and converting fused mel+encoder…") + mel_encoder = MelEncoderWrapper(preprocessor, encoder) + traced_mel_encoder = torch.jit.trace( + mel_encoder, (audio_tensor, audio_length), strict=False + ) + traced_mel_encoder.eval() + mel_encoder_inputs = [ + ct.TensorType( + name="audio_signal", shape=(1, max_samples), dtype=np.float32 + ), + ct.TensorType(name="audio_length", shape=(1,), dtype=np.int32), + ] + mel_encoder_outputs = [ + ct.TensorType(name="encoder", dtype=np.float32), + ct.TensorType(name="encoder_length", dtype=np.int32), + ct.TensorType(name="frame_times", dtype=np.float32), + ] + mel_encoder_model = _coreml_convert( + traced_mel_encoder, + mel_encoder_inputs, + mel_encoder_outputs, + export_settings, + compute_units_override=melenc_cu, + ) + mel_encoder_path = output_dir / "parakeet_eou_mel_encoder.mlpackage" + _save_mlpackage( + mel_encoder_model, + mel_encoder_path, + f"Parakeet EOU fused Mel+Encoder ({max_audio_seconds}s window)", + ) + + # === Export Decoder === + typer.echo("Tracing and converting decoder…") + traced_decoder = torch.jit.trace( + decoder, + (targets, target_lengths, zero_state, zero_state), + strict=False, + ) + traced_decoder.eval() + decoder_inputs = [ + ct.TensorType( + name="targets", shape=_tensor_shape(targets), dtype=np.int32 + ), + ct.TensorType(name="target_length", shape=(1,), dtype=np.int32), + ct.TensorType( + name="h_in", shape=_tensor_shape(zero_state), dtype=np.float32 + ), + ct.TensorType( + name="c_in", shape=_tensor_shape(zero_state), dtype=np.float32 + ), + ] + decoder_outputs = [ + ct.TensorType(name="decoder", dtype=np.float32), + ct.TensorType(name="h_out", dtype=np.float32), + ct.TensorType(name="c_out", dtype=np.float32), + ] + decoder_model = _coreml_convert( + traced_decoder, + decoder_inputs, + decoder_outputs, + export_settings, + compute_units_override=ct.ComputeUnit.CPU_ONLY, + ) + decoder_path = output_dir / "parakeet_eou_decoder.mlpackage" + _save_mlpackage( + decoder_model, + decoder_path, + "Parakeet EOU decoder (RNNT prediction network)", + ) + + # === Export Joint === + typer.echo("Tracing and converting joint…") + traced_joint = torch.jit.trace( + joint, + (encoder_ref, decoder_ref), + strict=False, + ) + traced_joint.eval() + joint_inputs = [ + ct.TensorType( + name="encoder", shape=_tensor_shape(encoder_ref), dtype=np.float32 + ), + ct.TensorType( + name="decoder", shape=_tensor_shape(decoder_ref), dtype=np.float32 + ), + ] + joint_outputs = [ + ct.TensorType(name="logits", dtype=np.float32), + ] + joint_model = _coreml_convert( + traced_joint, + joint_inputs, + joint_outputs, + export_settings, + compute_units_override=ct.ComputeUnit.CPU_ONLY, + ) + joint_path = output_dir / "parakeet_eou_joint.mlpackage" + _save_mlpackage( + joint_model, + joint_path, + "Parakeet EOU joint network (RNNT)", + ) + + # === Export Joint Decision Head === + typer.echo("Tracing and converting joint decision head…") + joint_decision = JointDecisionWrapper(joint, vocab_size=vocab_size) + traced_joint_decision = torch.jit.trace( + joint_decision, + (encoder_ref, decoder_ref), + strict=False, + ) + traced_joint_decision.eval() + joint_decision_inputs = [ + ct.TensorType( + name="encoder", shape=_tensor_shape(encoder_ref), dtype=np.float32 + ), + ct.TensorType( + name="decoder", shape=_tensor_shape(decoder_ref), dtype=np.float32 + ), + ] + joint_decision_outputs = [ + ct.TensorType(name="token_id", dtype=np.int32), + ct.TensorType(name="token_prob", dtype=np.float32), + ] + joint_decision_model = _coreml_convert( + traced_joint_decision, + joint_decision_inputs, + joint_decision_outputs, + export_settings, + compute_units_override=ct.ComputeUnit.CPU_ONLY, + ) + joint_decision_path = output_dir / "parakeet_eou_joint_decision.mlpackage" + _save_mlpackage( + joint_decision_model, + joint_decision_path, + "Parakeet EOU joint + decision head (softmax, argmax)", + ) + + # === Export Single-Step Joint Decision === + typer.echo("Tracing and converting single-step joint decision…") + jd_single = JointDecisionSingleStep(joint, vocab_size=vocab_size) + # Create single-step slices from refs + enc_step = encoder_ref[:, :, :1].contiguous() + dec_step = decoder_ref[:, :, :1].contiguous() + traced_jd_single = torch.jit.trace( + jd_single, + (enc_step, dec_step), + strict=False, + ) + traced_jd_single.eval() + jd_single_inputs = [ + ct.TensorType( + name="encoder_step", + shape=(1, enc_step.shape[1], 1), + dtype=np.float32, + ), + ct.TensorType( + name="decoder_step", + shape=(1, dec_step.shape[1], 1), + dtype=np.float32, + ), + ] + jd_single_outputs = [ + ct.TensorType(name="token_id", dtype=np.int32), + ct.TensorType(name="token_prob", dtype=np.float32), + ct.TensorType(name="top_k_ids", dtype=np.int32), + ct.TensorType(name="top_k_logits", dtype=np.float32), + ] + jd_single_model = _coreml_convert( + traced_jd_single, + jd_single_inputs, + jd_single_outputs, + export_settings, + compute_units_override=ct.ComputeUnit.CPU_ONLY, + ) + jd_single_path = output_dir / "parakeet_eou_joint_decision_single_step.mlpackage" + _save_mlpackage( + jd_single_model, + jd_single_path, + "Parakeet EOU single-step joint decision (current frame)", + ) + + # === Save Metadata === + metadata: Dict[str, object] = { + "model_id": model_id, + "model_name": "parakeet_realtime_eou_120m-v1", + "model_class": type(asr_model).__name__, + "encoder_class": type(asr_model.encoder).__name__, + "sample_rate": sample_rate, + "max_audio_seconds": export_settings.max_audio_seconds, + "max_audio_samples": max_samples, + "max_symbol_steps": export_settings.max_symbol_steps, + "vocab_size": vocab_size, + "vocab_with_blank": vocab_size + 1, + "decoder_hidden": decoder_hidden, + "decoder_layers": decoder_layers, + "num_extra_outputs": num_extra, + "has_eou_token": True, + "checkpoint": checkpoint_meta, + "coreml": { + "compute_units": export_settings.compute_units.name, + "compute_precision": ( + export_settings.compute_precision.name + if export_settings.compute_precision is not None + else "FLOAT32" + ), + }, + "components": { + "preprocessor": { + "inputs": { + "audio_signal": [1, max_samples], + "audio_length": [1], + }, + "outputs": { + "mel": list(_tensor_shape(mel_ref)), + "mel_length": [1], + }, + "path": preprocessor_path.name, + }, + "encoder": { + "inputs": { + "mel": list(_tensor_shape(mel_ref)), + "mel_length": [1], + }, + "outputs": { + "encoder": list(_tensor_shape(encoder_ref)), + "encoder_length": [1], + "frame_times": [1, _tensor_shape(encoder_ref)[2]], + }, + "path": encoder_path.name, + }, + "mel_encoder": { + "inputs": { + "audio_signal": [1, max_samples], + "audio_length": [1], + }, + "outputs": { + "encoder": list(_tensor_shape(encoder_ref)), + "encoder_length": [1], + "frame_times": [1, _tensor_shape(encoder_ref)[2]], + }, + "path": mel_encoder_path.name, + }, + "decoder": { + "inputs": { + "targets": list(_tensor_shape(targets)), + "target_length": [1], + "h_in": list(_tensor_shape(zero_state)), + "c_in": list(_tensor_shape(zero_state)), + }, + "outputs": { + "decoder": list(_tensor_shape(decoder_ref)), + "h_out": list(_tensor_shape(h_ref)), + "c_out": list(_tensor_shape(c_ref)), + }, + "path": decoder_path.name, + }, + "joint": { + "inputs": { + "encoder": list(_tensor_shape(encoder_ref)), + "decoder": list(_tensor_shape(decoder_ref)), + }, + "outputs": { + "logits": list(_tensor_shape(joint_ref)), + }, + "path": joint_path.name, + }, + "joint_decision": { + "inputs": { + "encoder": list(_tensor_shape(encoder_ref)), + "decoder": list(_tensor_shape(decoder_ref)), + }, + "outputs": { + "token_id": [ + _tensor_shape(encoder_ref)[0], + _tensor_shape(encoder_ref)[2], + _tensor_shape(decoder_ref)[2], + ], + "token_prob": [ + _tensor_shape(encoder_ref)[0], + _tensor_shape(encoder_ref)[2], + _tensor_shape(decoder_ref)[2], + ], + }, + "path": joint_decision_path.name, + }, + "joint_decision_single_step": { + "inputs": { + "encoder_step": [1, _tensor_shape(encoder_ref)[1], 1], + "decoder_step": [1, _tensor_shape(decoder_ref)[1], 1], + }, + "outputs": { + "token_id": [1, 1, 1], + "token_prob": [1, 1, 1], + "top_k_ids": [1, 1, 1, 64], + "top_k_logits": [1, 1, 1, 64], + }, + "path": jd_single_path.name, + }, + }, + } + + # Export tokenizer vocab if available + try: + tokenizer = asr_model.tokenizer + vocab = { + "blank_id": int(asr_model.decoder.blank_idx), + "vocab_size": vocab_size, + } + # Try to get special tokens + if hasattr(tokenizer, "tokenizer"): + inner_tokenizer = tokenizer.tokenizer + if hasattr(inner_tokenizer, "get_vocab"): + full_vocab = inner_tokenizer.get_vocab() + # Find EOU token + eou_token = None + for token, idx in full_vocab.items(): + if "" in token.upper() or "eou" in token.lower(): + eou_token = {"token": token, "id": idx} + break + if eou_token: + vocab["eou_token"] = eou_token + metadata["tokenizer"] = vocab + except Exception as e: + typer.echo(f"Warning: Could not export tokenizer info: {e}") + + metadata_path = output_dir / "metadata.json" + metadata_path.write_text(json.dumps(metadata, indent=2)) + typer.echo(f"\nExport complete. Metadata written to {metadata_path}") + typer.echo(f"Output directory: {output_dir}") + + finally: + asr_model.decoder._rnnt_export = decoder_export_flag + + +if __name__ == "__main__": + app() diff --git a/320ms/convert_streaming_encoder.py b/320ms/convert_streaming_encoder.py new file mode 100644 index 0000000000000000000000000000000000000000..a1c199c8a6c540dab48b973ab4232db2901b8e72 --- /dev/null +++ b/320ms/convert_streaming_encoder.py @@ -0,0 +1,193 @@ + +import torch +import torch.nn as nn +import coremltools as ct +import numpy as np +import typer +from pathlib import Path +from typing import Tuple, List, Optional +import json +import shutil + +# Iimport torch +import coremltools as ct +import numpy as np +import argparse +from nemo.collections.asr.models import EncDecRNNTBPEModel + +app = typer.Typer() + +class LoopbackEncoderWrapper(nn.Module): + """ + Wraps the entire Parakeet Encoder (PreEncode + Conformer) for CoreML Loopback Streaming. + + Inputs: + - audio_signal: [B, D, T] (Mel spectrogram chunk) + - audio_length: [B] + - pre_cache: [B, D, pre_cache_size] (Previous audio context) + - cache_last_channel: [layers, B, cache_size, hidden] + - cache_last_time: [layers, B, hidden, time_cache] + - cache_last_channel_len: [B] + + Outputs: + - encoded_output: [B, D_out, T_out] + - encoded_length: [B] + - new_pre_cache: [B, D, pre_cache_size] + - new_cache_last_channel + - new_cache_last_time + - new_cache_last_channel_len + """ + def __init__(self, encoder, pre_cache_size=16): + super().__init__() + self.encoder = encoder + self.pre_cache_size = pre_cache_size + + def forward( + self, + audio_signal: torch.Tensor, + audio_length: torch.Tensor, + pre_cache: torch.Tensor, + cache_last_channel: torch.Tensor, + cache_last_time: torch.Tensor, + cache_last_channel_len: torch.Tensor + ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor]: + + # 1. Prepend pre_cache to audio_signal + # audio_signal: [B, D, T] + # pre_cache: [B, D, T_cache] + full_input = torch.cat([pre_cache, audio_signal], dim=2) + full_length = audio_length + self.pre_cache_size + + # 2. Extract NEW pre_cache (last N frames of full_input) + # Note: We do this BEFORE processing because we want the raw audio context + new_pre_cache = full_input[:, :, -self.pre_cache_size:] + + # 3. Process with Encoder + # Reconstruct NeMo cache object + current_cache = [cache_last_channel, cache_last_time, cache_last_channel_len] + + encoded, encoded_len, new_cache_channel, new_cache_time, new_cache_len = self.encoder.cache_aware_stream_step( + processed_signal=full_input, + processed_signal_length=full_length, + cache_last_channel=cache_last_channel, + cache_last_time=cache_last_time, + cache_last_channel_len=cache_last_channel_len + ) + + # 4. Drop the first few frames corresponding to pre_cache? + # NeMo's cache_aware_stream_step usually handles the "valid" output frames. + # But since we manually prepended, we might get extra output frames. + # However, for streaming, we usually want the model to see the context but only output the new tokens. + # Let's trust NeMo's streaming logic for now, or check if we need to slice. + # Given we are using 'cache_aware_stream_step', it expects the full context window? + # Actually, standard usage is: input IS the new chunk, but internal convolution looks at past. + # But since we are stateless, we MUST provide the past. + # So passing (pre_cache + chunk) is correct. + + # Cast lengths to Int32 for CoreML + encoded_len_32 = encoded_len.to(dtype=torch.int32) + new_channel_len_32 = new_cache_len.to(dtype=torch.int32) + + return encoded, encoded_len_32, new_pre_cache, new_cache_channel, new_cache_time, new_channel_len_32 + +def _coreml_convert( + traced_model, + inputs, + outputs, + compute_units=ct.ComputeUnit.CPU_ONLY +): + return ct.convert( + traced_model, + inputs=inputs, + outputs=outputs, + compute_units=compute_units, + minimum_deployment_target=ct.target.macOS14, + ) + +def main(): + model_id: str = "nvidia/parakeet_realtime_eou_120m-v1" + output_dir: str = "temp_swift_models/StreamingLoopback" + output_path = Path(output_dir) + output_path.mkdir(parents=True, exist_ok=True) + + print(f"Loading model: {model_id}...") + asr_model = EncDecRNNTBPEModel.from_pretrained(model_name=model_id) + asr_model.eval() + + parser = argparse.ArgumentParser() + parser.add_argument("--chunk-frames", type=int, default=17, help="Number of frames in the input chunk (e.g. 17 for 160ms, 129 for 1.28s)") + args = parser.parse_args() + + encoder = asr_model.encoder + + # --- Configuration --- + # 160ms chunk = 16 frames (but preprocessor produces 17 with padding/centering) + # 1.28s chunk = 128 frames (preprocessor produces 129) + chunk_size_in = args.chunk_frames + mel_dim = 128 + hidden_dim = encoder.d_model # 512 + num_layers = len(encoder.layers) # 17 + + # Cache sizes + cache_channel_size = 70 + cache_time_size = 8 + pre_cache_size = 16 + + print(f"Config: Chunk={chunk_size_in}, Mel={mel_dim}, Hidden={hidden_dim}, Layers={num_layers}") + print(f"Cache: Channel={cache_channel_size}, Time={cache_time_size}, Pre={pre_cache_size}") + + # --- Wrapper --- + wrapper = LoopbackEncoderWrapper(encoder, pre_cache_size=pre_cache_size) + wrapper.eval() + + # --- Test Inputs (for Tracing) --- + batch_size = 1 + test_mel = torch.randn(batch_size, mel_dim, chunk_size_in) + test_mel_len = torch.tensor([chunk_size_in], dtype=torch.int32) + test_pre_cache = torch.zeros(batch_size, mel_dim, pre_cache_size) + + # Initial Cache (Zeros) + test_cache_channel = torch.zeros(num_layers, batch_size, cache_channel_size, hidden_dim) + test_cache_time = torch.zeros(num_layers, batch_size, hidden_dim, cache_time_size) + test_cache_len = torch.zeros(batch_size, dtype=torch.int32) + + print("Tracing model...") + traced_model = torch.jit.trace( + wrapper, + (test_mel, test_mel_len, test_pre_cache, test_cache_channel, test_cache_time, test_cache_len), + strict=False + ) + + # --- CoreML Conversion --- + print("Converting to CoreML...") + + inputs = [ + ct.TensorType(name="audio_signal", shape=(1, 128, chunk_size_in), dtype=np.float32), + ct.TensorType(name="audio_length", shape=(1,), dtype=np.int32), + ct.TensorType(name="pre_cache", shape=(1, 128, pre_cache_size), dtype=np.float32), + ct.TensorType(name="cache_last_channel", shape=(num_layers, 1, cache_channel_size, hidden_dim), dtype=np.float32), + ct.TensorType(name="cache_last_time", shape=(num_layers, 1, hidden_dim, cache_time_size), dtype=np.float32), + ct.TensorType(name="cache_last_channel_len", shape=(1,), dtype=np.int32), + ] + + outputs = [ + ct.TensorType(name="encoded_output", dtype=np.float32), + ct.TensorType(name="encoded_length", dtype=np.int32), + ct.TensorType(name="new_pre_cache", dtype=np.float32), + ct.TensorType(name="new_cache_last_channel", dtype=np.float32), + ct.TensorType(name="new_cache_last_time", dtype=np.float32), + ct.TensorType(name="new_cache_last_channel_len", dtype=np.int32), + ] + + mlmodel = _coreml_convert(traced_model, inputs, outputs) + + save_path = output_path / "streaming_encoder.mlpackage" + mlmodel.save(str(save_path)) + print(f"Saved: {save_path}") + + # Also export Preprocessor, Decoder, Joint for completeness? + # For now, let's assume we reuse the existing ones or export them separately if needed. + # But the user asked specifically for the Encoder loopback. + +if __name__ == "__main__": + main() diff --git a/320ms/convert_streaming_encoder_unified.py b/320ms/convert_streaming_encoder_unified.py new file mode 100644 index 0000000000000000000000000000000000000000..f457f84634ed52af0aeb3cb4907c121b884849d7 --- /dev/null +++ b/320ms/convert_streaming_encoder_unified.py @@ -0,0 +1,322 @@ +#!/usr/bin/env python3 +""" +Unified CoreML conversion script for Parakeet EOU streaming encoder. + +Supports 160ms, 320ms, and 1600ms chunk sizes by properly configuring NeMo's +streaming parameters before tracing. + +Usage: + # 160ms (default) + python convert_streaming_encoder_unified.py --chunk-ms 160 --output-dir Models/160ms/160ms + + # 320ms + python convert_streaming_encoder_unified.py --chunk-ms 320 --output-dir Models/320ms + + # 1600ms + python convert_streaming_encoder_unified.py --chunk-ms 1600 --output-dir Models/1600ms +""" + +import argparse +import json +from pathlib import Path +from typing import Tuple + +import coremltools as ct +import numpy as np +import torch +import torch.nn as nn +from nemo.collections.asr.models import EncDecRNNTBPEModel + + +class LoopbackEncoderWrapper(nn.Module): + """ + Wraps the Parakeet Encoder for CoreML Loopback Streaming. + + This wrapper handles the pre_cache concatenation and cache management + that NeMo does internally in its streaming pipeline. + + Inputs: + - audio_signal: [B, D, T] (Mel spectrogram chunk) + - audio_length: [B] + - pre_cache: [B, D, pre_cache_size] (Previous mel context) + - cache_last_channel: [layers, B, cache_size, hidden] + - cache_last_time: [layers, B, hidden, time_cache] + - cache_last_channel_len: [B] + + Outputs: + - encoded_output: [B, D_out, T_out] + - encoded_length: [B] + - new_pre_cache: [B, D, pre_cache_size] + - new_cache_last_channel + - new_cache_last_time + - new_cache_last_channel_len + """ + + def __init__(self, encoder, pre_cache_size: int): + super().__init__() + self.encoder = encoder + self.pre_cache_size = pre_cache_size + + def forward( + self, + audio_signal: torch.Tensor, + audio_length: torch.Tensor, + pre_cache: torch.Tensor, + cache_last_channel: torch.Tensor, + cache_last_time: torch.Tensor, + cache_last_channel_len: torch.Tensor, + ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor]: + + # 1. Prepend pre_cache to audio_signal + full_input = torch.cat([pre_cache, audio_signal], dim=2) + full_length = audio_length + self.pre_cache_size + + # 2. Extract NEW pre_cache (last N frames of full_input) + new_pre_cache = full_input[:, :, -self.pre_cache_size :] + + # 3. Process with Encoder using cache_aware_stream_step + encoded, encoded_len, new_cache_channel, new_cache_time, new_cache_len = ( + self.encoder.cache_aware_stream_step( + processed_signal=full_input, + processed_signal_length=full_length, + cache_last_channel=cache_last_channel, + cache_last_time=cache_last_time, + cache_last_channel_len=cache_last_channel_len, + ) + ) + + # Cast lengths to Int32 for CoreML + encoded_len_32 = encoded_len.to(dtype=torch.int32) + new_channel_len_32 = new_cache_len.to(dtype=torch.int32) + + return ( + encoded, + encoded_len_32, + new_pre_cache, + new_cache_channel, + new_cache_time, + new_channel_len_32, + ) + + +def get_streaming_config(encoder, chunk_ms: int): + """ + Get the correct streaming configuration for the given chunk size. + + Returns: + dict with: + - chunk_size: encoder output steps + - shift_size: shift steps + - mel_frames: input mel frames for the chunk + - pre_cache_size: pre-encode cache size in mel frames + - valid_out_len: number of valid output frames per chunk + """ + if chunk_ms == 160: + # Default 160ms config - no need to call setup_streaming_params + # Uses: chunk_size=[9, 16], valid_out_len=2 + # With pre_cache=16, input is 17 mel frames + return { + "chunk_size": 4, # encoder output steps + "shift_size": 2, + "mel_frames": 17, # 16 + 1 for padding + "pre_cache_size": 16, + "valid_out_len": 2, + "samples": 2560, # 160ms * 16000 + } + elif chunk_ms == 320: + # 320ms config - need to call setup_streaming_params(chunk_size=8, shift_size=4) + # After setup: chunk_size=[57, 64], shift_size=[25, 32], pre_encode_cache_size=[0, 9] + # valid_out_len=4 + encoder.setup_streaming_params(chunk_size=8, shift_size=4) + cfg = encoder.streaming_cfg + print(f"320ms streaming_cfg: {cfg}") + + # Use NeMo's exact values from streaming_cfg: + # - chunk_size[1] = 64 mel frames + # - shift_size[1] = 32 mel frames (320ms latency) + # - pre_encode_cache_size[1] = 9 + return { + "chunk_size": 8, # encoder output steps + "shift_size": 4, + "mel_frames": 64, # From NeMo cfg.chunk_size[1] = 64 mel frames + "pre_cache_size": 9, # From NeMo cfg.pre_encode_cache_size[1] = 9 + "valid_out_len": 4, + "samples": 10240, # 64 mel frames * 160 hop_length = 10240 samples (~640ms audio) + "shift_samples": 5120, # 32 mel frames * 160 = 5120 samples (320ms latency) + } + elif chunk_ms == 1600: + # 1600ms config - need to call setup_streaming_params(chunk_size=40, shift_size=20) + # After setup: chunk_size=[313, 320] mel frames, shift_size=[153, 160] mel frames + # The "1600ms" refers to the latency (shift), not the chunk size! + # - Chunk size: 320 mel frames (3183ms of audio) + # - Shift: 160 mel frames (1600ms latency) + # - Output: valid_out_len=20 encoder frames + encoder.setup_streaming_params(chunk_size=40, shift_size=20) + cfg = encoder.streaming_cfg + print(f"1600ms streaming_cfg: {cfg}") + + # Audio samples for 320 mel frames: (320-1)*160 + 400 - 512 = 50928 samples (~3183ms) + # Audio shift for 160 mel frames: 160*160 = 25600 samples (1600ms) + return { + "chunk_size": 40, # encoder output steps + "shift_size": 20, + "mel_frames": 320, # From NeMo cfg.chunk_size[1] = 320 mel frames + "pre_cache_size": 9, # From NeMo cfg.pre_encode_cache_size[1] = 9 + "valid_out_len": 20, + "samples": 50928, # (320-1)*160 + 400 - 512 = 50928 samples (~3183ms) + "shift_samples": 25600, # 160*160 = 25600 samples (1600ms) + } + else: + raise ValueError(f"Unsupported chunk size: {chunk_ms}ms. Use 160, 320, or 1600.") + + +def main(): + parser = argparse.ArgumentParser(description="Convert Parakeet EOU encoder to CoreML") + parser.add_argument( + "--chunk-ms", + type=int, + default=160, + choices=[160, 320, 1600], + help="Chunk size in milliseconds (160, 320, or 1600)", + ) + parser.add_argument( + "--output-dir", + type=str, + default=None, + help="Output directory for the CoreML model", + ) + parser.add_argument( + "--model-id", + type=str, + default="nvidia/parakeet_realtime_eou_120m-v1", + help="HuggingFace model ID", + ) + args = parser.parse_args() + + # Default output directory + if args.output_dir is None: + args.output_dir = f"Models/{args.chunk_ms}ms" + + output_path = Path(args.output_dir) + output_path.mkdir(parents=True, exist_ok=True) + + print(f"Loading model: {args.model_id}...") + asr_model = EncDecRNNTBPEModel.from_pretrained(model_name=args.model_id) + asr_model.eval() + + encoder = asr_model.encoder + + # Get streaming configuration + config = get_streaming_config(encoder, args.chunk_ms) + + mel_dim = 128 + hidden_dim = encoder.d_model # 512 + num_layers = len(encoder.layers) # 17 + + # Cache sizes + cache_channel_size = 70 + cache_time_size = 8 + pre_cache_size = config["pre_cache_size"] + chunk_size_in = config["mel_frames"] + + print(f"\n=== Configuration for {args.chunk_ms}ms ===") + print(f"Mel frames: {chunk_size_in}") + print(f"Pre-cache: {pre_cache_size}") + print(f"Valid output len: {config['valid_out_len']}") + print(f"Hidden dim: {hidden_dim}, Layers: {num_layers}") + print(f"Cache: Channel={cache_channel_size}, Time={cache_time_size}") + + # Create wrapper + wrapper = LoopbackEncoderWrapper(encoder, pre_cache_size=pre_cache_size) + wrapper.eval() + + # Test inputs for tracing + batch_size = 1 + test_mel = torch.randn(batch_size, mel_dim, chunk_size_in) + test_mel_len = torch.tensor([chunk_size_in], dtype=torch.int32) + test_pre_cache = torch.zeros(batch_size, mel_dim, pre_cache_size) + test_cache_channel = torch.zeros(num_layers, batch_size, cache_channel_size, hidden_dim) + test_cache_time = torch.zeros(num_layers, batch_size, hidden_dim, cache_time_size) + test_cache_len = torch.zeros(batch_size, dtype=torch.int32) + + print("\nTracing model...") + traced_model = torch.jit.trace( + wrapper, + (test_mel, test_mel_len, test_pre_cache, test_cache_channel, test_cache_time, test_cache_len), + strict=False, + ) + + # Test output shape + with torch.no_grad(): + out = traced_model( + test_mel, test_mel_len, test_pre_cache, test_cache_channel, test_cache_time, test_cache_len + ) + print(f"Encoder output shape: {out[0].shape}") # [B, D, T_out] + + # CoreML conversion + print("\nConverting to CoreML...") + inputs = [ + ct.TensorType(name="audio_signal", shape=(1, mel_dim, chunk_size_in), dtype=np.float32), + ct.TensorType(name="audio_length", shape=(1,), dtype=np.int32), + ct.TensorType(name="pre_cache", shape=(1, mel_dim, pre_cache_size), dtype=np.float32), + ct.TensorType( + name="cache_last_channel", + shape=(num_layers, 1, cache_channel_size, hidden_dim), + dtype=np.float32, + ), + ct.TensorType( + name="cache_last_time", + shape=(num_layers, 1, hidden_dim, cache_time_size), + dtype=np.float32, + ), + ct.TensorType(name="cache_last_channel_len", shape=(1,), dtype=np.int32), + ] + + outputs = [ + ct.TensorType(name="encoded_output", dtype=np.float32), + ct.TensorType(name="encoded_length", dtype=np.int32), + ct.TensorType(name="new_pre_cache", dtype=np.float32), + ct.TensorType(name="new_cache_last_channel", dtype=np.float32), + ct.TensorType(name="new_cache_last_time", dtype=np.float32), + ct.TensorType(name="new_cache_last_channel_len", dtype=np.int32), + ] + + mlmodel = ct.convert( + traced_model, + inputs=inputs, + outputs=outputs, + compute_units=ct.ComputeUnit.CPU_ONLY, + minimum_deployment_target=ct.target.macOS14, + ) + + save_path = output_path / "streaming_encoder.mlpackage" + mlmodel.save(str(save_path)) + print(f"Saved: {save_path}") + + # Save metadata + metadata = { + "model_id": args.model_id, + "chunk_ms": args.chunk_ms, + "mel_frames": chunk_size_in, + "pre_cache_size": pre_cache_size, + "valid_out_len": config["valid_out_len"], + "samples_per_chunk": config["samples"], + "hidden_dim": hidden_dim, + "num_layers": num_layers, + "cache_channel_size": cache_channel_size, + "cache_time_size": cache_time_size, + } + + metadata_path = output_path / "streaming_encoder_metadata.json" + with open(metadata_path, "w") as f: + json.dump(metadata, f, indent=2) + print(f"Saved metadata: {metadata_path}") + + print(f"\n=== Export complete for {args.chunk_ms}ms ===") + print(f"Output: {output_path}") + print("\nNote: Decoder and Joint models are shared between chunk sizes.") + print("Copy decoder.mlmodelc, joint_decision.mlmodelc, and vocab.json from 160ms directory.") + + +if __name__ == "__main__": + main() diff --git a/320ms/decoder.mlmodelc/analytics/coremldata.bin b/320ms/decoder.mlmodelc/analytics/coremldata.bin new file mode 100644 index 0000000000000000000000000000000000000000..c6f51dedd2638184c5a2d71512339ceea086abc0 --- /dev/null +++ b/320ms/decoder.mlmodelc/analytics/coremldata.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3996975a8cbc1949159c55605b3132b39b2484f51acbd55d796d93c70de02b49 +size 243 diff --git a/320ms/decoder.mlmodelc/coremldata.bin b/320ms/decoder.mlmodelc/coremldata.bin new file mode 100644 index 0000000000000000000000000000000000000000..71936e61ffe7697e4af55bae8870d95c1fbb66c5 --- /dev/null +++ b/320ms/decoder.mlmodelc/coremldata.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3ccbff963d8cf07e2be2bd56ea3384a89ea49628922c6bd95ff62e2ae57dc34 +size 497 diff --git a/320ms/decoder.mlmodelc/metadata.json b/320ms/decoder.mlmodelc/metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..3581673bcb062f7e0e1e0efa9f59edc5f218d50e --- /dev/null +++ b/320ms/decoder.mlmodelc/metadata.json @@ -0,0 +1,118 @@ +[ + { + "metadataOutputVersion" : "3.0", + "shortDescription" : "Parakeet EOU decoder (RNNT prediction network)", + "outputSchema" : [ + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 640 × 1)", + "shortDescription" : "", + "shape" : "[1, 640, 1]", + "name" : "decoder", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 1 × 640)", + "shortDescription" : "", + "shape" : "[1, 1, 640]", + "name" : "h_out", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 1 × 640)", + "shortDescription" : "", + "shape" : "[1, 1, 640]", + "name" : "c_out", + "type" : "MultiArray" + } + ], + "storagePrecision" : "Float16", + "modelParameters" : [ + + ], + "author" : "Fluid Inference", + "specificationVersion" : 8, + "mlProgramOperationTypeHistogram" : { + "Ios17.squeeze" : 2, + "Ios17.gather" : 1, + "Ios17.cast" : 6, + "Ios17.lstm" : 1, + "Ios17.transpose" : 2, + "Identity" : 1, + "Ios17.expandDims" : 2 + }, + "computePrecision" : "Mixed (Float16, Float32, Int16, Int32)", + "isUpdatable" : "0", + "stateSchema" : [ + + ], + "availability" : { + "macOS" : "14.0", + "tvOS" : "17.0", + "visionOS" : "1.0", + "watchOS" : "10.0", + "iOS" : "17.0", + "macCatalyst" : "17.0" + }, + "modelType" : { + "name" : "MLModelType_mlProgram" + }, + "inputSchema" : [ + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1 × 1)", + "shortDescription" : "", + "shape" : "[1, 1]", + "name" : "targets", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1)", + "shortDescription" : "", + "shape" : "[1]", + "name" : "target_length", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 1 × 640)", + "shortDescription" : "", + "shape" : "[1, 1, 640]", + "name" : "h_in", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 1 × 640)", + "shortDescription" : "", + "shape" : "[1, 1, 640]", + "name" : "c_in", + "type" : "MultiArray" + } + ], + "userDefinedMetadata" : { + "com.github.apple.coremltools.version" : "8.3.0", + "com.github.apple.coremltools.source" : "torch==2.4.0", + "com.github.apple.coremltools.source_dialect" : "TorchScript" + }, + "generatedClassName" : "parakeet_eou_decoder", + "method" : "predict" + } +] \ No newline at end of file diff --git a/320ms/decoder.mlmodelc/model.mil b/320ms/decoder.mlmodelc/model.mil new file mode 100644 index 0000000000000000000000000000000000000000..5ead417fbdc13f8bf0c05a26039b2828503d4eca --- /dev/null +++ b/320ms/decoder.mlmodelc/model.mil @@ -0,0 +1,45 @@ +program(1.0) +[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.4.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] +{ + func main(tensor c_in, tensor h_in, tensor target_length, tensor targets) { + tensor y_axis_0 = const()[name = tensor("y_axis_0"), val = tensor(0)]; + tensor y_batch_dims_0 = const()[name = tensor("y_batch_dims_0"), val = tensor(0)]; + tensor y_validate_indices_0 = const()[name = tensor("y_validate_indices_0"), val = tensor(false)]; + tensor module_prediction_embed_weight_to_fp16 = const()[name = tensor("module_prediction_embed_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; + tensor targets_to_int16_dtype_0 = const()[name = tensor("targets_to_int16_dtype_0"), val = tensor("int16")]; + tensor targets_to_int16 = cast(dtype = targets_to_int16_dtype_0, x = targets)[name = tensor("cast_8")]; + tensor y_cast_fp16_cast_uint16 = gather(axis = y_axis_0, batch_dims = y_batch_dims_0, indices = targets_to_int16, validate_indices = y_validate_indices_0, x = module_prediction_embed_weight_to_fp16)[name = tensor("y_cast_fp16_cast_uint16")]; + tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([1, 0, 2])]; + tensor input_lstm_h0_squeeze_axes_0 = const()[name = tensor("input_lstm_h0_squeeze_axes_0"), val = tensor([0])]; + tensor h_in_to_fp16_dtype_0 = const()[name = tensor("h_in_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor h_in_to_fp16 = cast(dtype = h_in_to_fp16_dtype_0, x = h_in)[name = tensor("cast_7")]; + tensor input_lstm_h0_squeeze_cast_fp16 = squeeze(axes = input_lstm_h0_squeeze_axes_0, x = h_in_to_fp16)[name = tensor("input_lstm_h0_squeeze_cast_fp16")]; + tensor input_lstm_c0_squeeze_axes_0 = const()[name = tensor("input_lstm_c0_squeeze_axes_0"), val = tensor([0])]; + tensor c_in_to_fp16_dtype_0 = const()[name = tensor("c_in_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor c_in_to_fp16 = cast(dtype = c_in_to_fp16_dtype_0, x = c_in)[name = tensor("cast_6")]; + tensor input_lstm_c0_squeeze_cast_fp16 = squeeze(axes = input_lstm_c0_squeeze_axes_0, x = c_in_to_fp16)[name = tensor("input_lstm_c0_squeeze_cast_fp16")]; + tensor input_direction_0 = const()[name = tensor("input_direction_0"), val = tensor("forward")]; + tensor input_output_sequence_0 = const()[name = tensor("input_output_sequence_0"), val = tensor(true)]; + tensor input_recurrent_activation_0 = const()[name = tensor("input_recurrent_activation_0"), val = tensor("sigmoid")]; + tensor input_cell_activation_0 = const()[name = tensor("input_cell_activation_0"), val = tensor("tanh")]; + tensor input_activation_0 = const()[name = tensor("input_activation_0"), val = tensor("tanh")]; + tensor concat_1_to_fp16 = const()[name = tensor("concat_1_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1314688)))]; + tensor concat_2_to_fp16 = const()[name = tensor("concat_2_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4591552)))]; + tensor concat_0_to_fp16 = const()[name = tensor("concat_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7868416)))]; + tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = y_cast_fp16_cast_uint16)[name = tensor("transpose_2")]; + tensor input_cast_fp16_0, tensor input_cast_fp16_1, tensor input_cast_fp16_2 = lstm(activation = input_activation_0, bias = concat_0_to_fp16, cell_activation = input_cell_activation_0, direction = input_direction_0, initial_c = input_lstm_c0_squeeze_cast_fp16, initial_h = input_lstm_h0_squeeze_cast_fp16, output_sequence = input_output_sequence_0, recurrent_activation = input_recurrent_activation_0, weight_hh = concat_2_to_fp16, weight_ih = concat_1_to_fp16, x = input_3_cast_fp16)[name = tensor("input_cast_fp16")]; + tensor obj_3_axes_0 = const()[name = tensor("obj_3_axes_0"), val = tensor([0])]; + tensor obj_3_cast_fp16 = expand_dims(axes = obj_3_axes_0, x = input_cast_fp16_1)[name = tensor("obj_3_cast_fp16")]; + tensor obj_3_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_3_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor obj_axes_0 = const()[name = tensor("obj_axes_0"), val = tensor([0])]; + tensor obj_cast_fp16 = expand_dims(axes = obj_axes_0, x = input_cast_fp16_2)[name = tensor("obj_cast_fp16")]; + tensor obj_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor transpose_0_perm_0 = const()[name = tensor("transpose_0_perm_0"), val = tensor([1, 2, 0])]; + tensor transpose_0_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("transpose_0_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor transpose_0_cast_fp16 = transpose(perm = transpose_0_perm_0, x = input_cast_fp16_0)[name = tensor("transpose_1")]; + tensor decoder = cast(dtype = transpose_0_cast_fp16_to_fp32_dtype_0, x = transpose_0_cast_fp16)[name = tensor("cast_3")]; + tensor c_out = cast(dtype = obj_cast_fp16_to_fp32_dtype_0, x = obj_cast_fp16)[name = tensor("cast_4")]; + tensor h_out = cast(dtype = obj_3_cast_fp16_to_fp32_dtype_0, x = obj_3_cast_fp16)[name = tensor("cast_5")]; + tensor target_length_tmp = identity(x = target_length)[name = tensor("target_length_tmp")]; + } -> (decoder, h_out, c_out); +} \ No newline at end of file diff --git a/320ms/decoder.mlmodelc/weights/weight.bin b/320ms/decoder.mlmodelc/weights/weight.bin new file mode 100644 index 0000000000000000000000000000000000000000..cde1618cbed235b421984a302289e3bdd7e3df02 --- /dev/null +++ b/320ms/decoder.mlmodelc/weights/weight.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b4cacecdcd9df79ab1e56de67230baf5a8664d2afe0bb8f3408eefa972cb2f4 +size 7873600 diff --git a/320ms/decoder.mlpackage/Data/com.apple.CoreML/model.mlmodel b/320ms/decoder.mlpackage/Data/com.apple.CoreML/model.mlmodel new file mode 100644 index 0000000000000000000000000000000000000000..0631f4ecb012ab42d13312be15475aff506055c5 --- /dev/null +++ b/320ms/decoder.mlpackage/Data/com.apple.CoreML/model.mlmodel @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:09f2dbd1f6a06faa6995f71d4b25d7c446996b6059cfac5ecc889853bdc7c6e5 +size 6728 diff --git a/320ms/decoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/320ms/decoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin new file mode 100644 index 0000000000000000000000000000000000000000..cde1618cbed235b421984a302289e3bdd7e3df02 --- /dev/null +++ b/320ms/decoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b4cacecdcd9df79ab1e56de67230baf5a8664d2afe0bb8f3408eefa972cb2f4 +size 7873600 diff --git a/320ms/decoder.mlpackage/Manifest.json b/320ms/decoder.mlpackage/Manifest.json new file mode 100644 index 0000000000000000000000000000000000000000..4e498499a0a4026095bbc46d4e1a47ecbc39b0ca --- /dev/null +++ b/320ms/decoder.mlpackage/Manifest.json @@ -0,0 +1,18 @@ +{ + "fileFormatVersion": "1.0.0", + "itemInfoEntries": { + "8201D73A-2B5D-488C-9C2B-7E2E75DF700D": { + "author": "com.apple.CoreML", + "description": "CoreML Model Weights", + "name": "weights", + "path": "com.apple.CoreML/weights" + }, + "F8EEBE8D-F17D-4556-B8DF-9BC11701B36D": { + "author": "com.apple.CoreML", + "description": "CoreML Model Specification", + "name": "model.mlmodel", + "path": "com.apple.CoreML/model.mlmodel" + } + }, + "rootModelIdentifier": "F8EEBE8D-F17D-4556-B8DF-9BC11701B36D" +} diff --git a/320ms/individual_components.py b/320ms/individual_components.py new file mode 100644 index 0000000000000000000000000000000000000000..47271397bc8d9d17cc0fabcf4bb63be7e7c2109c --- /dev/null +++ b/320ms/individual_components.py @@ -0,0 +1,250 @@ +#!/usr/bin/env python3 +"""Export Parakeet Realtime EOU 120M RNNT components into CoreML.""" +from __future__ import annotations + +from dataclasses import dataclass +from pathlib import Path +from typing import Optional, Tuple + +import coremltools as ct +import torch + + +@dataclass +class ExportSettings: + output_dir: Path + compute_units: ct.ComputeUnit + deployment_target: Optional[ct.target] + compute_precision: Optional[ct.precision] + max_audio_seconds: float + max_symbol_steps: int + + +class PreprocessorWrapper(torch.nn.Module): + """Wrapper for the audio preprocessor (mel spectrogram extraction).""" + + def __init__(self, module: torch.nn.Module) -> None: + super().__init__() + self.module = module + + def forward( + self, audio_signal: torch.Tensor, length: torch.Tensor + ) -> Tuple[torch.Tensor, torch.Tensor]: + mel, mel_length = self.module( + input_signal=audio_signal, length=length.to(dtype=torch.long) + ) + return mel, mel_length + + +class EncoderWrapper(torch.nn.Module): + """Wrapper for the cache-aware FastConformer encoder. + + Note: For the realtime EOU model, the encoder is cache-aware which means + it can operate in a streaming fashion. For CoreML export, we export + without cache state for simplicity (full-context mode). + """ + + def __init__(self, module: torch.nn.Module) -> None: + super().__init__() + self.module = module + + def forward( + self, features: torch.Tensor, length: torch.Tensor + ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: + encoded, encoded_lengths = self.module( + audio_signal=features, length=length.to(dtype=torch.long) + ) + # Synthesize per-frame timestamps (seconds) using the 80 ms encoder stride. + # Shape: [B, T_enc] + frame_times = ( + torch.arange(encoded.shape[-1], device=encoded.device, dtype=torch.float32) + * 0.08 + ) + return encoded, encoded_lengths, frame_times + + +class DecoderWrapper(torch.nn.Module): + """Wrapper for the RNNT prediction network (decoder).""" + + def __init__(self, module: torch.nn.Module) -> None: + super().__init__() + self.module = module + + def forward( + self, + targets: torch.Tensor, + target_lengths: torch.Tensor, + h_in: torch.Tensor, + c_in: torch.Tensor, + ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: + state = [h_in, c_in] + decoder_output, _, new_state = self.module( + targets=targets.to(dtype=torch.long), + target_length=target_lengths.to(dtype=torch.long), + states=state, + ) + return decoder_output, new_state[0], new_state[1] + + +class JointWrapper(torch.nn.Module): + """Wrapper for the RNNT joint network. + + Note: Unlike Parakeet TDT v3, the realtime EOU model does NOT have + duration outputs (num_extra_outputs). The joint network outputs only + token logits over the vocabulary + blank. + """ + + def __init__(self, module: torch.nn.Module) -> None: + super().__init__() + self.module = module + + def forward( + self, encoder_outputs: torch.Tensor, decoder_outputs: torch.Tensor + ) -> torch.Tensor: + # Input: encoder_outputs [B, D, T], decoder_outputs [B, D, U] + # Transpose to match what projection layers expect + encoder_outputs = encoder_outputs.transpose(1, 2) # [B, T, D] + decoder_outputs = decoder_outputs.transpose(1, 2) # [B, U, D] + + # Apply projections + enc_proj = self.module.enc(encoder_outputs) # [B, T, joint_hidden] + dec_proj = self.module.pred(decoder_outputs) # [B, U, joint_hidden] + + # Explicit broadcasting along T and U to avoid converter ambiguity + x = enc_proj.unsqueeze(2) + dec_proj.unsqueeze(1) # [B, T, U, joint_hidden] + x = self.module.joint_net[0](x) # ReLU + x = self.module.joint_net[1](x) # Dropout (no-op in eval) + out = self.module.joint_net[2](x) # Linear -> logits [B, T, U, vocab+blank] + return out + + +class MelEncoderWrapper(torch.nn.Module): + """Fused wrapper: waveform -> mel -> encoder. + + Inputs: + - audio_signal: [B, S] + - audio_length: [B] + + Outputs: + - encoder: [B, D, T_enc] + - encoder_length: [B] + - frame_times: [T_enc] + """ + + def __init__( + self, preprocessor: PreprocessorWrapper, encoder: EncoderWrapper + ) -> None: + super().__init__() + self.preprocessor = preprocessor + self.encoder = encoder + + def forward( + self, audio_signal: torch.Tensor, audio_length: torch.Tensor + ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: + mel, mel_length = self.preprocessor(audio_signal, audio_length) + encoded, enc_len, frame_times = self.encoder(mel, mel_length.to(dtype=torch.int32)) + return encoded, enc_len, frame_times + + +class JointDecisionWrapper(torch.nn.Module): + """Joint + decision head: outputs label id and label prob. + + Unlike Parakeet TDT v3, this model does NOT have duration outputs. + + Inputs: + - encoder_outputs: [B, D, T] + - decoder_outputs: [B, D, U] + + Returns: + - token_id: [B, T, U] int32 + - token_prob: [B, T, U] float32 + """ + + def __init__(self, joint: JointWrapper, vocab_size: int) -> None: + super().__init__() + self.joint = joint + self.vocab_with_blank = int(vocab_size) + 1 + + def forward(self, encoder_outputs: torch.Tensor, decoder_outputs: torch.Tensor): + logits = self.joint(encoder_outputs, decoder_outputs) + + # Token selection + token_ids = torch.argmax(logits, dim=-1).to(dtype=torch.int32) + token_probs_all = torch.softmax(logits, dim=-1) + # gather expects int64 (long) indices; cast only for gather + token_prob = torch.gather( + token_probs_all, dim=-1, index=token_ids.long().unsqueeze(-1) + ).squeeze(-1) + + return token_ids, token_prob + + +class JointDecisionSingleStep(torch.nn.Module): + """Single-step variant for streaming: encoder_step -> token decision. + + Inputs: + - encoder_step: [B=1, D, T=1] + - decoder_step: [B=1, D, U=1] + + Returns: + - token_id: [1, 1, 1] int32 + - token_prob: [1, 1, 1] float32 + - top_k_ids: [1, 1, 1, K] int32 + - top_k_logits: [1, 1, 1, K] float32 + """ + + def __init__(self, joint: JointWrapper, vocab_size: int, top_k: int = 64) -> None: + super().__init__() + self.joint = joint + self.vocab_with_blank = int(vocab_size) + 1 + self.top_k = int(top_k) + + def forward(self, encoder_step: torch.Tensor, decoder_step: torch.Tensor): + # Reuse JointWrapper which expects [B, D, T] and [B, D, U] + logits = self.joint(encoder_step, decoder_step) # [1, 1, 1, V+blank] + + token_ids = torch.argmax(logits, dim=-1, keepdim=False).to(dtype=torch.int32) + token_probs_all = torch.softmax(logits, dim=-1) + token_prob = torch.gather( + token_probs_all, dim=-1, index=token_ids.long().unsqueeze(-1) + ).squeeze(-1) + + # Also expose top-K candidates for host-side processing + topk_logits, topk_ids_long = torch.topk( + logits, k=min(self.top_k, logits.shape[-1]), dim=-1 + ) + topk_ids = topk_ids_long.to(dtype=torch.int32) + return token_ids, token_prob, topk_ids, topk_logits + + +def _coreml_convert( + traced: torch.jit.ScriptModule, + inputs, + outputs, + settings: ExportSettings, + compute_units_override: Optional[ct.ComputeUnit] = None, + compute_precision: Optional[ct.precision] = None, +) -> ct.models.MLModel: + cu = ( + compute_units_override + if compute_units_override is not None + else settings.compute_units + ) + kwargs = { + "convert_to": "mlprogram", + "inputs": inputs, + "outputs": outputs, + "compute_units": cu, + } + print("Converting:", traced.__class__.__name__) + print("Conversion kwargs:", kwargs) + if settings.deployment_target is not None: + kwargs["minimum_deployment_target"] = settings.deployment_target + + # Priority: explicit argument > settings + if compute_precision is not None: + kwargs["compute_precision"] = compute_precision + elif settings.compute_precision is not None: + kwargs["compute_precision"] = settings.compute_precision + + return ct.convert(traced, **kwargs) diff --git a/320ms/joint_decision.mlmodelc/analytics/coremldata.bin b/320ms/joint_decision.mlmodelc/analytics/coremldata.bin new file mode 100644 index 0000000000000000000000000000000000000000..ff9929e12bae25c509655bda0ebda6b9fe13fade --- /dev/null +++ b/320ms/joint_decision.mlmodelc/analytics/coremldata.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5bca32ad130dcad6605cc00044c752aa5b45ef57d14c17f2d1a2fa49d6cf55b5 +size 243 diff --git a/320ms/joint_decision.mlmodelc/coremldata.bin b/320ms/joint_decision.mlmodelc/coremldata.bin new file mode 100644 index 0000000000000000000000000000000000000000..1aa5afe7c02e93f9baba2757cad5c21f563157bf --- /dev/null +++ b/320ms/joint_decision.mlmodelc/coremldata.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22d4abc4625b935ee035b5f8ce7cb28d1041b9b01c12173e287bf4b5f5d99625 +size 493 diff --git a/320ms/joint_decision.mlmodelc/metadata.json b/320ms/joint_decision.mlmodelc/metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..a0a319857f429f5ec4b106397cd52ec354ef2e96 --- /dev/null +++ b/320ms/joint_decision.mlmodelc/metadata.json @@ -0,0 +1,112 @@ +[ + { + "metadataOutputVersion" : "3.0", + "shortDescription" : "Parakeet EOU single-step joint decision", + "outputSchema" : [ + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1 × 1 × 1)", + "shortDescription" : "", + "shape" : "[1, 1, 1]", + "name" : "token_id", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 1 × 1)", + "shortDescription" : "", + "shape" : "[1, 1, 1]", + "name" : "token_prob", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1 × 1 × 1 × 64)", + "shortDescription" : "", + "shape" : "[1, 1, 1, 64]", + "name" : "top_k_ids", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 1 × 1 × 64)", + "shortDescription" : "", + "shape" : "[1, 1, 1, 64]", + "name" : "top_k_logits", + "type" : "MultiArray" + } + ], + "storagePrecision" : "Float16", + "modelParameters" : [ + + ], + "author" : "Fluid Inference", + "specificationVersion" : 8, + "mlProgramOperationTypeHistogram" : { + "Ios17.reduceArgmax" : 1, + "Ios17.squeeze" : 1, + "Ios17.cast" : 6, + "Ios17.linear" : 3, + "Ios17.transpose" : 2, + "Ios17.add" : 1, + "Ios16.relu" : 1, + "Ios16.softmax" : 1, + "Ios17.gatherAlongAxis" : 1, + "Ios17.topk" : 1, + "Ios17.expandDims" : 3 + }, + "computePrecision" : "Mixed (Float16, Float32, Int16, Int32, UInt16)", + "isUpdatable" : "0", + "stateSchema" : [ + + ], + "availability" : { + "macOS" : "14.0", + "tvOS" : "17.0", + "visionOS" : "1.0", + "watchOS" : "10.0", + "iOS" : "17.0", + "macCatalyst" : "17.0" + }, + "modelType" : { + "name" : "MLModelType_mlProgram" + }, + "inputSchema" : [ + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 512 × 1)", + "shortDescription" : "", + "shape" : "[1, 512, 1]", + "name" : "encoder_step", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 640 × 1)", + "shortDescription" : "", + "shape" : "[1, 640, 1]", + "name" : "decoder_step", + "type" : "MultiArray" + } + ], + "userDefinedMetadata" : { + "com.github.apple.coremltools.source_dialect" : "TorchScript", + "com.github.apple.coremltools.version" : "8.3.0", + "com.github.apple.coremltools.source" : "torch==2.4.0" + }, + "generatedClassName" : "parakeet_eou_joint_decision_single_step", + "method" : "predict" + } +] \ No newline at end of file diff --git a/320ms/joint_decision.mlmodelc/model.mil b/320ms/joint_decision.mlmodelc/model.mil new file mode 100644 index 0000000000000000000000000000000000000000..172a7579866c84407b32b5f746b7f1ae132599d8 --- /dev/null +++ b/320ms/joint_decision.mlmodelc/model.mil @@ -0,0 +1,57 @@ +program(1.0) +[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.4.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] +{ + func main(tensor decoder_step, tensor encoder_step) { + tensor input_1_perm_0 = const()[name = tensor("input_1_perm_0"), val = tensor([0, 2, 1])]; + tensor encoder_step_to_fp16_dtype_0 = const()[name = tensor("encoder_step_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([0, 2, 1])]; + tensor decoder_step_to_fp16_dtype_0 = const()[name = tensor("decoder_step_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor joint_module_enc_weight_to_fp16 = const()[name = tensor("joint_module_enc_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; + tensor joint_module_enc_bias_to_fp16 = const()[name = tensor("joint_module_enc_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(655488)))]; + tensor encoder_step_to_fp16 = cast(dtype = encoder_step_to_fp16_dtype_0, x = encoder_step)[name = tensor("cast_8")]; + tensor input_1_cast_fp16 = transpose(perm = input_1_perm_0, x = encoder_step_to_fp16)[name = tensor("transpose_1")]; + tensor linear_0_cast_fp16 = linear(bias = joint_module_enc_bias_to_fp16, weight = joint_module_enc_weight_to_fp16, x = input_1_cast_fp16)[name = tensor("linear_0_cast_fp16")]; + tensor joint_module_pred_weight_to_fp16 = const()[name = tensor("joint_module_pred_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(656832)))]; + tensor joint_module_pred_bias_to_fp16 = const()[name = tensor("joint_module_pred_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1476096)))]; + tensor decoder_step_to_fp16 = cast(dtype = decoder_step_to_fp16_dtype_0, x = decoder_step)[name = tensor("cast_7")]; + tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = decoder_step_to_fp16)[name = tensor("transpose_0")]; + tensor linear_1_cast_fp16 = linear(bias = joint_module_pred_bias_to_fp16, weight = joint_module_pred_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("linear_1_cast_fp16")]; + tensor var_23_axes_0 = const()[name = tensor("op_23_axes_0"), val = tensor([2])]; + tensor var_23_cast_fp16 = expand_dims(axes = var_23_axes_0, x = linear_0_cast_fp16)[name = tensor("op_23_cast_fp16")]; + tensor var_24_axes_0 = const()[name = tensor("op_24_axes_0"), val = tensor([1])]; + tensor var_24_cast_fp16 = expand_dims(axes = var_24_axes_0, x = linear_1_cast_fp16)[name = tensor("op_24_cast_fp16")]; + tensor input_5_cast_fp16 = add(x = var_23_cast_fp16, y = var_24_cast_fp16)[name = tensor("input_5_cast_fp16")]; + tensor input_7_cast_fp16 = relu(x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; + tensor joint_module_joint_net_2_weight_to_fp16 = const()[name = tensor("joint_module_joint_net_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1477440)))]; + tensor joint_module_joint_net_2_bias_to_fp16 = const()[name = tensor("joint_module_joint_net_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2792064)))]; + tensor linear_2_cast_fp16 = linear(bias = joint_module_joint_net_2_bias_to_fp16, weight = joint_module_joint_net_2_weight_to_fp16, x = input_7_cast_fp16)[name = tensor("linear_2_cast_fp16")]; + tensor var_38_axis_0 = const()[name = tensor("op_38_axis_0"), val = tensor(-1)]; + tensor var_38_keep_dims_0 = const()[name = tensor("op_38_keep_dims_0"), val = tensor(false)]; + tensor var_38_output_dtype_0 = const()[name = tensor("op_38_output_dtype_0"), val = tensor("int32")]; + tensor token_id = reduce_argmax(axis = var_38_axis_0, keep_dims = var_38_keep_dims_0, output_dtype = var_38_output_dtype_0, x = linear_2_cast_fp16)[name = tensor("op_38_cast_fp16")]; + tensor var_44 = const()[name = tensor("op_44"), val = tensor(-1)]; + tensor token_probs_all_cast_fp16 = softmax(axis = var_44, x = linear_2_cast_fp16)[name = tensor("token_probs_all_cast_fp16")]; + tensor var_53_axes_0 = const()[name = tensor("op_53_axes_0"), val = tensor([-1])]; + tensor var_53 = expand_dims(axes = var_53_axes_0, x = token_id)[name = tensor("op_53")]; + tensor var_54 = const()[name = tensor("op_54"), val = tensor(-1)]; + tensor var_56_validate_indices_0 = const()[name = tensor("op_56_validate_indices_0"), val = tensor(false)]; + tensor var_53_to_int16_dtype_0 = const()[name = tensor("op_53_to_int16_dtype_0"), val = tensor("int16")]; + tensor var_53_to_int16 = cast(dtype = var_53_to_int16_dtype_0, x = var_53)[name = tensor("cast_6")]; + tensor var_56_cast_fp16_cast_int16 = gather_along_axis(axis = var_54, indices = var_53_to_int16, validate_indices = var_56_validate_indices_0, x = token_probs_all_cast_fp16)[name = tensor("op_56_cast_fp16_cast_int16")]; + tensor var_58_axes_0 = const()[name = tensor("op_58_axes_0"), val = tensor([-1])]; + tensor var_58_cast_fp16 = squeeze(axes = var_58_axes_0, x = var_56_cast_fp16_cast_int16)[name = tensor("op_58_cast_fp16")]; + tensor var_58_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_58_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor var_59 = const()[name = tensor("op_59"), val = tensor(64)]; + tensor var_63_axis_0 = const()[name = tensor("op_63_axis_0"), val = tensor(-1)]; + tensor var_63_ascending_0 = const()[name = tensor("op_63_ascending_0"), val = tensor(false)]; + tensor var_63_sort_0 = const()[name = tensor("op_63_sort_0"), val = tensor(true)]; + tensor var_63_return_indices_0 = const()[name = tensor("op_63_return_indices_0"), val = tensor(true)]; + tensor var_63_cast_fp16_cast_int16_output_indices_dtype_0 = const()[name = tensor("op_63_cast_fp16_cast_int16_output_indices_dtype_0"), val = tensor("uint16")]; + tensor var_63_cast_fp16_cast_int16_0, tensor var_63_cast_fp16_cast_int16_1 = topk(ascending = var_63_ascending_0, axis = var_63_axis_0, k = var_59, output_indices_dtype = var_63_cast_fp16_cast_int16_output_indices_dtype_0, return_indices = var_63_return_indices_0, sort = var_63_sort_0, x = linear_2_cast_fp16)[name = tensor("op_63_cast_fp16_cast_int16")]; + tensor var_63_cast_fp16_cast_int16_1_to_int32_dtype_0 = const()[name = tensor("op_63_cast_fp16_cast_int16_1_to_int32_dtype_0"), val = tensor("int32")]; + tensor var_63_cast_fp16_0_to_fp32_dtype_0 = const()[name = tensor("op_63_cast_fp16_0_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor top_k_logits = cast(dtype = var_63_cast_fp16_0_to_fp32_dtype_0, x = var_63_cast_fp16_cast_int16_0)[name = tensor("cast_3")]; + tensor top_k_ids = cast(dtype = var_63_cast_fp16_cast_int16_1_to_int32_dtype_0, x = var_63_cast_fp16_cast_int16_1)[name = tensor("cast_4")]; + tensor token_prob = cast(dtype = var_58_cast_fp16_to_fp32_dtype_0, x = var_58_cast_fp16)[name = tensor("cast_5")]; + } -> (token_id, token_prob, top_k_ids, top_k_logits); +} \ No newline at end of file diff --git a/320ms/joint_decision.mlmodelc/weights/weight.bin b/320ms/joint_decision.mlmodelc/weights/weight.bin new file mode 100644 index 0000000000000000000000000000000000000000..ec3d2c1dea18b332fa1a3cd8d31981bae4e3f649 --- /dev/null +++ b/320ms/joint_decision.mlmodelc/weights/weight.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7039b2010a269153f5a96edf28637f921a86ef8822f248f2d6712f7a6bce84b4 +size 2794182 diff --git a/320ms/joint_decision.mlpackage/Data/com.apple.CoreML/model.mlmodel b/320ms/joint_decision.mlpackage/Data/com.apple.CoreML/model.mlmodel new file mode 100644 index 0000000000000000000000000000000000000000..678be6fae0d594ff4bb923f973b0459139c8b370 --- /dev/null +++ b/320ms/joint_decision.mlpackage/Data/com.apple.CoreML/model.mlmodel @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25d4d7be6eeb60c7de1d3a1278a5a4700cbe34017e1a8c1cab33204ddb2e4d5e +size 8701 diff --git a/320ms/joint_decision.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/320ms/joint_decision.mlpackage/Data/com.apple.CoreML/weights/weight.bin new file mode 100644 index 0000000000000000000000000000000000000000..ec3d2c1dea18b332fa1a3cd8d31981bae4e3f649 --- /dev/null +++ b/320ms/joint_decision.mlpackage/Data/com.apple.CoreML/weights/weight.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7039b2010a269153f5a96edf28637f921a86ef8822f248f2d6712f7a6bce84b4 +size 2794182 diff --git a/320ms/joint_decision.mlpackage/Manifest.json b/320ms/joint_decision.mlpackage/Manifest.json new file mode 100644 index 0000000000000000000000000000000000000000..fffd90a77691d335c8153fd9c38d49adb725c949 --- /dev/null +++ b/320ms/joint_decision.mlpackage/Manifest.json @@ -0,0 +1,18 @@ +{ + "fileFormatVersion": "1.0.0", + "itemInfoEntries": { + "634E266B-4447-41D3-879E-F3611888F54B": { + "author": "com.apple.CoreML", + "description": "CoreML Model Weights", + "name": "weights", + "path": "com.apple.CoreML/weights" + }, + "C7F40527-180B-45CD-BC12-4F054F2E5D9A": { + "author": "com.apple.CoreML", + "description": "CoreML Model Specification", + "name": "model.mlmodel", + "path": "com.apple.CoreML/model.mlmodel" + } + }, + "rootModelIdentifier": "C7F40527-180B-45CD-BC12-4F054F2E5D9A" +} diff --git a/320ms/streaming_encoder.mlmodelc/analytics/coremldata.bin b/320ms/streaming_encoder.mlmodelc/analytics/coremldata.bin new file mode 100644 index 0000000000000000000000000000000000000000..eec5bd02b4d1075885ea77fb1d5526138526e198 --- /dev/null +++ b/320ms/streaming_encoder.mlmodelc/analytics/coremldata.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae0a559f8d53c43fe13c5c142b15ca237b97812febde71a36475b7b9612baec5 +size 243 diff --git a/320ms/streaming_encoder.mlmodelc/coremldata.bin b/320ms/streaming_encoder.mlmodelc/coremldata.bin new file mode 100644 index 0000000000000000000000000000000000000000..4bd36c23d7283099ead6c7857659a4e1a4f70dc5 --- /dev/null +++ b/320ms/streaming_encoder.mlmodelc/coremldata.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b57dfe32ed2acb0081f97d7a9e410117005823e3d2d995e21feb9e60775443c3 +size 670 diff --git a/320ms/streaming_encoder.mlmodelc/metadata.json b/320ms/streaming_encoder.mlmodelc/metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..4830f432c8a7079ceced0f4dbb661a775224824d --- /dev/null +++ b/320ms/streaming_encoder.mlmodelc/metadata.json @@ -0,0 +1,187 @@ +[ + { + "metadataOutputVersion" : "3.0", + "storagePrecision" : "Float16", + "outputSchema" : [ + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 512 × 8)", + "shortDescription" : "", + "shape" : "[1, 512, 8]", + "name" : "encoded_output", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1)", + "shortDescription" : "", + "shape" : "[1]", + "name" : "encoded_length", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 128 × 9)", + "shortDescription" : "", + "shape" : "[1, 128, 9]", + "name" : "new_pre_cache", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 17 × 1 × 70 × 512)", + "shortDescription" : "", + "shape" : "[17, 1, 70, 512]", + "name" : "new_cache_last_channel", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 17 × 1 × 512 × 8)", + "shortDescription" : "", + "shape" : "[17, 1, 512, 8]", + "name" : "new_cache_last_time", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1)", + "shortDescription" : "", + "shape" : "[1]", + "name" : "new_cache_last_channel_len", + "type" : "MultiArray" + } + ], + "modelParameters" : [ + + ], + "specificationVersion" : 8, + "mlProgramOperationTypeHistogram" : { + "Ios17.logicalAnd" : 3, + "Ios17.reshape" : 103, + "Ios16.softmax" : 17, + "Ios17.matmul" : 51, + "Ios17.transpose" : 157, + "Split" : 17, + "Ios17.expandDims" : 18, + "Select" : 51, + "Ios17.add" : 132, + "Tile" : 8, + "Ios17.sliceByIndex" : 140, + "Ios16.sigmoid" : 17, + "Pad" : 20, + "Ios17.logicalNot" : 2, + "Ios17.layerNorm" : 102, + "Ios17.less" : 5, + "Ios17.sub" : 4, + "Ios17.conv" : 56, + "Ios16.relu" : 3, + "Ios17.clip" : 2, + "Ios17.linear" : 137, + "Ios17.concat" : 52, + "Ios17.floorDiv" : 3, + "Ios17.greaterEqual" : 1, + "Ios17.cast" : 20, + "Ios16.silu" : 51, + "Stack" : 2, + "Ios17.mul" : 78 + }, + "computePrecision" : "Mixed (Float16, Float32, Int32)", + "isUpdatable" : "0", + "stateSchema" : [ + + ], + "availability" : { + "macOS" : "14.0", + "tvOS" : "17.0", + "visionOS" : "1.0", + "watchOS" : "10.0", + "iOS" : "17.0", + "macCatalyst" : "17.0" + }, + "modelType" : { + "name" : "MLModelType_mlProgram" + }, + "userDefinedMetadata" : { + "com.github.apple.coremltools.source_dialect" : "TorchScript", + "com.github.apple.coremltools.source" : "torch==2.4.0", + "com.github.apple.coremltools.version" : "8.3.0" + }, + "inputSchema" : [ + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 128 × 64)", + "shortDescription" : "", + "shape" : "[1, 128, 64]", + "name" : "audio_signal", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1)", + "shortDescription" : "", + "shape" : "[1]", + "name" : "audio_length", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 1 × 128 × 9)", + "shortDescription" : "", + "shape" : "[1, 128, 9]", + "name" : "pre_cache", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 17 × 1 × 70 × 512)", + "shortDescription" : "", + "shape" : "[17, 1, 70, 512]", + "name" : "cache_last_channel", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Float32", + "formattedType" : "MultiArray (Float32 17 × 1 × 512 × 8)", + "shortDescription" : "", + "shape" : "[17, 1, 512, 8]", + "name" : "cache_last_time", + "type" : "MultiArray" + }, + { + "hasShapeFlexibility" : "0", + "isOptional" : "0", + "dataType" : "Int32", + "formattedType" : "MultiArray (Int32 1)", + "shortDescription" : "", + "shape" : "[1]", + "name" : "cache_last_channel_len", + "type" : "MultiArray" + } + ], + "generatedClassName" : "streaming_encoder", + "method" : "predict" + } +] \ No newline at end of file diff --git a/320ms/streaming_encoder.mlmodelc/model.mil b/320ms/streaming_encoder.mlmodelc/model.mil new file mode 100644 index 0000000000000000000000000000000000000000..7902d6696813c64c694517418f9a9bdd2ee40119 --- /dev/null +++ b/320ms/streaming_encoder.mlmodelc/model.mil @@ -0,0 +1,3195 @@ +program(1.0) +[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.4.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] +{ + func main(tensor audio_length, tensor audio_signal, tensor cache_last_channel, tensor cache_last_channel_len, tensor cache_last_time, tensor pre_cache) { + tensor var_9 = const()[name = tensor("op_9"), val = tensor(2)]; + tensor full_input_interleave_0 = const()[name = tensor("full_input_interleave_0"), val = tensor(false)]; + tensor pre_cache_to_fp16_dtype_0 = const()[name = tensor("pre_cache_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor audio_signal_to_fp16_dtype_0 = const()[name = tensor("audio_signal_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor audio_signal_to_fp16 = cast(dtype = audio_signal_to_fp16_dtype_0, x = audio_signal)[name = tensor("cast_199")]; + tensor pre_cache_to_fp16 = cast(dtype = pre_cache_to_fp16_dtype_0, x = pre_cache)[name = tensor("cast_200")]; + tensor full_input_cast_fp16 = concat(axis = var_9, interleave = full_input_interleave_0, values = (pre_cache_to_fp16, audio_signal_to_fp16))[name = tensor("full_input_cast_fp16")]; + tensor var_12 = const()[name = tensor("op_12"), val = tensor(9)]; + tensor value_1 = add(x = audio_length, y = var_12)[name = tensor("value_1")]; + tensor var_28_begin_0 = const()[name = tensor("op_28_begin_0"), val = tensor([0, 0, 64])]; + tensor var_28_end_0 = const()[name = tensor("op_28_end_0"), val = tensor([1, 128, 73])]; + tensor var_28_end_mask_0 = const()[name = tensor("op_28_end_mask_0"), val = tensor([true, true, true])]; + tensor var_28_cast_fp16 = slice_by_index(begin = var_28_begin_0, end = var_28_end_0, end_mask = var_28_end_mask_0, x = full_input_cast_fp16)[name = tensor("op_28_cast_fp16")]; + tensor var_28_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_28_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor var_56 = const()[name = tensor("op_56"), val = tensor(-1)]; + tensor var_65 = const()[name = tensor("op_65"), val = tensor(1)]; + tensor x_1_perm_0 = const()[name = tensor("x_1_perm_0"), val = tensor([0, 2, 1])]; + tensor tensor_1_axes_0 = const()[name = tensor("tensor_1_axes_0"), val = tensor([1])]; + tensor x_1_cast_fp16 = transpose(perm = x_1_perm_0, x = full_input_cast_fp16)[name = tensor("transpose_241")]; + tensor tensor_1_cast_fp16 = expand_dims(axes = tensor_1_axes_0, x = x_1_cast_fp16)[name = tensor("tensor_1_cast_fp16")]; + tensor expand_dims_0 = const()[name = tensor("expand_dims_0"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72]])]; + tensor var_120_axes_0 = const()[name = tensor("op_120_axes_0"), val = tensor([1])]; + tensor var_120 = expand_dims(axes = var_120_axes_0, x = value_1)[name = tensor("op_120")]; + tensor time_mask_1 = less(x = expand_dims_0, y = var_120)[name = tensor("time_mask_1")]; + tensor var_122_axes_0 = const()[name = tensor("op_122_axes_0"), val = tensor([-1])]; + tensor var_122 = expand_dims(axes = var_122_axes_0, x = time_mask_1)[name = tensor("op_122")]; + tensor var_124_reps_0 = const()[name = tensor("op_124_reps_0"), val = tensor([1, 1, 128])]; + tensor var_124 = tile(reps = var_124_reps_0, x = var_122)[name = tensor("op_124")]; + tensor var_130_axes_0 = const()[name = tensor("op_130_axes_0"), val = tensor([1])]; + tensor cast_2_to_fp16_dtype_0 = const()[name = tensor("cast_2_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor var_124_to_fp16 = cast(dtype = cast_2_to_fp16_dtype_0, x = var_124)[name = tensor("cast_197")]; + tensor var_130_cast_fp16 = expand_dims(axes = var_130_axes_0, x = var_124_to_fp16)[name = tensor("op_130_cast_fp16")]; + tensor input_1_cast_fp16 = mul(x = tensor_1_cast_fp16, y = var_130_cast_fp16)[name = tensor("input_1_cast_fp16")]; + tensor input_3_pad_0 = const()[name = tensor("input_3_pad_0"), val = tensor([0, 0, 0, 0, 2, 1, 2, 1])]; + tensor input_3_mode_0 = const()[name = tensor("input_3_mode_0"), val = tensor("constant")]; + tensor const_9_to_fp16 = const()[name = tensor("const_9_to_fp16"), val = tensor(0x0p+0)]; + tensor input_3_cast_fp16 = pad(constant_val = const_9_to_fp16, mode = input_3_mode_0, pad = input_3_pad_0, x = input_1_cast_fp16)[name = tensor("input_3_cast_fp16")]; + tensor tensor_3_pad_type_0 = const()[name = tensor("tensor_3_pad_type_0"), val = tensor("valid")]; + tensor tensor_3_strides_0 = const()[name = tensor("tensor_3_strides_0"), val = tensor([2, 2])]; + tensor tensor_3_pad_0 = const()[name = tensor("tensor_3_pad_0"), val = tensor([0, 0, 0, 0])]; + tensor tensor_3_dilations_0 = const()[name = tensor("tensor_3_dilations_0"), val = tensor([1, 1])]; + tensor tensor_3_groups_0 = const()[name = tensor("tensor_3_groups_0"), val = tensor(1)]; + tensor encoder_pre_encode_conv_0_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; + tensor encoder_pre_encode_conv_0_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4736)))]; + tensor tensor_3_cast_fp16 = conv(bias = encoder_pre_encode_conv_0_bias_to_fp16, dilations = tensor_3_dilations_0, groups = tensor_3_groups_0, pad = tensor_3_pad_0, pad_type = tensor_3_pad_type_0, strides = tensor_3_strides_0, weight = encoder_pre_encode_conv_0_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("tensor_3_cast_fp16")]; + tensor cast_0_to_fp16_dtype_0 = const()[name = tensor("cast_0_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor var_143_promoted_to_fp16 = const()[name = tensor("op_143_promoted_to_fp16"), val = tensor(0x1p+1)]; + tensor value_1_to_fp16 = cast(dtype = cast_0_to_fp16_dtype_0, x = value_1)[name = tensor("cast_196")]; + tensor var_144_cast_fp16 = add(x = value_1_to_fp16, y = var_143_promoted_to_fp16)[name = tensor("op_144_cast_fp16")]; + tensor var_145_promoted_to_fp16 = const()[name = tensor("op_145_promoted_to_fp16"), val = tensor(0x1p+0)]; + tensor var_146_cast_fp16 = add(x = var_144_cast_fp16, y = var_145_promoted_to_fp16)[name = tensor("op_146_cast_fp16")]; + tensor var_147_promoted_to_fp16 = const()[name = tensor("op_147_promoted_to_fp16"), val = tensor(0x1.8p+1)]; + tensor var_148_cast_fp16 = sub(x = var_146_cast_fp16, y = var_147_promoted_to_fp16)[name = tensor("op_148_cast_fp16")]; + tensor var_53_promoted_to_fp16 = const()[name = tensor("op_53_promoted_to_fp16"), val = tensor(0x1p+1)]; + tensor floor_div_0_cast_fp16 = floor_div(x = var_148_cast_fp16, y = var_53_promoted_to_fp16)[name = tensor("floor_div_0_cast_fp16")]; + tensor var_150_promoted_to_fp16 = const()[name = tensor("op_150_promoted_to_fp16"), val = tensor(0x1p+0)]; + tensor current_lengths_3_cast_fp16 = add(x = floor_div_0_cast_fp16, y = var_150_promoted_to_fp16)[name = tensor("current_lengths_3_cast_fp16")]; + tensor cast_3_dtype_0 = const()[name = tensor("cast_3_dtype_0"), val = tensor("int32")]; + tensor expand_dims_1 = const()[name = tensor("expand_dims_1"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36]])]; + tensor var_159_axes_0 = const()[name = tensor("op_159_axes_0"), val = tensor([1])]; + tensor current_lengths_3_cast_fp16_to_int32 = cast(dtype = cast_3_dtype_0, x = current_lengths_3_cast_fp16)[name = tensor("cast_195")]; + tensor var_159 = expand_dims(axes = var_159_axes_0, x = current_lengths_3_cast_fp16_to_int32)[name = tensor("op_159")]; + tensor time_mask_3 = less(x = expand_dims_1, y = var_159)[name = tensor("time_mask_3")]; + tensor var_161_axes_0 = const()[name = tensor("op_161_axes_0"), val = tensor([-1])]; + tensor var_161 = expand_dims(axes = var_161_axes_0, x = time_mask_3)[name = tensor("op_161")]; + tensor var_163_reps_0 = const()[name = tensor("op_163_reps_0"), val = tensor([1, 1, 65])]; + tensor var_163 = tile(reps = var_163_reps_0, x = var_161)[name = tensor("op_163")]; + tensor var_169_axes_0 = const()[name = tensor("op_169_axes_0"), val = tensor([1])]; + tensor cast_4_to_fp16_dtype_0 = const()[name = tensor("cast_4_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor var_163_to_fp16 = cast(dtype = cast_4_to_fp16_dtype_0, x = var_163)[name = tensor("cast_194")]; + tensor var_169_cast_fp16 = expand_dims(axes = var_169_axes_0, x = var_163_to_fp16)[name = tensor("op_169_cast_fp16")]; + tensor expanded_mask_3_reps_0 = const()[name = tensor("expanded_mask_3_reps_0"), val = tensor([1, 256, 1, 1])]; + tensor expanded_mask_3_cast_fp16 = tile(reps = expanded_mask_3_reps_0, x = var_169_cast_fp16)[name = tensor("expanded_mask_3_cast_fp16")]; + tensor input_5_cast_fp16 = mul(x = tensor_3_cast_fp16, y = expanded_mask_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; + tensor tensor_5_cast_fp16 = relu(x = input_5_cast_fp16)[name = tensor("tensor_5_cast_fp16")]; + tensor input_7_cast_fp16 = mul(x = tensor_5_cast_fp16, y = expanded_mask_3_cast_fp16)[name = tensor("input_7_cast_fp16")]; + tensor input_9_pad_0 = const()[name = tensor("input_9_pad_0"), val = tensor([0, 0, 0, 0, 2, 1, 2, 1])]; + tensor input_9_mode_0 = const()[name = tensor("input_9_mode_0"), val = tensor("constant")]; + tensor const_23_to_fp16 = const()[name = tensor("const_23_to_fp16"), val = tensor(0x0p+0)]; + tensor input_9_cast_fp16 = pad(constant_val = const_23_to_fp16, mode = input_9_mode_0, pad = input_9_pad_0, x = input_7_cast_fp16)[name = tensor("input_9_cast_fp16")]; + tensor tensor_7_pad_type_0 = const()[name = tensor("tensor_7_pad_type_0"), val = tensor("valid")]; + tensor tensor_7_strides_0 = const()[name = tensor("tensor_7_strides_0"), val = tensor([2, 2])]; + tensor tensor_7_groups_0 = const()[name = tensor("tensor_7_groups_0"), val = tensor(256)]; + tensor tensor_7_pad_0 = const()[name = tensor("tensor_7_pad_0"), val = tensor([0, 0, 0, 0])]; + tensor tensor_7_dilations_0 = const()[name = tensor("tensor_7_dilations_0"), val = tensor([1, 1])]; + tensor encoder_pre_encode_conv_2_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5312)))]; + tensor encoder_pre_encode_conv_2_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9984)))]; + tensor tensor_7_cast_fp16 = conv(bias = encoder_pre_encode_conv_2_bias_to_fp16, dilations = tensor_7_dilations_0, groups = tensor_7_groups_0, pad = tensor_7_pad_0, pad_type = tensor_7_pad_type_0, strides = tensor_7_strides_0, weight = encoder_pre_encode_conv_2_weight_to_fp16, x = input_9_cast_fp16)[name = tensor("tensor_7_cast_fp16")]; + tensor var_191_promoted_to_fp16 = const()[name = tensor("op_191_promoted_to_fp16"), val = tensor(0x1p+1)]; + tensor var_192_cast_fp16 = add(x = current_lengths_3_cast_fp16, y = var_191_promoted_to_fp16)[name = tensor("op_192_cast_fp16")]; + tensor var_193_promoted_to_fp16 = const()[name = tensor("op_193_promoted_to_fp16"), val = tensor(0x1p+0)]; + tensor var_194_cast_fp16 = add(x = var_192_cast_fp16, y = var_193_promoted_to_fp16)[name = tensor("op_194_cast_fp16")]; + tensor var_195_promoted_to_fp16 = const()[name = tensor("op_195_promoted_to_fp16"), val = tensor(0x1.8p+1)]; + tensor var_196_cast_fp16 = sub(x = var_194_cast_fp16, y = var_195_promoted_to_fp16)[name = tensor("op_196_cast_fp16")]; + tensor var_53_promoted_1_to_fp16 = const()[name = tensor("op_53_promoted_1_to_fp16"), val = tensor(0x1p+1)]; + tensor floor_div_1_cast_fp16 = floor_div(x = var_196_cast_fp16, y = var_53_promoted_1_to_fp16)[name = tensor("floor_div_1_cast_fp16")]; + tensor var_198_promoted_to_fp16 = const()[name = tensor("op_198_promoted_to_fp16"), val = tensor(0x1p+0)]; + tensor current_lengths_5_cast_fp16 = add(x = floor_div_1_cast_fp16, y = var_198_promoted_to_fp16)[name = tensor("current_lengths_5_cast_fp16")]; + tensor cast_5_dtype_0 = const()[name = tensor("cast_5_dtype_0"), val = tensor("int32")]; + tensor expand_dims_2 = const()[name = tensor("expand_dims_2"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18]])]; + tensor var_207_axes_0 = const()[name = tensor("op_207_axes_0"), val = tensor([1])]; + tensor current_lengths_5_cast_fp16_to_int32 = cast(dtype = cast_5_dtype_0, x = current_lengths_5_cast_fp16)[name = tensor("cast_193")]; + tensor var_207 = expand_dims(axes = var_207_axes_0, x = current_lengths_5_cast_fp16_to_int32)[name = tensor("op_207")]; + tensor time_mask_5 = less(x = expand_dims_2, y = var_207)[name = tensor("time_mask_5")]; + tensor var_209_axes_0 = const()[name = tensor("op_209_axes_0"), val = tensor([-1])]; + tensor var_209 = expand_dims(axes = var_209_axes_0, x = time_mask_5)[name = tensor("op_209")]; + tensor var_211_reps_0 = const()[name = tensor("op_211_reps_0"), val = tensor([1, 1, 33])]; + tensor var_211 = tile(reps = var_211_reps_0, x = var_209)[name = tensor("op_211")]; + tensor var_217_axes_0 = const()[name = tensor("op_217_axes_0"), val = tensor([1])]; + tensor cast_6_to_fp16_dtype_0 = const()[name = tensor("cast_6_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor var_211_to_fp16 = cast(dtype = cast_6_to_fp16_dtype_0, x = var_211)[name = tensor("cast_192")]; + tensor var_217_cast_fp16 = expand_dims(axes = var_217_axes_0, x = var_211_to_fp16)[name = tensor("op_217_cast_fp16")]; + tensor expanded_mask_7_reps_0 = const()[name = tensor("expanded_mask_7_reps_0"), val = tensor([1, 256, 1, 1])]; + tensor expanded_mask_7_cast_fp16 = tile(reps = expanded_mask_7_reps_0, x = var_217_cast_fp16)[name = tensor("expanded_mask_7_cast_fp16")]; + tensor input_11_cast_fp16 = mul(x = tensor_7_cast_fp16, y = expanded_mask_7_cast_fp16)[name = tensor("input_11_cast_fp16")]; + tensor tensor_9_pad_type_0 = const()[name = tensor("tensor_9_pad_type_0"), val = tensor("valid")]; + tensor tensor_9_strides_0 = const()[name = tensor("tensor_9_strides_0"), val = tensor([1, 1])]; + tensor tensor_9_pad_0 = const()[name = tensor("tensor_9_pad_0"), val = tensor([0, 0, 0, 0])]; + tensor tensor_9_dilations_0 = const()[name = tensor("tensor_9_dilations_0"), val = tensor([1, 1])]; + tensor tensor_9_groups_0 = const()[name = tensor("tensor_9_groups_0"), val = tensor(1)]; + tensor encoder_pre_encode_conv_3_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_3_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10560)))]; + tensor encoder_pre_encode_conv_3_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_3_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141696)))]; + tensor tensor_9_cast_fp16 = conv(bias = encoder_pre_encode_conv_3_bias_to_fp16, dilations = tensor_9_dilations_0, groups = tensor_9_groups_0, pad = tensor_9_pad_0, pad_type = tensor_9_pad_type_0, strides = tensor_9_strides_0, weight = encoder_pre_encode_conv_3_weight_to_fp16, x = input_11_cast_fp16)[name = tensor("tensor_9_cast_fp16")]; + tensor input_13_cast_fp16 = mul(x = tensor_9_cast_fp16, y = expanded_mask_7_cast_fp16)[name = tensor("input_13_cast_fp16")]; + tensor tensor_11_cast_fp16 = relu(x = input_13_cast_fp16)[name = tensor("tensor_11_cast_fp16")]; + tensor input_15_cast_fp16 = mul(x = tensor_11_cast_fp16, y = expanded_mask_7_cast_fp16)[name = tensor("input_15_cast_fp16")]; + tensor input_17_pad_0 = const()[name = tensor("input_17_pad_0"), val = tensor([0, 0, 0, 0, 2, 1, 2, 1])]; + tensor input_17_mode_0 = const()[name = tensor("input_17_mode_0"), val = tensor("constant")]; + tensor const_41_to_fp16 = const()[name = tensor("const_41_to_fp16"), val = tensor(0x0p+0)]; + tensor input_17_cast_fp16 = pad(constant_val = const_41_to_fp16, mode = input_17_mode_0, pad = input_17_pad_0, x = input_15_cast_fp16)[name = tensor("input_17_cast_fp16")]; + tensor tensor_13_pad_type_0 = const()[name = tensor("tensor_13_pad_type_0"), val = tensor("valid")]; + tensor tensor_13_strides_0 = const()[name = tensor("tensor_13_strides_0"), val = tensor([2, 2])]; + tensor tensor_13_groups_0 = const()[name = tensor("tensor_13_groups_0"), val = tensor(256)]; + tensor tensor_13_pad_0 = const()[name = tensor("tensor_13_pad_0"), val = tensor([0, 0, 0, 0])]; + tensor tensor_13_dilations_0 = const()[name = tensor("tensor_13_dilations_0"), val = tensor([1, 1])]; + tensor encoder_pre_encode_conv_5_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_5_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142272)))]; + tensor encoder_pre_encode_conv_5_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_5_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146944)))]; + tensor tensor_13_cast_fp16 = conv(bias = encoder_pre_encode_conv_5_bias_to_fp16, dilations = tensor_13_dilations_0, groups = tensor_13_groups_0, pad = tensor_13_pad_0, pad_type = tensor_13_pad_type_0, strides = tensor_13_strides_0, weight = encoder_pre_encode_conv_5_weight_to_fp16, x = input_17_cast_fp16)[name = tensor("tensor_13_cast_fp16")]; + tensor var_254_promoted_to_fp16 = const()[name = tensor("op_254_promoted_to_fp16"), val = tensor(0x1p+1)]; + tensor var_255_cast_fp16 = add(x = current_lengths_5_cast_fp16, y = var_254_promoted_to_fp16)[name = tensor("op_255_cast_fp16")]; + tensor var_256_promoted_to_fp16 = const()[name = tensor("op_256_promoted_to_fp16"), val = tensor(0x1p+0)]; + tensor var_257_cast_fp16 = add(x = var_255_cast_fp16, y = var_256_promoted_to_fp16)[name = tensor("op_257_cast_fp16")]; + tensor var_258_promoted_to_fp16 = const()[name = tensor("op_258_promoted_to_fp16"), val = tensor(0x1.8p+1)]; + tensor var_259_cast_fp16 = sub(x = var_257_cast_fp16, y = var_258_promoted_to_fp16)[name = tensor("op_259_cast_fp16")]; + tensor var_53_promoted_2_to_fp16 = const()[name = tensor("op_53_promoted_2_to_fp16"), val = tensor(0x1p+1)]; + tensor floor_div_2_cast_fp16 = floor_div(x = var_259_cast_fp16, y = var_53_promoted_2_to_fp16)[name = tensor("floor_div_2_cast_fp16")]; + tensor var_261_promoted_to_fp16 = const()[name = tensor("op_261_promoted_to_fp16"), val = tensor(0x1p+0)]; + tensor current_lengths_cast_fp16 = add(x = floor_div_2_cast_fp16, y = var_261_promoted_to_fp16)[name = tensor("current_lengths_cast_fp16")]; + tensor cast_7_dtype_0 = const()[name = tensor("cast_7_dtype_0"), val = tensor("int32")]; + tensor expand_dims_3 = const()[name = tensor("expand_dims_3"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9]])]; + tensor var_270_axes_0 = const()[name = tensor("op_270_axes_0"), val = tensor([1])]; + tensor current_lengths_cast_fp16_to_int32 = cast(dtype = cast_7_dtype_0, x = current_lengths_cast_fp16)[name = tensor("cast_191")]; + tensor var_270 = expand_dims(axes = var_270_axes_0, x = current_lengths_cast_fp16_to_int32)[name = tensor("op_270")]; + tensor time_mask = less(x = expand_dims_3, y = var_270)[name = tensor("time_mask")]; + tensor var_272_axes_0 = const()[name = tensor("op_272_axes_0"), val = tensor([-1])]; + tensor var_272 = expand_dims(axes = var_272_axes_0, x = time_mask)[name = tensor("op_272")]; + tensor var_274_reps_0 = const()[name = tensor("op_274_reps_0"), val = tensor([1, 1, 17])]; + tensor var_274 = tile(reps = var_274_reps_0, x = var_272)[name = tensor("op_274")]; + tensor var_280_axes_0 = const()[name = tensor("op_280_axes_0"), val = tensor([1])]; + tensor cast_8_to_fp16_dtype_0 = const()[name = tensor("cast_8_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor var_274_to_fp16 = cast(dtype = cast_8_to_fp16_dtype_0, x = var_274)[name = tensor("cast_190")]; + tensor var_280_cast_fp16 = expand_dims(axes = var_280_axes_0, x = var_274_to_fp16)[name = tensor("op_280_cast_fp16")]; + tensor expanded_mask_13_reps_0 = const()[name = tensor("expanded_mask_13_reps_0"), val = tensor([1, 256, 1, 1])]; + tensor expanded_mask_13_cast_fp16 = tile(reps = expanded_mask_13_reps_0, x = var_280_cast_fp16)[name = tensor("expanded_mask_13_cast_fp16")]; + tensor input_19_cast_fp16 = mul(x = tensor_13_cast_fp16, y = expanded_mask_13_cast_fp16)[name = tensor("input_19_cast_fp16")]; + tensor tensor_15_pad_type_0 = const()[name = tensor("tensor_15_pad_type_0"), val = tensor("valid")]; + tensor tensor_15_strides_0 = const()[name = tensor("tensor_15_strides_0"), val = tensor([1, 1])]; + tensor tensor_15_pad_0 = const()[name = tensor("tensor_15_pad_0"), val = tensor([0, 0, 0, 0])]; + tensor tensor_15_dilations_0 = const()[name = tensor("tensor_15_dilations_0"), val = tensor([1, 1])]; + tensor tensor_15_groups_0 = const()[name = tensor("tensor_15_groups_0"), val = tensor(1)]; + tensor encoder_pre_encode_conv_6_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_6_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147520)))]; + tensor encoder_pre_encode_conv_6_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_conv_6_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(278656)))]; + tensor tensor_15_cast_fp16 = conv(bias = encoder_pre_encode_conv_6_bias_to_fp16, dilations = tensor_15_dilations_0, groups = tensor_15_groups_0, pad = tensor_15_pad_0, pad_type = tensor_15_pad_type_0, strides = tensor_15_strides_0, weight = encoder_pre_encode_conv_6_weight_to_fp16, x = input_19_cast_fp16)[name = tensor("tensor_15_cast_fp16")]; + tensor input_21_cast_fp16 = mul(x = tensor_15_cast_fp16, y = expanded_mask_13_cast_fp16)[name = tensor("input_21_cast_fp16")]; + tensor tensor_cast_fp16 = relu(x = input_21_cast_fp16)[name = tensor("tensor_cast_fp16")]; + tensor x_3_cast_fp16 = mul(x = tensor_cast_fp16, y = expanded_mask_13_cast_fp16)[name = tensor("x_3_cast_fp16")]; + tensor var_314_perm_0 = const()[name = tensor("op_314_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_315 = const()[name = tensor("op_315"), val = tensor([1, 10, -1])]; + tensor var_314_cast_fp16 = transpose(perm = var_314_perm_0, x = x_3_cast_fp16)[name = tensor("transpose_240")]; + tensor input_23_cast_fp16 = reshape(shape = var_315, x = var_314_cast_fp16)[name = tensor("input_23_cast_fp16")]; + tensor encoder_pre_encode_out_weight_to_fp16 = const()[name = tensor("encoder_pre_encode_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279232)))]; + tensor encoder_pre_encode_out_bias_to_fp16 = const()[name = tensor("encoder_pre_encode_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4735744)))]; + tensor linear_0_cast_fp16 = linear(bias = encoder_pre_encode_out_bias_to_fp16, weight = encoder_pre_encode_out_weight_to_fp16, x = input_23_cast_fp16)[name = tensor("linear_0_cast_fp16")]; + tensor var_325_begin_0 = const()[name = tensor("op_325_begin_0"), val = tensor([0, 2, 0])]; + tensor var_325_end_0 = const()[name = tensor("op_325_end_0"), val = tensor([1, 10, 512])]; + tensor var_325_end_mask_0 = const()[name = tensor("op_325_end_mask_0"), val = tensor([true, true, true])]; + tensor var_325_cast_fp16 = slice_by_index(begin = var_325_begin_0, end = var_325_end_0, end_mask = var_325_end_mask_0, x = linear_0_cast_fp16)[name = tensor("op_325_cast_fp16")]; + tensor var_327 = const()[name = tensor("op_327"), val = tensor(2)]; + tensor var_328 = sub(x = current_lengths_cast_fp16_to_int32, y = var_327)[name = tensor("op_328")]; + tensor var_328_promoted_to_fp16_dtype_0 = const()[name = tensor("op_328_promoted_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor var_59_promoted_to_fp16 = const()[name = tensor("op_59_promoted_to_fp16"), val = tensor(0x0p+0)]; + tensor const_61_to_fp16 = const()[name = tensor("const_61_to_fp16"), val = tensor(inf)]; + tensor var_328_to_fp16 = cast(dtype = var_328_promoted_to_fp16_dtype_0, x = var_328)[name = tensor("cast_189")]; + tensor clip_0_cast_fp16 = clip(alpha = var_59_promoted_to_fp16, beta = const_61_to_fp16, x = var_328_to_fp16)[name = tensor("clip_0_cast_fp16")]; + tensor cache_keep_size = const()[name = tensor("cache_keep_size"), val = tensor([4])]; + tensor var_344_promoted_to_fp16 = const()[name = tensor("op_344_promoted_to_fp16"), val = tensor(0x1.18p+6)]; + tensor padding_length_cast_fp16 = add(x = clip_0_cast_fp16, y = var_344_promoted_to_fp16)[name = tensor("padding_length_cast_fp16")]; + tensor const_63 = const()[name = tensor("const_63"), val = tensor(-1)]; + tensor var_346 = mul(x = cache_last_channel_len, y = const_63)[name = tensor("op_346")]; + tensor var_347 = const()[name = tensor("op_347"), val = tensor(70)]; + tensor offset = add(x = var_346, y = var_347)[name = tensor("offset")]; + tensor var_387_axes_0 = const()[name = tensor("op_387_axes_0"), val = tensor([-1])]; + tensor var_387_cast_fp16 = expand_dims(axes = var_387_axes_0, x = padding_length_cast_fp16)[name = tensor("op_387_cast_fp16")]; + tensor var_386_promoted_to_fp16 = const()[name = tensor("op_386_promoted_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4736832)))]; + tensor pad_mask_1_cast_fp16 = less(x = var_386_promoted_to_fp16, y = var_387_cast_fp16)[name = tensor("pad_mask_1_cast_fp16")]; + tensor expand_dims_5 = const()[name = tensor("expand_dims_5"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77]])]; + tensor var_393_axes_0 = const()[name = tensor("op_393_axes_0"), val = tensor([-1])]; + tensor var_393 = expand_dims(axes = var_393_axes_0, x = offset)[name = tensor("op_393")]; + tensor pad_mask_off = greater_equal(x = expand_dims_5, y = var_393)[name = tensor("pad_mask_off")]; + tensor pad_mask_3 = logical_and(x = pad_mask_off, y = pad_mask_1_cast_fp16)[name = tensor("pad_mask_3")]; + tensor var_396_axes_0 = const()[name = tensor("op_396_axes_0"), val = tensor([1])]; + tensor var_396 = expand_dims(axes = var_396_axes_0, x = pad_mask_3)[name = tensor("op_396")]; + tensor var_397 = const()[name = tensor("op_397"), val = tensor([1, 78, 1])]; + tensor pad_mask_for_att_mask_1 = tile(reps = var_397, x = var_396)[name = tensor("pad_mask_for_att_mask_1")]; + tensor var_399_perm_0 = const()[name = tensor("op_399_perm_0"), val = tensor([0, 2, 1])]; + tensor var_399 = transpose(perm = var_399_perm_0, x = pad_mask_for_att_mask_1)[name = tensor("transpose_239")]; + tensor pad_mask_for_att_mask = logical_and(x = pad_mask_for_att_mask_1, y = var_399)[name = tensor("pad_mask_for_att_mask")]; + tensor const_71 = const()[name = tensor("const_71"), val = tensor([[[true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false], [false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false], [false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false], [false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false], [false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false], [false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [false, false, false, false, false, false, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true]]])]; + tensor att_mask_9 = logical_and(x = pad_mask_for_att_mask, y = const_71)[name = tensor("att_mask_9")]; + tensor att_mask = logical_not(x = att_mask_9)[name = tensor("att_mask")]; + tensor pad_mask_5 = logical_not(x = pad_mask_3)[name = tensor("pad_mask_5")]; + tensor pad_mask_begin_0 = const()[name = tensor("pad_mask_begin_0"), val = tensor([0, 70])]; + tensor pad_mask_end_0 = const()[name = tensor("pad_mask_end_0"), val = tensor([1, 78])]; + tensor pad_mask_end_mask_0 = const()[name = tensor("pad_mask_end_mask_0"), val = tensor([true, true])]; + tensor pad_mask = slice_by_index(begin = pad_mask_begin_0, end = pad_mask_end_0, end_mask = pad_mask_end_mask_0, x = pad_mask_5)[name = tensor("pad_mask")]; + tensor mask_9_begin_0 = const()[name = tensor("mask_9_begin_0"), val = tensor([0, 70, 0])]; + tensor mask_9_end_0 = const()[name = tensor("mask_9_end_0"), val = tensor([1, 78, 78])]; + tensor mask_9_end_mask_0 = const()[name = tensor("mask_9_end_mask_0"), val = tensor([true, true, true])]; + tensor mask_9 = slice_by_index(begin = mask_9_begin_0, end = mask_9_end_0, end_mask = mask_9_end_mask_0, x = att_mask)[name = tensor("mask_9")]; + tensor cache_1_begin_0 = const()[name = tensor("cache_1_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor cache_1_end_0 = const()[name = tensor("cache_1_end_0"), val = tensor([1, 1, 70, 512])]; + tensor cache_1_end_mask_0 = const()[name = tensor("cache_1_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_1_squeeze_mask_0 = const()[name = tensor("cache_1_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_last_channel_to_fp16_dtype_0 = const()[name = tensor("cache_last_channel_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor cache_last_channel_to_fp16 = cast(dtype = cache_last_channel_to_fp16_dtype_0, x = cache_last_channel)[name = tensor("cast_188")]; + tensor cache_1_cast_fp16 = slice_by_index(begin = cache_1_begin_0, end = cache_1_end_0, end_mask = cache_1_end_mask_0, squeeze_mask = cache_1_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_1_cast_fp16")]; + tensor cache_3_begin_0 = const()[name = tensor("cache_3_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor cache_3_end_0 = const()[name = tensor("cache_3_end_0"), val = tensor([1, 1, 512, 8])]; + tensor cache_3_end_mask_0 = const()[name = tensor("cache_3_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_3_squeeze_mask_0 = const()[name = tensor("cache_3_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_last_time_to_fp16_dtype_0 = const()[name = tensor("cache_last_time_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor cache_last_time_to_fp16 = cast(dtype = cache_last_time_to_fp16_dtype_0, x = cache_last_time)[name = tensor("cast_187")]; + tensor cache_3_cast_fp16 = slice_by_index(begin = cache_3_begin_0, end = cache_3_end_0, end_mask = cache_3_end_mask_0, squeeze_mask = cache_3_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_3_cast_fp16")]; + tensor input_27_axes_0 = const()[name = tensor("input_27_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4737088)))]; + tensor encoder_layers_0_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4738176)))]; + tensor var_38_to_fp16 = const()[name = tensor("op_38_to_fp16"), val = tensor(0x1.5p-17)]; + tensor input_27_cast_fp16 = layer_norm(axes = input_27_axes_0, beta = encoder_layers_0_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_norm_feed_forward1_weight_to_fp16, x = var_325_cast_fp16)[name = tensor("input_27_cast_fp16")]; + tensor encoder_layers_0_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_0_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4739264)))]; + tensor linear_1_bias_0_to_fp16 = const()[name = tensor("linear_1_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6836480)))]; + tensor linear_1_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_0_feed_forward1_linear1_weight_to_fp16, x = input_27_cast_fp16)[name = tensor("linear_1_cast_fp16")]; + tensor input_31_cast_fp16 = silu(x = linear_1_cast_fp16)[name = tensor("input_31_cast_fp16")]; + tensor encoder_layers_0_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_0_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6840640)))]; + tensor linear_2_bias_0_to_fp16 = const()[name = tensor("linear_2_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8937856)))]; + tensor linear_2_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_feed_forward1_linear2_weight_to_fp16, x = input_31_cast_fp16)[name = tensor("linear_2_cast_fp16")]; + tensor var_436_to_fp16 = const()[name = tensor("op_436_to_fp16"), val = tensor(0x1p-1)]; + tensor var_437_cast_fp16 = mul(x = linear_2_cast_fp16, y = var_436_to_fp16)[name = tensor("op_437_cast_fp16")]; + tensor input_37_cast_fp16 = add(x = var_325_cast_fp16, y = var_437_cast_fp16)[name = tensor("input_37_cast_fp16")]; + tensor key_1_axes_0 = const()[name = tensor("key_1_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8938944)))]; + tensor encoder_layers_0_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8940032)))]; + tensor key_1_cast_fp16 = layer_norm(axes = key_1_axes_0, beta = encoder_layers_0_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_norm_self_att_weight_to_fp16, x = input_37_cast_fp16)[name = tensor("key_1_cast_fp16")]; + tensor input_39_interleave_0 = const()[name = tensor("input_39_interleave_0"), val = tensor(false)]; + tensor input_39_cast_fp16 = concat(axis = var_65, interleave = input_39_interleave_0, values = (cache_1_cast_fp16, key_1_cast_fp16))[name = tensor("input_39_cast_fp16")]; + tensor var_459_begin_0 = const()[name = tensor("op_459_begin_0"), val = tensor([0, 4, 0])]; + tensor var_459_end_0 = const()[name = tensor("op_459_end_0"), val = tensor([1, 70, 512])]; + tensor var_459_end_mask_0 = const()[name = tensor("op_459_end_mask_0"), val = tensor([true, true, true])]; + tensor var_459_cast_fp16 = slice_by_index(begin = var_459_begin_0, end = var_459_end_0, end_mask = var_459_end_mask_0, x = cache_1_cast_fp16)[name = tensor("op_459_cast_fp16")]; + tensor var_462_begin_0 = const()[name = tensor("op_462_begin_0"), val = tensor([0, 0, 0])]; + tensor var_462_end_0 = const()[name = tensor("op_462_end_0"), val = tensor([1, 4, 512])]; + tensor var_462_end_mask_0 = const()[name = tensor("op_462_end_mask_0"), val = tensor([true, false, true])]; + tensor var_462_cast_fp16 = slice_by_index(begin = var_462_begin_0, end = var_462_end_0, end_mask = var_462_end_mask_0, x = key_1_cast_fp16)[name = tensor("op_462_cast_fp16")]; + tensor var_465_interleave_0 = const()[name = tensor("op_465_interleave_0"), val = tensor(false)]; + tensor var_465_cast_fp16 = concat(axis = var_65, interleave = var_465_interleave_0, values = (var_459_cast_fp16, var_462_cast_fp16))[name = tensor("op_465_cast_fp16")]; + tensor encoder_layers_0_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8941120)))]; + tensor linear_3_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_self_attn_linear_q_weight_to_fp16, x = key_1_cast_fp16)[name = tensor("linear_3_cast_fp16")]; + tensor var_469 = const()[name = tensor("op_469"), val = tensor([1, -1, 8, 64])]; + tensor q_1_cast_fp16 = reshape(shape = var_469, x = linear_3_cast_fp16)[name = tensor("q_1_cast_fp16")]; + tensor encoder_layers_0_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9465472)))]; + tensor linear_4_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_self_attn_linear_k_weight_to_fp16, x = input_39_cast_fp16)[name = tensor("linear_4_cast_fp16")]; + tensor var_473 = const()[name = tensor("op_473"), val = tensor([1, -1, 8, 64])]; + tensor k_1_cast_fp16 = reshape(shape = var_473, x = linear_4_cast_fp16)[name = tensor("k_1_cast_fp16")]; + tensor encoder_layers_0_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9989824)))]; + tensor linear_5_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_self_attn_linear_v_weight_to_fp16, x = input_39_cast_fp16)[name = tensor("linear_5_cast_fp16")]; + tensor var_477 = const()[name = tensor("op_477"), val = tensor([1, -1, 8, 64])]; + tensor v_1_cast_fp16 = reshape(shape = var_477, x = linear_5_cast_fp16)[name = tensor("v_1_cast_fp16")]; + tensor value_3_perm_0 = const()[name = tensor("value_3_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_0_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10514176)))]; + tensor var_489_cast_fp16 = add(x = q_1_cast_fp16, y = encoder_layers_0_self_attn_pos_bias_u_to_fp16)[name = tensor("op_489_cast_fp16")]; + tensor encoder_layers_0_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10515264)))]; + tensor var_491_cast_fp16 = add(x = q_1_cast_fp16, y = encoder_layers_0_self_attn_pos_bias_v_to_fp16)[name = tensor("op_491_cast_fp16")]; + tensor q_with_bias_v_1_perm_0 = const()[name = tensor("q_with_bias_v_1_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_7_transpose_x_0 = const()[name = tensor("x_7_transpose_x_0"), val = tensor(false)]; + tensor x_7_transpose_y_0 = const()[name = tensor("x_7_transpose_y_0"), val = tensor(false)]; + tensor var_493_to_fp16 = const()[name = tensor("op_493_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10516352)))]; + tensor q_with_bias_v_1_cast_fp16 = transpose(perm = q_with_bias_v_1_perm_0, x = var_491_cast_fp16)[name = tensor("transpose_237")]; + tensor x_7_cast_fp16 = matmul(transpose_x = x_7_transpose_x_0, transpose_y = x_7_transpose_y_0, x = q_with_bias_v_1_cast_fp16, y = var_493_to_fp16)[name = tensor("x_7_cast_fp16")]; + tensor x_9_pad_0 = const()[name = tensor("x_9_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_9_mode_0 = const()[name = tensor("x_9_mode_0"), val = tensor("constant")]; + tensor const_79_to_fp16 = const()[name = tensor("const_79_to_fp16"), val = tensor(0x0p+0)]; + tensor x_9_cast_fp16 = pad(constant_val = const_79_to_fp16, mode = x_9_mode_0, pad = x_9_pad_0, x = x_7_cast_fp16)[name = tensor("x_9_cast_fp16")]; + tensor var_501 = const()[name = tensor("op_501"), val = tensor([1, 8, -1, 8])]; + tensor x_11_cast_fp16 = reshape(shape = var_501, x = x_9_cast_fp16)[name = tensor("x_11_cast_fp16")]; + tensor var_505_begin_0 = const()[name = tensor("op_505_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_505_end_0 = const()[name = tensor("op_505_end_0"), val = tensor([1, 8, 156, 8])]; + tensor var_505_end_mask_0 = const()[name = tensor("op_505_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_505_cast_fp16 = slice_by_index(begin = var_505_begin_0, end = var_505_end_0, end_mask = var_505_end_mask_0, x = x_11_cast_fp16)[name = tensor("op_505_cast_fp16")]; + tensor var_506 = const()[name = tensor("op_506"), val = tensor([1, 8, 8, 155])]; + tensor matrix_bd_1_cast_fp16 = reshape(shape = var_506, x = var_505_cast_fp16)[name = tensor("matrix_bd_1_cast_fp16")]; + tensor matrix_ac_1_transpose_x_0 = const()[name = tensor("matrix_ac_1_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_1_transpose_y_0 = const()[name = tensor("matrix_ac_1_transpose_y_0"), val = tensor(false)]; + tensor transpose_51_perm_0 = const()[name = tensor("transpose_51_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_52_perm_0 = const()[name = tensor("transpose_52_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_52 = transpose(perm = transpose_52_perm_0, x = k_1_cast_fp16)[name = tensor("transpose_235")]; + tensor transpose_51 = transpose(perm = transpose_51_perm_0, x = var_489_cast_fp16)[name = tensor("transpose_236")]; + tensor matrix_ac_1_cast_fp16 = matmul(transpose_x = matrix_ac_1_transpose_x_0, transpose_y = matrix_ac_1_transpose_y_0, x = transpose_51, y = transpose_52)[name = tensor("matrix_ac_1_cast_fp16")]; + tensor matrix_bd_3_begin_0 = const()[name = tensor("matrix_bd_3_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_3_end_0 = const()[name = tensor("matrix_bd_3_end_0"), val = tensor([1, 8, 8, 78])]; + tensor matrix_bd_3_end_mask_0 = const()[name = tensor("matrix_bd_3_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_3_cast_fp16 = slice_by_index(begin = matrix_bd_3_begin_0, end = matrix_bd_3_end_0, end_mask = matrix_bd_3_end_mask_0, x = matrix_bd_1_cast_fp16)[name = tensor("matrix_bd_3_cast_fp16")]; + tensor var_515_cast_fp16 = add(x = matrix_ac_1_cast_fp16, y = matrix_bd_3_cast_fp16)[name = tensor("op_515_cast_fp16")]; + tensor _inversed_scores_1_y_0_to_fp16 = const()[name = tensor("_inversed_scores_1_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_1_cast_fp16 = mul(x = var_515_cast_fp16, y = _inversed_scores_1_y_0_to_fp16)[name = tensor("_inversed_scores_1_cast_fp16")]; + tensor mask_11_axes_0 = const()[name = tensor("mask_11_axes_0"), val = tensor([1])]; + tensor mask_11 = expand_dims(axes = mask_11_axes_0, x = mask_9)[name = tensor("mask_11")]; + tensor var_41_to_fp16 = const()[name = tensor("op_41_to_fp16"), val = tensor(-0x1.388p+13)]; + tensor scores_3_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_1_cast_fp16, cond = mask_11)[name = tensor("scores_3_cast_fp16")]; + tensor var_521_cast_fp16 = softmax(axis = var_56, x = scores_3_cast_fp16)[name = tensor("op_521_cast_fp16")]; + tensor var_40_to_fp16 = const()[name = tensor("op_40_to_fp16"), val = tensor(0x0p+0)]; + tensor input_41_cast_fp16 = select(a = var_40_to_fp16, b = var_521_cast_fp16, cond = mask_11)[name = tensor("input_41_cast_fp16")]; + tensor x_13_transpose_x_0 = const()[name = tensor("x_13_transpose_x_0"), val = tensor(false)]; + tensor x_13_transpose_y_0 = const()[name = tensor("x_13_transpose_y_0"), val = tensor(false)]; + tensor value_3_cast_fp16 = transpose(perm = value_3_perm_0, x = v_1_cast_fp16)[name = tensor("transpose_238")]; + tensor x_13_cast_fp16 = matmul(transpose_x = x_13_transpose_x_0, transpose_y = x_13_transpose_y_0, x = input_41_cast_fp16, y = value_3_cast_fp16)[name = tensor("x_13_cast_fp16")]; + tensor var_525_perm_0 = const()[name = tensor("op_525_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_526 = const()[name = tensor("op_526"), val = tensor([1, -1, 512])]; + tensor var_525_cast_fp16 = transpose(perm = var_525_perm_0, x = x_13_cast_fp16)[name = tensor("transpose_234")]; + tensor input_43_cast_fp16 = reshape(shape = var_526, x = var_525_cast_fp16)[name = tensor("input_43_cast_fp16")]; + tensor encoder_layers_0_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_0_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10675136)))]; + tensor linear_7_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_self_attn_linear_out_weight_to_fp16, x = input_43_cast_fp16)[name = tensor("linear_7_cast_fp16")]; + tensor input_47_cast_fp16 = add(x = input_37_cast_fp16, y = linear_7_cast_fp16)[name = tensor("input_47_cast_fp16")]; + tensor x_17_axes_0 = const()[name = tensor("x_17_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11199488)))]; + tensor encoder_layers_0_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11200576)))]; + tensor x_17_cast_fp16 = layer_norm(axes = x_17_axes_0, beta = encoder_layers_0_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_norm_conv_weight_to_fp16, x = input_47_cast_fp16)[name = tensor("x_17_cast_fp16")]; + tensor input_49_perm_0 = const()[name = tensor("input_49_perm_0"), val = tensor([0, 2, 1])]; + tensor input_51_pad_type_0 = const()[name = tensor("input_51_pad_type_0"), val = tensor("valid")]; + tensor input_51_strides_0 = const()[name = tensor("input_51_strides_0"), val = tensor([1])]; + tensor input_51_pad_0 = const()[name = tensor("input_51_pad_0"), val = tensor([0, 0])]; + tensor input_51_dilations_0 = const()[name = tensor("input_51_dilations_0"), val = tensor([1])]; + tensor input_51_groups_0 = const()[name = tensor("input_51_groups_0"), val = tensor(1)]; + tensor encoder_layers_0_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_0_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11201664)))]; + tensor input_49_cast_fp16 = transpose(perm = input_49_perm_0, x = x_17_cast_fp16)[name = tensor("transpose_233")]; + tensor input_51_cast_fp16 = conv(dilations = input_51_dilations_0, groups = input_51_groups_0, pad = input_51_pad_0, pad_type = input_51_pad_type_0, strides = input_51_strides_0, weight = encoder_layers_0_conv_pointwise_conv1_weight_to_fp16, x = input_49_cast_fp16)[name = tensor("input_51_cast_fp16")]; + tensor x_19_split_num_splits_0 = const()[name = tensor("x_19_split_num_splits_0"), val = tensor(2)]; + tensor x_19_split_axis_0 = const()[name = tensor("x_19_split_axis_0"), val = tensor(1)]; + tensor x_19_split_cast_fp16_0, tensor x_19_split_cast_fp16_1 = split(axis = x_19_split_axis_0, num_splits = x_19_split_num_splits_0, x = input_51_cast_fp16)[name = tensor("x_19_split_cast_fp16")]; + tensor x_19_split_1_sigmoid_cast_fp16 = sigmoid(x = x_19_split_cast_fp16_1)[name = tensor("x_19_split_1_sigmoid_cast_fp16")]; + tensor x_19_cast_fp16 = mul(x = x_19_split_cast_fp16_0, y = x_19_split_1_sigmoid_cast_fp16)[name = tensor("x_19_cast_fp16")]; + tensor var_551_axes_0 = const()[name = tensor("op_551_axes_0"), val = tensor([1])]; + tensor var_551 = expand_dims(axes = var_551_axes_0, x = pad_mask)[name = tensor("op_551")]; + tensor input_53_cast_fp16 = select(a = var_40_to_fp16, b = x_19_cast_fp16, cond = var_551)[name = tensor("input_53_cast_fp16")]; + tensor new_x_3_interleave_0 = const()[name = tensor("new_x_3_interleave_0"), val = tensor(false)]; + tensor new_x_3_cast_fp16 = concat(axis = var_56, interleave = new_x_3_interleave_0, values = (cache_3_cast_fp16, input_53_cast_fp16))[name = tensor("new_x_3_cast_fp16")]; + tensor next_cache_1_begin_0 = const()[name = tensor("next_cache_1_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_1_end_0 = const()[name = tensor("next_cache_1_end_0"), val = tensor([1, 512, 12])]; + tensor next_cache_1_end_mask_0 = const()[name = tensor("next_cache_1_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_1_cast_fp16 = slice_by_index(begin = next_cache_1_begin_0, end = next_cache_1_end_0, end_mask = next_cache_1_end_mask_0, x = new_x_3_cast_fp16)[name = tensor("next_cache_1_cast_fp16")]; + tensor var_567_begin_0 = const()[name = tensor("op_567_begin_0"), val = tensor([0, 0, 4])]; + tensor var_567_end_0 = const()[name = tensor("op_567_end_0"), val = tensor([1, 512, 12])]; + tensor var_567_end_mask_0 = const()[name = tensor("op_567_end_mask_0"), val = tensor([true, true, true])]; + tensor var_567_cast_fp16 = slice_by_index(begin = var_567_begin_0, end = var_567_end_0, end_mask = var_567_end_mask_0, x = next_cache_1_cast_fp16)[name = tensor("op_567_cast_fp16")]; + tensor x_21_pad_type_0 = const()[name = tensor("x_21_pad_type_0"), val = tensor("valid")]; + tensor x_21_groups_0 = const()[name = tensor("x_21_groups_0"), val = tensor(512)]; + tensor x_21_strides_0 = const()[name = tensor("x_21_strides_0"), val = tensor([1])]; + tensor x_21_pad_0 = const()[name = tensor("x_21_pad_0"), val = tensor([0, 0])]; + tensor x_21_dilations_0 = const()[name = tensor("x_21_dilations_0"), val = tensor([1])]; + tensor encoder_layers_0_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_0_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12250304)))]; + tensor x_21_cast_fp16 = conv(dilations = x_21_dilations_0, groups = x_21_groups_0, pad = x_21_pad_0, pad_type = x_21_pad_type_0, strides = x_21_strides_0, weight = encoder_layers_0_conv_depthwise_conv_weight_to_fp16, x = new_x_3_cast_fp16)[name = tensor("x_21_cast_fp16")]; + tensor input_55_perm_0 = const()[name = tensor("input_55_perm_0"), val = tensor([0, 2, 1])]; + tensor x_23_axes_0 = const()[name = tensor("x_23_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_0_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12259584)))]; + tensor encoder_layers_0_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_0_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12260672)))]; + tensor input_55_cast_fp16 = transpose(perm = input_55_perm_0, x = x_21_cast_fp16)[name = tensor("transpose_232")]; + tensor x_23_cast_fp16 = layer_norm(axes = x_23_axes_0, beta = encoder_layers_0_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_conv_batch_norm_weight_to_fp16, x = input_55_cast_fp16)[name = tensor("x_23_cast_fp16")]; + tensor input_57_perm_0 = const()[name = tensor("input_57_perm_0"), val = tensor([0, 2, 1])]; + tensor input_57_cast_fp16 = transpose(perm = input_57_perm_0, x = x_23_cast_fp16)[name = tensor("transpose_231")]; + tensor input_59_cast_fp16 = silu(x = input_57_cast_fp16)[name = tensor("input_59_cast_fp16")]; + tensor x_25_pad_type_0 = const()[name = tensor("x_25_pad_type_0"), val = tensor("valid")]; + tensor x_25_strides_0 = const()[name = tensor("x_25_strides_0"), val = tensor([1])]; + tensor x_25_pad_0 = const()[name = tensor("x_25_pad_0"), val = tensor([0, 0])]; + tensor x_25_dilations_0 = const()[name = tensor("x_25_dilations_0"), val = tensor([1])]; + tensor x_25_groups_0 = const()[name = tensor("x_25_groups_0"), val = tensor(1)]; + tensor encoder_layers_0_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_0_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12261760)))]; + tensor x_25_cast_fp16 = conv(dilations = x_25_dilations_0, groups = x_25_groups_0, pad = x_25_pad_0, pad_type = x_25_pad_type_0, strides = x_25_strides_0, weight = encoder_layers_0_conv_pointwise_conv2_weight_to_fp16, x = input_59_cast_fp16)[name = tensor("x_25_cast_fp16")]; + tensor input_61_perm_0 = const()[name = tensor("input_61_perm_0"), val = tensor([0, 2, 1])]; + tensor input_61_cast_fp16 = transpose(perm = input_61_perm_0, x = x_25_cast_fp16)[name = tensor("transpose_230")]; + tensor input_63_cast_fp16 = add(x = input_47_cast_fp16, y = input_61_cast_fp16)[name = tensor("input_63_cast_fp16")]; + tensor input_65_axes_0 = const()[name = tensor("input_65_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12786112)))]; + tensor encoder_layers_0_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12787200)))]; + tensor input_65_cast_fp16 = layer_norm(axes = input_65_axes_0, beta = encoder_layers_0_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_norm_feed_forward2_weight_to_fp16, x = input_63_cast_fp16)[name = tensor("input_65_cast_fp16")]; + tensor encoder_layers_0_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_0_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(12788288)))]; + tensor linear_8_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_0_feed_forward2_linear1_weight_to_fp16, x = input_65_cast_fp16)[name = tensor("linear_8_cast_fp16")]; + tensor input_69_cast_fp16 = silu(x = linear_8_cast_fp16)[name = tensor("input_69_cast_fp16")]; + tensor encoder_layers_0_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_0_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14885504)))]; + tensor linear_9_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_0_feed_forward2_linear2_weight_to_fp16, x = input_69_cast_fp16)[name = tensor("linear_9_cast_fp16")]; + tensor var_608_to_fp16 = const()[name = tensor("op_608_to_fp16"), val = tensor(0x1p-1)]; + tensor var_609_cast_fp16 = mul(x = linear_9_cast_fp16, y = var_608_to_fp16)[name = tensor("op_609_cast_fp16")]; + tensor input_75_cast_fp16 = add(x = input_63_cast_fp16, y = var_609_cast_fp16)[name = tensor("input_75_cast_fp16")]; + tensor input_77_axes_0 = const()[name = tensor("input_77_axes_0"), val = tensor([-1])]; + tensor encoder_layers_0_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_0_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16982720)))]; + tensor encoder_layers_0_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_0_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16983808)))]; + tensor input_77_cast_fp16 = layer_norm(axes = input_77_axes_0, beta = encoder_layers_0_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_0_norm_out_weight_to_fp16, x = input_75_cast_fp16)[name = tensor("input_77_cast_fp16")]; + tensor cache_5_begin_0 = const()[name = tensor("cache_5_begin_0"), val = tensor([1, 0, 0, 0])]; + tensor cache_5_end_0 = const()[name = tensor("cache_5_end_0"), val = tensor([2, 1, 70, 512])]; + tensor cache_5_end_mask_0 = const()[name = tensor("cache_5_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_5_squeeze_mask_0 = const()[name = tensor("cache_5_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_5_cast_fp16 = slice_by_index(begin = cache_5_begin_0, end = cache_5_end_0, end_mask = cache_5_end_mask_0, squeeze_mask = cache_5_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_5_cast_fp16")]; + tensor cache_7_begin_0 = const()[name = tensor("cache_7_begin_0"), val = tensor([1, 0, 0, 0])]; + tensor cache_7_end_0 = const()[name = tensor("cache_7_end_0"), val = tensor([2, 1, 512, 8])]; + tensor cache_7_end_mask_0 = const()[name = tensor("cache_7_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_7_squeeze_mask_0 = const()[name = tensor("cache_7_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_7_cast_fp16 = slice_by_index(begin = cache_7_begin_0, end = cache_7_end_0, end_mask = cache_7_end_mask_0, squeeze_mask = cache_7_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_7_cast_fp16")]; + tensor input_79_axes_0 = const()[name = tensor("input_79_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16984896)))]; + tensor encoder_layers_1_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16985984)))]; + tensor input_79_cast_fp16 = layer_norm(axes = input_79_axes_0, beta = encoder_layers_1_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_norm_feed_forward1_weight_to_fp16, x = input_77_cast_fp16)[name = tensor("input_79_cast_fp16")]; + tensor encoder_layers_1_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_1_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16987072)))]; + tensor linear_10_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_1_feed_forward1_linear1_weight_to_fp16, x = input_79_cast_fp16)[name = tensor("linear_10_cast_fp16")]; + tensor input_83_cast_fp16 = silu(x = linear_10_cast_fp16)[name = tensor("input_83_cast_fp16")]; + tensor encoder_layers_1_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_1_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19084288)))]; + tensor linear_11_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_feed_forward1_linear2_weight_to_fp16, x = input_83_cast_fp16)[name = tensor("linear_11_cast_fp16")]; + tensor var_643_to_fp16 = const()[name = tensor("op_643_to_fp16"), val = tensor(0x1p-1)]; + tensor var_644_cast_fp16 = mul(x = linear_11_cast_fp16, y = var_643_to_fp16)[name = tensor("op_644_cast_fp16")]; + tensor input_89_cast_fp16 = add(x = input_77_cast_fp16, y = var_644_cast_fp16)[name = tensor("input_89_cast_fp16")]; + tensor key_3_axes_0 = const()[name = tensor("key_3_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21181504)))]; + tensor encoder_layers_1_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21182592)))]; + tensor key_3_cast_fp16 = layer_norm(axes = key_3_axes_0, beta = encoder_layers_1_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_norm_self_att_weight_to_fp16, x = input_89_cast_fp16)[name = tensor("key_3_cast_fp16")]; + tensor input_91_interleave_0 = const()[name = tensor("input_91_interleave_0"), val = tensor(false)]; + tensor input_91_cast_fp16 = concat(axis = var_65, interleave = input_91_interleave_0, values = (cache_5_cast_fp16, key_3_cast_fp16))[name = tensor("input_91_cast_fp16")]; + tensor var_666_begin_0 = const()[name = tensor("op_666_begin_0"), val = tensor([0, 4, 0])]; + tensor var_666_end_0 = const()[name = tensor("op_666_end_0"), val = tensor([1, 70, 512])]; + tensor var_666_end_mask_0 = const()[name = tensor("op_666_end_mask_0"), val = tensor([true, true, true])]; + tensor var_666_cast_fp16 = slice_by_index(begin = var_666_begin_0, end = var_666_end_0, end_mask = var_666_end_mask_0, x = cache_5_cast_fp16)[name = tensor("op_666_cast_fp16")]; + tensor var_669_begin_0 = const()[name = tensor("op_669_begin_0"), val = tensor([0, 0, 0])]; + tensor var_669_end_0 = const()[name = tensor("op_669_end_0"), val = tensor([1, 4, 512])]; + tensor var_669_end_mask_0 = const()[name = tensor("op_669_end_mask_0"), val = tensor([true, false, true])]; + tensor var_669_cast_fp16 = slice_by_index(begin = var_669_begin_0, end = var_669_end_0, end_mask = var_669_end_mask_0, x = key_3_cast_fp16)[name = tensor("op_669_cast_fp16")]; + tensor var_672_interleave_0 = const()[name = tensor("op_672_interleave_0"), val = tensor(false)]; + tensor var_672_cast_fp16 = concat(axis = var_65, interleave = var_672_interleave_0, values = (var_666_cast_fp16, var_669_cast_fp16))[name = tensor("op_672_cast_fp16")]; + tensor encoder_layers_1_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21183680)))]; + tensor linear_12_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_self_attn_linear_q_weight_to_fp16, x = key_3_cast_fp16)[name = tensor("linear_12_cast_fp16")]; + tensor var_676 = const()[name = tensor("op_676"), val = tensor([1, -1, 8, 64])]; + tensor q_7_cast_fp16 = reshape(shape = var_676, x = linear_12_cast_fp16)[name = tensor("q_7_cast_fp16")]; + tensor encoder_layers_1_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21708032)))]; + tensor linear_13_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_self_attn_linear_k_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("linear_13_cast_fp16")]; + tensor var_680 = const()[name = tensor("op_680"), val = tensor([1, -1, 8, 64])]; + tensor k_5_cast_fp16 = reshape(shape = var_680, x = linear_13_cast_fp16)[name = tensor("k_5_cast_fp16")]; + tensor encoder_layers_1_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22232384)))]; + tensor linear_14_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_self_attn_linear_v_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("linear_14_cast_fp16")]; + tensor var_684 = const()[name = tensor("op_684"), val = tensor([1, -1, 8, 64])]; + tensor v_3_cast_fp16 = reshape(shape = var_684, x = linear_14_cast_fp16)[name = tensor("v_3_cast_fp16")]; + tensor value_5_perm_0 = const()[name = tensor("value_5_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_1_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22756736)))]; + tensor var_696_cast_fp16 = add(x = q_7_cast_fp16, y = encoder_layers_1_self_attn_pos_bias_u_to_fp16)[name = tensor("op_696_cast_fp16")]; + tensor encoder_layers_1_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22757824)))]; + tensor var_698_cast_fp16 = add(x = q_7_cast_fp16, y = encoder_layers_1_self_attn_pos_bias_v_to_fp16)[name = tensor("op_698_cast_fp16")]; + tensor q_with_bias_v_3_perm_0 = const()[name = tensor("q_with_bias_v_3_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_33_transpose_x_0 = const()[name = tensor("x_33_transpose_x_0"), val = tensor(false)]; + tensor x_33_transpose_y_0 = const()[name = tensor("x_33_transpose_y_0"), val = tensor(false)]; + tensor var_700_to_fp16 = const()[name = tensor("op_700_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22758912)))]; + tensor q_with_bias_v_3_cast_fp16 = transpose(perm = q_with_bias_v_3_perm_0, x = var_698_cast_fp16)[name = tensor("transpose_228")]; + tensor x_33_cast_fp16 = matmul(transpose_x = x_33_transpose_x_0, transpose_y = x_33_transpose_y_0, x = q_with_bias_v_3_cast_fp16, y = var_700_to_fp16)[name = tensor("x_33_cast_fp16")]; + tensor x_35_pad_0 = const()[name = tensor("x_35_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_35_mode_0 = const()[name = tensor("x_35_mode_0"), val = tensor("constant")]; + tensor const_92_to_fp16 = const()[name = tensor("const_92_to_fp16"), val = tensor(0x0p+0)]; + tensor x_35_cast_fp16 = pad(constant_val = const_92_to_fp16, mode = x_35_mode_0, pad = x_35_pad_0, x = x_33_cast_fp16)[name = tensor("x_35_cast_fp16")]; + tensor var_708 = const()[name = tensor("op_708"), val = tensor([1, 8, -1, 8])]; + tensor x_37_cast_fp16 = reshape(shape = var_708, x = x_35_cast_fp16)[name = tensor("x_37_cast_fp16")]; + tensor var_712_begin_0 = const()[name = tensor("op_712_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_712_end_0 = const()[name = tensor("op_712_end_0"), val = tensor([1, 8, 156, 8])]; + tensor var_712_end_mask_0 = const()[name = tensor("op_712_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_712_cast_fp16 = slice_by_index(begin = var_712_begin_0, end = var_712_end_0, end_mask = var_712_end_mask_0, x = x_37_cast_fp16)[name = tensor("op_712_cast_fp16")]; + tensor var_713 = const()[name = tensor("op_713"), val = tensor([1, 8, 8, 155])]; + tensor matrix_bd_5_cast_fp16 = reshape(shape = var_713, x = var_712_cast_fp16)[name = tensor("matrix_bd_5_cast_fp16")]; + tensor matrix_ac_3_transpose_x_0 = const()[name = tensor("matrix_ac_3_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_3_transpose_y_0 = const()[name = tensor("matrix_ac_3_transpose_y_0"), val = tensor(false)]; + tensor transpose_53_perm_0 = const()[name = tensor("transpose_53_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_54_perm_0 = const()[name = tensor("transpose_54_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_54 = transpose(perm = transpose_54_perm_0, x = k_5_cast_fp16)[name = tensor("transpose_226")]; + tensor transpose_53 = transpose(perm = transpose_53_perm_0, x = var_696_cast_fp16)[name = tensor("transpose_227")]; + tensor matrix_ac_3_cast_fp16 = matmul(transpose_x = matrix_ac_3_transpose_x_0, transpose_y = matrix_ac_3_transpose_y_0, x = transpose_53, y = transpose_54)[name = tensor("matrix_ac_3_cast_fp16")]; + tensor matrix_bd_7_begin_0 = const()[name = tensor("matrix_bd_7_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_7_end_0 = const()[name = tensor("matrix_bd_7_end_0"), val = tensor([1, 8, 8, 78])]; + tensor matrix_bd_7_end_mask_0 = const()[name = tensor("matrix_bd_7_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_7_cast_fp16 = slice_by_index(begin = matrix_bd_7_begin_0, end = matrix_bd_7_end_0, end_mask = matrix_bd_7_end_mask_0, x = matrix_bd_5_cast_fp16)[name = tensor("matrix_bd_7_cast_fp16")]; + tensor var_722_cast_fp16 = add(x = matrix_ac_3_cast_fp16, y = matrix_bd_7_cast_fp16)[name = tensor("op_722_cast_fp16")]; + tensor _inversed_scores_5_y_0_to_fp16 = const()[name = tensor("_inversed_scores_5_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_5_cast_fp16 = mul(x = var_722_cast_fp16, y = _inversed_scores_5_y_0_to_fp16)[name = tensor("_inversed_scores_5_cast_fp16")]; + tensor scores_7_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_5_cast_fp16, cond = mask_11)[name = tensor("scores_7_cast_fp16")]; + tensor var_728_cast_fp16 = softmax(axis = var_56, x = scores_7_cast_fp16)[name = tensor("op_728_cast_fp16")]; + tensor input_93_cast_fp16 = select(a = var_40_to_fp16, b = var_728_cast_fp16, cond = mask_11)[name = tensor("input_93_cast_fp16")]; + tensor x_39_transpose_x_0 = const()[name = tensor("x_39_transpose_x_0"), val = tensor(false)]; + tensor x_39_transpose_y_0 = const()[name = tensor("x_39_transpose_y_0"), val = tensor(false)]; + tensor value_5_cast_fp16 = transpose(perm = value_5_perm_0, x = v_3_cast_fp16)[name = tensor("transpose_229")]; + tensor x_39_cast_fp16 = matmul(transpose_x = x_39_transpose_x_0, transpose_y = x_39_transpose_y_0, x = input_93_cast_fp16, y = value_5_cast_fp16)[name = tensor("x_39_cast_fp16")]; + tensor var_732_perm_0 = const()[name = tensor("op_732_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_733 = const()[name = tensor("op_733"), val = tensor([1, -1, 512])]; + tensor var_732_cast_fp16 = transpose(perm = var_732_perm_0, x = x_39_cast_fp16)[name = tensor("transpose_225")]; + tensor input_95_cast_fp16 = reshape(shape = var_733, x = var_732_cast_fp16)[name = tensor("input_95_cast_fp16")]; + tensor encoder_layers_1_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_1_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22917696)))]; + tensor linear_16_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_self_attn_linear_out_weight_to_fp16, x = input_95_cast_fp16)[name = tensor("linear_16_cast_fp16")]; + tensor input_99_cast_fp16 = add(x = input_89_cast_fp16, y = linear_16_cast_fp16)[name = tensor("input_99_cast_fp16")]; + tensor x_43_axes_0 = const()[name = tensor("x_43_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23442048)))]; + tensor encoder_layers_1_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23443136)))]; + tensor x_43_cast_fp16 = layer_norm(axes = x_43_axes_0, beta = encoder_layers_1_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_norm_conv_weight_to_fp16, x = input_99_cast_fp16)[name = tensor("x_43_cast_fp16")]; + tensor input_101_perm_0 = const()[name = tensor("input_101_perm_0"), val = tensor([0, 2, 1])]; + tensor input_103_pad_type_0 = const()[name = tensor("input_103_pad_type_0"), val = tensor("valid")]; + tensor input_103_strides_0 = const()[name = tensor("input_103_strides_0"), val = tensor([1])]; + tensor input_103_pad_0 = const()[name = tensor("input_103_pad_0"), val = tensor([0, 0])]; + tensor input_103_dilations_0 = const()[name = tensor("input_103_dilations_0"), val = tensor([1])]; + tensor input_103_groups_0 = const()[name = tensor("input_103_groups_0"), val = tensor(1)]; + tensor encoder_layers_1_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_1_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23444224)))]; + tensor input_101_cast_fp16 = transpose(perm = input_101_perm_0, x = x_43_cast_fp16)[name = tensor("transpose_224")]; + tensor input_103_cast_fp16 = conv(dilations = input_103_dilations_0, groups = input_103_groups_0, pad = input_103_pad_0, pad_type = input_103_pad_type_0, strides = input_103_strides_0, weight = encoder_layers_1_conv_pointwise_conv1_weight_to_fp16, x = input_101_cast_fp16)[name = tensor("input_103_cast_fp16")]; + tensor x_45_split_num_splits_0 = const()[name = tensor("x_45_split_num_splits_0"), val = tensor(2)]; + tensor x_45_split_axis_0 = const()[name = tensor("x_45_split_axis_0"), val = tensor(1)]; + tensor x_45_split_cast_fp16_0, tensor x_45_split_cast_fp16_1 = split(axis = x_45_split_axis_0, num_splits = x_45_split_num_splits_0, x = input_103_cast_fp16)[name = tensor("x_45_split_cast_fp16")]; + tensor x_45_split_1_sigmoid_cast_fp16 = sigmoid(x = x_45_split_cast_fp16_1)[name = tensor("x_45_split_1_sigmoid_cast_fp16")]; + tensor x_45_cast_fp16 = mul(x = x_45_split_cast_fp16_0, y = x_45_split_1_sigmoid_cast_fp16)[name = tensor("x_45_cast_fp16")]; + tensor input_105_cast_fp16 = select(a = var_40_to_fp16, b = x_45_cast_fp16, cond = var_551)[name = tensor("input_105_cast_fp16")]; + tensor new_x_7_interleave_0 = const()[name = tensor("new_x_7_interleave_0"), val = tensor(false)]; + tensor new_x_7_cast_fp16 = concat(axis = var_56, interleave = new_x_7_interleave_0, values = (cache_7_cast_fp16, input_105_cast_fp16))[name = tensor("new_x_7_cast_fp16")]; + tensor next_cache_3_begin_0 = const()[name = tensor("next_cache_3_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_3_end_0 = const()[name = tensor("next_cache_3_end_0"), val = tensor([1, 512, 12])]; + tensor next_cache_3_end_mask_0 = const()[name = tensor("next_cache_3_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_3_cast_fp16 = slice_by_index(begin = next_cache_3_begin_0, end = next_cache_3_end_0, end_mask = next_cache_3_end_mask_0, x = new_x_7_cast_fp16)[name = tensor("next_cache_3_cast_fp16")]; + tensor var_774_begin_0 = const()[name = tensor("op_774_begin_0"), val = tensor([0, 0, 4])]; + tensor var_774_end_0 = const()[name = tensor("op_774_end_0"), val = tensor([1, 512, 12])]; + tensor var_774_end_mask_0 = const()[name = tensor("op_774_end_mask_0"), val = tensor([true, true, true])]; + tensor var_774_cast_fp16 = slice_by_index(begin = var_774_begin_0, end = var_774_end_0, end_mask = var_774_end_mask_0, x = next_cache_3_cast_fp16)[name = tensor("op_774_cast_fp16")]; + tensor x_47_pad_type_0 = const()[name = tensor("x_47_pad_type_0"), val = tensor("valid")]; + tensor x_47_groups_0 = const()[name = tensor("x_47_groups_0"), val = tensor(512)]; + tensor x_47_strides_0 = const()[name = tensor("x_47_strides_0"), val = tensor([1])]; + tensor x_47_pad_0 = const()[name = tensor("x_47_pad_0"), val = tensor([0, 0])]; + tensor x_47_dilations_0 = const()[name = tensor("x_47_dilations_0"), val = tensor([1])]; + tensor encoder_layers_1_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_1_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24492864)))]; + tensor x_47_cast_fp16 = conv(dilations = x_47_dilations_0, groups = x_47_groups_0, pad = x_47_pad_0, pad_type = x_47_pad_type_0, strides = x_47_strides_0, weight = encoder_layers_1_conv_depthwise_conv_weight_to_fp16, x = new_x_7_cast_fp16)[name = tensor("x_47_cast_fp16")]; + tensor input_107_perm_0 = const()[name = tensor("input_107_perm_0"), val = tensor([0, 2, 1])]; + tensor x_49_axes_0 = const()[name = tensor("x_49_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_1_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24502144)))]; + tensor encoder_layers_1_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_1_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24503232)))]; + tensor input_107_cast_fp16 = transpose(perm = input_107_perm_0, x = x_47_cast_fp16)[name = tensor("transpose_223")]; + tensor x_49_cast_fp16 = layer_norm(axes = x_49_axes_0, beta = encoder_layers_1_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_conv_batch_norm_weight_to_fp16, x = input_107_cast_fp16)[name = tensor("x_49_cast_fp16")]; + tensor input_109_perm_0 = const()[name = tensor("input_109_perm_0"), val = tensor([0, 2, 1])]; + tensor input_109_cast_fp16 = transpose(perm = input_109_perm_0, x = x_49_cast_fp16)[name = tensor("transpose_222")]; + tensor input_111_cast_fp16 = silu(x = input_109_cast_fp16)[name = tensor("input_111_cast_fp16")]; + tensor x_51_pad_type_0 = const()[name = tensor("x_51_pad_type_0"), val = tensor("valid")]; + tensor x_51_strides_0 = const()[name = tensor("x_51_strides_0"), val = tensor([1])]; + tensor x_51_pad_0 = const()[name = tensor("x_51_pad_0"), val = tensor([0, 0])]; + tensor x_51_dilations_0 = const()[name = tensor("x_51_dilations_0"), val = tensor([1])]; + tensor x_51_groups_0 = const()[name = tensor("x_51_groups_0"), val = tensor(1)]; + tensor encoder_layers_1_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_1_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24504320)))]; + tensor x_51_cast_fp16 = conv(dilations = x_51_dilations_0, groups = x_51_groups_0, pad = x_51_pad_0, pad_type = x_51_pad_type_0, strides = x_51_strides_0, weight = encoder_layers_1_conv_pointwise_conv2_weight_to_fp16, x = input_111_cast_fp16)[name = tensor("x_51_cast_fp16")]; + tensor input_113_perm_0 = const()[name = tensor("input_113_perm_0"), val = tensor([0, 2, 1])]; + tensor input_113_cast_fp16 = transpose(perm = input_113_perm_0, x = x_51_cast_fp16)[name = tensor("transpose_221")]; + tensor input_115_cast_fp16 = add(x = input_99_cast_fp16, y = input_113_cast_fp16)[name = tensor("input_115_cast_fp16")]; + tensor input_117_axes_0 = const()[name = tensor("input_117_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25028672)))]; + tensor encoder_layers_1_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25029760)))]; + tensor input_117_cast_fp16 = layer_norm(axes = input_117_axes_0, beta = encoder_layers_1_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_norm_feed_forward2_weight_to_fp16, x = input_115_cast_fp16)[name = tensor("input_117_cast_fp16")]; + tensor encoder_layers_1_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_1_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25030848)))]; + tensor linear_17_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_1_feed_forward2_linear1_weight_to_fp16, x = input_117_cast_fp16)[name = tensor("linear_17_cast_fp16")]; + tensor input_121_cast_fp16 = silu(x = linear_17_cast_fp16)[name = tensor("input_121_cast_fp16")]; + tensor encoder_layers_1_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_1_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27128064)))]; + tensor linear_18_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_1_feed_forward2_linear2_weight_to_fp16, x = input_121_cast_fp16)[name = tensor("linear_18_cast_fp16")]; + tensor var_815_to_fp16 = const()[name = tensor("op_815_to_fp16"), val = tensor(0x1p-1)]; + tensor var_816_cast_fp16 = mul(x = linear_18_cast_fp16, y = var_815_to_fp16)[name = tensor("op_816_cast_fp16")]; + tensor input_127_cast_fp16 = add(x = input_115_cast_fp16, y = var_816_cast_fp16)[name = tensor("input_127_cast_fp16")]; + tensor input_129_axes_0 = const()[name = tensor("input_129_axes_0"), val = tensor([-1])]; + tensor encoder_layers_1_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_1_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29225280)))]; + tensor encoder_layers_1_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_1_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29226368)))]; + tensor input_129_cast_fp16 = layer_norm(axes = input_129_axes_0, beta = encoder_layers_1_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_1_norm_out_weight_to_fp16, x = input_127_cast_fp16)[name = tensor("input_129_cast_fp16")]; + tensor cache_9_begin_0 = const()[name = tensor("cache_9_begin_0"), val = tensor([2, 0, 0, 0])]; + tensor cache_9_end_0 = const()[name = tensor("cache_9_end_0"), val = tensor([3, 1, 70, 512])]; + tensor cache_9_end_mask_0 = const()[name = tensor("cache_9_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_9_squeeze_mask_0 = const()[name = tensor("cache_9_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_9_cast_fp16 = slice_by_index(begin = cache_9_begin_0, end = cache_9_end_0, end_mask = cache_9_end_mask_0, squeeze_mask = cache_9_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_9_cast_fp16")]; + tensor cache_11_begin_0 = const()[name = tensor("cache_11_begin_0"), val = tensor([2, 0, 0, 0])]; + tensor cache_11_end_0 = const()[name = tensor("cache_11_end_0"), val = tensor([3, 1, 512, 8])]; + tensor cache_11_end_mask_0 = const()[name = tensor("cache_11_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_11_squeeze_mask_0 = const()[name = tensor("cache_11_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_11_cast_fp16 = slice_by_index(begin = cache_11_begin_0, end = cache_11_end_0, end_mask = cache_11_end_mask_0, squeeze_mask = cache_11_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_11_cast_fp16")]; + tensor input_131_axes_0 = const()[name = tensor("input_131_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29227456)))]; + tensor encoder_layers_2_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29228544)))]; + tensor input_131_cast_fp16 = layer_norm(axes = input_131_axes_0, beta = encoder_layers_2_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_norm_feed_forward1_weight_to_fp16, x = input_129_cast_fp16)[name = tensor("input_131_cast_fp16")]; + tensor encoder_layers_2_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_2_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29229632)))]; + tensor linear_19_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_2_feed_forward1_linear1_weight_to_fp16, x = input_131_cast_fp16)[name = tensor("linear_19_cast_fp16")]; + tensor input_135_cast_fp16 = silu(x = linear_19_cast_fp16)[name = tensor("input_135_cast_fp16")]; + tensor encoder_layers_2_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_2_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(31326848)))]; + tensor linear_20_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_feed_forward1_linear2_weight_to_fp16, x = input_135_cast_fp16)[name = tensor("linear_20_cast_fp16")]; + tensor var_850_to_fp16 = const()[name = tensor("op_850_to_fp16"), val = tensor(0x1p-1)]; + tensor var_851_cast_fp16 = mul(x = linear_20_cast_fp16, y = var_850_to_fp16)[name = tensor("op_851_cast_fp16")]; + tensor input_141_cast_fp16 = add(x = input_129_cast_fp16, y = var_851_cast_fp16)[name = tensor("input_141_cast_fp16")]; + tensor key_5_axes_0 = const()[name = tensor("key_5_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33424064)))]; + tensor encoder_layers_2_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33425152)))]; + tensor key_5_cast_fp16 = layer_norm(axes = key_5_axes_0, beta = encoder_layers_2_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_norm_self_att_weight_to_fp16, x = input_141_cast_fp16)[name = tensor("key_5_cast_fp16")]; + tensor input_143_interleave_0 = const()[name = tensor("input_143_interleave_0"), val = tensor(false)]; + tensor input_143_cast_fp16 = concat(axis = var_65, interleave = input_143_interleave_0, values = (cache_9_cast_fp16, key_5_cast_fp16))[name = tensor("input_143_cast_fp16")]; + tensor var_873_begin_0 = const()[name = tensor("op_873_begin_0"), val = tensor([0, 4, 0])]; + tensor var_873_end_0 = const()[name = tensor("op_873_end_0"), val = tensor([1, 70, 512])]; + tensor var_873_end_mask_0 = const()[name = tensor("op_873_end_mask_0"), val = tensor([true, true, true])]; + tensor var_873_cast_fp16 = slice_by_index(begin = var_873_begin_0, end = var_873_end_0, end_mask = var_873_end_mask_0, x = cache_9_cast_fp16)[name = tensor("op_873_cast_fp16")]; + tensor var_876_begin_0 = const()[name = tensor("op_876_begin_0"), val = tensor([0, 0, 0])]; + tensor var_876_end_0 = const()[name = tensor("op_876_end_0"), val = tensor([1, 4, 512])]; + tensor var_876_end_mask_0 = const()[name = tensor("op_876_end_mask_0"), val = tensor([true, false, true])]; + tensor var_876_cast_fp16 = slice_by_index(begin = var_876_begin_0, end = var_876_end_0, end_mask = var_876_end_mask_0, x = key_5_cast_fp16)[name = tensor("op_876_cast_fp16")]; + tensor var_879_interleave_0 = const()[name = tensor("op_879_interleave_0"), val = tensor(false)]; + tensor var_879_cast_fp16 = concat(axis = var_65, interleave = var_879_interleave_0, values = (var_873_cast_fp16, var_876_cast_fp16))[name = tensor("op_879_cast_fp16")]; + tensor encoder_layers_2_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33426240)))]; + tensor linear_21_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_self_attn_linear_q_weight_to_fp16, x = key_5_cast_fp16)[name = tensor("linear_21_cast_fp16")]; + tensor var_883 = const()[name = tensor("op_883"), val = tensor([1, -1, 8, 64])]; + tensor q_13_cast_fp16 = reshape(shape = var_883, x = linear_21_cast_fp16)[name = tensor("q_13_cast_fp16")]; + tensor encoder_layers_2_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33950592)))]; + tensor linear_22_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_self_attn_linear_k_weight_to_fp16, x = input_143_cast_fp16)[name = tensor("linear_22_cast_fp16")]; + tensor var_887 = const()[name = tensor("op_887"), val = tensor([1, -1, 8, 64])]; + tensor k_9_cast_fp16 = reshape(shape = var_887, x = linear_22_cast_fp16)[name = tensor("k_9_cast_fp16")]; + tensor encoder_layers_2_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34474944)))]; + tensor linear_23_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_self_attn_linear_v_weight_to_fp16, x = input_143_cast_fp16)[name = tensor("linear_23_cast_fp16")]; + tensor var_891 = const()[name = tensor("op_891"), val = tensor([1, -1, 8, 64])]; + tensor v_5_cast_fp16 = reshape(shape = var_891, x = linear_23_cast_fp16)[name = tensor("v_5_cast_fp16")]; + tensor value_7_perm_0 = const()[name = tensor("value_7_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_2_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34999296)))]; + tensor var_903_cast_fp16 = add(x = q_13_cast_fp16, y = encoder_layers_2_self_attn_pos_bias_u_to_fp16)[name = tensor("op_903_cast_fp16")]; + tensor encoder_layers_2_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35000384)))]; + tensor var_905_cast_fp16 = add(x = q_13_cast_fp16, y = encoder_layers_2_self_attn_pos_bias_v_to_fp16)[name = tensor("op_905_cast_fp16")]; + tensor q_with_bias_v_5_perm_0 = const()[name = tensor("q_with_bias_v_5_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_59_transpose_x_0 = const()[name = tensor("x_59_transpose_x_0"), val = tensor(false)]; + tensor x_59_transpose_y_0 = const()[name = tensor("x_59_transpose_y_0"), val = tensor(false)]; + tensor var_907_to_fp16 = const()[name = tensor("op_907_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35001472)))]; + tensor q_with_bias_v_5_cast_fp16 = transpose(perm = q_with_bias_v_5_perm_0, x = var_905_cast_fp16)[name = tensor("transpose_219")]; + tensor x_59_cast_fp16 = matmul(transpose_x = x_59_transpose_x_0, transpose_y = x_59_transpose_y_0, x = q_with_bias_v_5_cast_fp16, y = var_907_to_fp16)[name = tensor("x_59_cast_fp16")]; + tensor x_61_pad_0 = const()[name = tensor("x_61_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_61_mode_0 = const()[name = tensor("x_61_mode_0"), val = tensor("constant")]; + tensor const_105_to_fp16 = const()[name = tensor("const_105_to_fp16"), val = tensor(0x0p+0)]; + tensor x_61_cast_fp16 = pad(constant_val = const_105_to_fp16, mode = x_61_mode_0, pad = x_61_pad_0, x = x_59_cast_fp16)[name = tensor("x_61_cast_fp16")]; + tensor var_915 = const()[name = tensor("op_915"), val = tensor([1, 8, -1, 8])]; + tensor x_63_cast_fp16 = reshape(shape = var_915, x = x_61_cast_fp16)[name = tensor("x_63_cast_fp16")]; + tensor var_919_begin_0 = const()[name = tensor("op_919_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_919_end_0 = const()[name = tensor("op_919_end_0"), val = tensor([1, 8, 156, 8])]; + tensor var_919_end_mask_0 = const()[name = tensor("op_919_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_919_cast_fp16 = slice_by_index(begin = var_919_begin_0, end = var_919_end_0, end_mask = var_919_end_mask_0, x = x_63_cast_fp16)[name = tensor("op_919_cast_fp16")]; + tensor var_920 = const()[name = tensor("op_920"), val = tensor([1, 8, 8, 155])]; + tensor matrix_bd_9_cast_fp16 = reshape(shape = var_920, x = var_919_cast_fp16)[name = tensor("matrix_bd_9_cast_fp16")]; + tensor matrix_ac_5_transpose_x_0 = const()[name = tensor("matrix_ac_5_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_5_transpose_y_0 = const()[name = tensor("matrix_ac_5_transpose_y_0"), val = tensor(false)]; + tensor transpose_55_perm_0 = const()[name = tensor("transpose_55_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_56_perm_0 = const()[name = tensor("transpose_56_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_56 = transpose(perm = transpose_56_perm_0, x = k_9_cast_fp16)[name = tensor("transpose_217")]; + tensor transpose_55 = transpose(perm = transpose_55_perm_0, x = var_903_cast_fp16)[name = tensor("transpose_218")]; + tensor matrix_ac_5_cast_fp16 = matmul(transpose_x = matrix_ac_5_transpose_x_0, transpose_y = matrix_ac_5_transpose_y_0, x = transpose_55, y = transpose_56)[name = tensor("matrix_ac_5_cast_fp16")]; + tensor matrix_bd_11_begin_0 = const()[name = tensor("matrix_bd_11_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_11_end_0 = const()[name = tensor("matrix_bd_11_end_0"), val = tensor([1, 8, 8, 78])]; + tensor matrix_bd_11_end_mask_0 = const()[name = tensor("matrix_bd_11_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_11_cast_fp16 = slice_by_index(begin = matrix_bd_11_begin_0, end = matrix_bd_11_end_0, end_mask = matrix_bd_11_end_mask_0, x = matrix_bd_9_cast_fp16)[name = tensor("matrix_bd_11_cast_fp16")]; + tensor var_929_cast_fp16 = add(x = matrix_ac_5_cast_fp16, y = matrix_bd_11_cast_fp16)[name = tensor("op_929_cast_fp16")]; + tensor _inversed_scores_9_y_0_to_fp16 = const()[name = tensor("_inversed_scores_9_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_9_cast_fp16 = mul(x = var_929_cast_fp16, y = _inversed_scores_9_y_0_to_fp16)[name = tensor("_inversed_scores_9_cast_fp16")]; + tensor scores_11_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_9_cast_fp16, cond = mask_11)[name = tensor("scores_11_cast_fp16")]; + tensor var_935_cast_fp16 = softmax(axis = var_56, x = scores_11_cast_fp16)[name = tensor("op_935_cast_fp16")]; + tensor input_145_cast_fp16 = select(a = var_40_to_fp16, b = var_935_cast_fp16, cond = mask_11)[name = tensor("input_145_cast_fp16")]; + tensor x_65_transpose_x_0 = const()[name = tensor("x_65_transpose_x_0"), val = tensor(false)]; + tensor x_65_transpose_y_0 = const()[name = tensor("x_65_transpose_y_0"), val = tensor(false)]; + tensor value_7_cast_fp16 = transpose(perm = value_7_perm_0, x = v_5_cast_fp16)[name = tensor("transpose_220")]; + tensor x_65_cast_fp16 = matmul(transpose_x = x_65_transpose_x_0, transpose_y = x_65_transpose_y_0, x = input_145_cast_fp16, y = value_7_cast_fp16)[name = tensor("x_65_cast_fp16")]; + tensor var_939_perm_0 = const()[name = tensor("op_939_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_940 = const()[name = tensor("op_940"), val = tensor([1, -1, 512])]; + tensor var_939_cast_fp16 = transpose(perm = var_939_perm_0, x = x_65_cast_fp16)[name = tensor("transpose_216")]; + tensor input_147_cast_fp16 = reshape(shape = var_940, x = var_939_cast_fp16)[name = tensor("input_147_cast_fp16")]; + tensor encoder_layers_2_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_2_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35160256)))]; + tensor linear_25_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_self_attn_linear_out_weight_to_fp16, x = input_147_cast_fp16)[name = tensor("linear_25_cast_fp16")]; + tensor input_151_cast_fp16 = add(x = input_141_cast_fp16, y = linear_25_cast_fp16)[name = tensor("input_151_cast_fp16")]; + tensor x_69_axes_0 = const()[name = tensor("x_69_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35684608)))]; + tensor encoder_layers_2_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35685696)))]; + tensor x_69_cast_fp16 = layer_norm(axes = x_69_axes_0, beta = encoder_layers_2_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_norm_conv_weight_to_fp16, x = input_151_cast_fp16)[name = tensor("x_69_cast_fp16")]; + tensor input_153_perm_0 = const()[name = tensor("input_153_perm_0"), val = tensor([0, 2, 1])]; + tensor input_155_pad_type_0 = const()[name = tensor("input_155_pad_type_0"), val = tensor("valid")]; + tensor input_155_strides_0 = const()[name = tensor("input_155_strides_0"), val = tensor([1])]; + tensor input_155_pad_0 = const()[name = tensor("input_155_pad_0"), val = tensor([0, 0])]; + tensor input_155_dilations_0 = const()[name = tensor("input_155_dilations_0"), val = tensor([1])]; + tensor input_155_groups_0 = const()[name = tensor("input_155_groups_0"), val = tensor(1)]; + tensor encoder_layers_2_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_2_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35686784)))]; + tensor input_153_cast_fp16 = transpose(perm = input_153_perm_0, x = x_69_cast_fp16)[name = tensor("transpose_215")]; + tensor input_155_cast_fp16 = conv(dilations = input_155_dilations_0, groups = input_155_groups_0, pad = input_155_pad_0, pad_type = input_155_pad_type_0, strides = input_155_strides_0, weight = encoder_layers_2_conv_pointwise_conv1_weight_to_fp16, x = input_153_cast_fp16)[name = tensor("input_155_cast_fp16")]; + tensor x_71_split_num_splits_0 = const()[name = tensor("x_71_split_num_splits_0"), val = tensor(2)]; + tensor x_71_split_axis_0 = const()[name = tensor("x_71_split_axis_0"), val = tensor(1)]; + tensor x_71_split_cast_fp16_0, tensor x_71_split_cast_fp16_1 = split(axis = x_71_split_axis_0, num_splits = x_71_split_num_splits_0, x = input_155_cast_fp16)[name = tensor("x_71_split_cast_fp16")]; + tensor x_71_split_1_sigmoid_cast_fp16 = sigmoid(x = x_71_split_cast_fp16_1)[name = tensor("x_71_split_1_sigmoid_cast_fp16")]; + tensor x_71_cast_fp16 = mul(x = x_71_split_cast_fp16_0, y = x_71_split_1_sigmoid_cast_fp16)[name = tensor("x_71_cast_fp16")]; + tensor input_157_cast_fp16 = select(a = var_40_to_fp16, b = x_71_cast_fp16, cond = var_551)[name = tensor("input_157_cast_fp16")]; + tensor new_x_11_interleave_0 = const()[name = tensor("new_x_11_interleave_0"), val = tensor(false)]; + tensor new_x_11_cast_fp16 = concat(axis = var_56, interleave = new_x_11_interleave_0, values = (cache_11_cast_fp16, input_157_cast_fp16))[name = tensor("new_x_11_cast_fp16")]; + tensor next_cache_5_begin_0 = const()[name = tensor("next_cache_5_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_5_end_0 = const()[name = tensor("next_cache_5_end_0"), val = tensor([1, 512, 12])]; + tensor next_cache_5_end_mask_0 = const()[name = tensor("next_cache_5_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_5_cast_fp16 = slice_by_index(begin = next_cache_5_begin_0, end = next_cache_5_end_0, end_mask = next_cache_5_end_mask_0, x = new_x_11_cast_fp16)[name = tensor("next_cache_5_cast_fp16")]; + tensor var_981_begin_0 = const()[name = tensor("op_981_begin_0"), val = tensor([0, 0, 4])]; + tensor var_981_end_0 = const()[name = tensor("op_981_end_0"), val = tensor([1, 512, 12])]; + tensor var_981_end_mask_0 = const()[name = tensor("op_981_end_mask_0"), val = tensor([true, true, true])]; + tensor var_981_cast_fp16 = slice_by_index(begin = var_981_begin_0, end = var_981_end_0, end_mask = var_981_end_mask_0, x = next_cache_5_cast_fp16)[name = tensor("op_981_cast_fp16")]; + tensor x_73_pad_type_0 = const()[name = tensor("x_73_pad_type_0"), val = tensor("valid")]; + tensor x_73_groups_0 = const()[name = tensor("x_73_groups_0"), val = tensor(512)]; + tensor x_73_strides_0 = const()[name = tensor("x_73_strides_0"), val = tensor([1])]; + tensor x_73_pad_0 = const()[name = tensor("x_73_pad_0"), val = tensor([0, 0])]; + tensor x_73_dilations_0 = const()[name = tensor("x_73_dilations_0"), val = tensor([1])]; + tensor encoder_layers_2_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_2_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36735424)))]; + tensor x_73_cast_fp16 = conv(dilations = x_73_dilations_0, groups = x_73_groups_0, pad = x_73_pad_0, pad_type = x_73_pad_type_0, strides = x_73_strides_0, weight = encoder_layers_2_conv_depthwise_conv_weight_to_fp16, x = new_x_11_cast_fp16)[name = tensor("x_73_cast_fp16")]; + tensor input_159_perm_0 = const()[name = tensor("input_159_perm_0"), val = tensor([0, 2, 1])]; + tensor x_75_axes_0 = const()[name = tensor("x_75_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_2_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36744704)))]; + tensor encoder_layers_2_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_2_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36745792)))]; + tensor input_159_cast_fp16 = transpose(perm = input_159_perm_0, x = x_73_cast_fp16)[name = tensor("transpose_214")]; + tensor x_75_cast_fp16 = layer_norm(axes = x_75_axes_0, beta = encoder_layers_2_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_conv_batch_norm_weight_to_fp16, x = input_159_cast_fp16)[name = tensor("x_75_cast_fp16")]; + tensor input_161_perm_0 = const()[name = tensor("input_161_perm_0"), val = tensor([0, 2, 1])]; + tensor input_161_cast_fp16 = transpose(perm = input_161_perm_0, x = x_75_cast_fp16)[name = tensor("transpose_213")]; + tensor input_163_cast_fp16 = silu(x = input_161_cast_fp16)[name = tensor("input_163_cast_fp16")]; + tensor x_77_pad_type_0 = const()[name = tensor("x_77_pad_type_0"), val = tensor("valid")]; + tensor x_77_strides_0 = const()[name = tensor("x_77_strides_0"), val = tensor([1])]; + tensor x_77_pad_0 = const()[name = tensor("x_77_pad_0"), val = tensor([0, 0])]; + tensor x_77_dilations_0 = const()[name = tensor("x_77_dilations_0"), val = tensor([1])]; + tensor x_77_groups_0 = const()[name = tensor("x_77_groups_0"), val = tensor(1)]; + tensor encoder_layers_2_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_2_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36746880)))]; + tensor x_77_cast_fp16 = conv(dilations = x_77_dilations_0, groups = x_77_groups_0, pad = x_77_pad_0, pad_type = x_77_pad_type_0, strides = x_77_strides_0, weight = encoder_layers_2_conv_pointwise_conv2_weight_to_fp16, x = input_163_cast_fp16)[name = tensor("x_77_cast_fp16")]; + tensor input_165_perm_0 = const()[name = tensor("input_165_perm_0"), val = tensor([0, 2, 1])]; + tensor input_165_cast_fp16 = transpose(perm = input_165_perm_0, x = x_77_cast_fp16)[name = tensor("transpose_212")]; + tensor input_167_cast_fp16 = add(x = input_151_cast_fp16, y = input_165_cast_fp16)[name = tensor("input_167_cast_fp16")]; + tensor input_169_axes_0 = const()[name = tensor("input_169_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37271232)))]; + tensor encoder_layers_2_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37272320)))]; + tensor input_169_cast_fp16 = layer_norm(axes = input_169_axes_0, beta = encoder_layers_2_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_norm_feed_forward2_weight_to_fp16, x = input_167_cast_fp16)[name = tensor("input_169_cast_fp16")]; + tensor encoder_layers_2_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_2_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37273408)))]; + tensor linear_26_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_2_feed_forward2_linear1_weight_to_fp16, x = input_169_cast_fp16)[name = tensor("linear_26_cast_fp16")]; + tensor input_173_cast_fp16 = silu(x = linear_26_cast_fp16)[name = tensor("input_173_cast_fp16")]; + tensor encoder_layers_2_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_2_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39370624)))]; + tensor linear_27_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_2_feed_forward2_linear2_weight_to_fp16, x = input_173_cast_fp16)[name = tensor("linear_27_cast_fp16")]; + tensor var_1022_to_fp16 = const()[name = tensor("op_1022_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1023_cast_fp16 = mul(x = linear_27_cast_fp16, y = var_1022_to_fp16)[name = tensor("op_1023_cast_fp16")]; + tensor input_179_cast_fp16 = add(x = input_167_cast_fp16, y = var_1023_cast_fp16)[name = tensor("input_179_cast_fp16")]; + tensor input_181_axes_0 = const()[name = tensor("input_181_axes_0"), val = tensor([-1])]; + tensor encoder_layers_2_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_2_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41467840)))]; + tensor encoder_layers_2_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_2_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41468928)))]; + tensor input_181_cast_fp16 = layer_norm(axes = input_181_axes_0, beta = encoder_layers_2_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_2_norm_out_weight_to_fp16, x = input_179_cast_fp16)[name = tensor("input_181_cast_fp16")]; + tensor cache_13_begin_0 = const()[name = tensor("cache_13_begin_0"), val = tensor([3, 0, 0, 0])]; + tensor cache_13_end_0 = const()[name = tensor("cache_13_end_0"), val = tensor([4, 1, 70, 512])]; + tensor cache_13_end_mask_0 = const()[name = tensor("cache_13_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_13_squeeze_mask_0 = const()[name = tensor("cache_13_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_13_cast_fp16 = slice_by_index(begin = cache_13_begin_0, end = cache_13_end_0, end_mask = cache_13_end_mask_0, squeeze_mask = cache_13_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_13_cast_fp16")]; + tensor cache_15_begin_0 = const()[name = tensor("cache_15_begin_0"), val = tensor([3, 0, 0, 0])]; + tensor cache_15_end_0 = const()[name = tensor("cache_15_end_0"), val = tensor([4, 1, 512, 8])]; + tensor cache_15_end_mask_0 = const()[name = tensor("cache_15_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_15_squeeze_mask_0 = const()[name = tensor("cache_15_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_15_cast_fp16 = slice_by_index(begin = cache_15_begin_0, end = cache_15_end_0, end_mask = cache_15_end_mask_0, squeeze_mask = cache_15_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_15_cast_fp16")]; + tensor input_183_axes_0 = const()[name = tensor("input_183_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41470016)))]; + tensor encoder_layers_3_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41471104)))]; + tensor input_183_cast_fp16 = layer_norm(axes = input_183_axes_0, beta = encoder_layers_3_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_norm_feed_forward1_weight_to_fp16, x = input_181_cast_fp16)[name = tensor("input_183_cast_fp16")]; + tensor encoder_layers_3_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_3_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41472192)))]; + tensor linear_28_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_3_feed_forward1_linear1_weight_to_fp16, x = input_183_cast_fp16)[name = tensor("linear_28_cast_fp16")]; + tensor input_187_cast_fp16 = silu(x = linear_28_cast_fp16)[name = tensor("input_187_cast_fp16")]; + tensor encoder_layers_3_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_3_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43569408)))]; + tensor linear_29_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_feed_forward1_linear2_weight_to_fp16, x = input_187_cast_fp16)[name = tensor("linear_29_cast_fp16")]; + tensor var_1057_to_fp16 = const()[name = tensor("op_1057_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1058_cast_fp16 = mul(x = linear_29_cast_fp16, y = var_1057_to_fp16)[name = tensor("op_1058_cast_fp16")]; + tensor input_193_cast_fp16 = add(x = input_181_cast_fp16, y = var_1058_cast_fp16)[name = tensor("input_193_cast_fp16")]; + tensor key_7_axes_0 = const()[name = tensor("key_7_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45666624)))]; + tensor encoder_layers_3_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45667712)))]; + tensor key_7_cast_fp16 = layer_norm(axes = key_7_axes_0, beta = encoder_layers_3_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_norm_self_att_weight_to_fp16, x = input_193_cast_fp16)[name = tensor("key_7_cast_fp16")]; + tensor input_195_interleave_0 = const()[name = tensor("input_195_interleave_0"), val = tensor(false)]; + tensor input_195_cast_fp16 = concat(axis = var_65, interleave = input_195_interleave_0, values = (cache_13_cast_fp16, key_7_cast_fp16))[name = tensor("input_195_cast_fp16")]; + tensor var_1080_begin_0 = const()[name = tensor("op_1080_begin_0"), val = tensor([0, 4, 0])]; + tensor var_1080_end_0 = const()[name = tensor("op_1080_end_0"), val = tensor([1, 70, 512])]; + tensor var_1080_end_mask_0 = const()[name = tensor("op_1080_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1080_cast_fp16 = slice_by_index(begin = var_1080_begin_0, end = var_1080_end_0, end_mask = var_1080_end_mask_0, x = cache_13_cast_fp16)[name = tensor("op_1080_cast_fp16")]; + tensor var_1083_begin_0 = const()[name = tensor("op_1083_begin_0"), val = tensor([0, 0, 0])]; + tensor var_1083_end_0 = const()[name = tensor("op_1083_end_0"), val = tensor([1, 4, 512])]; + tensor var_1083_end_mask_0 = const()[name = tensor("op_1083_end_mask_0"), val = tensor([true, false, true])]; + tensor var_1083_cast_fp16 = slice_by_index(begin = var_1083_begin_0, end = var_1083_end_0, end_mask = var_1083_end_mask_0, x = key_7_cast_fp16)[name = tensor("op_1083_cast_fp16")]; + tensor var_1086_interleave_0 = const()[name = tensor("op_1086_interleave_0"), val = tensor(false)]; + tensor var_1086_cast_fp16 = concat(axis = var_65, interleave = var_1086_interleave_0, values = (var_1080_cast_fp16, var_1083_cast_fp16))[name = tensor("op_1086_cast_fp16")]; + tensor encoder_layers_3_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45668800)))]; + tensor linear_30_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_self_attn_linear_q_weight_to_fp16, x = key_7_cast_fp16)[name = tensor("linear_30_cast_fp16")]; + tensor var_1090 = const()[name = tensor("op_1090"), val = tensor([1, -1, 8, 64])]; + tensor q_19_cast_fp16 = reshape(shape = var_1090, x = linear_30_cast_fp16)[name = tensor("q_19_cast_fp16")]; + tensor encoder_layers_3_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46193152)))]; + tensor linear_31_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_self_attn_linear_k_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("linear_31_cast_fp16")]; + tensor var_1094 = const()[name = tensor("op_1094"), val = tensor([1, -1, 8, 64])]; + tensor k_13_cast_fp16 = reshape(shape = var_1094, x = linear_31_cast_fp16)[name = tensor("k_13_cast_fp16")]; + tensor encoder_layers_3_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46717504)))]; + tensor linear_32_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_self_attn_linear_v_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("linear_32_cast_fp16")]; + tensor var_1098 = const()[name = tensor("op_1098"), val = tensor([1, -1, 8, 64])]; + tensor v_7_cast_fp16 = reshape(shape = var_1098, x = linear_32_cast_fp16)[name = tensor("v_7_cast_fp16")]; + tensor value_9_perm_0 = const()[name = tensor("value_9_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_3_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47241856)))]; + tensor var_1110_cast_fp16 = add(x = q_19_cast_fp16, y = encoder_layers_3_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1110_cast_fp16")]; + tensor encoder_layers_3_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47242944)))]; + tensor var_1112_cast_fp16 = add(x = q_19_cast_fp16, y = encoder_layers_3_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1112_cast_fp16")]; + tensor q_with_bias_v_7_perm_0 = const()[name = tensor("q_with_bias_v_7_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_85_transpose_x_0 = const()[name = tensor("x_85_transpose_x_0"), val = tensor(false)]; + tensor x_85_transpose_y_0 = const()[name = tensor("x_85_transpose_y_0"), val = tensor(false)]; + tensor var_1114_to_fp16 = const()[name = tensor("op_1114_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47244032)))]; + tensor q_with_bias_v_7_cast_fp16 = transpose(perm = q_with_bias_v_7_perm_0, x = var_1112_cast_fp16)[name = tensor("transpose_210")]; + tensor x_85_cast_fp16 = matmul(transpose_x = x_85_transpose_x_0, transpose_y = x_85_transpose_y_0, x = q_with_bias_v_7_cast_fp16, y = var_1114_to_fp16)[name = tensor("x_85_cast_fp16")]; + tensor x_87_pad_0 = const()[name = tensor("x_87_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_87_mode_0 = const()[name = tensor("x_87_mode_0"), val = tensor("constant")]; + tensor const_118_to_fp16 = const()[name = tensor("const_118_to_fp16"), val = tensor(0x0p+0)]; + tensor x_87_cast_fp16 = pad(constant_val = const_118_to_fp16, mode = x_87_mode_0, pad = x_87_pad_0, x = x_85_cast_fp16)[name = tensor("x_87_cast_fp16")]; + tensor var_1122 = const()[name = tensor("op_1122"), val = tensor([1, 8, -1, 8])]; + tensor x_89_cast_fp16 = reshape(shape = var_1122, x = x_87_cast_fp16)[name = tensor("x_89_cast_fp16")]; + tensor var_1126_begin_0 = const()[name = tensor("op_1126_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1126_end_0 = const()[name = tensor("op_1126_end_0"), val = tensor([1, 8, 156, 8])]; + tensor var_1126_end_mask_0 = const()[name = tensor("op_1126_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1126_cast_fp16 = slice_by_index(begin = var_1126_begin_0, end = var_1126_end_0, end_mask = var_1126_end_mask_0, x = x_89_cast_fp16)[name = tensor("op_1126_cast_fp16")]; + tensor var_1127 = const()[name = tensor("op_1127"), val = tensor([1, 8, 8, 155])]; + tensor matrix_bd_13_cast_fp16 = reshape(shape = var_1127, x = var_1126_cast_fp16)[name = tensor("matrix_bd_13_cast_fp16")]; + tensor matrix_ac_7_transpose_x_0 = const()[name = tensor("matrix_ac_7_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_7_transpose_y_0 = const()[name = tensor("matrix_ac_7_transpose_y_0"), val = tensor(false)]; + tensor transpose_57_perm_0 = const()[name = tensor("transpose_57_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_58_perm_0 = const()[name = tensor("transpose_58_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_58 = transpose(perm = transpose_58_perm_0, x = k_13_cast_fp16)[name = tensor("transpose_208")]; + tensor transpose_57 = transpose(perm = transpose_57_perm_0, x = var_1110_cast_fp16)[name = tensor("transpose_209")]; + tensor matrix_ac_7_cast_fp16 = matmul(transpose_x = matrix_ac_7_transpose_x_0, transpose_y = matrix_ac_7_transpose_y_0, x = transpose_57, y = transpose_58)[name = tensor("matrix_ac_7_cast_fp16")]; + tensor matrix_bd_15_begin_0 = const()[name = tensor("matrix_bd_15_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_15_end_0 = const()[name = tensor("matrix_bd_15_end_0"), val = tensor([1, 8, 8, 78])]; + tensor matrix_bd_15_end_mask_0 = const()[name = tensor("matrix_bd_15_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_15_cast_fp16 = slice_by_index(begin = matrix_bd_15_begin_0, end = matrix_bd_15_end_0, end_mask = matrix_bd_15_end_mask_0, x = matrix_bd_13_cast_fp16)[name = tensor("matrix_bd_15_cast_fp16")]; + tensor var_1136_cast_fp16 = add(x = matrix_ac_7_cast_fp16, y = matrix_bd_15_cast_fp16)[name = tensor("op_1136_cast_fp16")]; + tensor _inversed_scores_13_y_0_to_fp16 = const()[name = tensor("_inversed_scores_13_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_13_cast_fp16 = mul(x = var_1136_cast_fp16, y = _inversed_scores_13_y_0_to_fp16)[name = tensor("_inversed_scores_13_cast_fp16")]; + tensor scores_15_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_13_cast_fp16, cond = mask_11)[name = tensor("scores_15_cast_fp16")]; + tensor var_1142_cast_fp16 = softmax(axis = var_56, x = scores_15_cast_fp16)[name = tensor("op_1142_cast_fp16")]; + tensor input_197_cast_fp16 = select(a = var_40_to_fp16, b = var_1142_cast_fp16, cond = mask_11)[name = tensor("input_197_cast_fp16")]; + tensor x_91_transpose_x_0 = const()[name = tensor("x_91_transpose_x_0"), val = tensor(false)]; + tensor x_91_transpose_y_0 = const()[name = tensor("x_91_transpose_y_0"), val = tensor(false)]; + tensor value_9_cast_fp16 = transpose(perm = value_9_perm_0, x = v_7_cast_fp16)[name = tensor("transpose_211")]; + tensor x_91_cast_fp16 = matmul(transpose_x = x_91_transpose_x_0, transpose_y = x_91_transpose_y_0, x = input_197_cast_fp16, y = value_9_cast_fp16)[name = tensor("x_91_cast_fp16")]; + tensor var_1146_perm_0 = const()[name = tensor("op_1146_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1147 = const()[name = tensor("op_1147"), val = tensor([1, -1, 512])]; + tensor var_1146_cast_fp16 = transpose(perm = var_1146_perm_0, x = x_91_cast_fp16)[name = tensor("transpose_207")]; + tensor input_199_cast_fp16 = reshape(shape = var_1147, x = var_1146_cast_fp16)[name = tensor("input_199_cast_fp16")]; + tensor encoder_layers_3_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_3_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47402816)))]; + tensor linear_34_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_self_attn_linear_out_weight_to_fp16, x = input_199_cast_fp16)[name = tensor("linear_34_cast_fp16")]; + tensor input_203_cast_fp16 = add(x = input_193_cast_fp16, y = linear_34_cast_fp16)[name = tensor("input_203_cast_fp16")]; + tensor x_95_axes_0 = const()[name = tensor("x_95_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47927168)))]; + tensor encoder_layers_3_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47928256)))]; + tensor x_95_cast_fp16 = layer_norm(axes = x_95_axes_0, beta = encoder_layers_3_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_norm_conv_weight_to_fp16, x = input_203_cast_fp16)[name = tensor("x_95_cast_fp16")]; + tensor input_205_perm_0 = const()[name = tensor("input_205_perm_0"), val = tensor([0, 2, 1])]; + tensor input_207_pad_type_0 = const()[name = tensor("input_207_pad_type_0"), val = tensor("valid")]; + tensor input_207_strides_0 = const()[name = tensor("input_207_strides_0"), val = tensor([1])]; + tensor input_207_pad_0 = const()[name = tensor("input_207_pad_0"), val = tensor([0, 0])]; + tensor input_207_dilations_0 = const()[name = tensor("input_207_dilations_0"), val = tensor([1])]; + tensor input_207_groups_0 = const()[name = tensor("input_207_groups_0"), val = tensor(1)]; + tensor encoder_layers_3_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_3_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47929344)))]; + tensor input_205_cast_fp16 = transpose(perm = input_205_perm_0, x = x_95_cast_fp16)[name = tensor("transpose_206")]; + tensor input_207_cast_fp16 = conv(dilations = input_207_dilations_0, groups = input_207_groups_0, pad = input_207_pad_0, pad_type = input_207_pad_type_0, strides = input_207_strides_0, weight = encoder_layers_3_conv_pointwise_conv1_weight_to_fp16, x = input_205_cast_fp16)[name = tensor("input_207_cast_fp16")]; + tensor x_97_split_num_splits_0 = const()[name = tensor("x_97_split_num_splits_0"), val = tensor(2)]; + tensor x_97_split_axis_0 = const()[name = tensor("x_97_split_axis_0"), val = tensor(1)]; + tensor x_97_split_cast_fp16_0, tensor x_97_split_cast_fp16_1 = split(axis = x_97_split_axis_0, num_splits = x_97_split_num_splits_0, x = input_207_cast_fp16)[name = tensor("x_97_split_cast_fp16")]; + tensor x_97_split_1_sigmoid_cast_fp16 = sigmoid(x = x_97_split_cast_fp16_1)[name = tensor("x_97_split_1_sigmoid_cast_fp16")]; + tensor x_97_cast_fp16 = mul(x = x_97_split_cast_fp16_0, y = x_97_split_1_sigmoid_cast_fp16)[name = tensor("x_97_cast_fp16")]; + tensor input_209_cast_fp16 = select(a = var_40_to_fp16, b = x_97_cast_fp16, cond = var_551)[name = tensor("input_209_cast_fp16")]; + tensor new_x_15_interleave_0 = const()[name = tensor("new_x_15_interleave_0"), val = tensor(false)]; + tensor new_x_15_cast_fp16 = concat(axis = var_56, interleave = new_x_15_interleave_0, values = (cache_15_cast_fp16, input_209_cast_fp16))[name = tensor("new_x_15_cast_fp16")]; + tensor next_cache_7_begin_0 = const()[name = tensor("next_cache_7_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_7_end_0 = const()[name = tensor("next_cache_7_end_0"), val = tensor([1, 512, 12])]; + tensor next_cache_7_end_mask_0 = const()[name = tensor("next_cache_7_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_7_cast_fp16 = slice_by_index(begin = next_cache_7_begin_0, end = next_cache_7_end_0, end_mask = next_cache_7_end_mask_0, x = new_x_15_cast_fp16)[name = tensor("next_cache_7_cast_fp16")]; + tensor var_1188_begin_0 = const()[name = tensor("op_1188_begin_0"), val = tensor([0, 0, 4])]; + tensor var_1188_end_0 = const()[name = tensor("op_1188_end_0"), val = tensor([1, 512, 12])]; + tensor var_1188_end_mask_0 = const()[name = tensor("op_1188_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1188_cast_fp16 = slice_by_index(begin = var_1188_begin_0, end = var_1188_end_0, end_mask = var_1188_end_mask_0, x = next_cache_7_cast_fp16)[name = tensor("op_1188_cast_fp16")]; + tensor x_99_pad_type_0 = const()[name = tensor("x_99_pad_type_0"), val = tensor("valid")]; + tensor x_99_groups_0 = const()[name = tensor("x_99_groups_0"), val = tensor(512)]; + tensor x_99_strides_0 = const()[name = tensor("x_99_strides_0"), val = tensor([1])]; + tensor x_99_pad_0 = const()[name = tensor("x_99_pad_0"), val = tensor([0, 0])]; + tensor x_99_dilations_0 = const()[name = tensor("x_99_dilations_0"), val = tensor([1])]; + tensor encoder_layers_3_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_3_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48977984)))]; + tensor x_99_cast_fp16 = conv(dilations = x_99_dilations_0, groups = x_99_groups_0, pad = x_99_pad_0, pad_type = x_99_pad_type_0, strides = x_99_strides_0, weight = encoder_layers_3_conv_depthwise_conv_weight_to_fp16, x = new_x_15_cast_fp16)[name = tensor("x_99_cast_fp16")]; + tensor input_211_perm_0 = const()[name = tensor("input_211_perm_0"), val = tensor([0, 2, 1])]; + tensor x_101_axes_0 = const()[name = tensor("x_101_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_3_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48987264)))]; + tensor encoder_layers_3_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_3_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48988352)))]; + tensor input_211_cast_fp16 = transpose(perm = input_211_perm_0, x = x_99_cast_fp16)[name = tensor("transpose_205")]; + tensor x_101_cast_fp16 = layer_norm(axes = x_101_axes_0, beta = encoder_layers_3_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_conv_batch_norm_weight_to_fp16, x = input_211_cast_fp16)[name = tensor("x_101_cast_fp16")]; + tensor input_213_perm_0 = const()[name = tensor("input_213_perm_0"), val = tensor([0, 2, 1])]; + tensor input_213_cast_fp16 = transpose(perm = input_213_perm_0, x = x_101_cast_fp16)[name = tensor("transpose_204")]; + tensor input_215_cast_fp16 = silu(x = input_213_cast_fp16)[name = tensor("input_215_cast_fp16")]; + tensor x_103_pad_type_0 = const()[name = tensor("x_103_pad_type_0"), val = tensor("valid")]; + tensor x_103_strides_0 = const()[name = tensor("x_103_strides_0"), val = tensor([1])]; + tensor x_103_pad_0 = const()[name = tensor("x_103_pad_0"), val = tensor([0, 0])]; + tensor x_103_dilations_0 = const()[name = tensor("x_103_dilations_0"), val = tensor([1])]; + tensor x_103_groups_0 = const()[name = tensor("x_103_groups_0"), val = tensor(1)]; + tensor encoder_layers_3_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_3_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48989440)))]; + tensor x_103_cast_fp16 = conv(dilations = x_103_dilations_0, groups = x_103_groups_0, pad = x_103_pad_0, pad_type = x_103_pad_type_0, strides = x_103_strides_0, weight = encoder_layers_3_conv_pointwise_conv2_weight_to_fp16, x = input_215_cast_fp16)[name = tensor("x_103_cast_fp16")]; + tensor input_217_perm_0 = const()[name = tensor("input_217_perm_0"), val = tensor([0, 2, 1])]; + tensor input_217_cast_fp16 = transpose(perm = input_217_perm_0, x = x_103_cast_fp16)[name = tensor("transpose_203")]; + tensor input_219_cast_fp16 = add(x = input_203_cast_fp16, y = input_217_cast_fp16)[name = tensor("input_219_cast_fp16")]; + tensor input_221_axes_0 = const()[name = tensor("input_221_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49513792)))]; + tensor encoder_layers_3_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49514880)))]; + tensor input_221_cast_fp16 = layer_norm(axes = input_221_axes_0, beta = encoder_layers_3_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_norm_feed_forward2_weight_to_fp16, x = input_219_cast_fp16)[name = tensor("input_221_cast_fp16")]; + tensor encoder_layers_3_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_3_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49515968)))]; + tensor linear_35_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_3_feed_forward2_linear1_weight_to_fp16, x = input_221_cast_fp16)[name = tensor("linear_35_cast_fp16")]; + tensor input_225_cast_fp16 = silu(x = linear_35_cast_fp16)[name = tensor("input_225_cast_fp16")]; + tensor encoder_layers_3_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_3_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51613184)))]; + tensor linear_36_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_3_feed_forward2_linear2_weight_to_fp16, x = input_225_cast_fp16)[name = tensor("linear_36_cast_fp16")]; + tensor var_1229_to_fp16 = const()[name = tensor("op_1229_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1230_cast_fp16 = mul(x = linear_36_cast_fp16, y = var_1229_to_fp16)[name = tensor("op_1230_cast_fp16")]; + tensor input_231_cast_fp16 = add(x = input_219_cast_fp16, y = var_1230_cast_fp16)[name = tensor("input_231_cast_fp16")]; + tensor input_233_axes_0 = const()[name = tensor("input_233_axes_0"), val = tensor([-1])]; + tensor encoder_layers_3_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_3_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53710400)))]; + tensor encoder_layers_3_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_3_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53711488)))]; + tensor input_233_cast_fp16 = layer_norm(axes = input_233_axes_0, beta = encoder_layers_3_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_3_norm_out_weight_to_fp16, x = input_231_cast_fp16)[name = tensor("input_233_cast_fp16")]; + tensor cache_17_begin_0 = const()[name = tensor("cache_17_begin_0"), val = tensor([4, 0, 0, 0])]; + tensor cache_17_end_0 = const()[name = tensor("cache_17_end_0"), val = tensor([5, 1, 70, 512])]; + tensor cache_17_end_mask_0 = const()[name = tensor("cache_17_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_17_squeeze_mask_0 = const()[name = tensor("cache_17_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_17_cast_fp16 = slice_by_index(begin = cache_17_begin_0, end = cache_17_end_0, end_mask = cache_17_end_mask_0, squeeze_mask = cache_17_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_17_cast_fp16")]; + tensor cache_19_begin_0 = const()[name = tensor("cache_19_begin_0"), val = tensor([4, 0, 0, 0])]; + tensor cache_19_end_0 = const()[name = tensor("cache_19_end_0"), val = tensor([5, 1, 512, 8])]; + tensor cache_19_end_mask_0 = const()[name = tensor("cache_19_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_19_squeeze_mask_0 = const()[name = tensor("cache_19_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_19_cast_fp16 = slice_by_index(begin = cache_19_begin_0, end = cache_19_end_0, end_mask = cache_19_end_mask_0, squeeze_mask = cache_19_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_19_cast_fp16")]; + tensor input_235_axes_0 = const()[name = tensor("input_235_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53712576)))]; + tensor encoder_layers_4_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53713664)))]; + tensor input_235_cast_fp16 = layer_norm(axes = input_235_axes_0, beta = encoder_layers_4_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_norm_feed_forward1_weight_to_fp16, x = input_233_cast_fp16)[name = tensor("input_235_cast_fp16")]; + tensor encoder_layers_4_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_4_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53714752)))]; + tensor linear_37_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_4_feed_forward1_linear1_weight_to_fp16, x = input_235_cast_fp16)[name = tensor("linear_37_cast_fp16")]; + tensor input_239_cast_fp16 = silu(x = linear_37_cast_fp16)[name = tensor("input_239_cast_fp16")]; + tensor encoder_layers_4_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_4_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55811968)))]; + tensor linear_38_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_feed_forward1_linear2_weight_to_fp16, x = input_239_cast_fp16)[name = tensor("linear_38_cast_fp16")]; + tensor var_1264_to_fp16 = const()[name = tensor("op_1264_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1265_cast_fp16 = mul(x = linear_38_cast_fp16, y = var_1264_to_fp16)[name = tensor("op_1265_cast_fp16")]; + tensor input_245_cast_fp16 = add(x = input_233_cast_fp16, y = var_1265_cast_fp16)[name = tensor("input_245_cast_fp16")]; + tensor key_9_axes_0 = const()[name = tensor("key_9_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57909184)))]; + tensor encoder_layers_4_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57910272)))]; + tensor key_9_cast_fp16 = layer_norm(axes = key_9_axes_0, beta = encoder_layers_4_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_norm_self_att_weight_to_fp16, x = input_245_cast_fp16)[name = tensor("key_9_cast_fp16")]; + tensor input_247_interleave_0 = const()[name = tensor("input_247_interleave_0"), val = tensor(false)]; + tensor input_247_cast_fp16 = concat(axis = var_65, interleave = input_247_interleave_0, values = (cache_17_cast_fp16, key_9_cast_fp16))[name = tensor("input_247_cast_fp16")]; + tensor var_1287_begin_0 = const()[name = tensor("op_1287_begin_0"), val = tensor([0, 4, 0])]; + tensor var_1287_end_0 = const()[name = tensor("op_1287_end_0"), val = tensor([1, 70, 512])]; + tensor var_1287_end_mask_0 = const()[name = tensor("op_1287_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1287_cast_fp16 = slice_by_index(begin = var_1287_begin_0, end = var_1287_end_0, end_mask = var_1287_end_mask_0, x = cache_17_cast_fp16)[name = tensor("op_1287_cast_fp16")]; + tensor var_1290_begin_0 = const()[name = tensor("op_1290_begin_0"), val = tensor([0, 0, 0])]; + tensor var_1290_end_0 = const()[name = tensor("op_1290_end_0"), val = tensor([1, 4, 512])]; + tensor var_1290_end_mask_0 = const()[name = tensor("op_1290_end_mask_0"), val = tensor([true, false, true])]; + tensor var_1290_cast_fp16 = slice_by_index(begin = var_1290_begin_0, end = var_1290_end_0, end_mask = var_1290_end_mask_0, x = key_9_cast_fp16)[name = tensor("op_1290_cast_fp16")]; + tensor var_1293_interleave_0 = const()[name = tensor("op_1293_interleave_0"), val = tensor(false)]; + tensor var_1293_cast_fp16 = concat(axis = var_65, interleave = var_1293_interleave_0, values = (var_1287_cast_fp16, var_1290_cast_fp16))[name = tensor("op_1293_cast_fp16")]; + tensor encoder_layers_4_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57911360)))]; + tensor linear_39_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_self_attn_linear_q_weight_to_fp16, x = key_9_cast_fp16)[name = tensor("linear_39_cast_fp16")]; + tensor var_1297 = const()[name = tensor("op_1297"), val = tensor([1, -1, 8, 64])]; + tensor q_25_cast_fp16 = reshape(shape = var_1297, x = linear_39_cast_fp16)[name = tensor("q_25_cast_fp16")]; + tensor encoder_layers_4_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58435712)))]; + tensor linear_40_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_self_attn_linear_k_weight_to_fp16, x = input_247_cast_fp16)[name = tensor("linear_40_cast_fp16")]; + tensor var_1301 = const()[name = tensor("op_1301"), val = tensor([1, -1, 8, 64])]; + tensor k_17_cast_fp16 = reshape(shape = var_1301, x = linear_40_cast_fp16)[name = tensor("k_17_cast_fp16")]; + tensor encoder_layers_4_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58960064)))]; + tensor linear_41_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_self_attn_linear_v_weight_to_fp16, x = input_247_cast_fp16)[name = tensor("linear_41_cast_fp16")]; + tensor var_1305 = const()[name = tensor("op_1305"), val = tensor([1, -1, 8, 64])]; + tensor v_9_cast_fp16 = reshape(shape = var_1305, x = linear_41_cast_fp16)[name = tensor("v_9_cast_fp16")]; + tensor value_11_perm_0 = const()[name = tensor("value_11_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_4_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59484416)))]; + tensor var_1317_cast_fp16 = add(x = q_25_cast_fp16, y = encoder_layers_4_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1317_cast_fp16")]; + tensor encoder_layers_4_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59485504)))]; + tensor var_1319_cast_fp16 = add(x = q_25_cast_fp16, y = encoder_layers_4_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1319_cast_fp16")]; + tensor q_with_bias_v_9_perm_0 = const()[name = tensor("q_with_bias_v_9_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_111_transpose_x_0 = const()[name = tensor("x_111_transpose_x_0"), val = tensor(false)]; + tensor x_111_transpose_y_0 = const()[name = tensor("x_111_transpose_y_0"), val = tensor(false)]; + tensor var_1321_to_fp16 = const()[name = tensor("op_1321_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59486592)))]; + tensor q_with_bias_v_9_cast_fp16 = transpose(perm = q_with_bias_v_9_perm_0, x = var_1319_cast_fp16)[name = tensor("transpose_201")]; + tensor x_111_cast_fp16 = matmul(transpose_x = x_111_transpose_x_0, transpose_y = x_111_transpose_y_0, x = q_with_bias_v_9_cast_fp16, y = var_1321_to_fp16)[name = tensor("x_111_cast_fp16")]; + tensor x_113_pad_0 = const()[name = tensor("x_113_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_113_mode_0 = const()[name = tensor("x_113_mode_0"), val = tensor("constant")]; + tensor const_131_to_fp16 = const()[name = tensor("const_131_to_fp16"), val = tensor(0x0p+0)]; + tensor x_113_cast_fp16 = pad(constant_val = const_131_to_fp16, mode = x_113_mode_0, pad = x_113_pad_0, x = x_111_cast_fp16)[name = tensor("x_113_cast_fp16")]; + tensor var_1329 = const()[name = tensor("op_1329"), val = tensor([1, 8, -1, 8])]; + tensor x_115_cast_fp16 = reshape(shape = var_1329, x = x_113_cast_fp16)[name = tensor("x_115_cast_fp16")]; + tensor var_1333_begin_0 = const()[name = tensor("op_1333_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1333_end_0 = const()[name = tensor("op_1333_end_0"), val = tensor([1, 8, 156, 8])]; + tensor var_1333_end_mask_0 = const()[name = tensor("op_1333_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1333_cast_fp16 = slice_by_index(begin = var_1333_begin_0, end = var_1333_end_0, end_mask = var_1333_end_mask_0, x = x_115_cast_fp16)[name = tensor("op_1333_cast_fp16")]; + tensor var_1334 = const()[name = tensor("op_1334"), val = tensor([1, 8, 8, 155])]; + tensor matrix_bd_17_cast_fp16 = reshape(shape = var_1334, x = var_1333_cast_fp16)[name = tensor("matrix_bd_17_cast_fp16")]; + tensor matrix_ac_9_transpose_x_0 = const()[name = tensor("matrix_ac_9_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_9_transpose_y_0 = const()[name = tensor("matrix_ac_9_transpose_y_0"), val = tensor(false)]; + tensor transpose_59_perm_0 = const()[name = tensor("transpose_59_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_60_perm_0 = const()[name = tensor("transpose_60_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_60 = transpose(perm = transpose_60_perm_0, x = k_17_cast_fp16)[name = tensor("transpose_199")]; + tensor transpose_59 = transpose(perm = transpose_59_perm_0, x = var_1317_cast_fp16)[name = tensor("transpose_200")]; + tensor matrix_ac_9_cast_fp16 = matmul(transpose_x = matrix_ac_9_transpose_x_0, transpose_y = matrix_ac_9_transpose_y_0, x = transpose_59, y = transpose_60)[name = tensor("matrix_ac_9_cast_fp16")]; + tensor matrix_bd_19_begin_0 = const()[name = tensor("matrix_bd_19_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_19_end_0 = const()[name = tensor("matrix_bd_19_end_0"), val = tensor([1, 8, 8, 78])]; + tensor matrix_bd_19_end_mask_0 = const()[name = tensor("matrix_bd_19_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_19_cast_fp16 = slice_by_index(begin = matrix_bd_19_begin_0, end = matrix_bd_19_end_0, end_mask = matrix_bd_19_end_mask_0, x = matrix_bd_17_cast_fp16)[name = tensor("matrix_bd_19_cast_fp16")]; + tensor var_1343_cast_fp16 = add(x = matrix_ac_9_cast_fp16, y = matrix_bd_19_cast_fp16)[name = tensor("op_1343_cast_fp16")]; + tensor _inversed_scores_17_y_0_to_fp16 = const()[name = tensor("_inversed_scores_17_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_17_cast_fp16 = mul(x = var_1343_cast_fp16, y = _inversed_scores_17_y_0_to_fp16)[name = tensor("_inversed_scores_17_cast_fp16")]; + tensor scores_19_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_17_cast_fp16, cond = mask_11)[name = tensor("scores_19_cast_fp16")]; + tensor var_1349_cast_fp16 = softmax(axis = var_56, x = scores_19_cast_fp16)[name = tensor("op_1349_cast_fp16")]; + tensor input_249_cast_fp16 = select(a = var_40_to_fp16, b = var_1349_cast_fp16, cond = mask_11)[name = tensor("input_249_cast_fp16")]; + tensor x_117_transpose_x_0 = const()[name = tensor("x_117_transpose_x_0"), val = tensor(false)]; + tensor x_117_transpose_y_0 = const()[name = tensor("x_117_transpose_y_0"), val = tensor(false)]; + tensor value_11_cast_fp16 = transpose(perm = value_11_perm_0, x = v_9_cast_fp16)[name = tensor("transpose_202")]; + tensor x_117_cast_fp16 = matmul(transpose_x = x_117_transpose_x_0, transpose_y = x_117_transpose_y_0, x = input_249_cast_fp16, y = value_11_cast_fp16)[name = tensor("x_117_cast_fp16")]; + tensor var_1353_perm_0 = const()[name = tensor("op_1353_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1354 = const()[name = tensor("op_1354"), val = tensor([1, -1, 512])]; + tensor var_1353_cast_fp16 = transpose(perm = var_1353_perm_0, x = x_117_cast_fp16)[name = tensor("transpose_198")]; + tensor input_251_cast_fp16 = reshape(shape = var_1354, x = var_1353_cast_fp16)[name = tensor("input_251_cast_fp16")]; + tensor encoder_layers_4_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_4_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59645376)))]; + tensor linear_43_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_self_attn_linear_out_weight_to_fp16, x = input_251_cast_fp16)[name = tensor("linear_43_cast_fp16")]; + tensor input_255_cast_fp16 = add(x = input_245_cast_fp16, y = linear_43_cast_fp16)[name = tensor("input_255_cast_fp16")]; + tensor x_121_axes_0 = const()[name = tensor("x_121_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60169728)))]; + tensor encoder_layers_4_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60170816)))]; + tensor x_121_cast_fp16 = layer_norm(axes = x_121_axes_0, beta = encoder_layers_4_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_norm_conv_weight_to_fp16, x = input_255_cast_fp16)[name = tensor("x_121_cast_fp16")]; + tensor input_257_perm_0 = const()[name = tensor("input_257_perm_0"), val = tensor([0, 2, 1])]; + tensor input_259_pad_type_0 = const()[name = tensor("input_259_pad_type_0"), val = tensor("valid")]; + tensor input_259_strides_0 = const()[name = tensor("input_259_strides_0"), val = tensor([1])]; + tensor input_259_pad_0 = const()[name = tensor("input_259_pad_0"), val = tensor([0, 0])]; + tensor input_259_dilations_0 = const()[name = tensor("input_259_dilations_0"), val = tensor([1])]; + tensor input_259_groups_0 = const()[name = tensor("input_259_groups_0"), val = tensor(1)]; + tensor encoder_layers_4_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_4_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60171904)))]; + tensor input_257_cast_fp16 = transpose(perm = input_257_perm_0, x = x_121_cast_fp16)[name = tensor("transpose_197")]; + tensor input_259_cast_fp16 = conv(dilations = input_259_dilations_0, groups = input_259_groups_0, pad = input_259_pad_0, pad_type = input_259_pad_type_0, strides = input_259_strides_0, weight = encoder_layers_4_conv_pointwise_conv1_weight_to_fp16, x = input_257_cast_fp16)[name = tensor("input_259_cast_fp16")]; + tensor x_123_split_num_splits_0 = const()[name = tensor("x_123_split_num_splits_0"), val = tensor(2)]; + tensor x_123_split_axis_0 = const()[name = tensor("x_123_split_axis_0"), val = tensor(1)]; + tensor x_123_split_cast_fp16_0, tensor x_123_split_cast_fp16_1 = split(axis = x_123_split_axis_0, num_splits = x_123_split_num_splits_0, x = input_259_cast_fp16)[name = tensor("x_123_split_cast_fp16")]; + tensor x_123_split_1_sigmoid_cast_fp16 = sigmoid(x = x_123_split_cast_fp16_1)[name = tensor("x_123_split_1_sigmoid_cast_fp16")]; + tensor x_123_cast_fp16 = mul(x = x_123_split_cast_fp16_0, y = x_123_split_1_sigmoid_cast_fp16)[name = tensor("x_123_cast_fp16")]; + tensor input_261_cast_fp16 = select(a = var_40_to_fp16, b = x_123_cast_fp16, cond = var_551)[name = tensor("input_261_cast_fp16")]; + tensor new_x_19_interleave_0 = const()[name = tensor("new_x_19_interleave_0"), val = tensor(false)]; + tensor new_x_19_cast_fp16 = concat(axis = var_56, interleave = new_x_19_interleave_0, values = (cache_19_cast_fp16, input_261_cast_fp16))[name = tensor("new_x_19_cast_fp16")]; + tensor next_cache_9_begin_0 = const()[name = tensor("next_cache_9_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_9_end_0 = const()[name = tensor("next_cache_9_end_0"), val = tensor([1, 512, 12])]; + tensor next_cache_9_end_mask_0 = const()[name = tensor("next_cache_9_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_9_cast_fp16 = slice_by_index(begin = next_cache_9_begin_0, end = next_cache_9_end_0, end_mask = next_cache_9_end_mask_0, x = new_x_19_cast_fp16)[name = tensor("next_cache_9_cast_fp16")]; + tensor var_1395_begin_0 = const()[name = tensor("op_1395_begin_0"), val = tensor([0, 0, 4])]; + tensor var_1395_end_0 = const()[name = tensor("op_1395_end_0"), val = tensor([1, 512, 12])]; + tensor var_1395_end_mask_0 = const()[name = tensor("op_1395_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1395_cast_fp16 = slice_by_index(begin = var_1395_begin_0, end = var_1395_end_0, end_mask = var_1395_end_mask_0, x = next_cache_9_cast_fp16)[name = tensor("op_1395_cast_fp16")]; + tensor x_125_pad_type_0 = const()[name = tensor("x_125_pad_type_0"), val = tensor("valid")]; + tensor x_125_groups_0 = const()[name = tensor("x_125_groups_0"), val = tensor(512)]; + tensor x_125_strides_0 = const()[name = tensor("x_125_strides_0"), val = tensor([1])]; + tensor x_125_pad_0 = const()[name = tensor("x_125_pad_0"), val = tensor([0, 0])]; + tensor x_125_dilations_0 = const()[name = tensor("x_125_dilations_0"), val = tensor([1])]; + tensor encoder_layers_4_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_4_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61220544)))]; + tensor x_125_cast_fp16 = conv(dilations = x_125_dilations_0, groups = x_125_groups_0, pad = x_125_pad_0, pad_type = x_125_pad_type_0, strides = x_125_strides_0, weight = encoder_layers_4_conv_depthwise_conv_weight_to_fp16, x = new_x_19_cast_fp16)[name = tensor("x_125_cast_fp16")]; + tensor input_263_perm_0 = const()[name = tensor("input_263_perm_0"), val = tensor([0, 2, 1])]; + tensor x_127_axes_0 = const()[name = tensor("x_127_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_4_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61229824)))]; + tensor encoder_layers_4_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_4_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61230912)))]; + tensor input_263_cast_fp16 = transpose(perm = input_263_perm_0, x = x_125_cast_fp16)[name = tensor("transpose_196")]; + tensor x_127_cast_fp16 = layer_norm(axes = x_127_axes_0, beta = encoder_layers_4_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_conv_batch_norm_weight_to_fp16, x = input_263_cast_fp16)[name = tensor("x_127_cast_fp16")]; + tensor input_265_perm_0 = const()[name = tensor("input_265_perm_0"), val = tensor([0, 2, 1])]; + tensor input_265_cast_fp16 = transpose(perm = input_265_perm_0, x = x_127_cast_fp16)[name = tensor("transpose_195")]; + tensor input_267_cast_fp16 = silu(x = input_265_cast_fp16)[name = tensor("input_267_cast_fp16")]; + tensor x_129_pad_type_0 = const()[name = tensor("x_129_pad_type_0"), val = tensor("valid")]; + tensor x_129_strides_0 = const()[name = tensor("x_129_strides_0"), val = tensor([1])]; + tensor x_129_pad_0 = const()[name = tensor("x_129_pad_0"), val = tensor([0, 0])]; + tensor x_129_dilations_0 = const()[name = tensor("x_129_dilations_0"), val = tensor([1])]; + tensor x_129_groups_0 = const()[name = tensor("x_129_groups_0"), val = tensor(1)]; + tensor encoder_layers_4_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_4_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61232000)))]; + tensor x_129_cast_fp16 = conv(dilations = x_129_dilations_0, groups = x_129_groups_0, pad = x_129_pad_0, pad_type = x_129_pad_type_0, strides = x_129_strides_0, weight = encoder_layers_4_conv_pointwise_conv2_weight_to_fp16, x = input_267_cast_fp16)[name = tensor("x_129_cast_fp16")]; + tensor input_269_perm_0 = const()[name = tensor("input_269_perm_0"), val = tensor([0, 2, 1])]; + tensor input_269_cast_fp16 = transpose(perm = input_269_perm_0, x = x_129_cast_fp16)[name = tensor("transpose_194")]; + tensor input_271_cast_fp16 = add(x = input_255_cast_fp16, y = input_269_cast_fp16)[name = tensor("input_271_cast_fp16")]; + tensor input_273_axes_0 = const()[name = tensor("input_273_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61756352)))]; + tensor encoder_layers_4_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61757440)))]; + tensor input_273_cast_fp16 = layer_norm(axes = input_273_axes_0, beta = encoder_layers_4_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_norm_feed_forward2_weight_to_fp16, x = input_271_cast_fp16)[name = tensor("input_273_cast_fp16")]; + tensor encoder_layers_4_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_4_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61758528)))]; + tensor linear_44_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_4_feed_forward2_linear1_weight_to_fp16, x = input_273_cast_fp16)[name = tensor("linear_44_cast_fp16")]; + tensor input_277_cast_fp16 = silu(x = linear_44_cast_fp16)[name = tensor("input_277_cast_fp16")]; + tensor encoder_layers_4_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_4_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63855744)))]; + tensor linear_45_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_4_feed_forward2_linear2_weight_to_fp16, x = input_277_cast_fp16)[name = tensor("linear_45_cast_fp16")]; + tensor var_1436_to_fp16 = const()[name = tensor("op_1436_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1437_cast_fp16 = mul(x = linear_45_cast_fp16, y = var_1436_to_fp16)[name = tensor("op_1437_cast_fp16")]; + tensor input_283_cast_fp16 = add(x = input_271_cast_fp16, y = var_1437_cast_fp16)[name = tensor("input_283_cast_fp16")]; + tensor input_285_axes_0 = const()[name = tensor("input_285_axes_0"), val = tensor([-1])]; + tensor encoder_layers_4_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_4_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65952960)))]; + tensor encoder_layers_4_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_4_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65954048)))]; + tensor input_285_cast_fp16 = layer_norm(axes = input_285_axes_0, beta = encoder_layers_4_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_4_norm_out_weight_to_fp16, x = input_283_cast_fp16)[name = tensor("input_285_cast_fp16")]; + tensor cache_21_begin_0 = const()[name = tensor("cache_21_begin_0"), val = tensor([5, 0, 0, 0])]; + tensor cache_21_end_0 = const()[name = tensor("cache_21_end_0"), val = tensor([6, 1, 70, 512])]; + tensor cache_21_end_mask_0 = const()[name = tensor("cache_21_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_21_squeeze_mask_0 = const()[name = tensor("cache_21_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_21_cast_fp16 = slice_by_index(begin = cache_21_begin_0, end = cache_21_end_0, end_mask = cache_21_end_mask_0, squeeze_mask = cache_21_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_21_cast_fp16")]; + tensor cache_23_begin_0 = const()[name = tensor("cache_23_begin_0"), val = tensor([5, 0, 0, 0])]; + tensor cache_23_end_0 = const()[name = tensor("cache_23_end_0"), val = tensor([6, 1, 512, 8])]; + tensor cache_23_end_mask_0 = const()[name = tensor("cache_23_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_23_squeeze_mask_0 = const()[name = tensor("cache_23_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_23_cast_fp16 = slice_by_index(begin = cache_23_begin_0, end = cache_23_end_0, end_mask = cache_23_end_mask_0, squeeze_mask = cache_23_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_23_cast_fp16")]; + tensor input_287_axes_0 = const()[name = tensor("input_287_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65955136)))]; + tensor encoder_layers_5_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65956224)))]; + tensor input_287_cast_fp16 = layer_norm(axes = input_287_axes_0, beta = encoder_layers_5_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_norm_feed_forward1_weight_to_fp16, x = input_285_cast_fp16)[name = tensor("input_287_cast_fp16")]; + tensor encoder_layers_5_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_5_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65957312)))]; + tensor linear_46_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_5_feed_forward1_linear1_weight_to_fp16, x = input_287_cast_fp16)[name = tensor("linear_46_cast_fp16")]; + tensor input_291_cast_fp16 = silu(x = linear_46_cast_fp16)[name = tensor("input_291_cast_fp16")]; + tensor encoder_layers_5_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_5_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68054528)))]; + tensor linear_47_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_feed_forward1_linear2_weight_to_fp16, x = input_291_cast_fp16)[name = tensor("linear_47_cast_fp16")]; + tensor var_1471_to_fp16 = const()[name = tensor("op_1471_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1472_cast_fp16 = mul(x = linear_47_cast_fp16, y = var_1471_to_fp16)[name = tensor("op_1472_cast_fp16")]; + tensor input_297_cast_fp16 = add(x = input_285_cast_fp16, y = var_1472_cast_fp16)[name = tensor("input_297_cast_fp16")]; + tensor key_11_axes_0 = const()[name = tensor("key_11_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70151744)))]; + tensor encoder_layers_5_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70152832)))]; + tensor key_11_cast_fp16 = layer_norm(axes = key_11_axes_0, beta = encoder_layers_5_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_norm_self_att_weight_to_fp16, x = input_297_cast_fp16)[name = tensor("key_11_cast_fp16")]; + tensor input_299_interleave_0 = const()[name = tensor("input_299_interleave_0"), val = tensor(false)]; + tensor input_299_cast_fp16 = concat(axis = var_65, interleave = input_299_interleave_0, values = (cache_21_cast_fp16, key_11_cast_fp16))[name = tensor("input_299_cast_fp16")]; + tensor var_1494_begin_0 = const()[name = tensor("op_1494_begin_0"), val = tensor([0, 4, 0])]; + tensor var_1494_end_0 = const()[name = tensor("op_1494_end_0"), val = tensor([1, 70, 512])]; + tensor var_1494_end_mask_0 = const()[name = tensor("op_1494_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1494_cast_fp16 = slice_by_index(begin = var_1494_begin_0, end = var_1494_end_0, end_mask = var_1494_end_mask_0, x = cache_21_cast_fp16)[name = tensor("op_1494_cast_fp16")]; + tensor var_1497_begin_0 = const()[name = tensor("op_1497_begin_0"), val = tensor([0, 0, 0])]; + tensor var_1497_end_0 = const()[name = tensor("op_1497_end_0"), val = tensor([1, 4, 512])]; + tensor var_1497_end_mask_0 = const()[name = tensor("op_1497_end_mask_0"), val = tensor([true, false, true])]; + tensor var_1497_cast_fp16 = slice_by_index(begin = var_1497_begin_0, end = var_1497_end_0, end_mask = var_1497_end_mask_0, x = key_11_cast_fp16)[name = tensor("op_1497_cast_fp16")]; + tensor var_1500_interleave_0 = const()[name = tensor("op_1500_interleave_0"), val = tensor(false)]; + tensor var_1500_cast_fp16 = concat(axis = var_65, interleave = var_1500_interleave_0, values = (var_1494_cast_fp16, var_1497_cast_fp16))[name = tensor("op_1500_cast_fp16")]; + tensor encoder_layers_5_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70153920)))]; + tensor linear_48_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_self_attn_linear_q_weight_to_fp16, x = key_11_cast_fp16)[name = tensor("linear_48_cast_fp16")]; + tensor var_1504 = const()[name = tensor("op_1504"), val = tensor([1, -1, 8, 64])]; + tensor q_31_cast_fp16 = reshape(shape = var_1504, x = linear_48_cast_fp16)[name = tensor("q_31_cast_fp16")]; + tensor encoder_layers_5_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70678272)))]; + tensor linear_49_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_self_attn_linear_k_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("linear_49_cast_fp16")]; + tensor var_1508 = const()[name = tensor("op_1508"), val = tensor([1, -1, 8, 64])]; + tensor k_21_cast_fp16 = reshape(shape = var_1508, x = linear_49_cast_fp16)[name = tensor("k_21_cast_fp16")]; + tensor encoder_layers_5_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71202624)))]; + tensor linear_50_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_self_attn_linear_v_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("linear_50_cast_fp16")]; + tensor var_1512 = const()[name = tensor("op_1512"), val = tensor([1, -1, 8, 64])]; + tensor v_11_cast_fp16 = reshape(shape = var_1512, x = linear_50_cast_fp16)[name = tensor("v_11_cast_fp16")]; + tensor value_13_perm_0 = const()[name = tensor("value_13_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_5_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71726976)))]; + tensor var_1524_cast_fp16 = add(x = q_31_cast_fp16, y = encoder_layers_5_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1524_cast_fp16")]; + tensor encoder_layers_5_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71728064)))]; + tensor var_1526_cast_fp16 = add(x = q_31_cast_fp16, y = encoder_layers_5_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1526_cast_fp16")]; + tensor q_with_bias_v_11_perm_0 = const()[name = tensor("q_with_bias_v_11_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_137_transpose_x_0 = const()[name = tensor("x_137_transpose_x_0"), val = tensor(false)]; + tensor x_137_transpose_y_0 = const()[name = tensor("x_137_transpose_y_0"), val = tensor(false)]; + tensor var_1528_to_fp16 = const()[name = tensor("op_1528_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71729152)))]; + tensor q_with_bias_v_11_cast_fp16 = transpose(perm = q_with_bias_v_11_perm_0, x = var_1526_cast_fp16)[name = tensor("transpose_192")]; + tensor x_137_cast_fp16 = matmul(transpose_x = x_137_transpose_x_0, transpose_y = x_137_transpose_y_0, x = q_with_bias_v_11_cast_fp16, y = var_1528_to_fp16)[name = tensor("x_137_cast_fp16")]; + tensor x_139_pad_0 = const()[name = tensor("x_139_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_139_mode_0 = const()[name = tensor("x_139_mode_0"), val = tensor("constant")]; + tensor const_144_to_fp16 = const()[name = tensor("const_144_to_fp16"), val = tensor(0x0p+0)]; + tensor x_139_cast_fp16 = pad(constant_val = const_144_to_fp16, mode = x_139_mode_0, pad = x_139_pad_0, x = x_137_cast_fp16)[name = tensor("x_139_cast_fp16")]; + tensor var_1536 = const()[name = tensor("op_1536"), val = tensor([1, 8, -1, 8])]; + tensor x_141_cast_fp16 = reshape(shape = var_1536, x = x_139_cast_fp16)[name = tensor("x_141_cast_fp16")]; + tensor var_1540_begin_0 = const()[name = tensor("op_1540_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1540_end_0 = const()[name = tensor("op_1540_end_0"), val = tensor([1, 8, 156, 8])]; + tensor var_1540_end_mask_0 = const()[name = tensor("op_1540_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1540_cast_fp16 = slice_by_index(begin = var_1540_begin_0, end = var_1540_end_0, end_mask = var_1540_end_mask_0, x = x_141_cast_fp16)[name = tensor("op_1540_cast_fp16")]; + tensor var_1541 = const()[name = tensor("op_1541"), val = tensor([1, 8, 8, 155])]; + tensor matrix_bd_21_cast_fp16 = reshape(shape = var_1541, x = var_1540_cast_fp16)[name = tensor("matrix_bd_21_cast_fp16")]; + tensor matrix_ac_11_transpose_x_0 = const()[name = tensor("matrix_ac_11_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_11_transpose_y_0 = const()[name = tensor("matrix_ac_11_transpose_y_0"), val = tensor(false)]; + tensor transpose_61_perm_0 = const()[name = tensor("transpose_61_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_62_perm_0 = const()[name = tensor("transpose_62_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_62 = transpose(perm = transpose_62_perm_0, x = k_21_cast_fp16)[name = tensor("transpose_190")]; + tensor transpose_61 = transpose(perm = transpose_61_perm_0, x = var_1524_cast_fp16)[name = tensor("transpose_191")]; + tensor matrix_ac_11_cast_fp16 = matmul(transpose_x = matrix_ac_11_transpose_x_0, transpose_y = matrix_ac_11_transpose_y_0, x = transpose_61, y = transpose_62)[name = tensor("matrix_ac_11_cast_fp16")]; + tensor matrix_bd_23_begin_0 = const()[name = tensor("matrix_bd_23_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_23_end_0 = const()[name = tensor("matrix_bd_23_end_0"), val = tensor([1, 8, 8, 78])]; + tensor matrix_bd_23_end_mask_0 = const()[name = tensor("matrix_bd_23_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_23_cast_fp16 = slice_by_index(begin = matrix_bd_23_begin_0, end = matrix_bd_23_end_0, end_mask = matrix_bd_23_end_mask_0, x = matrix_bd_21_cast_fp16)[name = tensor("matrix_bd_23_cast_fp16")]; + tensor var_1550_cast_fp16 = add(x = matrix_ac_11_cast_fp16, y = matrix_bd_23_cast_fp16)[name = tensor("op_1550_cast_fp16")]; + tensor _inversed_scores_21_y_0_to_fp16 = const()[name = tensor("_inversed_scores_21_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_21_cast_fp16 = mul(x = var_1550_cast_fp16, y = _inversed_scores_21_y_0_to_fp16)[name = tensor("_inversed_scores_21_cast_fp16")]; + tensor scores_23_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_21_cast_fp16, cond = mask_11)[name = tensor("scores_23_cast_fp16")]; + tensor var_1556_cast_fp16 = softmax(axis = var_56, x = scores_23_cast_fp16)[name = tensor("op_1556_cast_fp16")]; + tensor input_301_cast_fp16 = select(a = var_40_to_fp16, b = var_1556_cast_fp16, cond = mask_11)[name = tensor("input_301_cast_fp16")]; + tensor x_143_transpose_x_0 = const()[name = tensor("x_143_transpose_x_0"), val = tensor(false)]; + tensor x_143_transpose_y_0 = const()[name = tensor("x_143_transpose_y_0"), val = tensor(false)]; + tensor value_13_cast_fp16 = transpose(perm = value_13_perm_0, x = v_11_cast_fp16)[name = tensor("transpose_193")]; + tensor x_143_cast_fp16 = matmul(transpose_x = x_143_transpose_x_0, transpose_y = x_143_transpose_y_0, x = input_301_cast_fp16, y = value_13_cast_fp16)[name = tensor("x_143_cast_fp16")]; + tensor var_1560_perm_0 = const()[name = tensor("op_1560_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1561 = const()[name = tensor("op_1561"), val = tensor([1, -1, 512])]; + tensor var_1560_cast_fp16 = transpose(perm = var_1560_perm_0, x = x_143_cast_fp16)[name = tensor("transpose_189")]; + tensor input_303_cast_fp16 = reshape(shape = var_1561, x = var_1560_cast_fp16)[name = tensor("input_303_cast_fp16")]; + tensor encoder_layers_5_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_5_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71887936)))]; + tensor linear_52_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_self_attn_linear_out_weight_to_fp16, x = input_303_cast_fp16)[name = tensor("linear_52_cast_fp16")]; + tensor input_307_cast_fp16 = add(x = input_297_cast_fp16, y = linear_52_cast_fp16)[name = tensor("input_307_cast_fp16")]; + tensor x_147_axes_0 = const()[name = tensor("x_147_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72412288)))]; + tensor encoder_layers_5_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72413376)))]; + tensor x_147_cast_fp16 = layer_norm(axes = x_147_axes_0, beta = encoder_layers_5_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_norm_conv_weight_to_fp16, x = input_307_cast_fp16)[name = tensor("x_147_cast_fp16")]; + tensor input_309_perm_0 = const()[name = tensor("input_309_perm_0"), val = tensor([0, 2, 1])]; + tensor input_311_pad_type_0 = const()[name = tensor("input_311_pad_type_0"), val = tensor("valid")]; + tensor input_311_strides_0 = const()[name = tensor("input_311_strides_0"), val = tensor([1])]; + tensor input_311_pad_0 = const()[name = tensor("input_311_pad_0"), val = tensor([0, 0])]; + tensor input_311_dilations_0 = const()[name = tensor("input_311_dilations_0"), val = tensor([1])]; + tensor input_311_groups_0 = const()[name = tensor("input_311_groups_0"), val = tensor(1)]; + tensor encoder_layers_5_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_5_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72414464)))]; + tensor input_309_cast_fp16 = transpose(perm = input_309_perm_0, x = x_147_cast_fp16)[name = tensor("transpose_188")]; + tensor input_311_cast_fp16 = conv(dilations = input_311_dilations_0, groups = input_311_groups_0, pad = input_311_pad_0, pad_type = input_311_pad_type_0, strides = input_311_strides_0, weight = encoder_layers_5_conv_pointwise_conv1_weight_to_fp16, x = input_309_cast_fp16)[name = tensor("input_311_cast_fp16")]; + tensor x_149_split_num_splits_0 = const()[name = tensor("x_149_split_num_splits_0"), val = tensor(2)]; + tensor x_149_split_axis_0 = const()[name = tensor("x_149_split_axis_0"), val = tensor(1)]; + tensor x_149_split_cast_fp16_0, tensor x_149_split_cast_fp16_1 = split(axis = x_149_split_axis_0, num_splits = x_149_split_num_splits_0, x = input_311_cast_fp16)[name = tensor("x_149_split_cast_fp16")]; + tensor x_149_split_1_sigmoid_cast_fp16 = sigmoid(x = x_149_split_cast_fp16_1)[name = tensor("x_149_split_1_sigmoid_cast_fp16")]; + tensor x_149_cast_fp16 = mul(x = x_149_split_cast_fp16_0, y = x_149_split_1_sigmoid_cast_fp16)[name = tensor("x_149_cast_fp16")]; + tensor input_313_cast_fp16 = select(a = var_40_to_fp16, b = x_149_cast_fp16, cond = var_551)[name = tensor("input_313_cast_fp16")]; + tensor new_x_23_interleave_0 = const()[name = tensor("new_x_23_interleave_0"), val = tensor(false)]; + tensor new_x_23_cast_fp16 = concat(axis = var_56, interleave = new_x_23_interleave_0, values = (cache_23_cast_fp16, input_313_cast_fp16))[name = tensor("new_x_23_cast_fp16")]; + tensor next_cache_11_begin_0 = const()[name = tensor("next_cache_11_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_11_end_0 = const()[name = tensor("next_cache_11_end_0"), val = tensor([1, 512, 12])]; + tensor next_cache_11_end_mask_0 = const()[name = tensor("next_cache_11_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_11_cast_fp16 = slice_by_index(begin = next_cache_11_begin_0, end = next_cache_11_end_0, end_mask = next_cache_11_end_mask_0, x = new_x_23_cast_fp16)[name = tensor("next_cache_11_cast_fp16")]; + tensor var_1602_begin_0 = const()[name = tensor("op_1602_begin_0"), val = tensor([0, 0, 4])]; + tensor var_1602_end_0 = const()[name = tensor("op_1602_end_0"), val = tensor([1, 512, 12])]; + tensor var_1602_end_mask_0 = const()[name = tensor("op_1602_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1602_cast_fp16 = slice_by_index(begin = var_1602_begin_0, end = var_1602_end_0, end_mask = var_1602_end_mask_0, x = next_cache_11_cast_fp16)[name = tensor("op_1602_cast_fp16")]; + tensor x_151_pad_type_0 = const()[name = tensor("x_151_pad_type_0"), val = tensor("valid")]; + tensor x_151_groups_0 = const()[name = tensor("x_151_groups_0"), val = tensor(512)]; + tensor x_151_strides_0 = const()[name = tensor("x_151_strides_0"), val = tensor([1])]; + tensor x_151_pad_0 = const()[name = tensor("x_151_pad_0"), val = tensor([0, 0])]; + tensor x_151_dilations_0 = const()[name = tensor("x_151_dilations_0"), val = tensor([1])]; + tensor encoder_layers_5_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_5_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73463104)))]; + tensor x_151_cast_fp16 = conv(dilations = x_151_dilations_0, groups = x_151_groups_0, pad = x_151_pad_0, pad_type = x_151_pad_type_0, strides = x_151_strides_0, weight = encoder_layers_5_conv_depthwise_conv_weight_to_fp16, x = new_x_23_cast_fp16)[name = tensor("x_151_cast_fp16")]; + tensor input_315_perm_0 = const()[name = tensor("input_315_perm_0"), val = tensor([0, 2, 1])]; + tensor x_153_axes_0 = const()[name = tensor("x_153_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_5_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73472384)))]; + tensor encoder_layers_5_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_5_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73473472)))]; + tensor input_315_cast_fp16 = transpose(perm = input_315_perm_0, x = x_151_cast_fp16)[name = tensor("transpose_187")]; + tensor x_153_cast_fp16 = layer_norm(axes = x_153_axes_0, beta = encoder_layers_5_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_conv_batch_norm_weight_to_fp16, x = input_315_cast_fp16)[name = tensor("x_153_cast_fp16")]; + tensor input_317_perm_0 = const()[name = tensor("input_317_perm_0"), val = tensor([0, 2, 1])]; + tensor input_317_cast_fp16 = transpose(perm = input_317_perm_0, x = x_153_cast_fp16)[name = tensor("transpose_186")]; + tensor input_319_cast_fp16 = silu(x = input_317_cast_fp16)[name = tensor("input_319_cast_fp16")]; + tensor x_155_pad_type_0 = const()[name = tensor("x_155_pad_type_0"), val = tensor("valid")]; + tensor x_155_strides_0 = const()[name = tensor("x_155_strides_0"), val = tensor([1])]; + tensor x_155_pad_0 = const()[name = tensor("x_155_pad_0"), val = tensor([0, 0])]; + tensor x_155_dilations_0 = const()[name = tensor("x_155_dilations_0"), val = tensor([1])]; + tensor x_155_groups_0 = const()[name = tensor("x_155_groups_0"), val = tensor(1)]; + tensor encoder_layers_5_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_5_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73474560)))]; + tensor x_155_cast_fp16 = conv(dilations = x_155_dilations_0, groups = x_155_groups_0, pad = x_155_pad_0, pad_type = x_155_pad_type_0, strides = x_155_strides_0, weight = encoder_layers_5_conv_pointwise_conv2_weight_to_fp16, x = input_319_cast_fp16)[name = tensor("x_155_cast_fp16")]; + tensor input_321_perm_0 = const()[name = tensor("input_321_perm_0"), val = tensor([0, 2, 1])]; + tensor input_321_cast_fp16 = transpose(perm = input_321_perm_0, x = x_155_cast_fp16)[name = tensor("transpose_185")]; + tensor input_323_cast_fp16 = add(x = input_307_cast_fp16, y = input_321_cast_fp16)[name = tensor("input_323_cast_fp16")]; + tensor input_325_axes_0 = const()[name = tensor("input_325_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73998912)))]; + tensor encoder_layers_5_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74000000)))]; + tensor input_325_cast_fp16 = layer_norm(axes = input_325_axes_0, beta = encoder_layers_5_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_norm_feed_forward2_weight_to_fp16, x = input_323_cast_fp16)[name = tensor("input_325_cast_fp16")]; + tensor encoder_layers_5_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_5_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74001088)))]; + tensor linear_53_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_5_feed_forward2_linear1_weight_to_fp16, x = input_325_cast_fp16)[name = tensor("linear_53_cast_fp16")]; + tensor input_329_cast_fp16 = silu(x = linear_53_cast_fp16)[name = tensor("input_329_cast_fp16")]; + tensor encoder_layers_5_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_5_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76098304)))]; + tensor linear_54_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_5_feed_forward2_linear2_weight_to_fp16, x = input_329_cast_fp16)[name = tensor("linear_54_cast_fp16")]; + tensor var_1643_to_fp16 = const()[name = tensor("op_1643_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1644_cast_fp16 = mul(x = linear_54_cast_fp16, y = var_1643_to_fp16)[name = tensor("op_1644_cast_fp16")]; + tensor input_335_cast_fp16 = add(x = input_323_cast_fp16, y = var_1644_cast_fp16)[name = tensor("input_335_cast_fp16")]; + tensor input_337_axes_0 = const()[name = tensor("input_337_axes_0"), val = tensor([-1])]; + tensor encoder_layers_5_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_5_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78195520)))]; + tensor encoder_layers_5_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_5_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78196608)))]; + tensor input_337_cast_fp16 = layer_norm(axes = input_337_axes_0, beta = encoder_layers_5_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_5_norm_out_weight_to_fp16, x = input_335_cast_fp16)[name = tensor("input_337_cast_fp16")]; + tensor cache_25_begin_0 = const()[name = tensor("cache_25_begin_0"), val = tensor([6, 0, 0, 0])]; + tensor cache_25_end_0 = const()[name = tensor("cache_25_end_0"), val = tensor([7, 1, 70, 512])]; + tensor cache_25_end_mask_0 = const()[name = tensor("cache_25_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_25_squeeze_mask_0 = const()[name = tensor("cache_25_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_25_cast_fp16 = slice_by_index(begin = cache_25_begin_0, end = cache_25_end_0, end_mask = cache_25_end_mask_0, squeeze_mask = cache_25_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_25_cast_fp16")]; + tensor cache_27_begin_0 = const()[name = tensor("cache_27_begin_0"), val = tensor([6, 0, 0, 0])]; + tensor cache_27_end_0 = const()[name = tensor("cache_27_end_0"), val = tensor([7, 1, 512, 8])]; + tensor cache_27_end_mask_0 = const()[name = tensor("cache_27_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_27_squeeze_mask_0 = const()[name = tensor("cache_27_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_27_cast_fp16 = slice_by_index(begin = cache_27_begin_0, end = cache_27_end_0, end_mask = cache_27_end_mask_0, squeeze_mask = cache_27_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_27_cast_fp16")]; + tensor input_339_axes_0 = const()[name = tensor("input_339_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78197696)))]; + tensor encoder_layers_6_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78198784)))]; + tensor input_339_cast_fp16 = layer_norm(axes = input_339_axes_0, beta = encoder_layers_6_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_norm_feed_forward1_weight_to_fp16, x = input_337_cast_fp16)[name = tensor("input_339_cast_fp16")]; + tensor encoder_layers_6_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_6_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78199872)))]; + tensor linear_55_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_6_feed_forward1_linear1_weight_to_fp16, x = input_339_cast_fp16)[name = tensor("linear_55_cast_fp16")]; + tensor input_343_cast_fp16 = silu(x = linear_55_cast_fp16)[name = tensor("input_343_cast_fp16")]; + tensor encoder_layers_6_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_6_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80297088)))]; + tensor linear_56_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_feed_forward1_linear2_weight_to_fp16, x = input_343_cast_fp16)[name = tensor("linear_56_cast_fp16")]; + tensor var_1678_to_fp16 = const()[name = tensor("op_1678_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1679_cast_fp16 = mul(x = linear_56_cast_fp16, y = var_1678_to_fp16)[name = tensor("op_1679_cast_fp16")]; + tensor input_349_cast_fp16 = add(x = input_337_cast_fp16, y = var_1679_cast_fp16)[name = tensor("input_349_cast_fp16")]; + tensor key_13_axes_0 = const()[name = tensor("key_13_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82394304)))]; + tensor encoder_layers_6_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82395392)))]; + tensor key_13_cast_fp16 = layer_norm(axes = key_13_axes_0, beta = encoder_layers_6_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_norm_self_att_weight_to_fp16, x = input_349_cast_fp16)[name = tensor("key_13_cast_fp16")]; + tensor input_351_interleave_0 = const()[name = tensor("input_351_interleave_0"), val = tensor(false)]; + tensor input_351_cast_fp16 = concat(axis = var_65, interleave = input_351_interleave_0, values = (cache_25_cast_fp16, key_13_cast_fp16))[name = tensor("input_351_cast_fp16")]; + tensor var_1701_begin_0 = const()[name = tensor("op_1701_begin_0"), val = tensor([0, 4, 0])]; + tensor var_1701_end_0 = const()[name = tensor("op_1701_end_0"), val = tensor([1, 70, 512])]; + tensor var_1701_end_mask_0 = const()[name = tensor("op_1701_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1701_cast_fp16 = slice_by_index(begin = var_1701_begin_0, end = var_1701_end_0, end_mask = var_1701_end_mask_0, x = cache_25_cast_fp16)[name = tensor("op_1701_cast_fp16")]; + tensor var_1704_begin_0 = const()[name = tensor("op_1704_begin_0"), val = tensor([0, 0, 0])]; + tensor var_1704_end_0 = const()[name = tensor("op_1704_end_0"), val = tensor([1, 4, 512])]; + tensor var_1704_end_mask_0 = const()[name = tensor("op_1704_end_mask_0"), val = tensor([true, false, true])]; + tensor var_1704_cast_fp16 = slice_by_index(begin = var_1704_begin_0, end = var_1704_end_0, end_mask = var_1704_end_mask_0, x = key_13_cast_fp16)[name = tensor("op_1704_cast_fp16")]; + tensor var_1707_interleave_0 = const()[name = tensor("op_1707_interleave_0"), val = tensor(false)]; + tensor var_1707_cast_fp16 = concat(axis = var_65, interleave = var_1707_interleave_0, values = (var_1701_cast_fp16, var_1704_cast_fp16))[name = tensor("op_1707_cast_fp16")]; + tensor encoder_layers_6_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82396480)))]; + tensor linear_57_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_self_attn_linear_q_weight_to_fp16, x = key_13_cast_fp16)[name = tensor("linear_57_cast_fp16")]; + tensor var_1711 = const()[name = tensor("op_1711"), val = tensor([1, -1, 8, 64])]; + tensor q_37_cast_fp16 = reshape(shape = var_1711, x = linear_57_cast_fp16)[name = tensor("q_37_cast_fp16")]; + tensor encoder_layers_6_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82920832)))]; + tensor linear_58_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_self_attn_linear_k_weight_to_fp16, x = input_351_cast_fp16)[name = tensor("linear_58_cast_fp16")]; + tensor var_1715 = const()[name = tensor("op_1715"), val = tensor([1, -1, 8, 64])]; + tensor k_25_cast_fp16 = reshape(shape = var_1715, x = linear_58_cast_fp16)[name = tensor("k_25_cast_fp16")]; + tensor encoder_layers_6_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83445184)))]; + tensor linear_59_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_self_attn_linear_v_weight_to_fp16, x = input_351_cast_fp16)[name = tensor("linear_59_cast_fp16")]; + tensor var_1719 = const()[name = tensor("op_1719"), val = tensor([1, -1, 8, 64])]; + tensor v_13_cast_fp16 = reshape(shape = var_1719, x = linear_59_cast_fp16)[name = tensor("v_13_cast_fp16")]; + tensor value_15_perm_0 = const()[name = tensor("value_15_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_6_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83969536)))]; + tensor var_1731_cast_fp16 = add(x = q_37_cast_fp16, y = encoder_layers_6_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1731_cast_fp16")]; + tensor encoder_layers_6_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83970624)))]; + tensor var_1733_cast_fp16 = add(x = q_37_cast_fp16, y = encoder_layers_6_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1733_cast_fp16")]; + tensor q_with_bias_v_13_perm_0 = const()[name = tensor("q_with_bias_v_13_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_163_transpose_x_0 = const()[name = tensor("x_163_transpose_x_0"), val = tensor(false)]; + tensor x_163_transpose_y_0 = const()[name = tensor("x_163_transpose_y_0"), val = tensor(false)]; + tensor var_1735_to_fp16 = const()[name = tensor("op_1735_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83971712)))]; + tensor q_with_bias_v_13_cast_fp16 = transpose(perm = q_with_bias_v_13_perm_0, x = var_1733_cast_fp16)[name = tensor("transpose_183")]; + tensor x_163_cast_fp16 = matmul(transpose_x = x_163_transpose_x_0, transpose_y = x_163_transpose_y_0, x = q_with_bias_v_13_cast_fp16, y = var_1735_to_fp16)[name = tensor("x_163_cast_fp16")]; + tensor x_165_pad_0 = const()[name = tensor("x_165_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_165_mode_0 = const()[name = tensor("x_165_mode_0"), val = tensor("constant")]; + tensor const_157_to_fp16 = const()[name = tensor("const_157_to_fp16"), val = tensor(0x0p+0)]; + tensor x_165_cast_fp16 = pad(constant_val = const_157_to_fp16, mode = x_165_mode_0, pad = x_165_pad_0, x = x_163_cast_fp16)[name = tensor("x_165_cast_fp16")]; + tensor var_1743 = const()[name = tensor("op_1743"), val = tensor([1, 8, -1, 8])]; + tensor x_167_cast_fp16 = reshape(shape = var_1743, x = x_165_cast_fp16)[name = tensor("x_167_cast_fp16")]; + tensor var_1747_begin_0 = const()[name = tensor("op_1747_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1747_end_0 = const()[name = tensor("op_1747_end_0"), val = tensor([1, 8, 156, 8])]; + tensor var_1747_end_mask_0 = const()[name = tensor("op_1747_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1747_cast_fp16 = slice_by_index(begin = var_1747_begin_0, end = var_1747_end_0, end_mask = var_1747_end_mask_0, x = x_167_cast_fp16)[name = tensor("op_1747_cast_fp16")]; + tensor var_1748 = const()[name = tensor("op_1748"), val = tensor([1, 8, 8, 155])]; + tensor matrix_bd_25_cast_fp16 = reshape(shape = var_1748, x = var_1747_cast_fp16)[name = tensor("matrix_bd_25_cast_fp16")]; + tensor matrix_ac_13_transpose_x_0 = const()[name = tensor("matrix_ac_13_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_13_transpose_y_0 = const()[name = tensor("matrix_ac_13_transpose_y_0"), val = tensor(false)]; + tensor transpose_63_perm_0 = const()[name = tensor("transpose_63_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_64_perm_0 = const()[name = tensor("transpose_64_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_64 = transpose(perm = transpose_64_perm_0, x = k_25_cast_fp16)[name = tensor("transpose_181")]; + tensor transpose_63 = transpose(perm = transpose_63_perm_0, x = var_1731_cast_fp16)[name = tensor("transpose_182")]; + tensor matrix_ac_13_cast_fp16 = matmul(transpose_x = matrix_ac_13_transpose_x_0, transpose_y = matrix_ac_13_transpose_y_0, x = transpose_63, y = transpose_64)[name = tensor("matrix_ac_13_cast_fp16")]; + tensor matrix_bd_27_begin_0 = const()[name = tensor("matrix_bd_27_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_27_end_0 = const()[name = tensor("matrix_bd_27_end_0"), val = tensor([1, 8, 8, 78])]; + tensor matrix_bd_27_end_mask_0 = const()[name = tensor("matrix_bd_27_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_27_cast_fp16 = slice_by_index(begin = matrix_bd_27_begin_0, end = matrix_bd_27_end_0, end_mask = matrix_bd_27_end_mask_0, x = matrix_bd_25_cast_fp16)[name = tensor("matrix_bd_27_cast_fp16")]; + tensor var_1757_cast_fp16 = add(x = matrix_ac_13_cast_fp16, y = matrix_bd_27_cast_fp16)[name = tensor("op_1757_cast_fp16")]; + tensor _inversed_scores_25_y_0_to_fp16 = const()[name = tensor("_inversed_scores_25_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_25_cast_fp16 = mul(x = var_1757_cast_fp16, y = _inversed_scores_25_y_0_to_fp16)[name = tensor("_inversed_scores_25_cast_fp16")]; + tensor scores_27_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_25_cast_fp16, cond = mask_11)[name = tensor("scores_27_cast_fp16")]; + tensor var_1763_cast_fp16 = softmax(axis = var_56, x = scores_27_cast_fp16)[name = tensor("op_1763_cast_fp16")]; + tensor input_353_cast_fp16 = select(a = var_40_to_fp16, b = var_1763_cast_fp16, cond = mask_11)[name = tensor("input_353_cast_fp16")]; + tensor x_169_transpose_x_0 = const()[name = tensor("x_169_transpose_x_0"), val = tensor(false)]; + tensor x_169_transpose_y_0 = const()[name = tensor("x_169_transpose_y_0"), val = tensor(false)]; + tensor value_15_cast_fp16 = transpose(perm = value_15_perm_0, x = v_13_cast_fp16)[name = tensor("transpose_184")]; + tensor x_169_cast_fp16 = matmul(transpose_x = x_169_transpose_x_0, transpose_y = x_169_transpose_y_0, x = input_353_cast_fp16, y = value_15_cast_fp16)[name = tensor("x_169_cast_fp16")]; + tensor var_1767_perm_0 = const()[name = tensor("op_1767_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1768 = const()[name = tensor("op_1768"), val = tensor([1, -1, 512])]; + tensor var_1767_cast_fp16 = transpose(perm = var_1767_perm_0, x = x_169_cast_fp16)[name = tensor("transpose_180")]; + tensor input_355_cast_fp16 = reshape(shape = var_1768, x = var_1767_cast_fp16)[name = tensor("input_355_cast_fp16")]; + tensor encoder_layers_6_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_6_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84130496)))]; + tensor linear_61_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_self_attn_linear_out_weight_to_fp16, x = input_355_cast_fp16)[name = tensor("linear_61_cast_fp16")]; + tensor input_359_cast_fp16 = add(x = input_349_cast_fp16, y = linear_61_cast_fp16)[name = tensor("input_359_cast_fp16")]; + tensor x_173_axes_0 = const()[name = tensor("x_173_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84654848)))]; + tensor encoder_layers_6_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84655936)))]; + tensor x_173_cast_fp16 = layer_norm(axes = x_173_axes_0, beta = encoder_layers_6_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_norm_conv_weight_to_fp16, x = input_359_cast_fp16)[name = tensor("x_173_cast_fp16")]; + tensor input_361_perm_0 = const()[name = tensor("input_361_perm_0"), val = tensor([0, 2, 1])]; + tensor input_363_pad_type_0 = const()[name = tensor("input_363_pad_type_0"), val = tensor("valid")]; + tensor input_363_strides_0 = const()[name = tensor("input_363_strides_0"), val = tensor([1])]; + tensor input_363_pad_0 = const()[name = tensor("input_363_pad_0"), val = tensor([0, 0])]; + tensor input_363_dilations_0 = const()[name = tensor("input_363_dilations_0"), val = tensor([1])]; + tensor input_363_groups_0 = const()[name = tensor("input_363_groups_0"), val = tensor(1)]; + tensor encoder_layers_6_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_6_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84657024)))]; + tensor input_361_cast_fp16 = transpose(perm = input_361_perm_0, x = x_173_cast_fp16)[name = tensor("transpose_179")]; + tensor input_363_cast_fp16 = conv(dilations = input_363_dilations_0, groups = input_363_groups_0, pad = input_363_pad_0, pad_type = input_363_pad_type_0, strides = input_363_strides_0, weight = encoder_layers_6_conv_pointwise_conv1_weight_to_fp16, x = input_361_cast_fp16)[name = tensor("input_363_cast_fp16")]; + tensor x_175_split_num_splits_0 = const()[name = tensor("x_175_split_num_splits_0"), val = tensor(2)]; + tensor x_175_split_axis_0 = const()[name = tensor("x_175_split_axis_0"), val = tensor(1)]; + tensor x_175_split_cast_fp16_0, tensor x_175_split_cast_fp16_1 = split(axis = x_175_split_axis_0, num_splits = x_175_split_num_splits_0, x = input_363_cast_fp16)[name = tensor("x_175_split_cast_fp16")]; + tensor x_175_split_1_sigmoid_cast_fp16 = sigmoid(x = x_175_split_cast_fp16_1)[name = tensor("x_175_split_1_sigmoid_cast_fp16")]; + tensor x_175_cast_fp16 = mul(x = x_175_split_cast_fp16_0, y = x_175_split_1_sigmoid_cast_fp16)[name = tensor("x_175_cast_fp16")]; + tensor input_365_cast_fp16 = select(a = var_40_to_fp16, b = x_175_cast_fp16, cond = var_551)[name = tensor("input_365_cast_fp16")]; + tensor new_x_27_interleave_0 = const()[name = tensor("new_x_27_interleave_0"), val = tensor(false)]; + tensor new_x_27_cast_fp16 = concat(axis = var_56, interleave = new_x_27_interleave_0, values = (cache_27_cast_fp16, input_365_cast_fp16))[name = tensor("new_x_27_cast_fp16")]; + tensor next_cache_13_begin_0 = const()[name = tensor("next_cache_13_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_13_end_0 = const()[name = tensor("next_cache_13_end_0"), val = tensor([1, 512, 12])]; + tensor next_cache_13_end_mask_0 = const()[name = tensor("next_cache_13_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_13_cast_fp16 = slice_by_index(begin = next_cache_13_begin_0, end = next_cache_13_end_0, end_mask = next_cache_13_end_mask_0, x = new_x_27_cast_fp16)[name = tensor("next_cache_13_cast_fp16")]; + tensor var_1809_begin_0 = const()[name = tensor("op_1809_begin_0"), val = tensor([0, 0, 4])]; + tensor var_1809_end_0 = const()[name = tensor("op_1809_end_0"), val = tensor([1, 512, 12])]; + tensor var_1809_end_mask_0 = const()[name = tensor("op_1809_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1809_cast_fp16 = slice_by_index(begin = var_1809_begin_0, end = var_1809_end_0, end_mask = var_1809_end_mask_0, x = next_cache_13_cast_fp16)[name = tensor("op_1809_cast_fp16")]; + tensor x_177_pad_type_0 = const()[name = tensor("x_177_pad_type_0"), val = tensor("valid")]; + tensor x_177_groups_0 = const()[name = tensor("x_177_groups_0"), val = tensor(512)]; + tensor x_177_strides_0 = const()[name = tensor("x_177_strides_0"), val = tensor([1])]; + tensor x_177_pad_0 = const()[name = tensor("x_177_pad_0"), val = tensor([0, 0])]; + tensor x_177_dilations_0 = const()[name = tensor("x_177_dilations_0"), val = tensor([1])]; + tensor encoder_layers_6_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_6_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85705664)))]; + tensor x_177_cast_fp16 = conv(dilations = x_177_dilations_0, groups = x_177_groups_0, pad = x_177_pad_0, pad_type = x_177_pad_type_0, strides = x_177_strides_0, weight = encoder_layers_6_conv_depthwise_conv_weight_to_fp16, x = new_x_27_cast_fp16)[name = tensor("x_177_cast_fp16")]; + tensor input_367_perm_0 = const()[name = tensor("input_367_perm_0"), val = tensor([0, 2, 1])]; + tensor x_179_axes_0 = const()[name = tensor("x_179_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_6_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85714944)))]; + tensor encoder_layers_6_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_6_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85716032)))]; + tensor input_367_cast_fp16 = transpose(perm = input_367_perm_0, x = x_177_cast_fp16)[name = tensor("transpose_178")]; + tensor x_179_cast_fp16 = layer_norm(axes = x_179_axes_0, beta = encoder_layers_6_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_conv_batch_norm_weight_to_fp16, x = input_367_cast_fp16)[name = tensor("x_179_cast_fp16")]; + tensor input_369_perm_0 = const()[name = tensor("input_369_perm_0"), val = tensor([0, 2, 1])]; + tensor input_369_cast_fp16 = transpose(perm = input_369_perm_0, x = x_179_cast_fp16)[name = tensor("transpose_177")]; + tensor input_371_cast_fp16 = silu(x = input_369_cast_fp16)[name = tensor("input_371_cast_fp16")]; + tensor x_181_pad_type_0 = const()[name = tensor("x_181_pad_type_0"), val = tensor("valid")]; + tensor x_181_strides_0 = const()[name = tensor("x_181_strides_0"), val = tensor([1])]; + tensor x_181_pad_0 = const()[name = tensor("x_181_pad_0"), val = tensor([0, 0])]; + tensor x_181_dilations_0 = const()[name = tensor("x_181_dilations_0"), val = tensor([1])]; + tensor x_181_groups_0 = const()[name = tensor("x_181_groups_0"), val = tensor(1)]; + tensor encoder_layers_6_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_6_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85717120)))]; + tensor x_181_cast_fp16 = conv(dilations = x_181_dilations_0, groups = x_181_groups_0, pad = x_181_pad_0, pad_type = x_181_pad_type_0, strides = x_181_strides_0, weight = encoder_layers_6_conv_pointwise_conv2_weight_to_fp16, x = input_371_cast_fp16)[name = tensor("x_181_cast_fp16")]; + tensor input_373_perm_0 = const()[name = tensor("input_373_perm_0"), val = tensor([0, 2, 1])]; + tensor input_373_cast_fp16 = transpose(perm = input_373_perm_0, x = x_181_cast_fp16)[name = tensor("transpose_176")]; + tensor input_375_cast_fp16 = add(x = input_359_cast_fp16, y = input_373_cast_fp16)[name = tensor("input_375_cast_fp16")]; + tensor input_377_axes_0 = const()[name = tensor("input_377_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86241472)))]; + tensor encoder_layers_6_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86242560)))]; + tensor input_377_cast_fp16 = layer_norm(axes = input_377_axes_0, beta = encoder_layers_6_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_norm_feed_forward2_weight_to_fp16, x = input_375_cast_fp16)[name = tensor("input_377_cast_fp16")]; + tensor encoder_layers_6_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_6_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86243648)))]; + tensor linear_62_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_6_feed_forward2_linear1_weight_to_fp16, x = input_377_cast_fp16)[name = tensor("linear_62_cast_fp16")]; + tensor input_381_cast_fp16 = silu(x = linear_62_cast_fp16)[name = tensor("input_381_cast_fp16")]; + tensor encoder_layers_6_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_6_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88340864)))]; + tensor linear_63_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_6_feed_forward2_linear2_weight_to_fp16, x = input_381_cast_fp16)[name = tensor("linear_63_cast_fp16")]; + tensor var_1850_to_fp16 = const()[name = tensor("op_1850_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1851_cast_fp16 = mul(x = linear_63_cast_fp16, y = var_1850_to_fp16)[name = tensor("op_1851_cast_fp16")]; + tensor input_387_cast_fp16 = add(x = input_375_cast_fp16, y = var_1851_cast_fp16)[name = tensor("input_387_cast_fp16")]; + tensor input_389_axes_0 = const()[name = tensor("input_389_axes_0"), val = tensor([-1])]; + tensor encoder_layers_6_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_6_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90438080)))]; + tensor encoder_layers_6_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_6_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90439168)))]; + tensor input_389_cast_fp16 = layer_norm(axes = input_389_axes_0, beta = encoder_layers_6_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_6_norm_out_weight_to_fp16, x = input_387_cast_fp16)[name = tensor("input_389_cast_fp16")]; + tensor cache_29_begin_0 = const()[name = tensor("cache_29_begin_0"), val = tensor([7, 0, 0, 0])]; + tensor cache_29_end_0 = const()[name = tensor("cache_29_end_0"), val = tensor([8, 1, 70, 512])]; + tensor cache_29_end_mask_0 = const()[name = tensor("cache_29_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_29_squeeze_mask_0 = const()[name = tensor("cache_29_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_29_cast_fp16 = slice_by_index(begin = cache_29_begin_0, end = cache_29_end_0, end_mask = cache_29_end_mask_0, squeeze_mask = cache_29_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_29_cast_fp16")]; + tensor cache_31_begin_0 = const()[name = tensor("cache_31_begin_0"), val = tensor([7, 0, 0, 0])]; + tensor cache_31_end_0 = const()[name = tensor("cache_31_end_0"), val = tensor([8, 1, 512, 8])]; + tensor cache_31_end_mask_0 = const()[name = tensor("cache_31_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_31_squeeze_mask_0 = const()[name = tensor("cache_31_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_31_cast_fp16 = slice_by_index(begin = cache_31_begin_0, end = cache_31_end_0, end_mask = cache_31_end_mask_0, squeeze_mask = cache_31_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_31_cast_fp16")]; + tensor input_391_axes_0 = const()[name = tensor("input_391_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90440256)))]; + tensor encoder_layers_7_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90441344)))]; + tensor input_391_cast_fp16 = layer_norm(axes = input_391_axes_0, beta = encoder_layers_7_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_norm_feed_forward1_weight_to_fp16, x = input_389_cast_fp16)[name = tensor("input_391_cast_fp16")]; + tensor encoder_layers_7_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_7_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90442432)))]; + tensor linear_64_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_7_feed_forward1_linear1_weight_to_fp16, x = input_391_cast_fp16)[name = tensor("linear_64_cast_fp16")]; + tensor input_395_cast_fp16 = silu(x = linear_64_cast_fp16)[name = tensor("input_395_cast_fp16")]; + tensor encoder_layers_7_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_7_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92539648)))]; + tensor linear_65_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_feed_forward1_linear2_weight_to_fp16, x = input_395_cast_fp16)[name = tensor("linear_65_cast_fp16")]; + tensor var_1885_to_fp16 = const()[name = tensor("op_1885_to_fp16"), val = tensor(0x1p-1)]; + tensor var_1886_cast_fp16 = mul(x = linear_65_cast_fp16, y = var_1885_to_fp16)[name = tensor("op_1886_cast_fp16")]; + tensor input_401_cast_fp16 = add(x = input_389_cast_fp16, y = var_1886_cast_fp16)[name = tensor("input_401_cast_fp16")]; + tensor key_15_axes_0 = const()[name = tensor("key_15_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94636864)))]; + tensor encoder_layers_7_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94637952)))]; + tensor key_15_cast_fp16 = layer_norm(axes = key_15_axes_0, beta = encoder_layers_7_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_norm_self_att_weight_to_fp16, x = input_401_cast_fp16)[name = tensor("key_15_cast_fp16")]; + tensor input_403_interleave_0 = const()[name = tensor("input_403_interleave_0"), val = tensor(false)]; + tensor input_403_cast_fp16 = concat(axis = var_65, interleave = input_403_interleave_0, values = (cache_29_cast_fp16, key_15_cast_fp16))[name = tensor("input_403_cast_fp16")]; + tensor var_1908_begin_0 = const()[name = tensor("op_1908_begin_0"), val = tensor([0, 4, 0])]; + tensor var_1908_end_0 = const()[name = tensor("op_1908_end_0"), val = tensor([1, 70, 512])]; + tensor var_1908_end_mask_0 = const()[name = tensor("op_1908_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1908_cast_fp16 = slice_by_index(begin = var_1908_begin_0, end = var_1908_end_0, end_mask = var_1908_end_mask_0, x = cache_29_cast_fp16)[name = tensor("op_1908_cast_fp16")]; + tensor var_1911_begin_0 = const()[name = tensor("op_1911_begin_0"), val = tensor([0, 0, 0])]; + tensor var_1911_end_0 = const()[name = tensor("op_1911_end_0"), val = tensor([1, 4, 512])]; + tensor var_1911_end_mask_0 = const()[name = tensor("op_1911_end_mask_0"), val = tensor([true, false, true])]; + tensor var_1911_cast_fp16 = slice_by_index(begin = var_1911_begin_0, end = var_1911_end_0, end_mask = var_1911_end_mask_0, x = key_15_cast_fp16)[name = tensor("op_1911_cast_fp16")]; + tensor var_1914_interleave_0 = const()[name = tensor("op_1914_interleave_0"), val = tensor(false)]; + tensor var_1914_cast_fp16 = concat(axis = var_65, interleave = var_1914_interleave_0, values = (var_1908_cast_fp16, var_1911_cast_fp16))[name = tensor("op_1914_cast_fp16")]; + tensor encoder_layers_7_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94639040)))]; + tensor linear_66_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_self_attn_linear_q_weight_to_fp16, x = key_15_cast_fp16)[name = tensor("linear_66_cast_fp16")]; + tensor var_1918 = const()[name = tensor("op_1918"), val = tensor([1, -1, 8, 64])]; + tensor q_43_cast_fp16 = reshape(shape = var_1918, x = linear_66_cast_fp16)[name = tensor("q_43_cast_fp16")]; + tensor encoder_layers_7_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95163392)))]; + tensor linear_67_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_self_attn_linear_k_weight_to_fp16, x = input_403_cast_fp16)[name = tensor("linear_67_cast_fp16")]; + tensor var_1922 = const()[name = tensor("op_1922"), val = tensor([1, -1, 8, 64])]; + tensor k_29_cast_fp16 = reshape(shape = var_1922, x = linear_67_cast_fp16)[name = tensor("k_29_cast_fp16")]; + tensor encoder_layers_7_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95687744)))]; + tensor linear_68_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_self_attn_linear_v_weight_to_fp16, x = input_403_cast_fp16)[name = tensor("linear_68_cast_fp16")]; + tensor var_1926 = const()[name = tensor("op_1926"), val = tensor([1, -1, 8, 64])]; + tensor v_15_cast_fp16 = reshape(shape = var_1926, x = linear_68_cast_fp16)[name = tensor("v_15_cast_fp16")]; + tensor value_17_perm_0 = const()[name = tensor("value_17_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_7_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96212096)))]; + tensor var_1938_cast_fp16 = add(x = q_43_cast_fp16, y = encoder_layers_7_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1938_cast_fp16")]; + tensor encoder_layers_7_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96213184)))]; + tensor var_1940_cast_fp16 = add(x = q_43_cast_fp16, y = encoder_layers_7_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1940_cast_fp16")]; + tensor q_with_bias_v_15_perm_0 = const()[name = tensor("q_with_bias_v_15_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_189_transpose_x_0 = const()[name = tensor("x_189_transpose_x_0"), val = tensor(false)]; + tensor x_189_transpose_y_0 = const()[name = tensor("x_189_transpose_y_0"), val = tensor(false)]; + tensor var_1942_to_fp16 = const()[name = tensor("op_1942_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96214272)))]; + tensor q_with_bias_v_15_cast_fp16 = transpose(perm = q_with_bias_v_15_perm_0, x = var_1940_cast_fp16)[name = tensor("transpose_174")]; + tensor x_189_cast_fp16 = matmul(transpose_x = x_189_transpose_x_0, transpose_y = x_189_transpose_y_0, x = q_with_bias_v_15_cast_fp16, y = var_1942_to_fp16)[name = tensor("x_189_cast_fp16")]; + tensor x_191_pad_0 = const()[name = tensor("x_191_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_191_mode_0 = const()[name = tensor("x_191_mode_0"), val = tensor("constant")]; + tensor const_170_to_fp16 = const()[name = tensor("const_170_to_fp16"), val = tensor(0x0p+0)]; + tensor x_191_cast_fp16 = pad(constant_val = const_170_to_fp16, mode = x_191_mode_0, pad = x_191_pad_0, x = x_189_cast_fp16)[name = tensor("x_191_cast_fp16")]; + tensor var_1950 = const()[name = tensor("op_1950"), val = tensor([1, 8, -1, 8])]; + tensor x_193_cast_fp16 = reshape(shape = var_1950, x = x_191_cast_fp16)[name = tensor("x_193_cast_fp16")]; + tensor var_1954_begin_0 = const()[name = tensor("op_1954_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1954_end_0 = const()[name = tensor("op_1954_end_0"), val = tensor([1, 8, 156, 8])]; + tensor var_1954_end_mask_0 = const()[name = tensor("op_1954_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1954_cast_fp16 = slice_by_index(begin = var_1954_begin_0, end = var_1954_end_0, end_mask = var_1954_end_mask_0, x = x_193_cast_fp16)[name = tensor("op_1954_cast_fp16")]; + tensor var_1955 = const()[name = tensor("op_1955"), val = tensor([1, 8, 8, 155])]; + tensor matrix_bd_29_cast_fp16 = reshape(shape = var_1955, x = var_1954_cast_fp16)[name = tensor("matrix_bd_29_cast_fp16")]; + tensor matrix_ac_15_transpose_x_0 = const()[name = tensor("matrix_ac_15_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_15_transpose_y_0 = const()[name = tensor("matrix_ac_15_transpose_y_0"), val = tensor(false)]; + tensor transpose_65_perm_0 = const()[name = tensor("transpose_65_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_66_perm_0 = const()[name = tensor("transpose_66_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_66 = transpose(perm = transpose_66_perm_0, x = k_29_cast_fp16)[name = tensor("transpose_172")]; + tensor transpose_65 = transpose(perm = transpose_65_perm_0, x = var_1938_cast_fp16)[name = tensor("transpose_173")]; + tensor matrix_ac_15_cast_fp16 = matmul(transpose_x = matrix_ac_15_transpose_x_0, transpose_y = matrix_ac_15_transpose_y_0, x = transpose_65, y = transpose_66)[name = tensor("matrix_ac_15_cast_fp16")]; + tensor matrix_bd_31_begin_0 = const()[name = tensor("matrix_bd_31_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_31_end_0 = const()[name = tensor("matrix_bd_31_end_0"), val = tensor([1, 8, 8, 78])]; + tensor matrix_bd_31_end_mask_0 = const()[name = tensor("matrix_bd_31_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_31_cast_fp16 = slice_by_index(begin = matrix_bd_31_begin_0, end = matrix_bd_31_end_0, end_mask = matrix_bd_31_end_mask_0, x = matrix_bd_29_cast_fp16)[name = tensor("matrix_bd_31_cast_fp16")]; + tensor var_1964_cast_fp16 = add(x = matrix_ac_15_cast_fp16, y = matrix_bd_31_cast_fp16)[name = tensor("op_1964_cast_fp16")]; + tensor _inversed_scores_29_y_0_to_fp16 = const()[name = tensor("_inversed_scores_29_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_29_cast_fp16 = mul(x = var_1964_cast_fp16, y = _inversed_scores_29_y_0_to_fp16)[name = tensor("_inversed_scores_29_cast_fp16")]; + tensor scores_31_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_29_cast_fp16, cond = mask_11)[name = tensor("scores_31_cast_fp16")]; + tensor var_1970_cast_fp16 = softmax(axis = var_56, x = scores_31_cast_fp16)[name = tensor("op_1970_cast_fp16")]; + tensor input_405_cast_fp16 = select(a = var_40_to_fp16, b = var_1970_cast_fp16, cond = mask_11)[name = tensor("input_405_cast_fp16")]; + tensor x_195_transpose_x_0 = const()[name = tensor("x_195_transpose_x_0"), val = tensor(false)]; + tensor x_195_transpose_y_0 = const()[name = tensor("x_195_transpose_y_0"), val = tensor(false)]; + tensor value_17_cast_fp16 = transpose(perm = value_17_perm_0, x = v_15_cast_fp16)[name = tensor("transpose_175")]; + tensor x_195_cast_fp16 = matmul(transpose_x = x_195_transpose_x_0, transpose_y = x_195_transpose_y_0, x = input_405_cast_fp16, y = value_17_cast_fp16)[name = tensor("x_195_cast_fp16")]; + tensor var_1974_perm_0 = const()[name = tensor("op_1974_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1975 = const()[name = tensor("op_1975"), val = tensor([1, -1, 512])]; + tensor var_1974_cast_fp16 = transpose(perm = var_1974_perm_0, x = x_195_cast_fp16)[name = tensor("transpose_171")]; + tensor input_407_cast_fp16 = reshape(shape = var_1975, x = var_1974_cast_fp16)[name = tensor("input_407_cast_fp16")]; + tensor encoder_layers_7_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_7_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96373056)))]; + tensor linear_70_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_self_attn_linear_out_weight_to_fp16, x = input_407_cast_fp16)[name = tensor("linear_70_cast_fp16")]; + tensor input_411_cast_fp16 = add(x = input_401_cast_fp16, y = linear_70_cast_fp16)[name = tensor("input_411_cast_fp16")]; + tensor x_199_axes_0 = const()[name = tensor("x_199_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96897408)))]; + tensor encoder_layers_7_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96898496)))]; + tensor x_199_cast_fp16 = layer_norm(axes = x_199_axes_0, beta = encoder_layers_7_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_norm_conv_weight_to_fp16, x = input_411_cast_fp16)[name = tensor("x_199_cast_fp16")]; + tensor input_413_perm_0 = const()[name = tensor("input_413_perm_0"), val = tensor([0, 2, 1])]; + tensor input_415_pad_type_0 = const()[name = tensor("input_415_pad_type_0"), val = tensor("valid")]; + tensor input_415_strides_0 = const()[name = tensor("input_415_strides_0"), val = tensor([1])]; + tensor input_415_pad_0 = const()[name = tensor("input_415_pad_0"), val = tensor([0, 0])]; + tensor input_415_dilations_0 = const()[name = tensor("input_415_dilations_0"), val = tensor([1])]; + tensor input_415_groups_0 = const()[name = tensor("input_415_groups_0"), val = tensor(1)]; + tensor encoder_layers_7_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_7_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96899584)))]; + tensor input_413_cast_fp16 = transpose(perm = input_413_perm_0, x = x_199_cast_fp16)[name = tensor("transpose_170")]; + tensor input_415_cast_fp16 = conv(dilations = input_415_dilations_0, groups = input_415_groups_0, pad = input_415_pad_0, pad_type = input_415_pad_type_0, strides = input_415_strides_0, weight = encoder_layers_7_conv_pointwise_conv1_weight_to_fp16, x = input_413_cast_fp16)[name = tensor("input_415_cast_fp16")]; + tensor x_201_split_num_splits_0 = const()[name = tensor("x_201_split_num_splits_0"), val = tensor(2)]; + tensor x_201_split_axis_0 = const()[name = tensor("x_201_split_axis_0"), val = tensor(1)]; + tensor x_201_split_cast_fp16_0, tensor x_201_split_cast_fp16_1 = split(axis = x_201_split_axis_0, num_splits = x_201_split_num_splits_0, x = input_415_cast_fp16)[name = tensor("x_201_split_cast_fp16")]; + tensor x_201_split_1_sigmoid_cast_fp16 = sigmoid(x = x_201_split_cast_fp16_1)[name = tensor("x_201_split_1_sigmoid_cast_fp16")]; + tensor x_201_cast_fp16 = mul(x = x_201_split_cast_fp16_0, y = x_201_split_1_sigmoid_cast_fp16)[name = tensor("x_201_cast_fp16")]; + tensor input_417_cast_fp16 = select(a = var_40_to_fp16, b = x_201_cast_fp16, cond = var_551)[name = tensor("input_417_cast_fp16")]; + tensor new_x_31_interleave_0 = const()[name = tensor("new_x_31_interleave_0"), val = tensor(false)]; + tensor new_x_31_cast_fp16 = concat(axis = var_56, interleave = new_x_31_interleave_0, values = (cache_31_cast_fp16, input_417_cast_fp16))[name = tensor("new_x_31_cast_fp16")]; + tensor next_cache_15_begin_0 = const()[name = tensor("next_cache_15_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_15_end_0 = const()[name = tensor("next_cache_15_end_0"), val = tensor([1, 512, 12])]; + tensor next_cache_15_end_mask_0 = const()[name = tensor("next_cache_15_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_15_cast_fp16 = slice_by_index(begin = next_cache_15_begin_0, end = next_cache_15_end_0, end_mask = next_cache_15_end_mask_0, x = new_x_31_cast_fp16)[name = tensor("next_cache_15_cast_fp16")]; + tensor var_2016_begin_0 = const()[name = tensor("op_2016_begin_0"), val = tensor([0, 0, 4])]; + tensor var_2016_end_0 = const()[name = tensor("op_2016_end_0"), val = tensor([1, 512, 12])]; + tensor var_2016_end_mask_0 = const()[name = tensor("op_2016_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2016_cast_fp16 = slice_by_index(begin = var_2016_begin_0, end = var_2016_end_0, end_mask = var_2016_end_mask_0, x = next_cache_15_cast_fp16)[name = tensor("op_2016_cast_fp16")]; + tensor x_203_pad_type_0 = const()[name = tensor("x_203_pad_type_0"), val = tensor("valid")]; + tensor x_203_groups_0 = const()[name = tensor("x_203_groups_0"), val = tensor(512)]; + tensor x_203_strides_0 = const()[name = tensor("x_203_strides_0"), val = tensor([1])]; + tensor x_203_pad_0 = const()[name = tensor("x_203_pad_0"), val = tensor([0, 0])]; + tensor x_203_dilations_0 = const()[name = tensor("x_203_dilations_0"), val = tensor([1])]; + tensor encoder_layers_7_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_7_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97948224)))]; + tensor x_203_cast_fp16 = conv(dilations = x_203_dilations_0, groups = x_203_groups_0, pad = x_203_pad_0, pad_type = x_203_pad_type_0, strides = x_203_strides_0, weight = encoder_layers_7_conv_depthwise_conv_weight_to_fp16, x = new_x_31_cast_fp16)[name = tensor("x_203_cast_fp16")]; + tensor input_419_perm_0 = const()[name = tensor("input_419_perm_0"), val = tensor([0, 2, 1])]; + tensor x_205_axes_0 = const()[name = tensor("x_205_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_7_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97957504)))]; + tensor encoder_layers_7_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_7_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97958592)))]; + tensor input_419_cast_fp16 = transpose(perm = input_419_perm_0, x = x_203_cast_fp16)[name = tensor("transpose_169")]; + tensor x_205_cast_fp16 = layer_norm(axes = x_205_axes_0, beta = encoder_layers_7_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_conv_batch_norm_weight_to_fp16, x = input_419_cast_fp16)[name = tensor("x_205_cast_fp16")]; + tensor input_421_perm_0 = const()[name = tensor("input_421_perm_0"), val = tensor([0, 2, 1])]; + tensor input_421_cast_fp16 = transpose(perm = input_421_perm_0, x = x_205_cast_fp16)[name = tensor("transpose_168")]; + tensor input_423_cast_fp16 = silu(x = input_421_cast_fp16)[name = tensor("input_423_cast_fp16")]; + tensor x_207_pad_type_0 = const()[name = tensor("x_207_pad_type_0"), val = tensor("valid")]; + tensor x_207_strides_0 = const()[name = tensor("x_207_strides_0"), val = tensor([1])]; + tensor x_207_pad_0 = const()[name = tensor("x_207_pad_0"), val = tensor([0, 0])]; + tensor x_207_dilations_0 = const()[name = tensor("x_207_dilations_0"), val = tensor([1])]; + tensor x_207_groups_0 = const()[name = tensor("x_207_groups_0"), val = tensor(1)]; + tensor encoder_layers_7_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_7_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97959680)))]; + tensor x_207_cast_fp16 = conv(dilations = x_207_dilations_0, groups = x_207_groups_0, pad = x_207_pad_0, pad_type = x_207_pad_type_0, strides = x_207_strides_0, weight = encoder_layers_7_conv_pointwise_conv2_weight_to_fp16, x = input_423_cast_fp16)[name = tensor("x_207_cast_fp16")]; + tensor input_425_perm_0 = const()[name = tensor("input_425_perm_0"), val = tensor([0, 2, 1])]; + tensor input_425_cast_fp16 = transpose(perm = input_425_perm_0, x = x_207_cast_fp16)[name = tensor("transpose_167")]; + tensor input_427_cast_fp16 = add(x = input_411_cast_fp16, y = input_425_cast_fp16)[name = tensor("input_427_cast_fp16")]; + tensor input_429_axes_0 = const()[name = tensor("input_429_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98484032)))]; + tensor encoder_layers_7_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98485120)))]; + tensor input_429_cast_fp16 = layer_norm(axes = input_429_axes_0, beta = encoder_layers_7_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_norm_feed_forward2_weight_to_fp16, x = input_427_cast_fp16)[name = tensor("input_429_cast_fp16")]; + tensor encoder_layers_7_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_7_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98486208)))]; + tensor linear_71_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_7_feed_forward2_linear1_weight_to_fp16, x = input_429_cast_fp16)[name = tensor("linear_71_cast_fp16")]; + tensor input_433_cast_fp16 = silu(x = linear_71_cast_fp16)[name = tensor("input_433_cast_fp16")]; + tensor encoder_layers_7_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_7_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100583424)))]; + tensor linear_72_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_7_feed_forward2_linear2_weight_to_fp16, x = input_433_cast_fp16)[name = tensor("linear_72_cast_fp16")]; + tensor var_2057_to_fp16 = const()[name = tensor("op_2057_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2058_cast_fp16 = mul(x = linear_72_cast_fp16, y = var_2057_to_fp16)[name = tensor("op_2058_cast_fp16")]; + tensor input_439_cast_fp16 = add(x = input_427_cast_fp16, y = var_2058_cast_fp16)[name = tensor("input_439_cast_fp16")]; + tensor input_441_axes_0 = const()[name = tensor("input_441_axes_0"), val = tensor([-1])]; + tensor encoder_layers_7_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_7_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102680640)))]; + tensor encoder_layers_7_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_7_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102681728)))]; + tensor input_441_cast_fp16 = layer_norm(axes = input_441_axes_0, beta = encoder_layers_7_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_7_norm_out_weight_to_fp16, x = input_439_cast_fp16)[name = tensor("input_441_cast_fp16")]; + tensor cache_33_begin_0 = const()[name = tensor("cache_33_begin_0"), val = tensor([8, 0, 0, 0])]; + tensor cache_33_end_0 = const()[name = tensor("cache_33_end_0"), val = tensor([9, 1, 70, 512])]; + tensor cache_33_end_mask_0 = const()[name = tensor("cache_33_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_33_squeeze_mask_0 = const()[name = tensor("cache_33_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_33_cast_fp16 = slice_by_index(begin = cache_33_begin_0, end = cache_33_end_0, end_mask = cache_33_end_mask_0, squeeze_mask = cache_33_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_33_cast_fp16")]; + tensor cache_35_begin_0 = const()[name = tensor("cache_35_begin_0"), val = tensor([8, 0, 0, 0])]; + tensor cache_35_end_0 = const()[name = tensor("cache_35_end_0"), val = tensor([9, 1, 512, 8])]; + tensor cache_35_end_mask_0 = const()[name = tensor("cache_35_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_35_squeeze_mask_0 = const()[name = tensor("cache_35_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_35_cast_fp16 = slice_by_index(begin = cache_35_begin_0, end = cache_35_end_0, end_mask = cache_35_end_mask_0, squeeze_mask = cache_35_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_35_cast_fp16")]; + tensor input_443_axes_0 = const()[name = tensor("input_443_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102682816)))]; + tensor encoder_layers_8_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102683904)))]; + tensor input_443_cast_fp16 = layer_norm(axes = input_443_axes_0, beta = encoder_layers_8_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_norm_feed_forward1_weight_to_fp16, x = input_441_cast_fp16)[name = tensor("input_443_cast_fp16")]; + tensor encoder_layers_8_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_8_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102684992)))]; + tensor linear_73_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_8_feed_forward1_linear1_weight_to_fp16, x = input_443_cast_fp16)[name = tensor("linear_73_cast_fp16")]; + tensor input_447_cast_fp16 = silu(x = linear_73_cast_fp16)[name = tensor("input_447_cast_fp16")]; + tensor encoder_layers_8_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_8_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(104782208)))]; + tensor linear_74_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_feed_forward1_linear2_weight_to_fp16, x = input_447_cast_fp16)[name = tensor("linear_74_cast_fp16")]; + tensor var_2092_to_fp16 = const()[name = tensor("op_2092_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2093_cast_fp16 = mul(x = linear_74_cast_fp16, y = var_2092_to_fp16)[name = tensor("op_2093_cast_fp16")]; + tensor input_453_cast_fp16 = add(x = input_441_cast_fp16, y = var_2093_cast_fp16)[name = tensor("input_453_cast_fp16")]; + tensor key_17_axes_0 = const()[name = tensor("key_17_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106879424)))]; + tensor encoder_layers_8_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106880512)))]; + tensor key_17_cast_fp16 = layer_norm(axes = key_17_axes_0, beta = encoder_layers_8_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_norm_self_att_weight_to_fp16, x = input_453_cast_fp16)[name = tensor("key_17_cast_fp16")]; + tensor input_455_interleave_0 = const()[name = tensor("input_455_interleave_0"), val = tensor(false)]; + tensor input_455_cast_fp16 = concat(axis = var_65, interleave = input_455_interleave_0, values = (cache_33_cast_fp16, key_17_cast_fp16))[name = tensor("input_455_cast_fp16")]; + tensor var_2115_begin_0 = const()[name = tensor("op_2115_begin_0"), val = tensor([0, 4, 0])]; + tensor var_2115_end_0 = const()[name = tensor("op_2115_end_0"), val = tensor([1, 70, 512])]; + tensor var_2115_end_mask_0 = const()[name = tensor("op_2115_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2115_cast_fp16 = slice_by_index(begin = var_2115_begin_0, end = var_2115_end_0, end_mask = var_2115_end_mask_0, x = cache_33_cast_fp16)[name = tensor("op_2115_cast_fp16")]; + tensor var_2118_begin_0 = const()[name = tensor("op_2118_begin_0"), val = tensor([0, 0, 0])]; + tensor var_2118_end_0 = const()[name = tensor("op_2118_end_0"), val = tensor([1, 4, 512])]; + tensor var_2118_end_mask_0 = const()[name = tensor("op_2118_end_mask_0"), val = tensor([true, false, true])]; + tensor var_2118_cast_fp16 = slice_by_index(begin = var_2118_begin_0, end = var_2118_end_0, end_mask = var_2118_end_mask_0, x = key_17_cast_fp16)[name = tensor("op_2118_cast_fp16")]; + tensor var_2121_interleave_0 = const()[name = tensor("op_2121_interleave_0"), val = tensor(false)]; + tensor var_2121_cast_fp16 = concat(axis = var_65, interleave = var_2121_interleave_0, values = (var_2115_cast_fp16, var_2118_cast_fp16))[name = tensor("op_2121_cast_fp16")]; + tensor encoder_layers_8_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106881600)))]; + tensor linear_75_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_self_attn_linear_q_weight_to_fp16, x = key_17_cast_fp16)[name = tensor("linear_75_cast_fp16")]; + tensor var_2125 = const()[name = tensor("op_2125"), val = tensor([1, -1, 8, 64])]; + tensor q_49_cast_fp16 = reshape(shape = var_2125, x = linear_75_cast_fp16)[name = tensor("q_49_cast_fp16")]; + tensor encoder_layers_8_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107405952)))]; + tensor linear_76_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_self_attn_linear_k_weight_to_fp16, x = input_455_cast_fp16)[name = tensor("linear_76_cast_fp16")]; + tensor var_2129 = const()[name = tensor("op_2129"), val = tensor([1, -1, 8, 64])]; + tensor k_33_cast_fp16 = reshape(shape = var_2129, x = linear_76_cast_fp16)[name = tensor("k_33_cast_fp16")]; + tensor encoder_layers_8_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107930304)))]; + tensor linear_77_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_self_attn_linear_v_weight_to_fp16, x = input_455_cast_fp16)[name = tensor("linear_77_cast_fp16")]; + tensor var_2133 = const()[name = tensor("op_2133"), val = tensor([1, -1, 8, 64])]; + tensor v_17_cast_fp16 = reshape(shape = var_2133, x = linear_77_cast_fp16)[name = tensor("v_17_cast_fp16")]; + tensor value_19_perm_0 = const()[name = tensor("value_19_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_8_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108454656)))]; + tensor var_2145_cast_fp16 = add(x = q_49_cast_fp16, y = encoder_layers_8_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2145_cast_fp16")]; + tensor encoder_layers_8_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108455744)))]; + tensor var_2147_cast_fp16 = add(x = q_49_cast_fp16, y = encoder_layers_8_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2147_cast_fp16")]; + tensor q_with_bias_v_17_perm_0 = const()[name = tensor("q_with_bias_v_17_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_215_transpose_x_0 = const()[name = tensor("x_215_transpose_x_0"), val = tensor(false)]; + tensor x_215_transpose_y_0 = const()[name = tensor("x_215_transpose_y_0"), val = tensor(false)]; + tensor var_2149_to_fp16 = const()[name = tensor("op_2149_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108456832)))]; + tensor q_with_bias_v_17_cast_fp16 = transpose(perm = q_with_bias_v_17_perm_0, x = var_2147_cast_fp16)[name = tensor("transpose_165")]; + tensor x_215_cast_fp16 = matmul(transpose_x = x_215_transpose_x_0, transpose_y = x_215_transpose_y_0, x = q_with_bias_v_17_cast_fp16, y = var_2149_to_fp16)[name = tensor("x_215_cast_fp16")]; + tensor x_217_pad_0 = const()[name = tensor("x_217_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_217_mode_0 = const()[name = tensor("x_217_mode_0"), val = tensor("constant")]; + tensor const_183_to_fp16 = const()[name = tensor("const_183_to_fp16"), val = tensor(0x0p+0)]; + tensor x_217_cast_fp16 = pad(constant_val = const_183_to_fp16, mode = x_217_mode_0, pad = x_217_pad_0, x = x_215_cast_fp16)[name = tensor("x_217_cast_fp16")]; + tensor var_2157 = const()[name = tensor("op_2157"), val = tensor([1, 8, -1, 8])]; + tensor x_219_cast_fp16 = reshape(shape = var_2157, x = x_217_cast_fp16)[name = tensor("x_219_cast_fp16")]; + tensor var_2161_begin_0 = const()[name = tensor("op_2161_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2161_end_0 = const()[name = tensor("op_2161_end_0"), val = tensor([1, 8, 156, 8])]; + tensor var_2161_end_mask_0 = const()[name = tensor("op_2161_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2161_cast_fp16 = slice_by_index(begin = var_2161_begin_0, end = var_2161_end_0, end_mask = var_2161_end_mask_0, x = x_219_cast_fp16)[name = tensor("op_2161_cast_fp16")]; + tensor var_2162 = const()[name = tensor("op_2162"), val = tensor([1, 8, 8, 155])]; + tensor matrix_bd_33_cast_fp16 = reshape(shape = var_2162, x = var_2161_cast_fp16)[name = tensor("matrix_bd_33_cast_fp16")]; + tensor matrix_ac_17_transpose_x_0 = const()[name = tensor("matrix_ac_17_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_17_transpose_y_0 = const()[name = tensor("matrix_ac_17_transpose_y_0"), val = tensor(false)]; + tensor transpose_67_perm_0 = const()[name = tensor("transpose_67_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_68_perm_0 = const()[name = tensor("transpose_68_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_68 = transpose(perm = transpose_68_perm_0, x = k_33_cast_fp16)[name = tensor("transpose_163")]; + tensor transpose_67 = transpose(perm = transpose_67_perm_0, x = var_2145_cast_fp16)[name = tensor("transpose_164")]; + tensor matrix_ac_17_cast_fp16 = matmul(transpose_x = matrix_ac_17_transpose_x_0, transpose_y = matrix_ac_17_transpose_y_0, x = transpose_67, y = transpose_68)[name = tensor("matrix_ac_17_cast_fp16")]; + tensor matrix_bd_35_begin_0 = const()[name = tensor("matrix_bd_35_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_35_end_0 = const()[name = tensor("matrix_bd_35_end_0"), val = tensor([1, 8, 8, 78])]; + tensor matrix_bd_35_end_mask_0 = const()[name = tensor("matrix_bd_35_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_35_cast_fp16 = slice_by_index(begin = matrix_bd_35_begin_0, end = matrix_bd_35_end_0, end_mask = matrix_bd_35_end_mask_0, x = matrix_bd_33_cast_fp16)[name = tensor("matrix_bd_35_cast_fp16")]; + tensor var_2171_cast_fp16 = add(x = matrix_ac_17_cast_fp16, y = matrix_bd_35_cast_fp16)[name = tensor("op_2171_cast_fp16")]; + tensor _inversed_scores_33_y_0_to_fp16 = const()[name = tensor("_inversed_scores_33_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_33_cast_fp16 = mul(x = var_2171_cast_fp16, y = _inversed_scores_33_y_0_to_fp16)[name = tensor("_inversed_scores_33_cast_fp16")]; + tensor scores_35_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_33_cast_fp16, cond = mask_11)[name = tensor("scores_35_cast_fp16")]; + tensor var_2177_cast_fp16 = softmax(axis = var_56, x = scores_35_cast_fp16)[name = tensor("op_2177_cast_fp16")]; + tensor input_457_cast_fp16 = select(a = var_40_to_fp16, b = var_2177_cast_fp16, cond = mask_11)[name = tensor("input_457_cast_fp16")]; + tensor x_221_transpose_x_0 = const()[name = tensor("x_221_transpose_x_0"), val = tensor(false)]; + tensor x_221_transpose_y_0 = const()[name = tensor("x_221_transpose_y_0"), val = tensor(false)]; + tensor value_19_cast_fp16 = transpose(perm = value_19_perm_0, x = v_17_cast_fp16)[name = tensor("transpose_166")]; + tensor x_221_cast_fp16 = matmul(transpose_x = x_221_transpose_x_0, transpose_y = x_221_transpose_y_0, x = input_457_cast_fp16, y = value_19_cast_fp16)[name = tensor("x_221_cast_fp16")]; + tensor var_2181_perm_0 = const()[name = tensor("op_2181_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2182 = const()[name = tensor("op_2182"), val = tensor([1, -1, 512])]; + tensor var_2181_cast_fp16 = transpose(perm = var_2181_perm_0, x = x_221_cast_fp16)[name = tensor("transpose_162")]; + tensor input_459_cast_fp16 = reshape(shape = var_2182, x = var_2181_cast_fp16)[name = tensor("input_459_cast_fp16")]; + tensor encoder_layers_8_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_8_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108615616)))]; + tensor linear_79_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_self_attn_linear_out_weight_to_fp16, x = input_459_cast_fp16)[name = tensor("linear_79_cast_fp16")]; + tensor input_463_cast_fp16 = add(x = input_453_cast_fp16, y = linear_79_cast_fp16)[name = tensor("input_463_cast_fp16")]; + tensor x_225_axes_0 = const()[name = tensor("x_225_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109139968)))]; + tensor encoder_layers_8_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109141056)))]; + tensor x_225_cast_fp16 = layer_norm(axes = x_225_axes_0, beta = encoder_layers_8_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_norm_conv_weight_to_fp16, x = input_463_cast_fp16)[name = tensor("x_225_cast_fp16")]; + tensor input_465_perm_0 = const()[name = tensor("input_465_perm_0"), val = tensor([0, 2, 1])]; + tensor input_467_pad_type_0 = const()[name = tensor("input_467_pad_type_0"), val = tensor("valid")]; + tensor input_467_strides_0 = const()[name = tensor("input_467_strides_0"), val = tensor([1])]; + tensor input_467_pad_0 = const()[name = tensor("input_467_pad_0"), val = tensor([0, 0])]; + tensor input_467_dilations_0 = const()[name = tensor("input_467_dilations_0"), val = tensor([1])]; + tensor input_467_groups_0 = const()[name = tensor("input_467_groups_0"), val = tensor(1)]; + tensor encoder_layers_8_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_8_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109142144)))]; + tensor input_465_cast_fp16 = transpose(perm = input_465_perm_0, x = x_225_cast_fp16)[name = tensor("transpose_161")]; + tensor input_467_cast_fp16 = conv(dilations = input_467_dilations_0, groups = input_467_groups_0, pad = input_467_pad_0, pad_type = input_467_pad_type_0, strides = input_467_strides_0, weight = encoder_layers_8_conv_pointwise_conv1_weight_to_fp16, x = input_465_cast_fp16)[name = tensor("input_467_cast_fp16")]; + tensor x_227_split_num_splits_0 = const()[name = tensor("x_227_split_num_splits_0"), val = tensor(2)]; + tensor x_227_split_axis_0 = const()[name = tensor("x_227_split_axis_0"), val = tensor(1)]; + tensor x_227_split_cast_fp16_0, tensor x_227_split_cast_fp16_1 = split(axis = x_227_split_axis_0, num_splits = x_227_split_num_splits_0, x = input_467_cast_fp16)[name = tensor("x_227_split_cast_fp16")]; + tensor x_227_split_1_sigmoid_cast_fp16 = sigmoid(x = x_227_split_cast_fp16_1)[name = tensor("x_227_split_1_sigmoid_cast_fp16")]; + tensor x_227_cast_fp16 = mul(x = x_227_split_cast_fp16_0, y = x_227_split_1_sigmoid_cast_fp16)[name = tensor("x_227_cast_fp16")]; + tensor input_469_cast_fp16 = select(a = var_40_to_fp16, b = x_227_cast_fp16, cond = var_551)[name = tensor("input_469_cast_fp16")]; + tensor new_x_35_interleave_0 = const()[name = tensor("new_x_35_interleave_0"), val = tensor(false)]; + tensor new_x_35_cast_fp16 = concat(axis = var_56, interleave = new_x_35_interleave_0, values = (cache_35_cast_fp16, input_469_cast_fp16))[name = tensor("new_x_35_cast_fp16")]; + tensor next_cache_17_begin_0 = const()[name = tensor("next_cache_17_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_17_end_0 = const()[name = tensor("next_cache_17_end_0"), val = tensor([1, 512, 12])]; + tensor next_cache_17_end_mask_0 = const()[name = tensor("next_cache_17_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_17_cast_fp16 = slice_by_index(begin = next_cache_17_begin_0, end = next_cache_17_end_0, end_mask = next_cache_17_end_mask_0, x = new_x_35_cast_fp16)[name = tensor("next_cache_17_cast_fp16")]; + tensor var_2223_begin_0 = const()[name = tensor("op_2223_begin_0"), val = tensor([0, 0, 4])]; + tensor var_2223_end_0 = const()[name = tensor("op_2223_end_0"), val = tensor([1, 512, 12])]; + tensor var_2223_end_mask_0 = const()[name = tensor("op_2223_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2223_cast_fp16 = slice_by_index(begin = var_2223_begin_0, end = var_2223_end_0, end_mask = var_2223_end_mask_0, x = next_cache_17_cast_fp16)[name = tensor("op_2223_cast_fp16")]; + tensor x_229_pad_type_0 = const()[name = tensor("x_229_pad_type_0"), val = tensor("valid")]; + tensor x_229_groups_0 = const()[name = tensor("x_229_groups_0"), val = tensor(512)]; + tensor x_229_strides_0 = const()[name = tensor("x_229_strides_0"), val = tensor([1])]; + tensor x_229_pad_0 = const()[name = tensor("x_229_pad_0"), val = tensor([0, 0])]; + tensor x_229_dilations_0 = const()[name = tensor("x_229_dilations_0"), val = tensor([1])]; + tensor encoder_layers_8_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_8_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110190784)))]; + tensor x_229_cast_fp16 = conv(dilations = x_229_dilations_0, groups = x_229_groups_0, pad = x_229_pad_0, pad_type = x_229_pad_type_0, strides = x_229_strides_0, weight = encoder_layers_8_conv_depthwise_conv_weight_to_fp16, x = new_x_35_cast_fp16)[name = tensor("x_229_cast_fp16")]; + tensor input_471_perm_0 = const()[name = tensor("input_471_perm_0"), val = tensor([0, 2, 1])]; + tensor x_231_axes_0 = const()[name = tensor("x_231_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_8_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110200064)))]; + tensor encoder_layers_8_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_8_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110201152)))]; + tensor input_471_cast_fp16 = transpose(perm = input_471_perm_0, x = x_229_cast_fp16)[name = tensor("transpose_160")]; + tensor x_231_cast_fp16 = layer_norm(axes = x_231_axes_0, beta = encoder_layers_8_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_conv_batch_norm_weight_to_fp16, x = input_471_cast_fp16)[name = tensor("x_231_cast_fp16")]; + tensor input_473_perm_0 = const()[name = tensor("input_473_perm_0"), val = tensor([0, 2, 1])]; + tensor input_473_cast_fp16 = transpose(perm = input_473_perm_0, x = x_231_cast_fp16)[name = tensor("transpose_159")]; + tensor input_475_cast_fp16 = silu(x = input_473_cast_fp16)[name = tensor("input_475_cast_fp16")]; + tensor x_233_pad_type_0 = const()[name = tensor("x_233_pad_type_0"), val = tensor("valid")]; + tensor x_233_strides_0 = const()[name = tensor("x_233_strides_0"), val = tensor([1])]; + tensor x_233_pad_0 = const()[name = tensor("x_233_pad_0"), val = tensor([0, 0])]; + tensor x_233_dilations_0 = const()[name = tensor("x_233_dilations_0"), val = tensor([1])]; + tensor x_233_groups_0 = const()[name = tensor("x_233_groups_0"), val = tensor(1)]; + tensor encoder_layers_8_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_8_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110202240)))]; + tensor x_233_cast_fp16 = conv(dilations = x_233_dilations_0, groups = x_233_groups_0, pad = x_233_pad_0, pad_type = x_233_pad_type_0, strides = x_233_strides_0, weight = encoder_layers_8_conv_pointwise_conv2_weight_to_fp16, x = input_475_cast_fp16)[name = tensor("x_233_cast_fp16")]; + tensor input_477_perm_0 = const()[name = tensor("input_477_perm_0"), val = tensor([0, 2, 1])]; + tensor input_477_cast_fp16 = transpose(perm = input_477_perm_0, x = x_233_cast_fp16)[name = tensor("transpose_158")]; + tensor input_479_cast_fp16 = add(x = input_463_cast_fp16, y = input_477_cast_fp16)[name = tensor("input_479_cast_fp16")]; + tensor input_481_axes_0 = const()[name = tensor("input_481_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110726592)))]; + tensor encoder_layers_8_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110727680)))]; + tensor input_481_cast_fp16 = layer_norm(axes = input_481_axes_0, beta = encoder_layers_8_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_norm_feed_forward2_weight_to_fp16, x = input_479_cast_fp16)[name = tensor("input_481_cast_fp16")]; + tensor encoder_layers_8_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_8_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110728768)))]; + tensor linear_80_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_8_feed_forward2_linear1_weight_to_fp16, x = input_481_cast_fp16)[name = tensor("linear_80_cast_fp16")]; + tensor input_485_cast_fp16 = silu(x = linear_80_cast_fp16)[name = tensor("input_485_cast_fp16")]; + tensor encoder_layers_8_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_8_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(112825984)))]; + tensor linear_81_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_8_feed_forward2_linear2_weight_to_fp16, x = input_485_cast_fp16)[name = tensor("linear_81_cast_fp16")]; + tensor var_2264_to_fp16 = const()[name = tensor("op_2264_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2265_cast_fp16 = mul(x = linear_81_cast_fp16, y = var_2264_to_fp16)[name = tensor("op_2265_cast_fp16")]; + tensor input_491_cast_fp16 = add(x = input_479_cast_fp16, y = var_2265_cast_fp16)[name = tensor("input_491_cast_fp16")]; + tensor input_493_axes_0 = const()[name = tensor("input_493_axes_0"), val = tensor([-1])]; + tensor encoder_layers_8_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_8_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114923200)))]; + tensor encoder_layers_8_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_8_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114924288)))]; + tensor input_493_cast_fp16 = layer_norm(axes = input_493_axes_0, beta = encoder_layers_8_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_8_norm_out_weight_to_fp16, x = input_491_cast_fp16)[name = tensor("input_493_cast_fp16")]; + tensor cache_37_begin_0 = const()[name = tensor("cache_37_begin_0"), val = tensor([9, 0, 0, 0])]; + tensor cache_37_end_0 = const()[name = tensor("cache_37_end_0"), val = tensor([10, 1, 70, 512])]; + tensor cache_37_end_mask_0 = const()[name = tensor("cache_37_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_37_squeeze_mask_0 = const()[name = tensor("cache_37_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_37_cast_fp16 = slice_by_index(begin = cache_37_begin_0, end = cache_37_end_0, end_mask = cache_37_end_mask_0, squeeze_mask = cache_37_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_37_cast_fp16")]; + tensor cache_39_begin_0 = const()[name = tensor("cache_39_begin_0"), val = tensor([9, 0, 0, 0])]; + tensor cache_39_end_0 = const()[name = tensor("cache_39_end_0"), val = tensor([10, 1, 512, 8])]; + tensor cache_39_end_mask_0 = const()[name = tensor("cache_39_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_39_squeeze_mask_0 = const()[name = tensor("cache_39_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_39_cast_fp16 = slice_by_index(begin = cache_39_begin_0, end = cache_39_end_0, end_mask = cache_39_end_mask_0, squeeze_mask = cache_39_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_39_cast_fp16")]; + tensor input_495_axes_0 = const()[name = tensor("input_495_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114925376)))]; + tensor encoder_layers_9_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114926464)))]; + tensor input_495_cast_fp16 = layer_norm(axes = input_495_axes_0, beta = encoder_layers_9_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_norm_feed_forward1_weight_to_fp16, x = input_493_cast_fp16)[name = tensor("input_495_cast_fp16")]; + tensor encoder_layers_9_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_9_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114927552)))]; + tensor linear_82_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_9_feed_forward1_linear1_weight_to_fp16, x = input_495_cast_fp16)[name = tensor("linear_82_cast_fp16")]; + tensor input_499_cast_fp16 = silu(x = linear_82_cast_fp16)[name = tensor("input_499_cast_fp16")]; + tensor encoder_layers_9_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_9_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117024768)))]; + tensor linear_83_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_feed_forward1_linear2_weight_to_fp16, x = input_499_cast_fp16)[name = tensor("linear_83_cast_fp16")]; + tensor var_2299_to_fp16 = const()[name = tensor("op_2299_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2300_cast_fp16 = mul(x = linear_83_cast_fp16, y = var_2299_to_fp16)[name = tensor("op_2300_cast_fp16")]; + tensor input_505_cast_fp16 = add(x = input_493_cast_fp16, y = var_2300_cast_fp16)[name = tensor("input_505_cast_fp16")]; + tensor key_19_axes_0 = const()[name = tensor("key_19_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119121984)))]; + tensor encoder_layers_9_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119123072)))]; + tensor key_19_cast_fp16 = layer_norm(axes = key_19_axes_0, beta = encoder_layers_9_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_norm_self_att_weight_to_fp16, x = input_505_cast_fp16)[name = tensor("key_19_cast_fp16")]; + tensor input_507_interleave_0 = const()[name = tensor("input_507_interleave_0"), val = tensor(false)]; + tensor input_507_cast_fp16 = concat(axis = var_65, interleave = input_507_interleave_0, values = (cache_37_cast_fp16, key_19_cast_fp16))[name = tensor("input_507_cast_fp16")]; + tensor var_2322_begin_0 = const()[name = tensor("op_2322_begin_0"), val = tensor([0, 4, 0])]; + tensor var_2322_end_0 = const()[name = tensor("op_2322_end_0"), val = tensor([1, 70, 512])]; + tensor var_2322_end_mask_0 = const()[name = tensor("op_2322_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2322_cast_fp16 = slice_by_index(begin = var_2322_begin_0, end = var_2322_end_0, end_mask = var_2322_end_mask_0, x = cache_37_cast_fp16)[name = tensor("op_2322_cast_fp16")]; + tensor var_2325_begin_0 = const()[name = tensor("op_2325_begin_0"), val = tensor([0, 0, 0])]; + tensor var_2325_end_0 = const()[name = tensor("op_2325_end_0"), val = tensor([1, 4, 512])]; + tensor var_2325_end_mask_0 = const()[name = tensor("op_2325_end_mask_0"), val = tensor([true, false, true])]; + tensor var_2325_cast_fp16 = slice_by_index(begin = var_2325_begin_0, end = var_2325_end_0, end_mask = var_2325_end_mask_0, x = key_19_cast_fp16)[name = tensor("op_2325_cast_fp16")]; + tensor var_2328_interleave_0 = const()[name = tensor("op_2328_interleave_0"), val = tensor(false)]; + tensor var_2328_cast_fp16 = concat(axis = var_65, interleave = var_2328_interleave_0, values = (var_2322_cast_fp16, var_2325_cast_fp16))[name = tensor("op_2328_cast_fp16")]; + tensor encoder_layers_9_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119124160)))]; + tensor linear_84_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_self_attn_linear_q_weight_to_fp16, x = key_19_cast_fp16)[name = tensor("linear_84_cast_fp16")]; + tensor var_2332 = const()[name = tensor("op_2332"), val = tensor([1, -1, 8, 64])]; + tensor q_55_cast_fp16 = reshape(shape = var_2332, x = linear_84_cast_fp16)[name = tensor("q_55_cast_fp16")]; + tensor encoder_layers_9_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119648512)))]; + tensor linear_85_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_self_attn_linear_k_weight_to_fp16, x = input_507_cast_fp16)[name = tensor("linear_85_cast_fp16")]; + tensor var_2336 = const()[name = tensor("op_2336"), val = tensor([1, -1, 8, 64])]; + tensor k_37_cast_fp16 = reshape(shape = var_2336, x = linear_85_cast_fp16)[name = tensor("k_37_cast_fp16")]; + tensor encoder_layers_9_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120172864)))]; + tensor linear_86_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_self_attn_linear_v_weight_to_fp16, x = input_507_cast_fp16)[name = tensor("linear_86_cast_fp16")]; + tensor var_2340 = const()[name = tensor("op_2340"), val = tensor([1, -1, 8, 64])]; + tensor v_19_cast_fp16 = reshape(shape = var_2340, x = linear_86_cast_fp16)[name = tensor("v_19_cast_fp16")]; + tensor value_21_perm_0 = const()[name = tensor("value_21_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_9_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120697216)))]; + tensor var_2352_cast_fp16 = add(x = q_55_cast_fp16, y = encoder_layers_9_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2352_cast_fp16")]; + tensor encoder_layers_9_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120698304)))]; + tensor var_2354_cast_fp16 = add(x = q_55_cast_fp16, y = encoder_layers_9_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2354_cast_fp16")]; + tensor q_with_bias_v_19_perm_0 = const()[name = tensor("q_with_bias_v_19_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_241_transpose_x_0 = const()[name = tensor("x_241_transpose_x_0"), val = tensor(false)]; + tensor x_241_transpose_y_0 = const()[name = tensor("x_241_transpose_y_0"), val = tensor(false)]; + tensor var_2356_to_fp16 = const()[name = tensor("op_2356_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120699392)))]; + tensor q_with_bias_v_19_cast_fp16 = transpose(perm = q_with_bias_v_19_perm_0, x = var_2354_cast_fp16)[name = tensor("transpose_156")]; + tensor x_241_cast_fp16 = matmul(transpose_x = x_241_transpose_x_0, transpose_y = x_241_transpose_y_0, x = q_with_bias_v_19_cast_fp16, y = var_2356_to_fp16)[name = tensor("x_241_cast_fp16")]; + tensor x_243_pad_0 = const()[name = tensor("x_243_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_243_mode_0 = const()[name = tensor("x_243_mode_0"), val = tensor("constant")]; + tensor const_196_to_fp16 = const()[name = tensor("const_196_to_fp16"), val = tensor(0x0p+0)]; + tensor x_243_cast_fp16 = pad(constant_val = const_196_to_fp16, mode = x_243_mode_0, pad = x_243_pad_0, x = x_241_cast_fp16)[name = tensor("x_243_cast_fp16")]; + tensor var_2364 = const()[name = tensor("op_2364"), val = tensor([1, 8, -1, 8])]; + tensor x_245_cast_fp16 = reshape(shape = var_2364, x = x_243_cast_fp16)[name = tensor("x_245_cast_fp16")]; + tensor var_2368_begin_0 = const()[name = tensor("op_2368_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2368_end_0 = const()[name = tensor("op_2368_end_0"), val = tensor([1, 8, 156, 8])]; + tensor var_2368_end_mask_0 = const()[name = tensor("op_2368_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2368_cast_fp16 = slice_by_index(begin = var_2368_begin_0, end = var_2368_end_0, end_mask = var_2368_end_mask_0, x = x_245_cast_fp16)[name = tensor("op_2368_cast_fp16")]; + tensor var_2369 = const()[name = tensor("op_2369"), val = tensor([1, 8, 8, 155])]; + tensor matrix_bd_37_cast_fp16 = reshape(shape = var_2369, x = var_2368_cast_fp16)[name = tensor("matrix_bd_37_cast_fp16")]; + tensor matrix_ac_19_transpose_x_0 = const()[name = tensor("matrix_ac_19_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_19_transpose_y_0 = const()[name = tensor("matrix_ac_19_transpose_y_0"), val = tensor(false)]; + tensor transpose_69_perm_0 = const()[name = tensor("transpose_69_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_70_perm_0 = const()[name = tensor("transpose_70_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_70 = transpose(perm = transpose_70_perm_0, x = k_37_cast_fp16)[name = tensor("transpose_154")]; + tensor transpose_69 = transpose(perm = transpose_69_perm_0, x = var_2352_cast_fp16)[name = tensor("transpose_155")]; + tensor matrix_ac_19_cast_fp16 = matmul(transpose_x = matrix_ac_19_transpose_x_0, transpose_y = matrix_ac_19_transpose_y_0, x = transpose_69, y = transpose_70)[name = tensor("matrix_ac_19_cast_fp16")]; + tensor matrix_bd_39_begin_0 = const()[name = tensor("matrix_bd_39_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_39_end_0 = const()[name = tensor("matrix_bd_39_end_0"), val = tensor([1, 8, 8, 78])]; + tensor matrix_bd_39_end_mask_0 = const()[name = tensor("matrix_bd_39_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_39_cast_fp16 = slice_by_index(begin = matrix_bd_39_begin_0, end = matrix_bd_39_end_0, end_mask = matrix_bd_39_end_mask_0, x = matrix_bd_37_cast_fp16)[name = tensor("matrix_bd_39_cast_fp16")]; + tensor var_2378_cast_fp16 = add(x = matrix_ac_19_cast_fp16, y = matrix_bd_39_cast_fp16)[name = tensor("op_2378_cast_fp16")]; + tensor _inversed_scores_37_y_0_to_fp16 = const()[name = tensor("_inversed_scores_37_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_37_cast_fp16 = mul(x = var_2378_cast_fp16, y = _inversed_scores_37_y_0_to_fp16)[name = tensor("_inversed_scores_37_cast_fp16")]; + tensor scores_39_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_37_cast_fp16, cond = mask_11)[name = tensor("scores_39_cast_fp16")]; + tensor var_2384_cast_fp16 = softmax(axis = var_56, x = scores_39_cast_fp16)[name = tensor("op_2384_cast_fp16")]; + tensor input_509_cast_fp16 = select(a = var_40_to_fp16, b = var_2384_cast_fp16, cond = mask_11)[name = tensor("input_509_cast_fp16")]; + tensor x_247_transpose_x_0 = const()[name = tensor("x_247_transpose_x_0"), val = tensor(false)]; + tensor x_247_transpose_y_0 = const()[name = tensor("x_247_transpose_y_0"), val = tensor(false)]; + tensor value_21_cast_fp16 = transpose(perm = value_21_perm_0, x = v_19_cast_fp16)[name = tensor("transpose_157")]; + tensor x_247_cast_fp16 = matmul(transpose_x = x_247_transpose_x_0, transpose_y = x_247_transpose_y_0, x = input_509_cast_fp16, y = value_21_cast_fp16)[name = tensor("x_247_cast_fp16")]; + tensor var_2388_perm_0 = const()[name = tensor("op_2388_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2389 = const()[name = tensor("op_2389"), val = tensor([1, -1, 512])]; + tensor var_2388_cast_fp16 = transpose(perm = var_2388_perm_0, x = x_247_cast_fp16)[name = tensor("transpose_153")]; + tensor input_511_cast_fp16 = reshape(shape = var_2389, x = var_2388_cast_fp16)[name = tensor("input_511_cast_fp16")]; + tensor encoder_layers_9_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_9_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120858176)))]; + tensor linear_88_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_self_attn_linear_out_weight_to_fp16, x = input_511_cast_fp16)[name = tensor("linear_88_cast_fp16")]; + tensor input_515_cast_fp16 = add(x = input_505_cast_fp16, y = linear_88_cast_fp16)[name = tensor("input_515_cast_fp16")]; + tensor x_251_axes_0 = const()[name = tensor("x_251_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121382528)))]; + tensor encoder_layers_9_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121383616)))]; + tensor x_251_cast_fp16 = layer_norm(axes = x_251_axes_0, beta = encoder_layers_9_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_norm_conv_weight_to_fp16, x = input_515_cast_fp16)[name = tensor("x_251_cast_fp16")]; + tensor input_517_perm_0 = const()[name = tensor("input_517_perm_0"), val = tensor([0, 2, 1])]; + tensor input_519_pad_type_0 = const()[name = tensor("input_519_pad_type_0"), val = tensor("valid")]; + tensor input_519_strides_0 = const()[name = tensor("input_519_strides_0"), val = tensor([1])]; + tensor input_519_pad_0 = const()[name = tensor("input_519_pad_0"), val = tensor([0, 0])]; + tensor input_519_dilations_0 = const()[name = tensor("input_519_dilations_0"), val = tensor([1])]; + tensor input_519_groups_0 = const()[name = tensor("input_519_groups_0"), val = tensor(1)]; + tensor encoder_layers_9_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_9_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121384704)))]; + tensor input_517_cast_fp16 = transpose(perm = input_517_perm_0, x = x_251_cast_fp16)[name = tensor("transpose_152")]; + tensor input_519_cast_fp16 = conv(dilations = input_519_dilations_0, groups = input_519_groups_0, pad = input_519_pad_0, pad_type = input_519_pad_type_0, strides = input_519_strides_0, weight = encoder_layers_9_conv_pointwise_conv1_weight_to_fp16, x = input_517_cast_fp16)[name = tensor("input_519_cast_fp16")]; + tensor x_253_split_num_splits_0 = const()[name = tensor("x_253_split_num_splits_0"), val = tensor(2)]; + tensor x_253_split_axis_0 = const()[name = tensor("x_253_split_axis_0"), val = tensor(1)]; + tensor x_253_split_cast_fp16_0, tensor x_253_split_cast_fp16_1 = split(axis = x_253_split_axis_0, num_splits = x_253_split_num_splits_0, x = input_519_cast_fp16)[name = tensor("x_253_split_cast_fp16")]; + tensor x_253_split_1_sigmoid_cast_fp16 = sigmoid(x = x_253_split_cast_fp16_1)[name = tensor("x_253_split_1_sigmoid_cast_fp16")]; + tensor x_253_cast_fp16 = mul(x = x_253_split_cast_fp16_0, y = x_253_split_1_sigmoid_cast_fp16)[name = tensor("x_253_cast_fp16")]; + tensor input_521_cast_fp16 = select(a = var_40_to_fp16, b = x_253_cast_fp16, cond = var_551)[name = tensor("input_521_cast_fp16")]; + tensor new_x_39_interleave_0 = const()[name = tensor("new_x_39_interleave_0"), val = tensor(false)]; + tensor new_x_39_cast_fp16 = concat(axis = var_56, interleave = new_x_39_interleave_0, values = (cache_39_cast_fp16, input_521_cast_fp16))[name = tensor("new_x_39_cast_fp16")]; + tensor next_cache_19_begin_0 = const()[name = tensor("next_cache_19_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_19_end_0 = const()[name = tensor("next_cache_19_end_0"), val = tensor([1, 512, 12])]; + tensor next_cache_19_end_mask_0 = const()[name = tensor("next_cache_19_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_19_cast_fp16 = slice_by_index(begin = next_cache_19_begin_0, end = next_cache_19_end_0, end_mask = next_cache_19_end_mask_0, x = new_x_39_cast_fp16)[name = tensor("next_cache_19_cast_fp16")]; + tensor var_2430_begin_0 = const()[name = tensor("op_2430_begin_0"), val = tensor([0, 0, 4])]; + tensor var_2430_end_0 = const()[name = tensor("op_2430_end_0"), val = tensor([1, 512, 12])]; + tensor var_2430_end_mask_0 = const()[name = tensor("op_2430_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2430_cast_fp16 = slice_by_index(begin = var_2430_begin_0, end = var_2430_end_0, end_mask = var_2430_end_mask_0, x = next_cache_19_cast_fp16)[name = tensor("op_2430_cast_fp16")]; + tensor x_255_pad_type_0 = const()[name = tensor("x_255_pad_type_0"), val = tensor("valid")]; + tensor x_255_groups_0 = const()[name = tensor("x_255_groups_0"), val = tensor(512)]; + tensor x_255_strides_0 = const()[name = tensor("x_255_strides_0"), val = tensor([1])]; + tensor x_255_pad_0 = const()[name = tensor("x_255_pad_0"), val = tensor([0, 0])]; + tensor x_255_dilations_0 = const()[name = tensor("x_255_dilations_0"), val = tensor([1])]; + tensor encoder_layers_9_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_9_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122433344)))]; + tensor x_255_cast_fp16 = conv(dilations = x_255_dilations_0, groups = x_255_groups_0, pad = x_255_pad_0, pad_type = x_255_pad_type_0, strides = x_255_strides_0, weight = encoder_layers_9_conv_depthwise_conv_weight_to_fp16, x = new_x_39_cast_fp16)[name = tensor("x_255_cast_fp16")]; + tensor input_523_perm_0 = const()[name = tensor("input_523_perm_0"), val = tensor([0, 2, 1])]; + tensor x_257_axes_0 = const()[name = tensor("x_257_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_9_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122442624)))]; + tensor encoder_layers_9_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_9_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122443712)))]; + tensor input_523_cast_fp16 = transpose(perm = input_523_perm_0, x = x_255_cast_fp16)[name = tensor("transpose_151")]; + tensor x_257_cast_fp16 = layer_norm(axes = x_257_axes_0, beta = encoder_layers_9_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_conv_batch_norm_weight_to_fp16, x = input_523_cast_fp16)[name = tensor("x_257_cast_fp16")]; + tensor input_525_perm_0 = const()[name = tensor("input_525_perm_0"), val = tensor([0, 2, 1])]; + tensor input_525_cast_fp16 = transpose(perm = input_525_perm_0, x = x_257_cast_fp16)[name = tensor("transpose_150")]; + tensor input_527_cast_fp16 = silu(x = input_525_cast_fp16)[name = tensor("input_527_cast_fp16")]; + tensor x_259_pad_type_0 = const()[name = tensor("x_259_pad_type_0"), val = tensor("valid")]; + tensor x_259_strides_0 = const()[name = tensor("x_259_strides_0"), val = tensor([1])]; + tensor x_259_pad_0 = const()[name = tensor("x_259_pad_0"), val = tensor([0, 0])]; + tensor x_259_dilations_0 = const()[name = tensor("x_259_dilations_0"), val = tensor([1])]; + tensor x_259_groups_0 = const()[name = tensor("x_259_groups_0"), val = tensor(1)]; + tensor encoder_layers_9_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_9_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122444800)))]; + tensor x_259_cast_fp16 = conv(dilations = x_259_dilations_0, groups = x_259_groups_0, pad = x_259_pad_0, pad_type = x_259_pad_type_0, strides = x_259_strides_0, weight = encoder_layers_9_conv_pointwise_conv2_weight_to_fp16, x = input_527_cast_fp16)[name = tensor("x_259_cast_fp16")]; + tensor input_529_perm_0 = const()[name = tensor("input_529_perm_0"), val = tensor([0, 2, 1])]; + tensor input_529_cast_fp16 = transpose(perm = input_529_perm_0, x = x_259_cast_fp16)[name = tensor("transpose_149")]; + tensor input_531_cast_fp16 = add(x = input_515_cast_fp16, y = input_529_cast_fp16)[name = tensor("input_531_cast_fp16")]; + tensor input_533_axes_0 = const()[name = tensor("input_533_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122969152)))]; + tensor encoder_layers_9_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122970240)))]; + tensor input_533_cast_fp16 = layer_norm(axes = input_533_axes_0, beta = encoder_layers_9_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_norm_feed_forward2_weight_to_fp16, x = input_531_cast_fp16)[name = tensor("input_533_cast_fp16")]; + tensor encoder_layers_9_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_9_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122971328)))]; + tensor linear_89_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_9_feed_forward2_linear1_weight_to_fp16, x = input_533_cast_fp16)[name = tensor("linear_89_cast_fp16")]; + tensor input_537_cast_fp16 = silu(x = linear_89_cast_fp16)[name = tensor("input_537_cast_fp16")]; + tensor encoder_layers_9_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_9_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125068544)))]; + tensor linear_90_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_9_feed_forward2_linear2_weight_to_fp16, x = input_537_cast_fp16)[name = tensor("linear_90_cast_fp16")]; + tensor var_2471_to_fp16 = const()[name = tensor("op_2471_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2472_cast_fp16 = mul(x = linear_90_cast_fp16, y = var_2471_to_fp16)[name = tensor("op_2472_cast_fp16")]; + tensor input_543_cast_fp16 = add(x = input_531_cast_fp16, y = var_2472_cast_fp16)[name = tensor("input_543_cast_fp16")]; + tensor input_545_axes_0 = const()[name = tensor("input_545_axes_0"), val = tensor([-1])]; + tensor encoder_layers_9_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_9_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127165760)))]; + tensor encoder_layers_9_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_9_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127166848)))]; + tensor input_545_cast_fp16 = layer_norm(axes = input_545_axes_0, beta = encoder_layers_9_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_9_norm_out_weight_to_fp16, x = input_543_cast_fp16)[name = tensor("input_545_cast_fp16")]; + tensor cache_41_begin_0 = const()[name = tensor("cache_41_begin_0"), val = tensor([10, 0, 0, 0])]; + tensor cache_41_end_0 = const()[name = tensor("cache_41_end_0"), val = tensor([11, 1, 70, 512])]; + tensor cache_41_end_mask_0 = const()[name = tensor("cache_41_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_41_squeeze_mask_0 = const()[name = tensor("cache_41_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_41_cast_fp16 = slice_by_index(begin = cache_41_begin_0, end = cache_41_end_0, end_mask = cache_41_end_mask_0, squeeze_mask = cache_41_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_41_cast_fp16")]; + tensor cache_43_begin_0 = const()[name = tensor("cache_43_begin_0"), val = tensor([10, 0, 0, 0])]; + tensor cache_43_end_0 = const()[name = tensor("cache_43_end_0"), val = tensor([11, 1, 512, 8])]; + tensor cache_43_end_mask_0 = const()[name = tensor("cache_43_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_43_squeeze_mask_0 = const()[name = tensor("cache_43_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_43_cast_fp16 = slice_by_index(begin = cache_43_begin_0, end = cache_43_end_0, end_mask = cache_43_end_mask_0, squeeze_mask = cache_43_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_43_cast_fp16")]; + tensor input_547_axes_0 = const()[name = tensor("input_547_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127167936)))]; + tensor encoder_layers_10_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127169024)))]; + tensor input_547_cast_fp16 = layer_norm(axes = input_547_axes_0, beta = encoder_layers_10_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_norm_feed_forward1_weight_to_fp16, x = input_545_cast_fp16)[name = tensor("input_547_cast_fp16")]; + tensor encoder_layers_10_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_10_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127170112)))]; + tensor linear_91_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_10_feed_forward1_linear1_weight_to_fp16, x = input_547_cast_fp16)[name = tensor("linear_91_cast_fp16")]; + tensor input_551_cast_fp16 = silu(x = linear_91_cast_fp16)[name = tensor("input_551_cast_fp16")]; + tensor encoder_layers_10_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_10_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129267328)))]; + tensor linear_92_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_feed_forward1_linear2_weight_to_fp16, x = input_551_cast_fp16)[name = tensor("linear_92_cast_fp16")]; + tensor var_2506_to_fp16 = const()[name = tensor("op_2506_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2507_cast_fp16 = mul(x = linear_92_cast_fp16, y = var_2506_to_fp16)[name = tensor("op_2507_cast_fp16")]; + tensor input_557_cast_fp16 = add(x = input_545_cast_fp16, y = var_2507_cast_fp16)[name = tensor("input_557_cast_fp16")]; + tensor key_21_axes_0 = const()[name = tensor("key_21_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131364544)))]; + tensor encoder_layers_10_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131365632)))]; + tensor key_21_cast_fp16 = layer_norm(axes = key_21_axes_0, beta = encoder_layers_10_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_norm_self_att_weight_to_fp16, x = input_557_cast_fp16)[name = tensor("key_21_cast_fp16")]; + tensor input_559_interleave_0 = const()[name = tensor("input_559_interleave_0"), val = tensor(false)]; + tensor input_559_cast_fp16 = concat(axis = var_65, interleave = input_559_interleave_0, values = (cache_41_cast_fp16, key_21_cast_fp16))[name = tensor("input_559_cast_fp16")]; + tensor var_2529_begin_0 = const()[name = tensor("op_2529_begin_0"), val = tensor([0, 4, 0])]; + tensor var_2529_end_0 = const()[name = tensor("op_2529_end_0"), val = tensor([1, 70, 512])]; + tensor var_2529_end_mask_0 = const()[name = tensor("op_2529_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2529_cast_fp16 = slice_by_index(begin = var_2529_begin_0, end = var_2529_end_0, end_mask = var_2529_end_mask_0, x = cache_41_cast_fp16)[name = tensor("op_2529_cast_fp16")]; + tensor var_2532_begin_0 = const()[name = tensor("op_2532_begin_0"), val = tensor([0, 0, 0])]; + tensor var_2532_end_0 = const()[name = tensor("op_2532_end_0"), val = tensor([1, 4, 512])]; + tensor var_2532_end_mask_0 = const()[name = tensor("op_2532_end_mask_0"), val = tensor([true, false, true])]; + tensor var_2532_cast_fp16 = slice_by_index(begin = var_2532_begin_0, end = var_2532_end_0, end_mask = var_2532_end_mask_0, x = key_21_cast_fp16)[name = tensor("op_2532_cast_fp16")]; + tensor var_2535_interleave_0 = const()[name = tensor("op_2535_interleave_0"), val = tensor(false)]; + tensor var_2535_cast_fp16 = concat(axis = var_65, interleave = var_2535_interleave_0, values = (var_2529_cast_fp16, var_2532_cast_fp16))[name = tensor("op_2535_cast_fp16")]; + tensor encoder_layers_10_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131366720)))]; + tensor linear_93_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_self_attn_linear_q_weight_to_fp16, x = key_21_cast_fp16)[name = tensor("linear_93_cast_fp16")]; + tensor var_2539 = const()[name = tensor("op_2539"), val = tensor([1, -1, 8, 64])]; + tensor q_61_cast_fp16 = reshape(shape = var_2539, x = linear_93_cast_fp16)[name = tensor("q_61_cast_fp16")]; + tensor encoder_layers_10_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131891072)))]; + tensor linear_94_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_self_attn_linear_k_weight_to_fp16, x = input_559_cast_fp16)[name = tensor("linear_94_cast_fp16")]; + tensor var_2543 = const()[name = tensor("op_2543"), val = tensor([1, -1, 8, 64])]; + tensor k_41_cast_fp16 = reshape(shape = var_2543, x = linear_94_cast_fp16)[name = tensor("k_41_cast_fp16")]; + tensor encoder_layers_10_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132415424)))]; + tensor linear_95_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_self_attn_linear_v_weight_to_fp16, x = input_559_cast_fp16)[name = tensor("linear_95_cast_fp16")]; + tensor var_2547 = const()[name = tensor("op_2547"), val = tensor([1, -1, 8, 64])]; + tensor v_21_cast_fp16 = reshape(shape = var_2547, x = linear_95_cast_fp16)[name = tensor("v_21_cast_fp16")]; + tensor value_23_perm_0 = const()[name = tensor("value_23_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_10_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132939776)))]; + tensor var_2559_cast_fp16 = add(x = q_61_cast_fp16, y = encoder_layers_10_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2559_cast_fp16")]; + tensor encoder_layers_10_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132940864)))]; + tensor var_2561_cast_fp16 = add(x = q_61_cast_fp16, y = encoder_layers_10_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2561_cast_fp16")]; + tensor q_with_bias_v_21_perm_0 = const()[name = tensor("q_with_bias_v_21_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_267_transpose_x_0 = const()[name = tensor("x_267_transpose_x_0"), val = tensor(false)]; + tensor x_267_transpose_y_0 = const()[name = tensor("x_267_transpose_y_0"), val = tensor(false)]; + tensor var_2563_to_fp16 = const()[name = tensor("op_2563_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132941952)))]; + tensor q_with_bias_v_21_cast_fp16 = transpose(perm = q_with_bias_v_21_perm_0, x = var_2561_cast_fp16)[name = tensor("transpose_147")]; + tensor x_267_cast_fp16 = matmul(transpose_x = x_267_transpose_x_0, transpose_y = x_267_transpose_y_0, x = q_with_bias_v_21_cast_fp16, y = var_2563_to_fp16)[name = tensor("x_267_cast_fp16")]; + tensor x_269_pad_0 = const()[name = tensor("x_269_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_269_mode_0 = const()[name = tensor("x_269_mode_0"), val = tensor("constant")]; + tensor const_209_to_fp16 = const()[name = tensor("const_209_to_fp16"), val = tensor(0x0p+0)]; + tensor x_269_cast_fp16 = pad(constant_val = const_209_to_fp16, mode = x_269_mode_0, pad = x_269_pad_0, x = x_267_cast_fp16)[name = tensor("x_269_cast_fp16")]; + tensor var_2571 = const()[name = tensor("op_2571"), val = tensor([1, 8, -1, 8])]; + tensor x_271_cast_fp16 = reshape(shape = var_2571, x = x_269_cast_fp16)[name = tensor("x_271_cast_fp16")]; + tensor var_2575_begin_0 = const()[name = tensor("op_2575_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2575_end_0 = const()[name = tensor("op_2575_end_0"), val = tensor([1, 8, 156, 8])]; + tensor var_2575_end_mask_0 = const()[name = tensor("op_2575_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2575_cast_fp16 = slice_by_index(begin = var_2575_begin_0, end = var_2575_end_0, end_mask = var_2575_end_mask_0, x = x_271_cast_fp16)[name = tensor("op_2575_cast_fp16")]; + tensor var_2576 = const()[name = tensor("op_2576"), val = tensor([1, 8, 8, 155])]; + tensor matrix_bd_41_cast_fp16 = reshape(shape = var_2576, x = var_2575_cast_fp16)[name = tensor("matrix_bd_41_cast_fp16")]; + tensor matrix_ac_21_transpose_x_0 = const()[name = tensor("matrix_ac_21_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_21_transpose_y_0 = const()[name = tensor("matrix_ac_21_transpose_y_0"), val = tensor(false)]; + tensor transpose_71_perm_0 = const()[name = tensor("transpose_71_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_72_perm_0 = const()[name = tensor("transpose_72_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_72 = transpose(perm = transpose_72_perm_0, x = k_41_cast_fp16)[name = tensor("transpose_145")]; + tensor transpose_71 = transpose(perm = transpose_71_perm_0, x = var_2559_cast_fp16)[name = tensor("transpose_146")]; + tensor matrix_ac_21_cast_fp16 = matmul(transpose_x = matrix_ac_21_transpose_x_0, transpose_y = matrix_ac_21_transpose_y_0, x = transpose_71, y = transpose_72)[name = tensor("matrix_ac_21_cast_fp16")]; + tensor matrix_bd_43_begin_0 = const()[name = tensor("matrix_bd_43_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_43_end_0 = const()[name = tensor("matrix_bd_43_end_0"), val = tensor([1, 8, 8, 78])]; + tensor matrix_bd_43_end_mask_0 = const()[name = tensor("matrix_bd_43_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_43_cast_fp16 = slice_by_index(begin = matrix_bd_43_begin_0, end = matrix_bd_43_end_0, end_mask = matrix_bd_43_end_mask_0, x = matrix_bd_41_cast_fp16)[name = tensor("matrix_bd_43_cast_fp16")]; + tensor var_2585_cast_fp16 = add(x = matrix_ac_21_cast_fp16, y = matrix_bd_43_cast_fp16)[name = tensor("op_2585_cast_fp16")]; + tensor _inversed_scores_41_y_0_to_fp16 = const()[name = tensor("_inversed_scores_41_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_41_cast_fp16 = mul(x = var_2585_cast_fp16, y = _inversed_scores_41_y_0_to_fp16)[name = tensor("_inversed_scores_41_cast_fp16")]; + tensor scores_43_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_41_cast_fp16, cond = mask_11)[name = tensor("scores_43_cast_fp16")]; + tensor var_2591_cast_fp16 = softmax(axis = var_56, x = scores_43_cast_fp16)[name = tensor("op_2591_cast_fp16")]; + tensor input_561_cast_fp16 = select(a = var_40_to_fp16, b = var_2591_cast_fp16, cond = mask_11)[name = tensor("input_561_cast_fp16")]; + tensor x_273_transpose_x_0 = const()[name = tensor("x_273_transpose_x_0"), val = tensor(false)]; + tensor x_273_transpose_y_0 = const()[name = tensor("x_273_transpose_y_0"), val = tensor(false)]; + tensor value_23_cast_fp16 = transpose(perm = value_23_perm_0, x = v_21_cast_fp16)[name = tensor("transpose_148")]; + tensor x_273_cast_fp16 = matmul(transpose_x = x_273_transpose_x_0, transpose_y = x_273_transpose_y_0, x = input_561_cast_fp16, y = value_23_cast_fp16)[name = tensor("x_273_cast_fp16")]; + tensor var_2595_perm_0 = const()[name = tensor("op_2595_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2596 = const()[name = tensor("op_2596"), val = tensor([1, -1, 512])]; + tensor var_2595_cast_fp16 = transpose(perm = var_2595_perm_0, x = x_273_cast_fp16)[name = tensor("transpose_144")]; + tensor input_563_cast_fp16 = reshape(shape = var_2596, x = var_2595_cast_fp16)[name = tensor("input_563_cast_fp16")]; + tensor encoder_layers_10_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_10_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133100736)))]; + tensor linear_97_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_self_attn_linear_out_weight_to_fp16, x = input_563_cast_fp16)[name = tensor("linear_97_cast_fp16")]; + tensor input_567_cast_fp16 = add(x = input_557_cast_fp16, y = linear_97_cast_fp16)[name = tensor("input_567_cast_fp16")]; + tensor x_277_axes_0 = const()[name = tensor("x_277_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133625088)))]; + tensor encoder_layers_10_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133626176)))]; + tensor x_277_cast_fp16 = layer_norm(axes = x_277_axes_0, beta = encoder_layers_10_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_norm_conv_weight_to_fp16, x = input_567_cast_fp16)[name = tensor("x_277_cast_fp16")]; + tensor input_569_perm_0 = const()[name = tensor("input_569_perm_0"), val = tensor([0, 2, 1])]; + tensor input_571_pad_type_0 = const()[name = tensor("input_571_pad_type_0"), val = tensor("valid")]; + tensor input_571_strides_0 = const()[name = tensor("input_571_strides_0"), val = tensor([1])]; + tensor input_571_pad_0 = const()[name = tensor("input_571_pad_0"), val = tensor([0, 0])]; + tensor input_571_dilations_0 = const()[name = tensor("input_571_dilations_0"), val = tensor([1])]; + tensor input_571_groups_0 = const()[name = tensor("input_571_groups_0"), val = tensor(1)]; + tensor encoder_layers_10_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_10_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133627264)))]; + tensor input_569_cast_fp16 = transpose(perm = input_569_perm_0, x = x_277_cast_fp16)[name = tensor("transpose_143")]; + tensor input_571_cast_fp16 = conv(dilations = input_571_dilations_0, groups = input_571_groups_0, pad = input_571_pad_0, pad_type = input_571_pad_type_0, strides = input_571_strides_0, weight = encoder_layers_10_conv_pointwise_conv1_weight_to_fp16, x = input_569_cast_fp16)[name = tensor("input_571_cast_fp16")]; + tensor x_279_split_num_splits_0 = const()[name = tensor("x_279_split_num_splits_0"), val = tensor(2)]; + tensor x_279_split_axis_0 = const()[name = tensor("x_279_split_axis_0"), val = tensor(1)]; + tensor x_279_split_cast_fp16_0, tensor x_279_split_cast_fp16_1 = split(axis = x_279_split_axis_0, num_splits = x_279_split_num_splits_0, x = input_571_cast_fp16)[name = tensor("x_279_split_cast_fp16")]; + tensor x_279_split_1_sigmoid_cast_fp16 = sigmoid(x = x_279_split_cast_fp16_1)[name = tensor("x_279_split_1_sigmoid_cast_fp16")]; + tensor x_279_cast_fp16 = mul(x = x_279_split_cast_fp16_0, y = x_279_split_1_sigmoid_cast_fp16)[name = tensor("x_279_cast_fp16")]; + tensor input_573_cast_fp16 = select(a = var_40_to_fp16, b = x_279_cast_fp16, cond = var_551)[name = tensor("input_573_cast_fp16")]; + tensor new_x_43_interleave_0 = const()[name = tensor("new_x_43_interleave_0"), val = tensor(false)]; + tensor new_x_43_cast_fp16 = concat(axis = var_56, interleave = new_x_43_interleave_0, values = (cache_43_cast_fp16, input_573_cast_fp16))[name = tensor("new_x_43_cast_fp16")]; + tensor next_cache_21_begin_0 = const()[name = tensor("next_cache_21_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_21_end_0 = const()[name = tensor("next_cache_21_end_0"), val = tensor([1, 512, 12])]; + tensor next_cache_21_end_mask_0 = const()[name = tensor("next_cache_21_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_21_cast_fp16 = slice_by_index(begin = next_cache_21_begin_0, end = next_cache_21_end_0, end_mask = next_cache_21_end_mask_0, x = new_x_43_cast_fp16)[name = tensor("next_cache_21_cast_fp16")]; + tensor var_2637_begin_0 = const()[name = tensor("op_2637_begin_0"), val = tensor([0, 0, 4])]; + tensor var_2637_end_0 = const()[name = tensor("op_2637_end_0"), val = tensor([1, 512, 12])]; + tensor var_2637_end_mask_0 = const()[name = tensor("op_2637_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2637_cast_fp16 = slice_by_index(begin = var_2637_begin_0, end = var_2637_end_0, end_mask = var_2637_end_mask_0, x = next_cache_21_cast_fp16)[name = tensor("op_2637_cast_fp16")]; + tensor x_281_pad_type_0 = const()[name = tensor("x_281_pad_type_0"), val = tensor("valid")]; + tensor x_281_groups_0 = const()[name = tensor("x_281_groups_0"), val = tensor(512)]; + tensor x_281_strides_0 = const()[name = tensor("x_281_strides_0"), val = tensor([1])]; + tensor x_281_pad_0 = const()[name = tensor("x_281_pad_0"), val = tensor([0, 0])]; + tensor x_281_dilations_0 = const()[name = tensor("x_281_dilations_0"), val = tensor([1])]; + tensor encoder_layers_10_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_10_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134675904)))]; + tensor x_281_cast_fp16 = conv(dilations = x_281_dilations_0, groups = x_281_groups_0, pad = x_281_pad_0, pad_type = x_281_pad_type_0, strides = x_281_strides_0, weight = encoder_layers_10_conv_depthwise_conv_weight_to_fp16, x = new_x_43_cast_fp16)[name = tensor("x_281_cast_fp16")]; + tensor input_575_perm_0 = const()[name = tensor("input_575_perm_0"), val = tensor([0, 2, 1])]; + tensor x_283_axes_0 = const()[name = tensor("x_283_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_10_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134685184)))]; + tensor encoder_layers_10_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_10_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134686272)))]; + tensor input_575_cast_fp16 = transpose(perm = input_575_perm_0, x = x_281_cast_fp16)[name = tensor("transpose_142")]; + tensor x_283_cast_fp16 = layer_norm(axes = x_283_axes_0, beta = encoder_layers_10_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_conv_batch_norm_weight_to_fp16, x = input_575_cast_fp16)[name = tensor("x_283_cast_fp16")]; + tensor input_577_perm_0 = const()[name = tensor("input_577_perm_0"), val = tensor([0, 2, 1])]; + tensor input_577_cast_fp16 = transpose(perm = input_577_perm_0, x = x_283_cast_fp16)[name = tensor("transpose_141")]; + tensor input_579_cast_fp16 = silu(x = input_577_cast_fp16)[name = tensor("input_579_cast_fp16")]; + tensor x_285_pad_type_0 = const()[name = tensor("x_285_pad_type_0"), val = tensor("valid")]; + tensor x_285_strides_0 = const()[name = tensor("x_285_strides_0"), val = tensor([1])]; + tensor x_285_pad_0 = const()[name = tensor("x_285_pad_0"), val = tensor([0, 0])]; + tensor x_285_dilations_0 = const()[name = tensor("x_285_dilations_0"), val = tensor([1])]; + tensor x_285_groups_0 = const()[name = tensor("x_285_groups_0"), val = tensor(1)]; + tensor encoder_layers_10_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_10_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134687360)))]; + tensor x_285_cast_fp16 = conv(dilations = x_285_dilations_0, groups = x_285_groups_0, pad = x_285_pad_0, pad_type = x_285_pad_type_0, strides = x_285_strides_0, weight = encoder_layers_10_conv_pointwise_conv2_weight_to_fp16, x = input_579_cast_fp16)[name = tensor("x_285_cast_fp16")]; + tensor input_581_perm_0 = const()[name = tensor("input_581_perm_0"), val = tensor([0, 2, 1])]; + tensor input_581_cast_fp16 = transpose(perm = input_581_perm_0, x = x_285_cast_fp16)[name = tensor("transpose_140")]; + tensor input_583_cast_fp16 = add(x = input_567_cast_fp16, y = input_581_cast_fp16)[name = tensor("input_583_cast_fp16")]; + tensor input_585_axes_0 = const()[name = tensor("input_585_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135211712)))]; + tensor encoder_layers_10_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135212800)))]; + tensor input_585_cast_fp16 = layer_norm(axes = input_585_axes_0, beta = encoder_layers_10_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_norm_feed_forward2_weight_to_fp16, x = input_583_cast_fp16)[name = tensor("input_585_cast_fp16")]; + tensor encoder_layers_10_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_10_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135213888)))]; + tensor linear_98_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_10_feed_forward2_linear1_weight_to_fp16, x = input_585_cast_fp16)[name = tensor("linear_98_cast_fp16")]; + tensor input_589_cast_fp16 = silu(x = linear_98_cast_fp16)[name = tensor("input_589_cast_fp16")]; + tensor encoder_layers_10_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_10_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(137311104)))]; + tensor linear_99_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_10_feed_forward2_linear2_weight_to_fp16, x = input_589_cast_fp16)[name = tensor("linear_99_cast_fp16")]; + tensor var_2678_to_fp16 = const()[name = tensor("op_2678_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2679_cast_fp16 = mul(x = linear_99_cast_fp16, y = var_2678_to_fp16)[name = tensor("op_2679_cast_fp16")]; + tensor input_595_cast_fp16 = add(x = input_583_cast_fp16, y = var_2679_cast_fp16)[name = tensor("input_595_cast_fp16")]; + tensor input_597_axes_0 = const()[name = tensor("input_597_axes_0"), val = tensor([-1])]; + tensor encoder_layers_10_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_10_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139408320)))]; + tensor encoder_layers_10_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_10_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139409408)))]; + tensor input_597_cast_fp16 = layer_norm(axes = input_597_axes_0, beta = encoder_layers_10_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_10_norm_out_weight_to_fp16, x = input_595_cast_fp16)[name = tensor("input_597_cast_fp16")]; + tensor cache_45_begin_0 = const()[name = tensor("cache_45_begin_0"), val = tensor([11, 0, 0, 0])]; + tensor cache_45_end_0 = const()[name = tensor("cache_45_end_0"), val = tensor([12, 1, 70, 512])]; + tensor cache_45_end_mask_0 = const()[name = tensor("cache_45_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_45_squeeze_mask_0 = const()[name = tensor("cache_45_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_45_cast_fp16 = slice_by_index(begin = cache_45_begin_0, end = cache_45_end_0, end_mask = cache_45_end_mask_0, squeeze_mask = cache_45_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_45_cast_fp16")]; + tensor cache_47_begin_0 = const()[name = tensor("cache_47_begin_0"), val = tensor([11, 0, 0, 0])]; + tensor cache_47_end_0 = const()[name = tensor("cache_47_end_0"), val = tensor([12, 1, 512, 8])]; + tensor cache_47_end_mask_0 = const()[name = tensor("cache_47_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_47_squeeze_mask_0 = const()[name = tensor("cache_47_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_47_cast_fp16 = slice_by_index(begin = cache_47_begin_0, end = cache_47_end_0, end_mask = cache_47_end_mask_0, squeeze_mask = cache_47_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_47_cast_fp16")]; + tensor input_599_axes_0 = const()[name = tensor("input_599_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139410496)))]; + tensor encoder_layers_11_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139411584)))]; + tensor input_599_cast_fp16 = layer_norm(axes = input_599_axes_0, beta = encoder_layers_11_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_norm_feed_forward1_weight_to_fp16, x = input_597_cast_fp16)[name = tensor("input_599_cast_fp16")]; + tensor encoder_layers_11_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_11_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139412672)))]; + tensor linear_100_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_11_feed_forward1_linear1_weight_to_fp16, x = input_599_cast_fp16)[name = tensor("linear_100_cast_fp16")]; + tensor input_603_cast_fp16 = silu(x = linear_100_cast_fp16)[name = tensor("input_603_cast_fp16")]; + tensor encoder_layers_11_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_11_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141509888)))]; + tensor linear_101_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_feed_forward1_linear2_weight_to_fp16, x = input_603_cast_fp16)[name = tensor("linear_101_cast_fp16")]; + tensor var_2713_to_fp16 = const()[name = tensor("op_2713_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2714_cast_fp16 = mul(x = linear_101_cast_fp16, y = var_2713_to_fp16)[name = tensor("op_2714_cast_fp16")]; + tensor input_609_cast_fp16 = add(x = input_597_cast_fp16, y = var_2714_cast_fp16)[name = tensor("input_609_cast_fp16")]; + tensor key_23_axes_0 = const()[name = tensor("key_23_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143607104)))]; + tensor encoder_layers_11_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143608192)))]; + tensor key_23_cast_fp16 = layer_norm(axes = key_23_axes_0, beta = encoder_layers_11_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_norm_self_att_weight_to_fp16, x = input_609_cast_fp16)[name = tensor("key_23_cast_fp16")]; + tensor input_611_interleave_0 = const()[name = tensor("input_611_interleave_0"), val = tensor(false)]; + tensor input_611_cast_fp16 = concat(axis = var_65, interleave = input_611_interleave_0, values = (cache_45_cast_fp16, key_23_cast_fp16))[name = tensor("input_611_cast_fp16")]; + tensor var_2736_begin_0 = const()[name = tensor("op_2736_begin_0"), val = tensor([0, 4, 0])]; + tensor var_2736_end_0 = const()[name = tensor("op_2736_end_0"), val = tensor([1, 70, 512])]; + tensor var_2736_end_mask_0 = const()[name = tensor("op_2736_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2736_cast_fp16 = slice_by_index(begin = var_2736_begin_0, end = var_2736_end_0, end_mask = var_2736_end_mask_0, x = cache_45_cast_fp16)[name = tensor("op_2736_cast_fp16")]; + tensor var_2739_begin_0 = const()[name = tensor("op_2739_begin_0"), val = tensor([0, 0, 0])]; + tensor var_2739_end_0 = const()[name = tensor("op_2739_end_0"), val = tensor([1, 4, 512])]; + tensor var_2739_end_mask_0 = const()[name = tensor("op_2739_end_mask_0"), val = tensor([true, false, true])]; + tensor var_2739_cast_fp16 = slice_by_index(begin = var_2739_begin_0, end = var_2739_end_0, end_mask = var_2739_end_mask_0, x = key_23_cast_fp16)[name = tensor("op_2739_cast_fp16")]; + tensor var_2742_interleave_0 = const()[name = tensor("op_2742_interleave_0"), val = tensor(false)]; + tensor var_2742_cast_fp16 = concat(axis = var_65, interleave = var_2742_interleave_0, values = (var_2736_cast_fp16, var_2739_cast_fp16))[name = tensor("op_2742_cast_fp16")]; + tensor encoder_layers_11_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143609280)))]; + tensor linear_102_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_self_attn_linear_q_weight_to_fp16, x = key_23_cast_fp16)[name = tensor("linear_102_cast_fp16")]; + tensor var_2746 = const()[name = tensor("op_2746"), val = tensor([1, -1, 8, 64])]; + tensor q_67_cast_fp16 = reshape(shape = var_2746, x = linear_102_cast_fp16)[name = tensor("q_67_cast_fp16")]; + tensor encoder_layers_11_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144133632)))]; + tensor linear_103_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_self_attn_linear_k_weight_to_fp16, x = input_611_cast_fp16)[name = tensor("linear_103_cast_fp16")]; + tensor var_2750 = const()[name = tensor("op_2750"), val = tensor([1, -1, 8, 64])]; + tensor k_45_cast_fp16 = reshape(shape = var_2750, x = linear_103_cast_fp16)[name = tensor("k_45_cast_fp16")]; + tensor encoder_layers_11_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144657984)))]; + tensor linear_104_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_self_attn_linear_v_weight_to_fp16, x = input_611_cast_fp16)[name = tensor("linear_104_cast_fp16")]; + tensor var_2754 = const()[name = tensor("op_2754"), val = tensor([1, -1, 8, 64])]; + tensor v_23_cast_fp16 = reshape(shape = var_2754, x = linear_104_cast_fp16)[name = tensor("v_23_cast_fp16")]; + tensor value_25_perm_0 = const()[name = tensor("value_25_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_11_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145182336)))]; + tensor var_2766_cast_fp16 = add(x = q_67_cast_fp16, y = encoder_layers_11_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2766_cast_fp16")]; + tensor encoder_layers_11_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145183424)))]; + tensor var_2768_cast_fp16 = add(x = q_67_cast_fp16, y = encoder_layers_11_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2768_cast_fp16")]; + tensor q_with_bias_v_23_perm_0 = const()[name = tensor("q_with_bias_v_23_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_293_transpose_x_0 = const()[name = tensor("x_293_transpose_x_0"), val = tensor(false)]; + tensor x_293_transpose_y_0 = const()[name = tensor("x_293_transpose_y_0"), val = tensor(false)]; + tensor var_2770_to_fp16 = const()[name = tensor("op_2770_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145184512)))]; + tensor q_with_bias_v_23_cast_fp16 = transpose(perm = q_with_bias_v_23_perm_0, x = var_2768_cast_fp16)[name = tensor("transpose_138")]; + tensor x_293_cast_fp16 = matmul(transpose_x = x_293_transpose_x_0, transpose_y = x_293_transpose_y_0, x = q_with_bias_v_23_cast_fp16, y = var_2770_to_fp16)[name = tensor("x_293_cast_fp16")]; + tensor x_295_pad_0 = const()[name = tensor("x_295_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_295_mode_0 = const()[name = tensor("x_295_mode_0"), val = tensor("constant")]; + tensor const_222_to_fp16 = const()[name = tensor("const_222_to_fp16"), val = tensor(0x0p+0)]; + tensor x_295_cast_fp16 = pad(constant_val = const_222_to_fp16, mode = x_295_mode_0, pad = x_295_pad_0, x = x_293_cast_fp16)[name = tensor("x_295_cast_fp16")]; + tensor var_2778 = const()[name = tensor("op_2778"), val = tensor([1, 8, -1, 8])]; + tensor x_297_cast_fp16 = reshape(shape = var_2778, x = x_295_cast_fp16)[name = tensor("x_297_cast_fp16")]; + tensor var_2782_begin_0 = const()[name = tensor("op_2782_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2782_end_0 = const()[name = tensor("op_2782_end_0"), val = tensor([1, 8, 156, 8])]; + tensor var_2782_end_mask_0 = const()[name = tensor("op_2782_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2782_cast_fp16 = slice_by_index(begin = var_2782_begin_0, end = var_2782_end_0, end_mask = var_2782_end_mask_0, x = x_297_cast_fp16)[name = tensor("op_2782_cast_fp16")]; + tensor var_2783 = const()[name = tensor("op_2783"), val = tensor([1, 8, 8, 155])]; + tensor matrix_bd_45_cast_fp16 = reshape(shape = var_2783, x = var_2782_cast_fp16)[name = tensor("matrix_bd_45_cast_fp16")]; + tensor matrix_ac_23_transpose_x_0 = const()[name = tensor("matrix_ac_23_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_23_transpose_y_0 = const()[name = tensor("matrix_ac_23_transpose_y_0"), val = tensor(false)]; + tensor transpose_73_perm_0 = const()[name = tensor("transpose_73_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_74_perm_0 = const()[name = tensor("transpose_74_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_74 = transpose(perm = transpose_74_perm_0, x = k_45_cast_fp16)[name = tensor("transpose_136")]; + tensor transpose_73 = transpose(perm = transpose_73_perm_0, x = var_2766_cast_fp16)[name = tensor("transpose_137")]; + tensor matrix_ac_23_cast_fp16 = matmul(transpose_x = matrix_ac_23_transpose_x_0, transpose_y = matrix_ac_23_transpose_y_0, x = transpose_73, y = transpose_74)[name = tensor("matrix_ac_23_cast_fp16")]; + tensor matrix_bd_47_begin_0 = const()[name = tensor("matrix_bd_47_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_47_end_0 = const()[name = tensor("matrix_bd_47_end_0"), val = tensor([1, 8, 8, 78])]; + tensor matrix_bd_47_end_mask_0 = const()[name = tensor("matrix_bd_47_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_47_cast_fp16 = slice_by_index(begin = matrix_bd_47_begin_0, end = matrix_bd_47_end_0, end_mask = matrix_bd_47_end_mask_0, x = matrix_bd_45_cast_fp16)[name = tensor("matrix_bd_47_cast_fp16")]; + tensor var_2792_cast_fp16 = add(x = matrix_ac_23_cast_fp16, y = matrix_bd_47_cast_fp16)[name = tensor("op_2792_cast_fp16")]; + tensor _inversed_scores_45_y_0_to_fp16 = const()[name = tensor("_inversed_scores_45_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_45_cast_fp16 = mul(x = var_2792_cast_fp16, y = _inversed_scores_45_y_0_to_fp16)[name = tensor("_inversed_scores_45_cast_fp16")]; + tensor scores_47_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_45_cast_fp16, cond = mask_11)[name = tensor("scores_47_cast_fp16")]; + tensor var_2798_cast_fp16 = softmax(axis = var_56, x = scores_47_cast_fp16)[name = tensor("op_2798_cast_fp16")]; + tensor input_613_cast_fp16 = select(a = var_40_to_fp16, b = var_2798_cast_fp16, cond = mask_11)[name = tensor("input_613_cast_fp16")]; + tensor x_299_transpose_x_0 = const()[name = tensor("x_299_transpose_x_0"), val = tensor(false)]; + tensor x_299_transpose_y_0 = const()[name = tensor("x_299_transpose_y_0"), val = tensor(false)]; + tensor value_25_cast_fp16 = transpose(perm = value_25_perm_0, x = v_23_cast_fp16)[name = tensor("transpose_139")]; + tensor x_299_cast_fp16 = matmul(transpose_x = x_299_transpose_x_0, transpose_y = x_299_transpose_y_0, x = input_613_cast_fp16, y = value_25_cast_fp16)[name = tensor("x_299_cast_fp16")]; + tensor var_2802_perm_0 = const()[name = tensor("op_2802_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2803 = const()[name = tensor("op_2803"), val = tensor([1, -1, 512])]; + tensor var_2802_cast_fp16 = transpose(perm = var_2802_perm_0, x = x_299_cast_fp16)[name = tensor("transpose_135")]; + tensor input_615_cast_fp16 = reshape(shape = var_2803, x = var_2802_cast_fp16)[name = tensor("input_615_cast_fp16")]; + tensor encoder_layers_11_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_11_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145343296)))]; + tensor linear_106_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_self_attn_linear_out_weight_to_fp16, x = input_615_cast_fp16)[name = tensor("linear_106_cast_fp16")]; + tensor input_619_cast_fp16 = add(x = input_609_cast_fp16, y = linear_106_cast_fp16)[name = tensor("input_619_cast_fp16")]; + tensor x_303_axes_0 = const()[name = tensor("x_303_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145867648)))]; + tensor encoder_layers_11_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145868736)))]; + tensor x_303_cast_fp16 = layer_norm(axes = x_303_axes_0, beta = encoder_layers_11_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_norm_conv_weight_to_fp16, x = input_619_cast_fp16)[name = tensor("x_303_cast_fp16")]; + tensor input_621_perm_0 = const()[name = tensor("input_621_perm_0"), val = tensor([0, 2, 1])]; + tensor input_623_pad_type_0 = const()[name = tensor("input_623_pad_type_0"), val = tensor("valid")]; + tensor input_623_strides_0 = const()[name = tensor("input_623_strides_0"), val = tensor([1])]; + tensor input_623_pad_0 = const()[name = tensor("input_623_pad_0"), val = tensor([0, 0])]; + tensor input_623_dilations_0 = const()[name = tensor("input_623_dilations_0"), val = tensor([1])]; + tensor input_623_groups_0 = const()[name = tensor("input_623_groups_0"), val = tensor(1)]; + tensor encoder_layers_11_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_11_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145869824)))]; + tensor input_621_cast_fp16 = transpose(perm = input_621_perm_0, x = x_303_cast_fp16)[name = tensor("transpose_134")]; + tensor input_623_cast_fp16 = conv(dilations = input_623_dilations_0, groups = input_623_groups_0, pad = input_623_pad_0, pad_type = input_623_pad_type_0, strides = input_623_strides_0, weight = encoder_layers_11_conv_pointwise_conv1_weight_to_fp16, x = input_621_cast_fp16)[name = tensor("input_623_cast_fp16")]; + tensor x_305_split_num_splits_0 = const()[name = tensor("x_305_split_num_splits_0"), val = tensor(2)]; + tensor x_305_split_axis_0 = const()[name = tensor("x_305_split_axis_0"), val = tensor(1)]; + tensor x_305_split_cast_fp16_0, tensor x_305_split_cast_fp16_1 = split(axis = x_305_split_axis_0, num_splits = x_305_split_num_splits_0, x = input_623_cast_fp16)[name = tensor("x_305_split_cast_fp16")]; + tensor x_305_split_1_sigmoid_cast_fp16 = sigmoid(x = x_305_split_cast_fp16_1)[name = tensor("x_305_split_1_sigmoid_cast_fp16")]; + tensor x_305_cast_fp16 = mul(x = x_305_split_cast_fp16_0, y = x_305_split_1_sigmoid_cast_fp16)[name = tensor("x_305_cast_fp16")]; + tensor input_625_cast_fp16 = select(a = var_40_to_fp16, b = x_305_cast_fp16, cond = var_551)[name = tensor("input_625_cast_fp16")]; + tensor new_x_47_interleave_0 = const()[name = tensor("new_x_47_interleave_0"), val = tensor(false)]; + tensor new_x_47_cast_fp16 = concat(axis = var_56, interleave = new_x_47_interleave_0, values = (cache_47_cast_fp16, input_625_cast_fp16))[name = tensor("new_x_47_cast_fp16")]; + tensor next_cache_23_begin_0 = const()[name = tensor("next_cache_23_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_23_end_0 = const()[name = tensor("next_cache_23_end_0"), val = tensor([1, 512, 12])]; + tensor next_cache_23_end_mask_0 = const()[name = tensor("next_cache_23_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_23_cast_fp16 = slice_by_index(begin = next_cache_23_begin_0, end = next_cache_23_end_0, end_mask = next_cache_23_end_mask_0, x = new_x_47_cast_fp16)[name = tensor("next_cache_23_cast_fp16")]; + tensor var_2844_begin_0 = const()[name = tensor("op_2844_begin_0"), val = tensor([0, 0, 4])]; + tensor var_2844_end_0 = const()[name = tensor("op_2844_end_0"), val = tensor([1, 512, 12])]; + tensor var_2844_end_mask_0 = const()[name = tensor("op_2844_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2844_cast_fp16 = slice_by_index(begin = var_2844_begin_0, end = var_2844_end_0, end_mask = var_2844_end_mask_0, x = next_cache_23_cast_fp16)[name = tensor("op_2844_cast_fp16")]; + tensor x_307_pad_type_0 = const()[name = tensor("x_307_pad_type_0"), val = tensor("valid")]; + tensor x_307_groups_0 = const()[name = tensor("x_307_groups_0"), val = tensor(512)]; + tensor x_307_strides_0 = const()[name = tensor("x_307_strides_0"), val = tensor([1])]; + tensor x_307_pad_0 = const()[name = tensor("x_307_pad_0"), val = tensor([0, 0])]; + tensor x_307_dilations_0 = const()[name = tensor("x_307_dilations_0"), val = tensor([1])]; + tensor encoder_layers_11_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_11_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146918464)))]; + tensor x_307_cast_fp16 = conv(dilations = x_307_dilations_0, groups = x_307_groups_0, pad = x_307_pad_0, pad_type = x_307_pad_type_0, strides = x_307_strides_0, weight = encoder_layers_11_conv_depthwise_conv_weight_to_fp16, x = new_x_47_cast_fp16)[name = tensor("x_307_cast_fp16")]; + tensor input_627_perm_0 = const()[name = tensor("input_627_perm_0"), val = tensor([0, 2, 1])]; + tensor x_309_axes_0 = const()[name = tensor("x_309_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_11_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146927744)))]; + tensor encoder_layers_11_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_11_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146928832)))]; + tensor input_627_cast_fp16 = transpose(perm = input_627_perm_0, x = x_307_cast_fp16)[name = tensor("transpose_133")]; + tensor x_309_cast_fp16 = layer_norm(axes = x_309_axes_0, beta = encoder_layers_11_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_conv_batch_norm_weight_to_fp16, x = input_627_cast_fp16)[name = tensor("x_309_cast_fp16")]; + tensor input_629_perm_0 = const()[name = tensor("input_629_perm_0"), val = tensor([0, 2, 1])]; + tensor input_629_cast_fp16 = transpose(perm = input_629_perm_0, x = x_309_cast_fp16)[name = tensor("transpose_132")]; + tensor input_631_cast_fp16 = silu(x = input_629_cast_fp16)[name = tensor("input_631_cast_fp16")]; + tensor x_311_pad_type_0 = const()[name = tensor("x_311_pad_type_0"), val = tensor("valid")]; + tensor x_311_strides_0 = const()[name = tensor("x_311_strides_0"), val = tensor([1])]; + tensor x_311_pad_0 = const()[name = tensor("x_311_pad_0"), val = tensor([0, 0])]; + tensor x_311_dilations_0 = const()[name = tensor("x_311_dilations_0"), val = tensor([1])]; + tensor x_311_groups_0 = const()[name = tensor("x_311_groups_0"), val = tensor(1)]; + tensor encoder_layers_11_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_11_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146929920)))]; + tensor x_311_cast_fp16 = conv(dilations = x_311_dilations_0, groups = x_311_groups_0, pad = x_311_pad_0, pad_type = x_311_pad_type_0, strides = x_311_strides_0, weight = encoder_layers_11_conv_pointwise_conv2_weight_to_fp16, x = input_631_cast_fp16)[name = tensor("x_311_cast_fp16")]; + tensor input_633_perm_0 = const()[name = tensor("input_633_perm_0"), val = tensor([0, 2, 1])]; + tensor input_633_cast_fp16 = transpose(perm = input_633_perm_0, x = x_311_cast_fp16)[name = tensor("transpose_131")]; + tensor input_635_cast_fp16 = add(x = input_619_cast_fp16, y = input_633_cast_fp16)[name = tensor("input_635_cast_fp16")]; + tensor input_637_axes_0 = const()[name = tensor("input_637_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147454272)))]; + tensor encoder_layers_11_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147455360)))]; + tensor input_637_cast_fp16 = layer_norm(axes = input_637_axes_0, beta = encoder_layers_11_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_norm_feed_forward2_weight_to_fp16, x = input_635_cast_fp16)[name = tensor("input_637_cast_fp16")]; + tensor encoder_layers_11_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_11_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147456448)))]; + tensor linear_107_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_11_feed_forward2_linear1_weight_to_fp16, x = input_637_cast_fp16)[name = tensor("linear_107_cast_fp16")]; + tensor input_641_cast_fp16 = silu(x = linear_107_cast_fp16)[name = tensor("input_641_cast_fp16")]; + tensor encoder_layers_11_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_11_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(149553664)))]; + tensor linear_108_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_11_feed_forward2_linear2_weight_to_fp16, x = input_641_cast_fp16)[name = tensor("linear_108_cast_fp16")]; + tensor var_2885_to_fp16 = const()[name = tensor("op_2885_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2886_cast_fp16 = mul(x = linear_108_cast_fp16, y = var_2885_to_fp16)[name = tensor("op_2886_cast_fp16")]; + tensor input_647_cast_fp16 = add(x = input_635_cast_fp16, y = var_2886_cast_fp16)[name = tensor("input_647_cast_fp16")]; + tensor input_649_axes_0 = const()[name = tensor("input_649_axes_0"), val = tensor([-1])]; + tensor encoder_layers_11_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_11_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151650880)))]; + tensor encoder_layers_11_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_11_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151651968)))]; + tensor input_649_cast_fp16 = layer_norm(axes = input_649_axes_0, beta = encoder_layers_11_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_11_norm_out_weight_to_fp16, x = input_647_cast_fp16)[name = tensor("input_649_cast_fp16")]; + tensor cache_49_begin_0 = const()[name = tensor("cache_49_begin_0"), val = tensor([12, 0, 0, 0])]; + tensor cache_49_end_0 = const()[name = tensor("cache_49_end_0"), val = tensor([13, 1, 70, 512])]; + tensor cache_49_end_mask_0 = const()[name = tensor("cache_49_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_49_squeeze_mask_0 = const()[name = tensor("cache_49_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_49_cast_fp16 = slice_by_index(begin = cache_49_begin_0, end = cache_49_end_0, end_mask = cache_49_end_mask_0, squeeze_mask = cache_49_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_49_cast_fp16")]; + tensor cache_51_begin_0 = const()[name = tensor("cache_51_begin_0"), val = tensor([12, 0, 0, 0])]; + tensor cache_51_end_0 = const()[name = tensor("cache_51_end_0"), val = tensor([13, 1, 512, 8])]; + tensor cache_51_end_mask_0 = const()[name = tensor("cache_51_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_51_squeeze_mask_0 = const()[name = tensor("cache_51_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_51_cast_fp16 = slice_by_index(begin = cache_51_begin_0, end = cache_51_end_0, end_mask = cache_51_end_mask_0, squeeze_mask = cache_51_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_51_cast_fp16")]; + tensor input_651_axes_0 = const()[name = tensor("input_651_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151653056)))]; + tensor encoder_layers_12_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151654144)))]; + tensor input_651_cast_fp16 = layer_norm(axes = input_651_axes_0, beta = encoder_layers_12_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_norm_feed_forward1_weight_to_fp16, x = input_649_cast_fp16)[name = tensor("input_651_cast_fp16")]; + tensor encoder_layers_12_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_12_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151655232)))]; + tensor linear_109_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_12_feed_forward1_linear1_weight_to_fp16, x = input_651_cast_fp16)[name = tensor("linear_109_cast_fp16")]; + tensor input_655_cast_fp16 = silu(x = linear_109_cast_fp16)[name = tensor("input_655_cast_fp16")]; + tensor encoder_layers_12_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_12_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(153752448)))]; + tensor linear_110_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_feed_forward1_linear2_weight_to_fp16, x = input_655_cast_fp16)[name = tensor("linear_110_cast_fp16")]; + tensor var_2920_to_fp16 = const()[name = tensor("op_2920_to_fp16"), val = tensor(0x1p-1)]; + tensor var_2921_cast_fp16 = mul(x = linear_110_cast_fp16, y = var_2920_to_fp16)[name = tensor("op_2921_cast_fp16")]; + tensor input_661_cast_fp16 = add(x = input_649_cast_fp16, y = var_2921_cast_fp16)[name = tensor("input_661_cast_fp16")]; + tensor key_25_axes_0 = const()[name = tensor("key_25_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(155849664)))]; + tensor encoder_layers_12_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(155850752)))]; + tensor key_25_cast_fp16 = layer_norm(axes = key_25_axes_0, beta = encoder_layers_12_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_norm_self_att_weight_to_fp16, x = input_661_cast_fp16)[name = tensor("key_25_cast_fp16")]; + tensor input_663_interleave_0 = const()[name = tensor("input_663_interleave_0"), val = tensor(false)]; + tensor input_663_cast_fp16 = concat(axis = var_65, interleave = input_663_interleave_0, values = (cache_49_cast_fp16, key_25_cast_fp16))[name = tensor("input_663_cast_fp16")]; + tensor var_2943_begin_0 = const()[name = tensor("op_2943_begin_0"), val = tensor([0, 4, 0])]; + tensor var_2943_end_0 = const()[name = tensor("op_2943_end_0"), val = tensor([1, 70, 512])]; + tensor var_2943_end_mask_0 = const()[name = tensor("op_2943_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2943_cast_fp16 = slice_by_index(begin = var_2943_begin_0, end = var_2943_end_0, end_mask = var_2943_end_mask_0, x = cache_49_cast_fp16)[name = tensor("op_2943_cast_fp16")]; + tensor var_2946_begin_0 = const()[name = tensor("op_2946_begin_0"), val = tensor([0, 0, 0])]; + tensor var_2946_end_0 = const()[name = tensor("op_2946_end_0"), val = tensor([1, 4, 512])]; + tensor var_2946_end_mask_0 = const()[name = tensor("op_2946_end_mask_0"), val = tensor([true, false, true])]; + tensor var_2946_cast_fp16 = slice_by_index(begin = var_2946_begin_0, end = var_2946_end_0, end_mask = var_2946_end_mask_0, x = key_25_cast_fp16)[name = tensor("op_2946_cast_fp16")]; + tensor var_2949_interleave_0 = const()[name = tensor("op_2949_interleave_0"), val = tensor(false)]; + tensor var_2949_cast_fp16 = concat(axis = var_65, interleave = var_2949_interleave_0, values = (var_2943_cast_fp16, var_2946_cast_fp16))[name = tensor("op_2949_cast_fp16")]; + tensor encoder_layers_12_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(155851840)))]; + tensor linear_111_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_self_attn_linear_q_weight_to_fp16, x = key_25_cast_fp16)[name = tensor("linear_111_cast_fp16")]; + tensor var_2953 = const()[name = tensor("op_2953"), val = tensor([1, -1, 8, 64])]; + tensor q_73_cast_fp16 = reshape(shape = var_2953, x = linear_111_cast_fp16)[name = tensor("q_73_cast_fp16")]; + tensor encoder_layers_12_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156376192)))]; + tensor linear_112_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_self_attn_linear_k_weight_to_fp16, x = input_663_cast_fp16)[name = tensor("linear_112_cast_fp16")]; + tensor var_2957 = const()[name = tensor("op_2957"), val = tensor([1, -1, 8, 64])]; + tensor k_49_cast_fp16 = reshape(shape = var_2957, x = linear_112_cast_fp16)[name = tensor("k_49_cast_fp16")]; + tensor encoder_layers_12_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156900544)))]; + tensor linear_113_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_self_attn_linear_v_weight_to_fp16, x = input_663_cast_fp16)[name = tensor("linear_113_cast_fp16")]; + tensor var_2961 = const()[name = tensor("op_2961"), val = tensor([1, -1, 8, 64])]; + tensor v_25_cast_fp16 = reshape(shape = var_2961, x = linear_113_cast_fp16)[name = tensor("v_25_cast_fp16")]; + tensor value_27_perm_0 = const()[name = tensor("value_27_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_12_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157424896)))]; + tensor var_2973_cast_fp16 = add(x = q_73_cast_fp16, y = encoder_layers_12_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2973_cast_fp16")]; + tensor encoder_layers_12_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157425984)))]; + tensor var_2975_cast_fp16 = add(x = q_73_cast_fp16, y = encoder_layers_12_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2975_cast_fp16")]; + tensor q_with_bias_v_25_perm_0 = const()[name = tensor("q_with_bias_v_25_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_319_transpose_x_0 = const()[name = tensor("x_319_transpose_x_0"), val = tensor(false)]; + tensor x_319_transpose_y_0 = const()[name = tensor("x_319_transpose_y_0"), val = tensor(false)]; + tensor var_2977_to_fp16 = const()[name = tensor("op_2977_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157427072)))]; + tensor q_with_bias_v_25_cast_fp16 = transpose(perm = q_with_bias_v_25_perm_0, x = var_2975_cast_fp16)[name = tensor("transpose_129")]; + tensor x_319_cast_fp16 = matmul(transpose_x = x_319_transpose_x_0, transpose_y = x_319_transpose_y_0, x = q_with_bias_v_25_cast_fp16, y = var_2977_to_fp16)[name = tensor("x_319_cast_fp16")]; + tensor x_321_pad_0 = const()[name = tensor("x_321_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_321_mode_0 = const()[name = tensor("x_321_mode_0"), val = tensor("constant")]; + tensor const_235_to_fp16 = const()[name = tensor("const_235_to_fp16"), val = tensor(0x0p+0)]; + tensor x_321_cast_fp16 = pad(constant_val = const_235_to_fp16, mode = x_321_mode_0, pad = x_321_pad_0, x = x_319_cast_fp16)[name = tensor("x_321_cast_fp16")]; + tensor var_2985 = const()[name = tensor("op_2985"), val = tensor([1, 8, -1, 8])]; + tensor x_323_cast_fp16 = reshape(shape = var_2985, x = x_321_cast_fp16)[name = tensor("x_323_cast_fp16")]; + tensor var_2989_begin_0 = const()[name = tensor("op_2989_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2989_end_0 = const()[name = tensor("op_2989_end_0"), val = tensor([1, 8, 156, 8])]; + tensor var_2989_end_mask_0 = const()[name = tensor("op_2989_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2989_cast_fp16 = slice_by_index(begin = var_2989_begin_0, end = var_2989_end_0, end_mask = var_2989_end_mask_0, x = x_323_cast_fp16)[name = tensor("op_2989_cast_fp16")]; + tensor var_2990 = const()[name = tensor("op_2990"), val = tensor([1, 8, 8, 155])]; + tensor matrix_bd_49_cast_fp16 = reshape(shape = var_2990, x = var_2989_cast_fp16)[name = tensor("matrix_bd_49_cast_fp16")]; + tensor matrix_ac_25_transpose_x_0 = const()[name = tensor("matrix_ac_25_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_25_transpose_y_0 = const()[name = tensor("matrix_ac_25_transpose_y_0"), val = tensor(false)]; + tensor transpose_75_perm_0 = const()[name = tensor("transpose_75_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_76_perm_0 = const()[name = tensor("transpose_76_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_76 = transpose(perm = transpose_76_perm_0, x = k_49_cast_fp16)[name = tensor("transpose_127")]; + tensor transpose_75 = transpose(perm = transpose_75_perm_0, x = var_2973_cast_fp16)[name = tensor("transpose_128")]; + tensor matrix_ac_25_cast_fp16 = matmul(transpose_x = matrix_ac_25_transpose_x_0, transpose_y = matrix_ac_25_transpose_y_0, x = transpose_75, y = transpose_76)[name = tensor("matrix_ac_25_cast_fp16")]; + tensor matrix_bd_51_begin_0 = const()[name = tensor("matrix_bd_51_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_51_end_0 = const()[name = tensor("matrix_bd_51_end_0"), val = tensor([1, 8, 8, 78])]; + tensor matrix_bd_51_end_mask_0 = const()[name = tensor("matrix_bd_51_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_51_cast_fp16 = slice_by_index(begin = matrix_bd_51_begin_0, end = matrix_bd_51_end_0, end_mask = matrix_bd_51_end_mask_0, x = matrix_bd_49_cast_fp16)[name = tensor("matrix_bd_51_cast_fp16")]; + tensor var_2999_cast_fp16 = add(x = matrix_ac_25_cast_fp16, y = matrix_bd_51_cast_fp16)[name = tensor("op_2999_cast_fp16")]; + tensor _inversed_scores_49_y_0_to_fp16 = const()[name = tensor("_inversed_scores_49_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_49_cast_fp16 = mul(x = var_2999_cast_fp16, y = _inversed_scores_49_y_0_to_fp16)[name = tensor("_inversed_scores_49_cast_fp16")]; + tensor scores_51_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_49_cast_fp16, cond = mask_11)[name = tensor("scores_51_cast_fp16")]; + tensor var_3005_cast_fp16 = softmax(axis = var_56, x = scores_51_cast_fp16)[name = tensor("op_3005_cast_fp16")]; + tensor input_665_cast_fp16 = select(a = var_40_to_fp16, b = var_3005_cast_fp16, cond = mask_11)[name = tensor("input_665_cast_fp16")]; + tensor x_325_transpose_x_0 = const()[name = tensor("x_325_transpose_x_0"), val = tensor(false)]; + tensor x_325_transpose_y_0 = const()[name = tensor("x_325_transpose_y_0"), val = tensor(false)]; + tensor value_27_cast_fp16 = transpose(perm = value_27_perm_0, x = v_25_cast_fp16)[name = tensor("transpose_130")]; + tensor x_325_cast_fp16 = matmul(transpose_x = x_325_transpose_x_0, transpose_y = x_325_transpose_y_0, x = input_665_cast_fp16, y = value_27_cast_fp16)[name = tensor("x_325_cast_fp16")]; + tensor var_3009_perm_0 = const()[name = tensor("op_3009_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_3010 = const()[name = tensor("op_3010"), val = tensor([1, -1, 512])]; + tensor var_3009_cast_fp16 = transpose(perm = var_3009_perm_0, x = x_325_cast_fp16)[name = tensor("transpose_126")]; + tensor input_667_cast_fp16 = reshape(shape = var_3010, x = var_3009_cast_fp16)[name = tensor("input_667_cast_fp16")]; + tensor encoder_layers_12_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_12_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157585856)))]; + tensor linear_115_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_self_attn_linear_out_weight_to_fp16, x = input_667_cast_fp16)[name = tensor("linear_115_cast_fp16")]; + tensor input_671_cast_fp16 = add(x = input_661_cast_fp16, y = linear_115_cast_fp16)[name = tensor("input_671_cast_fp16")]; + tensor x_329_axes_0 = const()[name = tensor("x_329_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158110208)))]; + tensor encoder_layers_12_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158111296)))]; + tensor x_329_cast_fp16 = layer_norm(axes = x_329_axes_0, beta = encoder_layers_12_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_norm_conv_weight_to_fp16, x = input_671_cast_fp16)[name = tensor("x_329_cast_fp16")]; + tensor input_673_perm_0 = const()[name = tensor("input_673_perm_0"), val = tensor([0, 2, 1])]; + tensor input_675_pad_type_0 = const()[name = tensor("input_675_pad_type_0"), val = tensor("valid")]; + tensor input_675_strides_0 = const()[name = tensor("input_675_strides_0"), val = tensor([1])]; + tensor input_675_pad_0 = const()[name = tensor("input_675_pad_0"), val = tensor([0, 0])]; + tensor input_675_dilations_0 = const()[name = tensor("input_675_dilations_0"), val = tensor([1])]; + tensor input_675_groups_0 = const()[name = tensor("input_675_groups_0"), val = tensor(1)]; + tensor encoder_layers_12_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_12_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158112384)))]; + tensor input_673_cast_fp16 = transpose(perm = input_673_perm_0, x = x_329_cast_fp16)[name = tensor("transpose_125")]; + tensor input_675_cast_fp16 = conv(dilations = input_675_dilations_0, groups = input_675_groups_0, pad = input_675_pad_0, pad_type = input_675_pad_type_0, strides = input_675_strides_0, weight = encoder_layers_12_conv_pointwise_conv1_weight_to_fp16, x = input_673_cast_fp16)[name = tensor("input_675_cast_fp16")]; + tensor x_331_split_num_splits_0 = const()[name = tensor("x_331_split_num_splits_0"), val = tensor(2)]; + tensor x_331_split_axis_0 = const()[name = tensor("x_331_split_axis_0"), val = tensor(1)]; + tensor x_331_split_cast_fp16_0, tensor x_331_split_cast_fp16_1 = split(axis = x_331_split_axis_0, num_splits = x_331_split_num_splits_0, x = input_675_cast_fp16)[name = tensor("x_331_split_cast_fp16")]; + tensor x_331_split_1_sigmoid_cast_fp16 = sigmoid(x = x_331_split_cast_fp16_1)[name = tensor("x_331_split_1_sigmoid_cast_fp16")]; + tensor x_331_cast_fp16 = mul(x = x_331_split_cast_fp16_0, y = x_331_split_1_sigmoid_cast_fp16)[name = tensor("x_331_cast_fp16")]; + tensor input_677_cast_fp16 = select(a = var_40_to_fp16, b = x_331_cast_fp16, cond = var_551)[name = tensor("input_677_cast_fp16")]; + tensor new_x_51_interleave_0 = const()[name = tensor("new_x_51_interleave_0"), val = tensor(false)]; + tensor new_x_51_cast_fp16 = concat(axis = var_56, interleave = new_x_51_interleave_0, values = (cache_51_cast_fp16, input_677_cast_fp16))[name = tensor("new_x_51_cast_fp16")]; + tensor next_cache_25_begin_0 = const()[name = tensor("next_cache_25_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_25_end_0 = const()[name = tensor("next_cache_25_end_0"), val = tensor([1, 512, 12])]; + tensor next_cache_25_end_mask_0 = const()[name = tensor("next_cache_25_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_25_cast_fp16 = slice_by_index(begin = next_cache_25_begin_0, end = next_cache_25_end_0, end_mask = next_cache_25_end_mask_0, x = new_x_51_cast_fp16)[name = tensor("next_cache_25_cast_fp16")]; + tensor var_3051_begin_0 = const()[name = tensor("op_3051_begin_0"), val = tensor([0, 0, 4])]; + tensor var_3051_end_0 = const()[name = tensor("op_3051_end_0"), val = tensor([1, 512, 12])]; + tensor var_3051_end_mask_0 = const()[name = tensor("op_3051_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3051_cast_fp16 = slice_by_index(begin = var_3051_begin_0, end = var_3051_end_0, end_mask = var_3051_end_mask_0, x = next_cache_25_cast_fp16)[name = tensor("op_3051_cast_fp16")]; + tensor x_333_pad_type_0 = const()[name = tensor("x_333_pad_type_0"), val = tensor("valid")]; + tensor x_333_groups_0 = const()[name = tensor("x_333_groups_0"), val = tensor(512)]; + tensor x_333_strides_0 = const()[name = tensor("x_333_strides_0"), val = tensor([1])]; + tensor x_333_pad_0 = const()[name = tensor("x_333_pad_0"), val = tensor([0, 0])]; + tensor x_333_dilations_0 = const()[name = tensor("x_333_dilations_0"), val = tensor([1])]; + tensor encoder_layers_12_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_12_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159161024)))]; + tensor x_333_cast_fp16 = conv(dilations = x_333_dilations_0, groups = x_333_groups_0, pad = x_333_pad_0, pad_type = x_333_pad_type_0, strides = x_333_strides_0, weight = encoder_layers_12_conv_depthwise_conv_weight_to_fp16, x = new_x_51_cast_fp16)[name = tensor("x_333_cast_fp16")]; + tensor input_679_perm_0 = const()[name = tensor("input_679_perm_0"), val = tensor([0, 2, 1])]; + tensor x_335_axes_0 = const()[name = tensor("x_335_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_12_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159170304)))]; + tensor encoder_layers_12_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_12_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159171392)))]; + tensor input_679_cast_fp16 = transpose(perm = input_679_perm_0, x = x_333_cast_fp16)[name = tensor("transpose_124")]; + tensor x_335_cast_fp16 = layer_norm(axes = x_335_axes_0, beta = encoder_layers_12_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_conv_batch_norm_weight_to_fp16, x = input_679_cast_fp16)[name = tensor("x_335_cast_fp16")]; + tensor input_681_perm_0 = const()[name = tensor("input_681_perm_0"), val = tensor([0, 2, 1])]; + tensor input_681_cast_fp16 = transpose(perm = input_681_perm_0, x = x_335_cast_fp16)[name = tensor("transpose_123")]; + tensor input_683_cast_fp16 = silu(x = input_681_cast_fp16)[name = tensor("input_683_cast_fp16")]; + tensor x_337_pad_type_0 = const()[name = tensor("x_337_pad_type_0"), val = tensor("valid")]; + tensor x_337_strides_0 = const()[name = tensor("x_337_strides_0"), val = tensor([1])]; + tensor x_337_pad_0 = const()[name = tensor("x_337_pad_0"), val = tensor([0, 0])]; + tensor x_337_dilations_0 = const()[name = tensor("x_337_dilations_0"), val = tensor([1])]; + tensor x_337_groups_0 = const()[name = tensor("x_337_groups_0"), val = tensor(1)]; + tensor encoder_layers_12_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_12_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159172480)))]; + tensor x_337_cast_fp16 = conv(dilations = x_337_dilations_0, groups = x_337_groups_0, pad = x_337_pad_0, pad_type = x_337_pad_type_0, strides = x_337_strides_0, weight = encoder_layers_12_conv_pointwise_conv2_weight_to_fp16, x = input_683_cast_fp16)[name = tensor("x_337_cast_fp16")]; + tensor input_685_perm_0 = const()[name = tensor("input_685_perm_0"), val = tensor([0, 2, 1])]; + tensor input_685_cast_fp16 = transpose(perm = input_685_perm_0, x = x_337_cast_fp16)[name = tensor("transpose_122")]; + tensor input_687_cast_fp16 = add(x = input_671_cast_fp16, y = input_685_cast_fp16)[name = tensor("input_687_cast_fp16")]; + tensor input_689_axes_0 = const()[name = tensor("input_689_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159696832)))]; + tensor encoder_layers_12_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159697920)))]; + tensor input_689_cast_fp16 = layer_norm(axes = input_689_axes_0, beta = encoder_layers_12_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_norm_feed_forward2_weight_to_fp16, x = input_687_cast_fp16)[name = tensor("input_689_cast_fp16")]; + tensor encoder_layers_12_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_12_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159699008)))]; + tensor linear_116_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_12_feed_forward2_linear1_weight_to_fp16, x = input_689_cast_fp16)[name = tensor("linear_116_cast_fp16")]; + tensor input_693_cast_fp16 = silu(x = linear_116_cast_fp16)[name = tensor("input_693_cast_fp16")]; + tensor encoder_layers_12_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_12_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161796224)))]; + tensor linear_117_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_12_feed_forward2_linear2_weight_to_fp16, x = input_693_cast_fp16)[name = tensor("linear_117_cast_fp16")]; + tensor var_3092_to_fp16 = const()[name = tensor("op_3092_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3093_cast_fp16 = mul(x = linear_117_cast_fp16, y = var_3092_to_fp16)[name = tensor("op_3093_cast_fp16")]; + tensor input_699_cast_fp16 = add(x = input_687_cast_fp16, y = var_3093_cast_fp16)[name = tensor("input_699_cast_fp16")]; + tensor input_701_axes_0 = const()[name = tensor("input_701_axes_0"), val = tensor([-1])]; + tensor encoder_layers_12_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_12_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163893440)))]; + tensor encoder_layers_12_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_12_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163894528)))]; + tensor input_701_cast_fp16 = layer_norm(axes = input_701_axes_0, beta = encoder_layers_12_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_12_norm_out_weight_to_fp16, x = input_699_cast_fp16)[name = tensor("input_701_cast_fp16")]; + tensor cache_53_begin_0 = const()[name = tensor("cache_53_begin_0"), val = tensor([13, 0, 0, 0])]; + tensor cache_53_end_0 = const()[name = tensor("cache_53_end_0"), val = tensor([14, 1, 70, 512])]; + tensor cache_53_end_mask_0 = const()[name = tensor("cache_53_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_53_squeeze_mask_0 = const()[name = tensor("cache_53_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_53_cast_fp16 = slice_by_index(begin = cache_53_begin_0, end = cache_53_end_0, end_mask = cache_53_end_mask_0, squeeze_mask = cache_53_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_53_cast_fp16")]; + tensor cache_55_begin_0 = const()[name = tensor("cache_55_begin_0"), val = tensor([13, 0, 0, 0])]; + tensor cache_55_end_0 = const()[name = tensor("cache_55_end_0"), val = tensor([14, 1, 512, 8])]; + tensor cache_55_end_mask_0 = const()[name = tensor("cache_55_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_55_squeeze_mask_0 = const()[name = tensor("cache_55_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_55_cast_fp16 = slice_by_index(begin = cache_55_begin_0, end = cache_55_end_0, end_mask = cache_55_end_mask_0, squeeze_mask = cache_55_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_55_cast_fp16")]; + tensor input_703_axes_0 = const()[name = tensor("input_703_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163895616)))]; + tensor encoder_layers_13_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163896704)))]; + tensor input_703_cast_fp16 = layer_norm(axes = input_703_axes_0, beta = encoder_layers_13_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_norm_feed_forward1_weight_to_fp16, x = input_701_cast_fp16)[name = tensor("input_703_cast_fp16")]; + tensor encoder_layers_13_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_13_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163897792)))]; + tensor linear_118_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_13_feed_forward1_linear1_weight_to_fp16, x = input_703_cast_fp16)[name = tensor("linear_118_cast_fp16")]; + tensor input_707_cast_fp16 = silu(x = linear_118_cast_fp16)[name = tensor("input_707_cast_fp16")]; + tensor encoder_layers_13_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_13_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165995008)))]; + tensor linear_119_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_feed_forward1_linear2_weight_to_fp16, x = input_707_cast_fp16)[name = tensor("linear_119_cast_fp16")]; + tensor var_3127_to_fp16 = const()[name = tensor("op_3127_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3128_cast_fp16 = mul(x = linear_119_cast_fp16, y = var_3127_to_fp16)[name = tensor("op_3128_cast_fp16")]; + tensor input_713_cast_fp16 = add(x = input_701_cast_fp16, y = var_3128_cast_fp16)[name = tensor("input_713_cast_fp16")]; + tensor key_27_axes_0 = const()[name = tensor("key_27_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168092224)))]; + tensor encoder_layers_13_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168093312)))]; + tensor key_27_cast_fp16 = layer_norm(axes = key_27_axes_0, beta = encoder_layers_13_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_norm_self_att_weight_to_fp16, x = input_713_cast_fp16)[name = tensor("key_27_cast_fp16")]; + tensor input_715_interleave_0 = const()[name = tensor("input_715_interleave_0"), val = tensor(false)]; + tensor input_715_cast_fp16 = concat(axis = var_65, interleave = input_715_interleave_0, values = (cache_53_cast_fp16, key_27_cast_fp16))[name = tensor("input_715_cast_fp16")]; + tensor var_3150_begin_0 = const()[name = tensor("op_3150_begin_0"), val = tensor([0, 4, 0])]; + tensor var_3150_end_0 = const()[name = tensor("op_3150_end_0"), val = tensor([1, 70, 512])]; + tensor var_3150_end_mask_0 = const()[name = tensor("op_3150_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3150_cast_fp16 = slice_by_index(begin = var_3150_begin_0, end = var_3150_end_0, end_mask = var_3150_end_mask_0, x = cache_53_cast_fp16)[name = tensor("op_3150_cast_fp16")]; + tensor var_3153_begin_0 = const()[name = tensor("op_3153_begin_0"), val = tensor([0, 0, 0])]; + tensor var_3153_end_0 = const()[name = tensor("op_3153_end_0"), val = tensor([1, 4, 512])]; + tensor var_3153_end_mask_0 = const()[name = tensor("op_3153_end_mask_0"), val = tensor([true, false, true])]; + tensor var_3153_cast_fp16 = slice_by_index(begin = var_3153_begin_0, end = var_3153_end_0, end_mask = var_3153_end_mask_0, x = key_27_cast_fp16)[name = tensor("op_3153_cast_fp16")]; + tensor var_3156_interleave_0 = const()[name = tensor("op_3156_interleave_0"), val = tensor(false)]; + tensor var_3156_cast_fp16 = concat(axis = var_65, interleave = var_3156_interleave_0, values = (var_3150_cast_fp16, var_3153_cast_fp16))[name = tensor("op_3156_cast_fp16")]; + tensor encoder_layers_13_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168094400)))]; + tensor linear_120_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_self_attn_linear_q_weight_to_fp16, x = key_27_cast_fp16)[name = tensor("linear_120_cast_fp16")]; + tensor var_3160 = const()[name = tensor("op_3160"), val = tensor([1, -1, 8, 64])]; + tensor q_79_cast_fp16 = reshape(shape = var_3160, x = linear_120_cast_fp16)[name = tensor("q_79_cast_fp16")]; + tensor encoder_layers_13_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168618752)))]; + tensor linear_121_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_self_attn_linear_k_weight_to_fp16, x = input_715_cast_fp16)[name = tensor("linear_121_cast_fp16")]; + tensor var_3164 = const()[name = tensor("op_3164"), val = tensor([1, -1, 8, 64])]; + tensor k_53_cast_fp16 = reshape(shape = var_3164, x = linear_121_cast_fp16)[name = tensor("k_53_cast_fp16")]; + tensor encoder_layers_13_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169143104)))]; + tensor linear_122_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_self_attn_linear_v_weight_to_fp16, x = input_715_cast_fp16)[name = tensor("linear_122_cast_fp16")]; + tensor var_3168 = const()[name = tensor("op_3168"), val = tensor([1, -1, 8, 64])]; + tensor v_27_cast_fp16 = reshape(shape = var_3168, x = linear_122_cast_fp16)[name = tensor("v_27_cast_fp16")]; + tensor value_29_perm_0 = const()[name = tensor("value_29_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_13_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169667456)))]; + tensor var_3180_cast_fp16 = add(x = q_79_cast_fp16, y = encoder_layers_13_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3180_cast_fp16")]; + tensor encoder_layers_13_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169668544)))]; + tensor var_3182_cast_fp16 = add(x = q_79_cast_fp16, y = encoder_layers_13_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3182_cast_fp16")]; + tensor q_with_bias_v_27_perm_0 = const()[name = tensor("q_with_bias_v_27_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_345_transpose_x_0 = const()[name = tensor("x_345_transpose_x_0"), val = tensor(false)]; + tensor x_345_transpose_y_0 = const()[name = tensor("x_345_transpose_y_0"), val = tensor(false)]; + tensor var_3184_to_fp16 = const()[name = tensor("op_3184_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169669632)))]; + tensor q_with_bias_v_27_cast_fp16 = transpose(perm = q_with_bias_v_27_perm_0, x = var_3182_cast_fp16)[name = tensor("transpose_120")]; + tensor x_345_cast_fp16 = matmul(transpose_x = x_345_transpose_x_0, transpose_y = x_345_transpose_y_0, x = q_with_bias_v_27_cast_fp16, y = var_3184_to_fp16)[name = tensor("x_345_cast_fp16")]; + tensor x_347_pad_0 = const()[name = tensor("x_347_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_347_mode_0 = const()[name = tensor("x_347_mode_0"), val = tensor("constant")]; + tensor const_248_to_fp16 = const()[name = tensor("const_248_to_fp16"), val = tensor(0x0p+0)]; + tensor x_347_cast_fp16 = pad(constant_val = const_248_to_fp16, mode = x_347_mode_0, pad = x_347_pad_0, x = x_345_cast_fp16)[name = tensor("x_347_cast_fp16")]; + tensor var_3192 = const()[name = tensor("op_3192"), val = tensor([1, 8, -1, 8])]; + tensor x_349_cast_fp16 = reshape(shape = var_3192, x = x_347_cast_fp16)[name = tensor("x_349_cast_fp16")]; + tensor var_3196_begin_0 = const()[name = tensor("op_3196_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_3196_end_0 = const()[name = tensor("op_3196_end_0"), val = tensor([1, 8, 156, 8])]; + tensor var_3196_end_mask_0 = const()[name = tensor("op_3196_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_3196_cast_fp16 = slice_by_index(begin = var_3196_begin_0, end = var_3196_end_0, end_mask = var_3196_end_mask_0, x = x_349_cast_fp16)[name = tensor("op_3196_cast_fp16")]; + tensor var_3197 = const()[name = tensor("op_3197"), val = tensor([1, 8, 8, 155])]; + tensor matrix_bd_53_cast_fp16 = reshape(shape = var_3197, x = var_3196_cast_fp16)[name = tensor("matrix_bd_53_cast_fp16")]; + tensor matrix_ac_27_transpose_x_0 = const()[name = tensor("matrix_ac_27_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_27_transpose_y_0 = const()[name = tensor("matrix_ac_27_transpose_y_0"), val = tensor(false)]; + tensor transpose_77_perm_0 = const()[name = tensor("transpose_77_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_78_perm_0 = const()[name = tensor("transpose_78_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_78 = transpose(perm = transpose_78_perm_0, x = k_53_cast_fp16)[name = tensor("transpose_118")]; + tensor transpose_77 = transpose(perm = transpose_77_perm_0, x = var_3180_cast_fp16)[name = tensor("transpose_119")]; + tensor matrix_ac_27_cast_fp16 = matmul(transpose_x = matrix_ac_27_transpose_x_0, transpose_y = matrix_ac_27_transpose_y_0, x = transpose_77, y = transpose_78)[name = tensor("matrix_ac_27_cast_fp16")]; + tensor matrix_bd_55_begin_0 = const()[name = tensor("matrix_bd_55_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_55_end_0 = const()[name = tensor("matrix_bd_55_end_0"), val = tensor([1, 8, 8, 78])]; + tensor matrix_bd_55_end_mask_0 = const()[name = tensor("matrix_bd_55_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_55_cast_fp16 = slice_by_index(begin = matrix_bd_55_begin_0, end = matrix_bd_55_end_0, end_mask = matrix_bd_55_end_mask_0, x = matrix_bd_53_cast_fp16)[name = tensor("matrix_bd_55_cast_fp16")]; + tensor var_3206_cast_fp16 = add(x = matrix_ac_27_cast_fp16, y = matrix_bd_55_cast_fp16)[name = tensor("op_3206_cast_fp16")]; + tensor _inversed_scores_53_y_0_to_fp16 = const()[name = tensor("_inversed_scores_53_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_53_cast_fp16 = mul(x = var_3206_cast_fp16, y = _inversed_scores_53_y_0_to_fp16)[name = tensor("_inversed_scores_53_cast_fp16")]; + tensor scores_55_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_53_cast_fp16, cond = mask_11)[name = tensor("scores_55_cast_fp16")]; + tensor var_3212_cast_fp16 = softmax(axis = var_56, x = scores_55_cast_fp16)[name = tensor("op_3212_cast_fp16")]; + tensor input_717_cast_fp16 = select(a = var_40_to_fp16, b = var_3212_cast_fp16, cond = mask_11)[name = tensor("input_717_cast_fp16")]; + tensor x_351_transpose_x_0 = const()[name = tensor("x_351_transpose_x_0"), val = tensor(false)]; + tensor x_351_transpose_y_0 = const()[name = tensor("x_351_transpose_y_0"), val = tensor(false)]; + tensor value_29_cast_fp16 = transpose(perm = value_29_perm_0, x = v_27_cast_fp16)[name = tensor("transpose_121")]; + tensor x_351_cast_fp16 = matmul(transpose_x = x_351_transpose_x_0, transpose_y = x_351_transpose_y_0, x = input_717_cast_fp16, y = value_29_cast_fp16)[name = tensor("x_351_cast_fp16")]; + tensor var_3216_perm_0 = const()[name = tensor("op_3216_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_3217 = const()[name = tensor("op_3217"), val = tensor([1, -1, 512])]; + tensor var_3216_cast_fp16 = transpose(perm = var_3216_perm_0, x = x_351_cast_fp16)[name = tensor("transpose_117")]; + tensor input_719_cast_fp16 = reshape(shape = var_3217, x = var_3216_cast_fp16)[name = tensor("input_719_cast_fp16")]; + tensor encoder_layers_13_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_13_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169828416)))]; + tensor linear_124_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_self_attn_linear_out_weight_to_fp16, x = input_719_cast_fp16)[name = tensor("linear_124_cast_fp16")]; + tensor input_723_cast_fp16 = add(x = input_713_cast_fp16, y = linear_124_cast_fp16)[name = tensor("input_723_cast_fp16")]; + tensor x_355_axes_0 = const()[name = tensor("x_355_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170352768)))]; + tensor encoder_layers_13_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170353856)))]; + tensor x_355_cast_fp16 = layer_norm(axes = x_355_axes_0, beta = encoder_layers_13_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_norm_conv_weight_to_fp16, x = input_723_cast_fp16)[name = tensor("x_355_cast_fp16")]; + tensor input_725_perm_0 = const()[name = tensor("input_725_perm_0"), val = tensor([0, 2, 1])]; + tensor input_727_pad_type_0 = const()[name = tensor("input_727_pad_type_0"), val = tensor("valid")]; + tensor input_727_strides_0 = const()[name = tensor("input_727_strides_0"), val = tensor([1])]; + tensor input_727_pad_0 = const()[name = tensor("input_727_pad_0"), val = tensor([0, 0])]; + tensor input_727_dilations_0 = const()[name = tensor("input_727_dilations_0"), val = tensor([1])]; + tensor input_727_groups_0 = const()[name = tensor("input_727_groups_0"), val = tensor(1)]; + tensor encoder_layers_13_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_13_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170354944)))]; + tensor input_725_cast_fp16 = transpose(perm = input_725_perm_0, x = x_355_cast_fp16)[name = tensor("transpose_116")]; + tensor input_727_cast_fp16 = conv(dilations = input_727_dilations_0, groups = input_727_groups_0, pad = input_727_pad_0, pad_type = input_727_pad_type_0, strides = input_727_strides_0, weight = encoder_layers_13_conv_pointwise_conv1_weight_to_fp16, x = input_725_cast_fp16)[name = tensor("input_727_cast_fp16")]; + tensor x_357_split_num_splits_0 = const()[name = tensor("x_357_split_num_splits_0"), val = tensor(2)]; + tensor x_357_split_axis_0 = const()[name = tensor("x_357_split_axis_0"), val = tensor(1)]; + tensor x_357_split_cast_fp16_0, tensor x_357_split_cast_fp16_1 = split(axis = x_357_split_axis_0, num_splits = x_357_split_num_splits_0, x = input_727_cast_fp16)[name = tensor("x_357_split_cast_fp16")]; + tensor x_357_split_1_sigmoid_cast_fp16 = sigmoid(x = x_357_split_cast_fp16_1)[name = tensor("x_357_split_1_sigmoid_cast_fp16")]; + tensor x_357_cast_fp16 = mul(x = x_357_split_cast_fp16_0, y = x_357_split_1_sigmoid_cast_fp16)[name = tensor("x_357_cast_fp16")]; + tensor input_729_cast_fp16 = select(a = var_40_to_fp16, b = x_357_cast_fp16, cond = var_551)[name = tensor("input_729_cast_fp16")]; + tensor new_x_55_interleave_0 = const()[name = tensor("new_x_55_interleave_0"), val = tensor(false)]; + tensor new_x_55_cast_fp16 = concat(axis = var_56, interleave = new_x_55_interleave_0, values = (cache_55_cast_fp16, input_729_cast_fp16))[name = tensor("new_x_55_cast_fp16")]; + tensor next_cache_27_begin_0 = const()[name = tensor("next_cache_27_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_27_end_0 = const()[name = tensor("next_cache_27_end_0"), val = tensor([1, 512, 12])]; + tensor next_cache_27_end_mask_0 = const()[name = tensor("next_cache_27_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_27_cast_fp16 = slice_by_index(begin = next_cache_27_begin_0, end = next_cache_27_end_0, end_mask = next_cache_27_end_mask_0, x = new_x_55_cast_fp16)[name = tensor("next_cache_27_cast_fp16")]; + tensor var_3258_begin_0 = const()[name = tensor("op_3258_begin_0"), val = tensor([0, 0, 4])]; + tensor var_3258_end_0 = const()[name = tensor("op_3258_end_0"), val = tensor([1, 512, 12])]; + tensor var_3258_end_mask_0 = const()[name = tensor("op_3258_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3258_cast_fp16 = slice_by_index(begin = var_3258_begin_0, end = var_3258_end_0, end_mask = var_3258_end_mask_0, x = next_cache_27_cast_fp16)[name = tensor("op_3258_cast_fp16")]; + tensor x_359_pad_type_0 = const()[name = tensor("x_359_pad_type_0"), val = tensor("valid")]; + tensor x_359_groups_0 = const()[name = tensor("x_359_groups_0"), val = tensor(512)]; + tensor x_359_strides_0 = const()[name = tensor("x_359_strides_0"), val = tensor([1])]; + tensor x_359_pad_0 = const()[name = tensor("x_359_pad_0"), val = tensor([0, 0])]; + tensor x_359_dilations_0 = const()[name = tensor("x_359_dilations_0"), val = tensor([1])]; + tensor encoder_layers_13_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_13_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171403584)))]; + tensor x_359_cast_fp16 = conv(dilations = x_359_dilations_0, groups = x_359_groups_0, pad = x_359_pad_0, pad_type = x_359_pad_type_0, strides = x_359_strides_0, weight = encoder_layers_13_conv_depthwise_conv_weight_to_fp16, x = new_x_55_cast_fp16)[name = tensor("x_359_cast_fp16")]; + tensor input_731_perm_0 = const()[name = tensor("input_731_perm_0"), val = tensor([0, 2, 1])]; + tensor x_361_axes_0 = const()[name = tensor("x_361_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_13_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171412864)))]; + tensor encoder_layers_13_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_13_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171413952)))]; + tensor input_731_cast_fp16 = transpose(perm = input_731_perm_0, x = x_359_cast_fp16)[name = tensor("transpose_115")]; + tensor x_361_cast_fp16 = layer_norm(axes = x_361_axes_0, beta = encoder_layers_13_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_conv_batch_norm_weight_to_fp16, x = input_731_cast_fp16)[name = tensor("x_361_cast_fp16")]; + tensor input_733_perm_0 = const()[name = tensor("input_733_perm_0"), val = tensor([0, 2, 1])]; + tensor input_733_cast_fp16 = transpose(perm = input_733_perm_0, x = x_361_cast_fp16)[name = tensor("transpose_114")]; + tensor input_735_cast_fp16 = silu(x = input_733_cast_fp16)[name = tensor("input_735_cast_fp16")]; + tensor x_363_pad_type_0 = const()[name = tensor("x_363_pad_type_0"), val = tensor("valid")]; + tensor x_363_strides_0 = const()[name = tensor("x_363_strides_0"), val = tensor([1])]; + tensor x_363_pad_0 = const()[name = tensor("x_363_pad_0"), val = tensor([0, 0])]; + tensor x_363_dilations_0 = const()[name = tensor("x_363_dilations_0"), val = tensor([1])]; + tensor x_363_groups_0 = const()[name = tensor("x_363_groups_0"), val = tensor(1)]; + tensor encoder_layers_13_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_13_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171415040)))]; + tensor x_363_cast_fp16 = conv(dilations = x_363_dilations_0, groups = x_363_groups_0, pad = x_363_pad_0, pad_type = x_363_pad_type_0, strides = x_363_strides_0, weight = encoder_layers_13_conv_pointwise_conv2_weight_to_fp16, x = input_735_cast_fp16)[name = tensor("x_363_cast_fp16")]; + tensor input_737_perm_0 = const()[name = tensor("input_737_perm_0"), val = tensor([0, 2, 1])]; + tensor input_737_cast_fp16 = transpose(perm = input_737_perm_0, x = x_363_cast_fp16)[name = tensor("transpose_113")]; + tensor input_739_cast_fp16 = add(x = input_723_cast_fp16, y = input_737_cast_fp16)[name = tensor("input_739_cast_fp16")]; + tensor input_741_axes_0 = const()[name = tensor("input_741_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171939392)))]; + tensor encoder_layers_13_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171940480)))]; + tensor input_741_cast_fp16 = layer_norm(axes = input_741_axes_0, beta = encoder_layers_13_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_norm_feed_forward2_weight_to_fp16, x = input_739_cast_fp16)[name = tensor("input_741_cast_fp16")]; + tensor encoder_layers_13_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_13_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171941568)))]; + tensor linear_125_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_13_feed_forward2_linear1_weight_to_fp16, x = input_741_cast_fp16)[name = tensor("linear_125_cast_fp16")]; + tensor input_745_cast_fp16 = silu(x = linear_125_cast_fp16)[name = tensor("input_745_cast_fp16")]; + tensor encoder_layers_13_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_13_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174038784)))]; + tensor linear_126_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_13_feed_forward2_linear2_weight_to_fp16, x = input_745_cast_fp16)[name = tensor("linear_126_cast_fp16")]; + tensor var_3299_to_fp16 = const()[name = tensor("op_3299_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3300_cast_fp16 = mul(x = linear_126_cast_fp16, y = var_3299_to_fp16)[name = tensor("op_3300_cast_fp16")]; + tensor input_751_cast_fp16 = add(x = input_739_cast_fp16, y = var_3300_cast_fp16)[name = tensor("input_751_cast_fp16")]; + tensor input_753_axes_0 = const()[name = tensor("input_753_axes_0"), val = tensor([-1])]; + tensor encoder_layers_13_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_13_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176136000)))]; + tensor encoder_layers_13_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_13_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176137088)))]; + tensor input_753_cast_fp16 = layer_norm(axes = input_753_axes_0, beta = encoder_layers_13_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_13_norm_out_weight_to_fp16, x = input_751_cast_fp16)[name = tensor("input_753_cast_fp16")]; + tensor cache_57_begin_0 = const()[name = tensor("cache_57_begin_0"), val = tensor([14, 0, 0, 0])]; + tensor cache_57_end_0 = const()[name = tensor("cache_57_end_0"), val = tensor([15, 1, 70, 512])]; + tensor cache_57_end_mask_0 = const()[name = tensor("cache_57_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_57_squeeze_mask_0 = const()[name = tensor("cache_57_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_57_cast_fp16 = slice_by_index(begin = cache_57_begin_0, end = cache_57_end_0, end_mask = cache_57_end_mask_0, squeeze_mask = cache_57_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_57_cast_fp16")]; + tensor cache_59_begin_0 = const()[name = tensor("cache_59_begin_0"), val = tensor([14, 0, 0, 0])]; + tensor cache_59_end_0 = const()[name = tensor("cache_59_end_0"), val = tensor([15, 1, 512, 8])]; + tensor cache_59_end_mask_0 = const()[name = tensor("cache_59_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_59_squeeze_mask_0 = const()[name = tensor("cache_59_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_59_cast_fp16 = slice_by_index(begin = cache_59_begin_0, end = cache_59_end_0, end_mask = cache_59_end_mask_0, squeeze_mask = cache_59_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_59_cast_fp16")]; + tensor input_755_axes_0 = const()[name = tensor("input_755_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176138176)))]; + tensor encoder_layers_14_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176139264)))]; + tensor input_755_cast_fp16 = layer_norm(axes = input_755_axes_0, beta = encoder_layers_14_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_norm_feed_forward1_weight_to_fp16, x = input_753_cast_fp16)[name = tensor("input_755_cast_fp16")]; + tensor encoder_layers_14_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_14_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176140352)))]; + tensor linear_127_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_14_feed_forward1_linear1_weight_to_fp16, x = input_755_cast_fp16)[name = tensor("linear_127_cast_fp16")]; + tensor input_759_cast_fp16 = silu(x = linear_127_cast_fp16)[name = tensor("input_759_cast_fp16")]; + tensor encoder_layers_14_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_14_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178237568)))]; + tensor linear_128_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_feed_forward1_linear2_weight_to_fp16, x = input_759_cast_fp16)[name = tensor("linear_128_cast_fp16")]; + tensor var_3334_to_fp16 = const()[name = tensor("op_3334_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3335_cast_fp16 = mul(x = linear_128_cast_fp16, y = var_3334_to_fp16)[name = tensor("op_3335_cast_fp16")]; + tensor input_765_cast_fp16 = add(x = input_753_cast_fp16, y = var_3335_cast_fp16)[name = tensor("input_765_cast_fp16")]; + tensor key_29_axes_0 = const()[name = tensor("key_29_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180334784)))]; + tensor encoder_layers_14_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180335872)))]; + tensor key_29_cast_fp16 = layer_norm(axes = key_29_axes_0, beta = encoder_layers_14_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_norm_self_att_weight_to_fp16, x = input_765_cast_fp16)[name = tensor("key_29_cast_fp16")]; + tensor input_767_interleave_0 = const()[name = tensor("input_767_interleave_0"), val = tensor(false)]; + tensor input_767_cast_fp16 = concat(axis = var_65, interleave = input_767_interleave_0, values = (cache_57_cast_fp16, key_29_cast_fp16))[name = tensor("input_767_cast_fp16")]; + tensor var_3357_begin_0 = const()[name = tensor("op_3357_begin_0"), val = tensor([0, 4, 0])]; + tensor var_3357_end_0 = const()[name = tensor("op_3357_end_0"), val = tensor([1, 70, 512])]; + tensor var_3357_end_mask_0 = const()[name = tensor("op_3357_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3357_cast_fp16 = slice_by_index(begin = var_3357_begin_0, end = var_3357_end_0, end_mask = var_3357_end_mask_0, x = cache_57_cast_fp16)[name = tensor("op_3357_cast_fp16")]; + tensor var_3360_begin_0 = const()[name = tensor("op_3360_begin_0"), val = tensor([0, 0, 0])]; + tensor var_3360_end_0 = const()[name = tensor("op_3360_end_0"), val = tensor([1, 4, 512])]; + tensor var_3360_end_mask_0 = const()[name = tensor("op_3360_end_mask_0"), val = tensor([true, false, true])]; + tensor var_3360_cast_fp16 = slice_by_index(begin = var_3360_begin_0, end = var_3360_end_0, end_mask = var_3360_end_mask_0, x = key_29_cast_fp16)[name = tensor("op_3360_cast_fp16")]; + tensor var_3363_interleave_0 = const()[name = tensor("op_3363_interleave_0"), val = tensor(false)]; + tensor var_3363_cast_fp16 = concat(axis = var_65, interleave = var_3363_interleave_0, values = (var_3357_cast_fp16, var_3360_cast_fp16))[name = tensor("op_3363_cast_fp16")]; + tensor encoder_layers_14_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180336960)))]; + tensor linear_129_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_self_attn_linear_q_weight_to_fp16, x = key_29_cast_fp16)[name = tensor("linear_129_cast_fp16")]; + tensor var_3367 = const()[name = tensor("op_3367"), val = tensor([1, -1, 8, 64])]; + tensor q_85_cast_fp16 = reshape(shape = var_3367, x = linear_129_cast_fp16)[name = tensor("q_85_cast_fp16")]; + tensor encoder_layers_14_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180861312)))]; + tensor linear_130_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_self_attn_linear_k_weight_to_fp16, x = input_767_cast_fp16)[name = tensor("linear_130_cast_fp16")]; + tensor var_3371 = const()[name = tensor("op_3371"), val = tensor([1, -1, 8, 64])]; + tensor k_57_cast_fp16 = reshape(shape = var_3371, x = linear_130_cast_fp16)[name = tensor("k_57_cast_fp16")]; + tensor encoder_layers_14_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181385664)))]; + tensor linear_131_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_self_attn_linear_v_weight_to_fp16, x = input_767_cast_fp16)[name = tensor("linear_131_cast_fp16")]; + tensor var_3375 = const()[name = tensor("op_3375"), val = tensor([1, -1, 8, 64])]; + tensor v_29_cast_fp16 = reshape(shape = var_3375, x = linear_131_cast_fp16)[name = tensor("v_29_cast_fp16")]; + tensor value_31_perm_0 = const()[name = tensor("value_31_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_14_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181910016)))]; + tensor var_3387_cast_fp16 = add(x = q_85_cast_fp16, y = encoder_layers_14_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3387_cast_fp16")]; + tensor encoder_layers_14_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181911104)))]; + tensor var_3389_cast_fp16 = add(x = q_85_cast_fp16, y = encoder_layers_14_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3389_cast_fp16")]; + tensor q_with_bias_v_29_perm_0 = const()[name = tensor("q_with_bias_v_29_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_371_transpose_x_0 = const()[name = tensor("x_371_transpose_x_0"), val = tensor(false)]; + tensor x_371_transpose_y_0 = const()[name = tensor("x_371_transpose_y_0"), val = tensor(false)]; + tensor var_3391_to_fp16 = const()[name = tensor("op_3391_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181912192)))]; + tensor q_with_bias_v_29_cast_fp16 = transpose(perm = q_with_bias_v_29_perm_0, x = var_3389_cast_fp16)[name = tensor("transpose_111")]; + tensor x_371_cast_fp16 = matmul(transpose_x = x_371_transpose_x_0, transpose_y = x_371_transpose_y_0, x = q_with_bias_v_29_cast_fp16, y = var_3391_to_fp16)[name = tensor("x_371_cast_fp16")]; + tensor x_373_pad_0 = const()[name = tensor("x_373_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_373_mode_0 = const()[name = tensor("x_373_mode_0"), val = tensor("constant")]; + tensor const_261_to_fp16 = const()[name = tensor("const_261_to_fp16"), val = tensor(0x0p+0)]; + tensor x_373_cast_fp16 = pad(constant_val = const_261_to_fp16, mode = x_373_mode_0, pad = x_373_pad_0, x = x_371_cast_fp16)[name = tensor("x_373_cast_fp16")]; + tensor var_3399 = const()[name = tensor("op_3399"), val = tensor([1, 8, -1, 8])]; + tensor x_375_cast_fp16 = reshape(shape = var_3399, x = x_373_cast_fp16)[name = tensor("x_375_cast_fp16")]; + tensor var_3403_begin_0 = const()[name = tensor("op_3403_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_3403_end_0 = const()[name = tensor("op_3403_end_0"), val = tensor([1, 8, 156, 8])]; + tensor var_3403_end_mask_0 = const()[name = tensor("op_3403_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_3403_cast_fp16 = slice_by_index(begin = var_3403_begin_0, end = var_3403_end_0, end_mask = var_3403_end_mask_0, x = x_375_cast_fp16)[name = tensor("op_3403_cast_fp16")]; + tensor var_3404 = const()[name = tensor("op_3404"), val = tensor([1, 8, 8, 155])]; + tensor matrix_bd_57_cast_fp16 = reshape(shape = var_3404, x = var_3403_cast_fp16)[name = tensor("matrix_bd_57_cast_fp16")]; + tensor matrix_ac_29_transpose_x_0 = const()[name = tensor("matrix_ac_29_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_29_transpose_y_0 = const()[name = tensor("matrix_ac_29_transpose_y_0"), val = tensor(false)]; + tensor transpose_79_perm_0 = const()[name = tensor("transpose_79_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_80_perm_0 = const()[name = tensor("transpose_80_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_80 = transpose(perm = transpose_80_perm_0, x = k_57_cast_fp16)[name = tensor("transpose_109")]; + tensor transpose_79 = transpose(perm = transpose_79_perm_0, x = var_3387_cast_fp16)[name = tensor("transpose_110")]; + tensor matrix_ac_29_cast_fp16 = matmul(transpose_x = matrix_ac_29_transpose_x_0, transpose_y = matrix_ac_29_transpose_y_0, x = transpose_79, y = transpose_80)[name = tensor("matrix_ac_29_cast_fp16")]; + tensor matrix_bd_59_begin_0 = const()[name = tensor("matrix_bd_59_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_59_end_0 = const()[name = tensor("matrix_bd_59_end_0"), val = tensor([1, 8, 8, 78])]; + tensor matrix_bd_59_end_mask_0 = const()[name = tensor("matrix_bd_59_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_59_cast_fp16 = slice_by_index(begin = matrix_bd_59_begin_0, end = matrix_bd_59_end_0, end_mask = matrix_bd_59_end_mask_0, x = matrix_bd_57_cast_fp16)[name = tensor("matrix_bd_59_cast_fp16")]; + tensor var_3413_cast_fp16 = add(x = matrix_ac_29_cast_fp16, y = matrix_bd_59_cast_fp16)[name = tensor("op_3413_cast_fp16")]; + tensor _inversed_scores_57_y_0_to_fp16 = const()[name = tensor("_inversed_scores_57_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_57_cast_fp16 = mul(x = var_3413_cast_fp16, y = _inversed_scores_57_y_0_to_fp16)[name = tensor("_inversed_scores_57_cast_fp16")]; + tensor scores_59_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_57_cast_fp16, cond = mask_11)[name = tensor("scores_59_cast_fp16")]; + tensor var_3419_cast_fp16 = softmax(axis = var_56, x = scores_59_cast_fp16)[name = tensor("op_3419_cast_fp16")]; + tensor input_769_cast_fp16 = select(a = var_40_to_fp16, b = var_3419_cast_fp16, cond = mask_11)[name = tensor("input_769_cast_fp16")]; + tensor x_377_transpose_x_0 = const()[name = tensor("x_377_transpose_x_0"), val = tensor(false)]; + tensor x_377_transpose_y_0 = const()[name = tensor("x_377_transpose_y_0"), val = tensor(false)]; + tensor value_31_cast_fp16 = transpose(perm = value_31_perm_0, x = v_29_cast_fp16)[name = tensor("transpose_112")]; + tensor x_377_cast_fp16 = matmul(transpose_x = x_377_transpose_x_0, transpose_y = x_377_transpose_y_0, x = input_769_cast_fp16, y = value_31_cast_fp16)[name = tensor("x_377_cast_fp16")]; + tensor var_3423_perm_0 = const()[name = tensor("op_3423_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_3424 = const()[name = tensor("op_3424"), val = tensor([1, -1, 512])]; + tensor var_3423_cast_fp16 = transpose(perm = var_3423_perm_0, x = x_377_cast_fp16)[name = tensor("transpose_108")]; + tensor input_771_cast_fp16 = reshape(shape = var_3424, x = var_3423_cast_fp16)[name = tensor("input_771_cast_fp16")]; + tensor encoder_layers_14_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_14_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182070976)))]; + tensor linear_133_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_self_attn_linear_out_weight_to_fp16, x = input_771_cast_fp16)[name = tensor("linear_133_cast_fp16")]; + tensor input_775_cast_fp16 = add(x = input_765_cast_fp16, y = linear_133_cast_fp16)[name = tensor("input_775_cast_fp16")]; + tensor x_381_axes_0 = const()[name = tensor("x_381_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182595328)))]; + tensor encoder_layers_14_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182596416)))]; + tensor x_381_cast_fp16 = layer_norm(axes = x_381_axes_0, beta = encoder_layers_14_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_norm_conv_weight_to_fp16, x = input_775_cast_fp16)[name = tensor("x_381_cast_fp16")]; + tensor input_777_perm_0 = const()[name = tensor("input_777_perm_0"), val = tensor([0, 2, 1])]; + tensor input_779_pad_type_0 = const()[name = tensor("input_779_pad_type_0"), val = tensor("valid")]; + tensor input_779_strides_0 = const()[name = tensor("input_779_strides_0"), val = tensor([1])]; + tensor input_779_pad_0 = const()[name = tensor("input_779_pad_0"), val = tensor([0, 0])]; + tensor input_779_dilations_0 = const()[name = tensor("input_779_dilations_0"), val = tensor([1])]; + tensor input_779_groups_0 = const()[name = tensor("input_779_groups_0"), val = tensor(1)]; + tensor encoder_layers_14_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_14_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182597504)))]; + tensor input_777_cast_fp16 = transpose(perm = input_777_perm_0, x = x_381_cast_fp16)[name = tensor("transpose_107")]; + tensor input_779_cast_fp16 = conv(dilations = input_779_dilations_0, groups = input_779_groups_0, pad = input_779_pad_0, pad_type = input_779_pad_type_0, strides = input_779_strides_0, weight = encoder_layers_14_conv_pointwise_conv1_weight_to_fp16, x = input_777_cast_fp16)[name = tensor("input_779_cast_fp16")]; + tensor x_383_split_num_splits_0 = const()[name = tensor("x_383_split_num_splits_0"), val = tensor(2)]; + tensor x_383_split_axis_0 = const()[name = tensor("x_383_split_axis_0"), val = tensor(1)]; + tensor x_383_split_cast_fp16_0, tensor x_383_split_cast_fp16_1 = split(axis = x_383_split_axis_0, num_splits = x_383_split_num_splits_0, x = input_779_cast_fp16)[name = tensor("x_383_split_cast_fp16")]; + tensor x_383_split_1_sigmoid_cast_fp16 = sigmoid(x = x_383_split_cast_fp16_1)[name = tensor("x_383_split_1_sigmoid_cast_fp16")]; + tensor x_383_cast_fp16 = mul(x = x_383_split_cast_fp16_0, y = x_383_split_1_sigmoid_cast_fp16)[name = tensor("x_383_cast_fp16")]; + tensor input_781_cast_fp16 = select(a = var_40_to_fp16, b = x_383_cast_fp16, cond = var_551)[name = tensor("input_781_cast_fp16")]; + tensor new_x_59_interleave_0 = const()[name = tensor("new_x_59_interleave_0"), val = tensor(false)]; + tensor new_x_59_cast_fp16 = concat(axis = var_56, interleave = new_x_59_interleave_0, values = (cache_59_cast_fp16, input_781_cast_fp16))[name = tensor("new_x_59_cast_fp16")]; + tensor next_cache_29_begin_0 = const()[name = tensor("next_cache_29_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_29_end_0 = const()[name = tensor("next_cache_29_end_0"), val = tensor([1, 512, 12])]; + tensor next_cache_29_end_mask_0 = const()[name = tensor("next_cache_29_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_29_cast_fp16 = slice_by_index(begin = next_cache_29_begin_0, end = next_cache_29_end_0, end_mask = next_cache_29_end_mask_0, x = new_x_59_cast_fp16)[name = tensor("next_cache_29_cast_fp16")]; + tensor var_3465_begin_0 = const()[name = tensor("op_3465_begin_0"), val = tensor([0, 0, 4])]; + tensor var_3465_end_0 = const()[name = tensor("op_3465_end_0"), val = tensor([1, 512, 12])]; + tensor var_3465_end_mask_0 = const()[name = tensor("op_3465_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3465_cast_fp16 = slice_by_index(begin = var_3465_begin_0, end = var_3465_end_0, end_mask = var_3465_end_mask_0, x = next_cache_29_cast_fp16)[name = tensor("op_3465_cast_fp16")]; + tensor x_385_pad_type_0 = const()[name = tensor("x_385_pad_type_0"), val = tensor("valid")]; + tensor x_385_groups_0 = const()[name = tensor("x_385_groups_0"), val = tensor(512)]; + tensor x_385_strides_0 = const()[name = tensor("x_385_strides_0"), val = tensor([1])]; + tensor x_385_pad_0 = const()[name = tensor("x_385_pad_0"), val = tensor([0, 0])]; + tensor x_385_dilations_0 = const()[name = tensor("x_385_dilations_0"), val = tensor([1])]; + tensor encoder_layers_14_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_14_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183646144)))]; + tensor x_385_cast_fp16 = conv(dilations = x_385_dilations_0, groups = x_385_groups_0, pad = x_385_pad_0, pad_type = x_385_pad_type_0, strides = x_385_strides_0, weight = encoder_layers_14_conv_depthwise_conv_weight_to_fp16, x = new_x_59_cast_fp16)[name = tensor("x_385_cast_fp16")]; + tensor input_783_perm_0 = const()[name = tensor("input_783_perm_0"), val = tensor([0, 2, 1])]; + tensor x_387_axes_0 = const()[name = tensor("x_387_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_14_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183655424)))]; + tensor encoder_layers_14_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_14_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183656512)))]; + tensor input_783_cast_fp16 = transpose(perm = input_783_perm_0, x = x_385_cast_fp16)[name = tensor("transpose_106")]; + tensor x_387_cast_fp16 = layer_norm(axes = x_387_axes_0, beta = encoder_layers_14_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_conv_batch_norm_weight_to_fp16, x = input_783_cast_fp16)[name = tensor("x_387_cast_fp16")]; + tensor input_785_perm_0 = const()[name = tensor("input_785_perm_0"), val = tensor([0, 2, 1])]; + tensor input_785_cast_fp16 = transpose(perm = input_785_perm_0, x = x_387_cast_fp16)[name = tensor("transpose_105")]; + tensor input_787_cast_fp16 = silu(x = input_785_cast_fp16)[name = tensor("input_787_cast_fp16")]; + tensor x_389_pad_type_0 = const()[name = tensor("x_389_pad_type_0"), val = tensor("valid")]; + tensor x_389_strides_0 = const()[name = tensor("x_389_strides_0"), val = tensor([1])]; + tensor x_389_pad_0 = const()[name = tensor("x_389_pad_0"), val = tensor([0, 0])]; + tensor x_389_dilations_0 = const()[name = tensor("x_389_dilations_0"), val = tensor([1])]; + tensor x_389_groups_0 = const()[name = tensor("x_389_groups_0"), val = tensor(1)]; + tensor encoder_layers_14_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_14_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183657600)))]; + tensor x_389_cast_fp16 = conv(dilations = x_389_dilations_0, groups = x_389_groups_0, pad = x_389_pad_0, pad_type = x_389_pad_type_0, strides = x_389_strides_0, weight = encoder_layers_14_conv_pointwise_conv2_weight_to_fp16, x = input_787_cast_fp16)[name = tensor("x_389_cast_fp16")]; + tensor input_789_perm_0 = const()[name = tensor("input_789_perm_0"), val = tensor([0, 2, 1])]; + tensor input_789_cast_fp16 = transpose(perm = input_789_perm_0, x = x_389_cast_fp16)[name = tensor("transpose_104")]; + tensor input_791_cast_fp16 = add(x = input_775_cast_fp16, y = input_789_cast_fp16)[name = tensor("input_791_cast_fp16")]; + tensor input_793_axes_0 = const()[name = tensor("input_793_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184181952)))]; + tensor encoder_layers_14_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184183040)))]; + tensor input_793_cast_fp16 = layer_norm(axes = input_793_axes_0, beta = encoder_layers_14_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_norm_feed_forward2_weight_to_fp16, x = input_791_cast_fp16)[name = tensor("input_793_cast_fp16")]; + tensor encoder_layers_14_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_14_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184184128)))]; + tensor linear_134_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_14_feed_forward2_linear1_weight_to_fp16, x = input_793_cast_fp16)[name = tensor("linear_134_cast_fp16")]; + tensor input_797_cast_fp16 = silu(x = linear_134_cast_fp16)[name = tensor("input_797_cast_fp16")]; + tensor encoder_layers_14_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_14_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186281344)))]; + tensor linear_135_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_14_feed_forward2_linear2_weight_to_fp16, x = input_797_cast_fp16)[name = tensor("linear_135_cast_fp16")]; + tensor var_3506_to_fp16 = const()[name = tensor("op_3506_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3507_cast_fp16 = mul(x = linear_135_cast_fp16, y = var_3506_to_fp16)[name = tensor("op_3507_cast_fp16")]; + tensor input_803_cast_fp16 = add(x = input_791_cast_fp16, y = var_3507_cast_fp16)[name = tensor("input_803_cast_fp16")]; + tensor input_805_axes_0 = const()[name = tensor("input_805_axes_0"), val = tensor([-1])]; + tensor encoder_layers_14_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_14_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188378560)))]; + tensor encoder_layers_14_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_14_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188379648)))]; + tensor input_805_cast_fp16 = layer_norm(axes = input_805_axes_0, beta = encoder_layers_14_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_14_norm_out_weight_to_fp16, x = input_803_cast_fp16)[name = tensor("input_805_cast_fp16")]; + tensor cache_61_begin_0 = const()[name = tensor("cache_61_begin_0"), val = tensor([15, 0, 0, 0])]; + tensor cache_61_end_0 = const()[name = tensor("cache_61_end_0"), val = tensor([16, 1, 70, 512])]; + tensor cache_61_end_mask_0 = const()[name = tensor("cache_61_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_61_squeeze_mask_0 = const()[name = tensor("cache_61_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_61_cast_fp16 = slice_by_index(begin = cache_61_begin_0, end = cache_61_end_0, end_mask = cache_61_end_mask_0, squeeze_mask = cache_61_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_61_cast_fp16")]; + tensor cache_63_begin_0 = const()[name = tensor("cache_63_begin_0"), val = tensor([15, 0, 0, 0])]; + tensor cache_63_end_0 = const()[name = tensor("cache_63_end_0"), val = tensor([16, 1, 512, 8])]; + tensor cache_63_end_mask_0 = const()[name = tensor("cache_63_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_63_squeeze_mask_0 = const()[name = tensor("cache_63_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_63_cast_fp16 = slice_by_index(begin = cache_63_begin_0, end = cache_63_end_0, end_mask = cache_63_end_mask_0, squeeze_mask = cache_63_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_63_cast_fp16")]; + tensor input_807_axes_0 = const()[name = tensor("input_807_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188380736)))]; + tensor encoder_layers_15_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188381824)))]; + tensor input_807_cast_fp16 = layer_norm(axes = input_807_axes_0, beta = encoder_layers_15_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_norm_feed_forward1_weight_to_fp16, x = input_805_cast_fp16)[name = tensor("input_807_cast_fp16")]; + tensor encoder_layers_15_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_15_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188382912)))]; + tensor linear_136_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_15_feed_forward1_linear1_weight_to_fp16, x = input_807_cast_fp16)[name = tensor("linear_136_cast_fp16")]; + tensor input_811_cast_fp16 = silu(x = linear_136_cast_fp16)[name = tensor("input_811_cast_fp16")]; + tensor encoder_layers_15_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_15_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190480128)))]; + tensor linear_137_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_feed_forward1_linear2_weight_to_fp16, x = input_811_cast_fp16)[name = tensor("linear_137_cast_fp16")]; + tensor var_3541_to_fp16 = const()[name = tensor("op_3541_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3542_cast_fp16 = mul(x = linear_137_cast_fp16, y = var_3541_to_fp16)[name = tensor("op_3542_cast_fp16")]; + tensor input_817_cast_fp16 = add(x = input_805_cast_fp16, y = var_3542_cast_fp16)[name = tensor("input_817_cast_fp16")]; + tensor key_31_axes_0 = const()[name = tensor("key_31_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192577344)))]; + tensor encoder_layers_15_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192578432)))]; + tensor key_31_cast_fp16 = layer_norm(axes = key_31_axes_0, beta = encoder_layers_15_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_norm_self_att_weight_to_fp16, x = input_817_cast_fp16)[name = tensor("key_31_cast_fp16")]; + tensor input_819_interleave_0 = const()[name = tensor("input_819_interleave_0"), val = tensor(false)]; + tensor input_819_cast_fp16 = concat(axis = var_65, interleave = input_819_interleave_0, values = (cache_61_cast_fp16, key_31_cast_fp16))[name = tensor("input_819_cast_fp16")]; + tensor var_3564_begin_0 = const()[name = tensor("op_3564_begin_0"), val = tensor([0, 4, 0])]; + tensor var_3564_end_0 = const()[name = tensor("op_3564_end_0"), val = tensor([1, 70, 512])]; + tensor var_3564_end_mask_0 = const()[name = tensor("op_3564_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3564_cast_fp16 = slice_by_index(begin = var_3564_begin_0, end = var_3564_end_0, end_mask = var_3564_end_mask_0, x = cache_61_cast_fp16)[name = tensor("op_3564_cast_fp16")]; + tensor var_3567_begin_0 = const()[name = tensor("op_3567_begin_0"), val = tensor([0, 0, 0])]; + tensor var_3567_end_0 = const()[name = tensor("op_3567_end_0"), val = tensor([1, 4, 512])]; + tensor var_3567_end_mask_0 = const()[name = tensor("op_3567_end_mask_0"), val = tensor([true, false, true])]; + tensor var_3567_cast_fp16 = slice_by_index(begin = var_3567_begin_0, end = var_3567_end_0, end_mask = var_3567_end_mask_0, x = key_31_cast_fp16)[name = tensor("op_3567_cast_fp16")]; + tensor var_3570_interleave_0 = const()[name = tensor("op_3570_interleave_0"), val = tensor(false)]; + tensor var_3570_cast_fp16 = concat(axis = var_65, interleave = var_3570_interleave_0, values = (var_3564_cast_fp16, var_3567_cast_fp16))[name = tensor("op_3570_cast_fp16")]; + tensor encoder_layers_15_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192579520)))]; + tensor linear_138_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_self_attn_linear_q_weight_to_fp16, x = key_31_cast_fp16)[name = tensor("linear_138_cast_fp16")]; + tensor var_3574 = const()[name = tensor("op_3574"), val = tensor([1, -1, 8, 64])]; + tensor q_91_cast_fp16 = reshape(shape = var_3574, x = linear_138_cast_fp16)[name = tensor("q_91_cast_fp16")]; + tensor encoder_layers_15_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(193103872)))]; + tensor linear_139_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_self_attn_linear_k_weight_to_fp16, x = input_819_cast_fp16)[name = tensor("linear_139_cast_fp16")]; + tensor var_3578 = const()[name = tensor("op_3578"), val = tensor([1, -1, 8, 64])]; + tensor k_61_cast_fp16 = reshape(shape = var_3578, x = linear_139_cast_fp16)[name = tensor("k_61_cast_fp16")]; + tensor encoder_layers_15_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(193628224)))]; + tensor linear_140_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_self_attn_linear_v_weight_to_fp16, x = input_819_cast_fp16)[name = tensor("linear_140_cast_fp16")]; + tensor var_3582 = const()[name = tensor("op_3582"), val = tensor([1, -1, 8, 64])]; + tensor v_31_cast_fp16 = reshape(shape = var_3582, x = linear_140_cast_fp16)[name = tensor("v_31_cast_fp16")]; + tensor value_33_perm_0 = const()[name = tensor("value_33_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_15_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194152576)))]; + tensor var_3594_cast_fp16 = add(x = q_91_cast_fp16, y = encoder_layers_15_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3594_cast_fp16")]; + tensor encoder_layers_15_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194153664)))]; + tensor var_3596_cast_fp16 = add(x = q_91_cast_fp16, y = encoder_layers_15_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3596_cast_fp16")]; + tensor q_with_bias_v_31_perm_0 = const()[name = tensor("q_with_bias_v_31_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_397_transpose_x_0 = const()[name = tensor("x_397_transpose_x_0"), val = tensor(false)]; + tensor x_397_transpose_y_0 = const()[name = tensor("x_397_transpose_y_0"), val = tensor(false)]; + tensor var_3598_to_fp16 = const()[name = tensor("op_3598_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194154752)))]; + tensor q_with_bias_v_31_cast_fp16 = transpose(perm = q_with_bias_v_31_perm_0, x = var_3596_cast_fp16)[name = tensor("transpose_102")]; + tensor x_397_cast_fp16 = matmul(transpose_x = x_397_transpose_x_0, transpose_y = x_397_transpose_y_0, x = q_with_bias_v_31_cast_fp16, y = var_3598_to_fp16)[name = tensor("x_397_cast_fp16")]; + tensor x_399_pad_0 = const()[name = tensor("x_399_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_399_mode_0 = const()[name = tensor("x_399_mode_0"), val = tensor("constant")]; + tensor const_274_to_fp16 = const()[name = tensor("const_274_to_fp16"), val = tensor(0x0p+0)]; + tensor x_399_cast_fp16 = pad(constant_val = const_274_to_fp16, mode = x_399_mode_0, pad = x_399_pad_0, x = x_397_cast_fp16)[name = tensor("x_399_cast_fp16")]; + tensor var_3606 = const()[name = tensor("op_3606"), val = tensor([1, 8, -1, 8])]; + tensor x_401_cast_fp16 = reshape(shape = var_3606, x = x_399_cast_fp16)[name = tensor("x_401_cast_fp16")]; + tensor var_3610_begin_0 = const()[name = tensor("op_3610_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_3610_end_0 = const()[name = tensor("op_3610_end_0"), val = tensor([1, 8, 156, 8])]; + tensor var_3610_end_mask_0 = const()[name = tensor("op_3610_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_3610_cast_fp16 = slice_by_index(begin = var_3610_begin_0, end = var_3610_end_0, end_mask = var_3610_end_mask_0, x = x_401_cast_fp16)[name = tensor("op_3610_cast_fp16")]; + tensor var_3611 = const()[name = tensor("op_3611"), val = tensor([1, 8, 8, 155])]; + tensor matrix_bd_61_cast_fp16 = reshape(shape = var_3611, x = var_3610_cast_fp16)[name = tensor("matrix_bd_61_cast_fp16")]; + tensor matrix_ac_31_transpose_x_0 = const()[name = tensor("matrix_ac_31_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_31_transpose_y_0 = const()[name = tensor("matrix_ac_31_transpose_y_0"), val = tensor(false)]; + tensor transpose_81_perm_0 = const()[name = tensor("transpose_81_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_82_perm_0 = const()[name = tensor("transpose_82_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_82 = transpose(perm = transpose_82_perm_0, x = k_61_cast_fp16)[name = tensor("transpose_100")]; + tensor transpose_81 = transpose(perm = transpose_81_perm_0, x = var_3594_cast_fp16)[name = tensor("transpose_101")]; + tensor matrix_ac_31_cast_fp16 = matmul(transpose_x = matrix_ac_31_transpose_x_0, transpose_y = matrix_ac_31_transpose_y_0, x = transpose_81, y = transpose_82)[name = tensor("matrix_ac_31_cast_fp16")]; + tensor matrix_bd_63_begin_0 = const()[name = tensor("matrix_bd_63_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_63_end_0 = const()[name = tensor("matrix_bd_63_end_0"), val = tensor([1, 8, 8, 78])]; + tensor matrix_bd_63_end_mask_0 = const()[name = tensor("matrix_bd_63_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_63_cast_fp16 = slice_by_index(begin = matrix_bd_63_begin_0, end = matrix_bd_63_end_0, end_mask = matrix_bd_63_end_mask_0, x = matrix_bd_61_cast_fp16)[name = tensor("matrix_bd_63_cast_fp16")]; + tensor var_3620_cast_fp16 = add(x = matrix_ac_31_cast_fp16, y = matrix_bd_63_cast_fp16)[name = tensor("op_3620_cast_fp16")]; + tensor _inversed_scores_61_y_0_to_fp16 = const()[name = tensor("_inversed_scores_61_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_61_cast_fp16 = mul(x = var_3620_cast_fp16, y = _inversed_scores_61_y_0_to_fp16)[name = tensor("_inversed_scores_61_cast_fp16")]; + tensor scores_63_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_61_cast_fp16, cond = mask_11)[name = tensor("scores_63_cast_fp16")]; + tensor var_3626_cast_fp16 = softmax(axis = var_56, x = scores_63_cast_fp16)[name = tensor("op_3626_cast_fp16")]; + tensor input_821_cast_fp16 = select(a = var_40_to_fp16, b = var_3626_cast_fp16, cond = mask_11)[name = tensor("input_821_cast_fp16")]; + tensor x_403_transpose_x_0 = const()[name = tensor("x_403_transpose_x_0"), val = tensor(false)]; + tensor x_403_transpose_y_0 = const()[name = tensor("x_403_transpose_y_0"), val = tensor(false)]; + tensor value_33_cast_fp16 = transpose(perm = value_33_perm_0, x = v_31_cast_fp16)[name = tensor("transpose_103")]; + tensor x_403_cast_fp16 = matmul(transpose_x = x_403_transpose_x_0, transpose_y = x_403_transpose_y_0, x = input_821_cast_fp16, y = value_33_cast_fp16)[name = tensor("x_403_cast_fp16")]; + tensor var_3630_perm_0 = const()[name = tensor("op_3630_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_3631 = const()[name = tensor("op_3631"), val = tensor([1, -1, 512])]; + tensor var_3630_cast_fp16 = transpose(perm = var_3630_perm_0, x = x_403_cast_fp16)[name = tensor("transpose_99")]; + tensor input_823_cast_fp16 = reshape(shape = var_3631, x = var_3630_cast_fp16)[name = tensor("input_823_cast_fp16")]; + tensor encoder_layers_15_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_15_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194313536)))]; + tensor linear_142_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_self_attn_linear_out_weight_to_fp16, x = input_823_cast_fp16)[name = tensor("linear_142_cast_fp16")]; + tensor input_827_cast_fp16 = add(x = input_817_cast_fp16, y = linear_142_cast_fp16)[name = tensor("input_827_cast_fp16")]; + tensor x_407_axes_0 = const()[name = tensor("x_407_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194837888)))]; + tensor encoder_layers_15_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194838976)))]; + tensor x_407_cast_fp16 = layer_norm(axes = x_407_axes_0, beta = encoder_layers_15_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_norm_conv_weight_to_fp16, x = input_827_cast_fp16)[name = tensor("x_407_cast_fp16")]; + tensor input_829_perm_0 = const()[name = tensor("input_829_perm_0"), val = tensor([0, 2, 1])]; + tensor input_831_pad_type_0 = const()[name = tensor("input_831_pad_type_0"), val = tensor("valid")]; + tensor input_831_strides_0 = const()[name = tensor("input_831_strides_0"), val = tensor([1])]; + tensor input_831_pad_0 = const()[name = tensor("input_831_pad_0"), val = tensor([0, 0])]; + tensor input_831_dilations_0 = const()[name = tensor("input_831_dilations_0"), val = tensor([1])]; + tensor input_831_groups_0 = const()[name = tensor("input_831_groups_0"), val = tensor(1)]; + tensor encoder_layers_15_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_15_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194840064)))]; + tensor input_829_cast_fp16 = transpose(perm = input_829_perm_0, x = x_407_cast_fp16)[name = tensor("transpose_98")]; + tensor input_831_cast_fp16 = conv(dilations = input_831_dilations_0, groups = input_831_groups_0, pad = input_831_pad_0, pad_type = input_831_pad_type_0, strides = input_831_strides_0, weight = encoder_layers_15_conv_pointwise_conv1_weight_to_fp16, x = input_829_cast_fp16)[name = tensor("input_831_cast_fp16")]; + tensor x_409_split_num_splits_0 = const()[name = tensor("x_409_split_num_splits_0"), val = tensor(2)]; + tensor x_409_split_axis_0 = const()[name = tensor("x_409_split_axis_0"), val = tensor(1)]; + tensor x_409_split_cast_fp16_0, tensor x_409_split_cast_fp16_1 = split(axis = x_409_split_axis_0, num_splits = x_409_split_num_splits_0, x = input_831_cast_fp16)[name = tensor("x_409_split_cast_fp16")]; + tensor x_409_split_1_sigmoid_cast_fp16 = sigmoid(x = x_409_split_cast_fp16_1)[name = tensor("x_409_split_1_sigmoid_cast_fp16")]; + tensor x_409_cast_fp16 = mul(x = x_409_split_cast_fp16_0, y = x_409_split_1_sigmoid_cast_fp16)[name = tensor("x_409_cast_fp16")]; + tensor input_833_cast_fp16 = select(a = var_40_to_fp16, b = x_409_cast_fp16, cond = var_551)[name = tensor("input_833_cast_fp16")]; + tensor new_x_63_interleave_0 = const()[name = tensor("new_x_63_interleave_0"), val = tensor(false)]; + tensor new_x_63_cast_fp16 = concat(axis = var_56, interleave = new_x_63_interleave_0, values = (cache_63_cast_fp16, input_833_cast_fp16))[name = tensor("new_x_63_cast_fp16")]; + tensor next_cache_31_begin_0 = const()[name = tensor("next_cache_31_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_31_end_0 = const()[name = tensor("next_cache_31_end_0"), val = tensor([1, 512, 12])]; + tensor next_cache_31_end_mask_0 = const()[name = tensor("next_cache_31_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_31_cast_fp16 = slice_by_index(begin = next_cache_31_begin_0, end = next_cache_31_end_0, end_mask = next_cache_31_end_mask_0, x = new_x_63_cast_fp16)[name = tensor("next_cache_31_cast_fp16")]; + tensor var_3672_begin_0 = const()[name = tensor("op_3672_begin_0"), val = tensor([0, 0, 4])]; + tensor var_3672_end_0 = const()[name = tensor("op_3672_end_0"), val = tensor([1, 512, 12])]; + tensor var_3672_end_mask_0 = const()[name = tensor("op_3672_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3672_cast_fp16 = slice_by_index(begin = var_3672_begin_0, end = var_3672_end_0, end_mask = var_3672_end_mask_0, x = next_cache_31_cast_fp16)[name = tensor("op_3672_cast_fp16")]; + tensor x_411_pad_type_0 = const()[name = tensor("x_411_pad_type_0"), val = tensor("valid")]; + tensor x_411_groups_0 = const()[name = tensor("x_411_groups_0"), val = tensor(512)]; + tensor x_411_strides_0 = const()[name = tensor("x_411_strides_0"), val = tensor([1])]; + tensor x_411_pad_0 = const()[name = tensor("x_411_pad_0"), val = tensor([0, 0])]; + tensor x_411_dilations_0 = const()[name = tensor("x_411_dilations_0"), val = tensor([1])]; + tensor encoder_layers_15_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_15_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195888704)))]; + tensor x_411_cast_fp16 = conv(dilations = x_411_dilations_0, groups = x_411_groups_0, pad = x_411_pad_0, pad_type = x_411_pad_type_0, strides = x_411_strides_0, weight = encoder_layers_15_conv_depthwise_conv_weight_to_fp16, x = new_x_63_cast_fp16)[name = tensor("x_411_cast_fp16")]; + tensor input_835_perm_0 = const()[name = tensor("input_835_perm_0"), val = tensor([0, 2, 1])]; + tensor x_413_axes_0 = const()[name = tensor("x_413_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_15_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195897984)))]; + tensor encoder_layers_15_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_15_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195899072)))]; + tensor input_835_cast_fp16 = transpose(perm = input_835_perm_0, x = x_411_cast_fp16)[name = tensor("transpose_97")]; + tensor x_413_cast_fp16 = layer_norm(axes = x_413_axes_0, beta = encoder_layers_15_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_conv_batch_norm_weight_to_fp16, x = input_835_cast_fp16)[name = tensor("x_413_cast_fp16")]; + tensor input_837_perm_0 = const()[name = tensor("input_837_perm_0"), val = tensor([0, 2, 1])]; + tensor input_837_cast_fp16 = transpose(perm = input_837_perm_0, x = x_413_cast_fp16)[name = tensor("transpose_96")]; + tensor input_839_cast_fp16 = silu(x = input_837_cast_fp16)[name = tensor("input_839_cast_fp16")]; + tensor x_415_pad_type_0 = const()[name = tensor("x_415_pad_type_0"), val = tensor("valid")]; + tensor x_415_strides_0 = const()[name = tensor("x_415_strides_0"), val = tensor([1])]; + tensor x_415_pad_0 = const()[name = tensor("x_415_pad_0"), val = tensor([0, 0])]; + tensor x_415_dilations_0 = const()[name = tensor("x_415_dilations_0"), val = tensor([1])]; + tensor x_415_groups_0 = const()[name = tensor("x_415_groups_0"), val = tensor(1)]; + tensor encoder_layers_15_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_15_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195900160)))]; + tensor x_415_cast_fp16 = conv(dilations = x_415_dilations_0, groups = x_415_groups_0, pad = x_415_pad_0, pad_type = x_415_pad_type_0, strides = x_415_strides_0, weight = encoder_layers_15_conv_pointwise_conv2_weight_to_fp16, x = input_839_cast_fp16)[name = tensor("x_415_cast_fp16")]; + tensor input_841_perm_0 = const()[name = tensor("input_841_perm_0"), val = tensor([0, 2, 1])]; + tensor input_841_cast_fp16 = transpose(perm = input_841_perm_0, x = x_415_cast_fp16)[name = tensor("transpose_95")]; + tensor input_843_cast_fp16 = add(x = input_827_cast_fp16, y = input_841_cast_fp16)[name = tensor("input_843_cast_fp16")]; + tensor input_845_axes_0 = const()[name = tensor("input_845_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196424512)))]; + tensor encoder_layers_15_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196425600)))]; + tensor input_845_cast_fp16 = layer_norm(axes = input_845_axes_0, beta = encoder_layers_15_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_norm_feed_forward2_weight_to_fp16, x = input_843_cast_fp16)[name = tensor("input_845_cast_fp16")]; + tensor encoder_layers_15_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_15_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196426688)))]; + tensor linear_143_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_15_feed_forward2_linear1_weight_to_fp16, x = input_845_cast_fp16)[name = tensor("linear_143_cast_fp16")]; + tensor input_849_cast_fp16 = silu(x = linear_143_cast_fp16)[name = tensor("input_849_cast_fp16")]; + tensor encoder_layers_15_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_15_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198523904)))]; + tensor linear_144_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_15_feed_forward2_linear2_weight_to_fp16, x = input_849_cast_fp16)[name = tensor("linear_144_cast_fp16")]; + tensor var_3713_to_fp16 = const()[name = tensor("op_3713_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3714_cast_fp16 = mul(x = linear_144_cast_fp16, y = var_3713_to_fp16)[name = tensor("op_3714_cast_fp16")]; + tensor input_855_cast_fp16 = add(x = input_843_cast_fp16, y = var_3714_cast_fp16)[name = tensor("input_855_cast_fp16")]; + tensor input_857_axes_0 = const()[name = tensor("input_857_axes_0"), val = tensor([-1])]; + tensor encoder_layers_15_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_15_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200621120)))]; + tensor encoder_layers_15_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_15_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200622208)))]; + tensor input_857_cast_fp16 = layer_norm(axes = input_857_axes_0, beta = encoder_layers_15_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_15_norm_out_weight_to_fp16, x = input_855_cast_fp16)[name = tensor("input_857_cast_fp16")]; + tensor cache_65_begin_0 = const()[name = tensor("cache_65_begin_0"), val = tensor([16, 0, 0, 0])]; + tensor cache_65_end_0 = const()[name = tensor("cache_65_end_0"), val = tensor([17, 1, 70, 512])]; + tensor cache_65_end_mask_0 = const()[name = tensor("cache_65_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_65_squeeze_mask_0 = const()[name = tensor("cache_65_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_65_cast_fp16 = slice_by_index(begin = cache_65_begin_0, end = cache_65_end_0, end_mask = cache_65_end_mask_0, squeeze_mask = cache_65_squeeze_mask_0, x = cache_last_channel_to_fp16)[name = tensor("cache_65_cast_fp16")]; + tensor cache_begin_0 = const()[name = tensor("cache_begin_0"), val = tensor([16, 0, 0, 0])]; + tensor cache_end_0 = const()[name = tensor("cache_end_0"), val = tensor([17, 1, 512, 8])]; + tensor cache_end_mask_0 = const()[name = tensor("cache_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_squeeze_mask_0 = const()[name = tensor("cache_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_cast_fp16 = slice_by_index(begin = cache_begin_0, end = cache_end_0, end_mask = cache_end_mask_0, squeeze_mask = cache_squeeze_mask_0, x = cache_last_time_to_fp16)[name = tensor("cache_cast_fp16")]; + tensor input_859_axes_0 = const()[name = tensor("input_859_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200623296)))]; + tensor encoder_layers_16_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200624384)))]; + tensor input_859_cast_fp16 = layer_norm(axes = input_859_axes_0, beta = encoder_layers_16_norm_feed_forward1_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_norm_feed_forward1_weight_to_fp16, x = input_857_cast_fp16)[name = tensor("input_859_cast_fp16")]; + tensor encoder_layers_16_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_16_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200625472)))]; + tensor linear_145_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_16_feed_forward1_linear1_weight_to_fp16, x = input_859_cast_fp16)[name = tensor("linear_145_cast_fp16")]; + tensor input_863_cast_fp16 = silu(x = linear_145_cast_fp16)[name = tensor("input_863_cast_fp16")]; + tensor encoder_layers_16_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_16_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202722688)))]; + tensor linear_146_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_feed_forward1_linear2_weight_to_fp16, x = input_863_cast_fp16)[name = tensor("linear_146_cast_fp16")]; + tensor var_3748_to_fp16 = const()[name = tensor("op_3748_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3749_cast_fp16 = mul(x = linear_146_cast_fp16, y = var_3748_to_fp16)[name = tensor("op_3749_cast_fp16")]; + tensor input_869_cast_fp16 = add(x = input_857_cast_fp16, y = var_3749_cast_fp16)[name = tensor("input_869_cast_fp16")]; + tensor key_axes_0 = const()[name = tensor("key_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204819904)))]; + tensor encoder_layers_16_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204820992)))]; + tensor key_cast_fp16 = layer_norm(axes = key_axes_0, beta = encoder_layers_16_norm_self_att_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_norm_self_att_weight_to_fp16, x = input_869_cast_fp16)[name = tensor("key_cast_fp16")]; + tensor input_871_interleave_0 = const()[name = tensor("input_871_interleave_0"), val = tensor(false)]; + tensor input_871_cast_fp16 = concat(axis = var_65, interleave = input_871_interleave_0, values = (cache_65_cast_fp16, key_cast_fp16))[name = tensor("input_871_cast_fp16")]; + tensor var_3771_begin_0 = const()[name = tensor("op_3771_begin_0"), val = tensor([0, 4, 0])]; + tensor var_3771_end_0 = const()[name = tensor("op_3771_end_0"), val = tensor([1, 70, 512])]; + tensor var_3771_end_mask_0 = const()[name = tensor("op_3771_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3771_cast_fp16 = slice_by_index(begin = var_3771_begin_0, end = var_3771_end_0, end_mask = var_3771_end_mask_0, x = cache_65_cast_fp16)[name = tensor("op_3771_cast_fp16")]; + tensor var_3774_begin_0 = const()[name = tensor("op_3774_begin_0"), val = tensor([0, 0, 0])]; + tensor var_3774_end_0 = const()[name = tensor("op_3774_end_0"), val = tensor([1, 4, 512])]; + tensor var_3774_end_mask_0 = const()[name = tensor("op_3774_end_mask_0"), val = tensor([true, false, true])]; + tensor var_3774_cast_fp16 = slice_by_index(begin = var_3774_begin_0, end = var_3774_end_0, end_mask = var_3774_end_mask_0, x = key_cast_fp16)[name = tensor("op_3774_cast_fp16")]; + tensor cache_last_channel_cur_interleave_0 = const()[name = tensor("cache_last_channel_cur_interleave_0"), val = tensor(false)]; + tensor cache_last_channel_cur_cast_fp16 = concat(axis = var_65, interleave = cache_last_channel_cur_interleave_0, values = (var_3771_cast_fp16, var_3774_cast_fp16))[name = tensor("cache_last_channel_cur_cast_fp16")]; + tensor encoder_layers_16_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204822080)))]; + tensor linear_147_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_self_attn_linear_q_weight_to_fp16, x = key_cast_fp16)[name = tensor("linear_147_cast_fp16")]; + tensor var_3781 = const()[name = tensor("op_3781"), val = tensor([1, -1, 8, 64])]; + tensor q_97_cast_fp16 = reshape(shape = var_3781, x = linear_147_cast_fp16)[name = tensor("q_97_cast_fp16")]; + tensor encoder_layers_16_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205346432)))]; + tensor linear_148_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_self_attn_linear_k_weight_to_fp16, x = input_871_cast_fp16)[name = tensor("linear_148_cast_fp16")]; + tensor var_3785 = const()[name = tensor("op_3785"), val = tensor([1, -1, 8, 64])]; + tensor k_65_cast_fp16 = reshape(shape = var_3785, x = linear_148_cast_fp16)[name = tensor("k_65_cast_fp16")]; + tensor encoder_layers_16_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205870784)))]; + tensor linear_149_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_self_attn_linear_v_weight_to_fp16, x = input_871_cast_fp16)[name = tensor("linear_149_cast_fp16")]; + tensor var_3789 = const()[name = tensor("op_3789"), val = tensor([1, -1, 8, 64])]; + tensor v_cast_fp16 = reshape(shape = var_3789, x = linear_149_cast_fp16)[name = tensor("v_cast_fp16")]; + tensor value_perm_0 = const()[name = tensor("value_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor encoder_layers_16_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206395136)))]; + tensor var_3801_cast_fp16 = add(x = q_97_cast_fp16, y = encoder_layers_16_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3801_cast_fp16")]; + tensor encoder_layers_16_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206396224)))]; + tensor var_3803_cast_fp16 = add(x = q_97_cast_fp16, y = encoder_layers_16_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3803_cast_fp16")]; + tensor q_with_bias_v_perm_0 = const()[name = tensor("q_with_bias_v_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor x_423_transpose_x_0 = const()[name = tensor("x_423_transpose_x_0"), val = tensor(false)]; + tensor x_423_transpose_y_0 = const()[name = tensor("x_423_transpose_y_0"), val = tensor(false)]; + tensor var_3805_to_fp16 = const()[name = tensor("op_3805_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206397312)))]; + tensor q_with_bias_v_cast_fp16 = transpose(perm = q_with_bias_v_perm_0, x = var_3803_cast_fp16)[name = tensor("transpose_93")]; + tensor x_423_cast_fp16 = matmul(transpose_x = x_423_transpose_x_0, transpose_y = x_423_transpose_y_0, x = q_with_bias_v_cast_fp16, y = var_3805_to_fp16)[name = tensor("x_423_cast_fp16")]; + tensor x_425_pad_0 = const()[name = tensor("x_425_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_425_mode_0 = const()[name = tensor("x_425_mode_0"), val = tensor("constant")]; + tensor const_287_to_fp16 = const()[name = tensor("const_287_to_fp16"), val = tensor(0x0p+0)]; + tensor x_425_cast_fp16 = pad(constant_val = const_287_to_fp16, mode = x_425_mode_0, pad = x_425_pad_0, x = x_423_cast_fp16)[name = tensor("x_425_cast_fp16")]; + tensor var_3813 = const()[name = tensor("op_3813"), val = tensor([1, 8, -1, 8])]; + tensor x_427_cast_fp16 = reshape(shape = var_3813, x = x_425_cast_fp16)[name = tensor("x_427_cast_fp16")]; + tensor var_3817_begin_0 = const()[name = tensor("op_3817_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_3817_end_0 = const()[name = tensor("op_3817_end_0"), val = tensor([1, 8, 156, 8])]; + tensor var_3817_end_mask_0 = const()[name = tensor("op_3817_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_3817_cast_fp16 = slice_by_index(begin = var_3817_begin_0, end = var_3817_end_0, end_mask = var_3817_end_mask_0, x = x_427_cast_fp16)[name = tensor("op_3817_cast_fp16")]; + tensor var_3818 = const()[name = tensor("op_3818"), val = tensor([1, 8, 8, 155])]; + tensor matrix_bd_65_cast_fp16 = reshape(shape = var_3818, x = var_3817_cast_fp16)[name = tensor("matrix_bd_65_cast_fp16")]; + tensor matrix_ac_transpose_x_0 = const()[name = tensor("matrix_ac_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_transpose_y_0 = const()[name = tensor("matrix_ac_transpose_y_0"), val = tensor(false)]; + tensor transpose_83_perm_0 = const()[name = tensor("transpose_83_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_84_perm_0 = const()[name = tensor("transpose_84_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_84 = transpose(perm = transpose_84_perm_0, x = k_65_cast_fp16)[name = tensor("transpose_91")]; + tensor transpose_83 = transpose(perm = transpose_83_perm_0, x = var_3801_cast_fp16)[name = tensor("transpose_92")]; + tensor matrix_ac_cast_fp16 = matmul(transpose_x = matrix_ac_transpose_x_0, transpose_y = matrix_ac_transpose_y_0, x = transpose_83, y = transpose_84)[name = tensor("matrix_ac_cast_fp16")]; + tensor matrix_bd_begin_0 = const()[name = tensor("matrix_bd_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_end_0 = const()[name = tensor("matrix_bd_end_0"), val = tensor([1, 8, 8, 78])]; + tensor matrix_bd_end_mask_0 = const()[name = tensor("matrix_bd_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_cast_fp16 = slice_by_index(begin = matrix_bd_begin_0, end = matrix_bd_end_0, end_mask = matrix_bd_end_mask_0, x = matrix_bd_65_cast_fp16)[name = tensor("matrix_bd_cast_fp16")]; + tensor var_3827_cast_fp16 = add(x = matrix_ac_cast_fp16, y = matrix_bd_cast_fp16)[name = tensor("op_3827_cast_fp16")]; + tensor _inversed_scores_65_y_0_to_fp16 = const()[name = tensor("_inversed_scores_65_y_0_to_fp16"), val = tensor(0x1p-3)]; + tensor _inversed_scores_65_cast_fp16 = mul(x = var_3827_cast_fp16, y = _inversed_scores_65_y_0_to_fp16)[name = tensor("_inversed_scores_65_cast_fp16")]; + tensor scores_cast_fp16 = select(a = var_41_to_fp16, b = _inversed_scores_65_cast_fp16, cond = mask_11)[name = tensor("scores_cast_fp16")]; + tensor var_3833_cast_fp16 = softmax(axis = var_56, x = scores_cast_fp16)[name = tensor("op_3833_cast_fp16")]; + tensor input_873_cast_fp16 = select(a = var_40_to_fp16, b = var_3833_cast_fp16, cond = mask_11)[name = tensor("input_873_cast_fp16")]; + tensor x_429_transpose_x_0 = const()[name = tensor("x_429_transpose_x_0"), val = tensor(false)]; + tensor x_429_transpose_y_0 = const()[name = tensor("x_429_transpose_y_0"), val = tensor(false)]; + tensor value_cast_fp16 = transpose(perm = value_perm_0, x = v_cast_fp16)[name = tensor("transpose_94")]; + tensor x_429_cast_fp16 = matmul(transpose_x = x_429_transpose_x_0, transpose_y = x_429_transpose_y_0, x = input_873_cast_fp16, y = value_cast_fp16)[name = tensor("x_429_cast_fp16")]; + tensor var_3837_perm_0 = const()[name = tensor("op_3837_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_3838 = const()[name = tensor("op_3838"), val = tensor([1, -1, 512])]; + tensor var_3837_cast_fp16 = transpose(perm = var_3837_perm_0, x = x_429_cast_fp16)[name = tensor("transpose_90")]; + tensor input_875_cast_fp16 = reshape(shape = var_3838, x = var_3837_cast_fp16)[name = tensor("input_875_cast_fp16")]; + tensor encoder_layers_16_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_layers_16_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206556096)))]; + tensor linear_151_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_self_attn_linear_out_weight_to_fp16, x = input_875_cast_fp16)[name = tensor("linear_151_cast_fp16")]; + tensor input_879_cast_fp16 = add(x = input_869_cast_fp16, y = linear_151_cast_fp16)[name = tensor("input_879_cast_fp16")]; + tensor x_433_axes_0 = const()[name = tensor("x_433_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207080448)))]; + tensor encoder_layers_16_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207081536)))]; + tensor x_433_cast_fp16 = layer_norm(axes = x_433_axes_0, beta = encoder_layers_16_norm_conv_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_norm_conv_weight_to_fp16, x = input_879_cast_fp16)[name = tensor("x_433_cast_fp16")]; + tensor input_881_perm_0 = const()[name = tensor("input_881_perm_0"), val = tensor([0, 2, 1])]; + tensor input_883_pad_type_0 = const()[name = tensor("input_883_pad_type_0"), val = tensor("valid")]; + tensor input_883_strides_0 = const()[name = tensor("input_883_strides_0"), val = tensor([1])]; + tensor input_883_pad_0 = const()[name = tensor("input_883_pad_0"), val = tensor([0, 0])]; + tensor input_883_dilations_0 = const()[name = tensor("input_883_dilations_0"), val = tensor([1])]; + tensor input_883_groups_0 = const()[name = tensor("input_883_groups_0"), val = tensor(1)]; + tensor encoder_layers_16_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_layers_16_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207082624)))]; + tensor input_881_cast_fp16 = transpose(perm = input_881_perm_0, x = x_433_cast_fp16)[name = tensor("transpose_89")]; + tensor input_883_cast_fp16 = conv(dilations = input_883_dilations_0, groups = input_883_groups_0, pad = input_883_pad_0, pad_type = input_883_pad_type_0, strides = input_883_strides_0, weight = encoder_layers_16_conv_pointwise_conv1_weight_to_fp16, x = input_881_cast_fp16)[name = tensor("input_883_cast_fp16")]; + tensor x_435_split_num_splits_0 = const()[name = tensor("x_435_split_num_splits_0"), val = tensor(2)]; + tensor x_435_split_axis_0 = const()[name = tensor("x_435_split_axis_0"), val = tensor(1)]; + tensor x_435_split_cast_fp16_0, tensor x_435_split_cast_fp16_1 = split(axis = x_435_split_axis_0, num_splits = x_435_split_num_splits_0, x = input_883_cast_fp16)[name = tensor("x_435_split_cast_fp16")]; + tensor x_435_split_1_sigmoid_cast_fp16 = sigmoid(x = x_435_split_cast_fp16_1)[name = tensor("x_435_split_1_sigmoid_cast_fp16")]; + tensor x_435_cast_fp16 = mul(x = x_435_split_cast_fp16_0, y = x_435_split_1_sigmoid_cast_fp16)[name = tensor("x_435_cast_fp16")]; + tensor input_885_cast_fp16 = select(a = var_40_to_fp16, b = x_435_cast_fp16, cond = var_551)[name = tensor("input_885_cast_fp16")]; + tensor new_x_interleave_0 = const()[name = tensor("new_x_interleave_0"), val = tensor(false)]; + tensor new_x_cast_fp16 = concat(axis = var_56, interleave = new_x_interleave_0, values = (cache_cast_fp16, input_885_cast_fp16))[name = tensor("new_x_cast_fp16")]; + tensor next_cache_begin_0 = const()[name = tensor("next_cache_begin_0"), val = tensor([0, 0, 0])]; + tensor next_cache_end_0 = const()[name = tensor("next_cache_end_0"), val = tensor([1, 512, 12])]; + tensor next_cache_end_mask_0 = const()[name = tensor("next_cache_end_mask_0"), val = tensor([true, true, false])]; + tensor next_cache_cast_fp16 = slice_by_index(begin = next_cache_begin_0, end = next_cache_end_0, end_mask = next_cache_end_mask_0, x = new_x_cast_fp16)[name = tensor("next_cache_cast_fp16")]; + tensor cache_last_time_cur_begin_0 = const()[name = tensor("cache_last_time_cur_begin_0"), val = tensor([0, 0, 4])]; + tensor cache_last_time_cur_end_0 = const()[name = tensor("cache_last_time_cur_end_0"), val = tensor([1, 512, 12])]; + tensor cache_last_time_cur_end_mask_0 = const()[name = tensor("cache_last_time_cur_end_mask_0"), val = tensor([true, true, true])]; + tensor cache_last_time_cur_cast_fp16 = slice_by_index(begin = cache_last_time_cur_begin_0, end = cache_last_time_cur_end_0, end_mask = cache_last_time_cur_end_mask_0, x = next_cache_cast_fp16)[name = tensor("cache_last_time_cur_cast_fp16")]; + tensor x_437_pad_type_0 = const()[name = tensor("x_437_pad_type_0"), val = tensor("valid")]; + tensor x_437_groups_0 = const()[name = tensor("x_437_groups_0"), val = tensor(512)]; + tensor x_437_strides_0 = const()[name = tensor("x_437_strides_0"), val = tensor([1])]; + tensor x_437_pad_0 = const()[name = tensor("x_437_pad_0"), val = tensor([0, 0])]; + tensor x_437_dilations_0 = const()[name = tensor("x_437_dilations_0"), val = tensor([1])]; + tensor encoder_layers_16_conv_depthwise_conv_weight_to_fp16 = const()[name = tensor("encoder_layers_16_conv_depthwise_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208131264)))]; + tensor x_437_cast_fp16 = conv(dilations = x_437_dilations_0, groups = x_437_groups_0, pad = x_437_pad_0, pad_type = x_437_pad_type_0, strides = x_437_strides_0, weight = encoder_layers_16_conv_depthwise_conv_weight_to_fp16, x = new_x_cast_fp16)[name = tensor("x_437_cast_fp16")]; + tensor input_887_perm_0 = const()[name = tensor("input_887_perm_0"), val = tensor([0, 2, 1])]; + tensor x_439_axes_0 = const()[name = tensor("x_439_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_conv_batch_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_16_conv_batch_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208140544)))]; + tensor encoder_layers_16_conv_batch_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_16_conv_batch_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208141632)))]; + tensor input_887_cast_fp16 = transpose(perm = input_887_perm_0, x = x_437_cast_fp16)[name = tensor("transpose_88")]; + tensor x_439_cast_fp16 = layer_norm(axes = x_439_axes_0, beta = encoder_layers_16_conv_batch_norm_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_conv_batch_norm_weight_to_fp16, x = input_887_cast_fp16)[name = tensor("x_439_cast_fp16")]; + tensor input_889_perm_0 = const()[name = tensor("input_889_perm_0"), val = tensor([0, 2, 1])]; + tensor input_889_cast_fp16 = transpose(perm = input_889_perm_0, x = x_439_cast_fp16)[name = tensor("transpose_87")]; + tensor input_891_cast_fp16 = silu(x = input_889_cast_fp16)[name = tensor("input_891_cast_fp16")]; + tensor x_441_pad_type_0 = const()[name = tensor("x_441_pad_type_0"), val = tensor("valid")]; + tensor x_441_strides_0 = const()[name = tensor("x_441_strides_0"), val = tensor([1])]; + tensor x_441_pad_0 = const()[name = tensor("x_441_pad_0"), val = tensor([0, 0])]; + tensor x_441_dilations_0 = const()[name = tensor("x_441_dilations_0"), val = tensor([1])]; + tensor x_441_groups_0 = const()[name = tensor("x_441_groups_0"), val = tensor(1)]; + tensor encoder_layers_16_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_layers_16_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208142720)))]; + tensor x_441_cast_fp16 = conv(dilations = x_441_dilations_0, groups = x_441_groups_0, pad = x_441_pad_0, pad_type = x_441_pad_type_0, strides = x_441_strides_0, weight = encoder_layers_16_conv_pointwise_conv2_weight_to_fp16, x = input_891_cast_fp16)[name = tensor("x_441_cast_fp16")]; + tensor input_893_perm_0 = const()[name = tensor("input_893_perm_0"), val = tensor([0, 2, 1])]; + tensor input_893_cast_fp16 = transpose(perm = input_893_perm_0, x = x_441_cast_fp16)[name = tensor("transpose_86")]; + tensor input_895_cast_fp16 = add(x = input_879_cast_fp16, y = input_893_cast_fp16)[name = tensor("input_895_cast_fp16")]; + tensor input_897_axes_0 = const()[name = tensor("input_897_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208667072)))]; + tensor encoder_layers_16_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208668160)))]; + tensor input_897_cast_fp16 = layer_norm(axes = input_897_axes_0, beta = encoder_layers_16_norm_feed_forward2_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_norm_feed_forward2_weight_to_fp16, x = input_895_cast_fp16)[name = tensor("input_897_cast_fp16")]; + tensor encoder_layers_16_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_layers_16_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208669248)))]; + tensor linear_152_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = encoder_layers_16_feed_forward2_linear1_weight_to_fp16, x = input_897_cast_fp16)[name = tensor("linear_152_cast_fp16")]; + tensor input_901_cast_fp16 = silu(x = linear_152_cast_fp16)[name = tensor("input_901_cast_fp16")]; + tensor encoder_layers_16_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_layers_16_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210766464)))]; + tensor linear_153_cast_fp16 = linear(bias = linear_2_bias_0_to_fp16, weight = encoder_layers_16_feed_forward2_linear2_weight_to_fp16, x = input_901_cast_fp16)[name = tensor("linear_153_cast_fp16")]; + tensor var_3920_to_fp16 = const()[name = tensor("op_3920_to_fp16"), val = tensor(0x1p-1)]; + tensor var_3921_cast_fp16 = mul(x = linear_153_cast_fp16, y = var_3920_to_fp16)[name = tensor("op_3921_cast_fp16")]; + tensor input_cast_fp16 = add(x = input_895_cast_fp16, y = var_3921_cast_fp16)[name = tensor("input_cast_fp16")]; + tensor audio_signal_axes_0 = const()[name = tensor("audio_signal_axes_0"), val = tensor([-1])]; + tensor encoder_layers_16_norm_out_weight_to_fp16 = const()[name = tensor("encoder_layers_16_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212863680)))]; + tensor encoder_layers_16_norm_out_bias_to_fp16 = const()[name = tensor("encoder_layers_16_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(212864768)))]; + tensor audio_signal_cast_fp16 = layer_norm(axes = audio_signal_axes_0, beta = encoder_layers_16_norm_out_bias_to_fp16, epsilon = var_38_to_fp16, gamma = encoder_layers_16_norm_out_weight_to_fp16, x = input_cast_fp16)[name = tensor("audio_signal_cast_fp16")]; + tensor obj_1_perm_0 = const()[name = tensor("obj_1_perm_0"), val = tensor([0, 2, 1])]; + tensor obj_1_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_1_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor obj_5_axis_0 = const()[name = tensor("obj_5_axis_0"), val = tensor(0)]; + tensor obj_5_cast_fp16 = stack(axis = obj_5_axis_0, values = (var_465_cast_fp16, var_672_cast_fp16, var_879_cast_fp16, var_1086_cast_fp16, var_1293_cast_fp16, var_1500_cast_fp16, var_1707_cast_fp16, var_1914_cast_fp16, var_2121_cast_fp16, var_2328_cast_fp16, var_2535_cast_fp16, var_2742_cast_fp16, var_2949_cast_fp16, var_3156_cast_fp16, var_3363_cast_fp16, var_3570_cast_fp16, cache_last_channel_cur_cast_fp16))[name = tensor("obj_5_cast_fp16")]; + tensor obj_7_axis_0 = const()[name = tensor("obj_7_axis_0"), val = tensor(0)]; + tensor obj_7_cast_fp16 = stack(axis = obj_7_axis_0, values = (var_567_cast_fp16, var_774_cast_fp16, var_981_cast_fp16, var_1188_cast_fp16, var_1395_cast_fp16, var_1602_cast_fp16, var_1809_cast_fp16, var_2016_cast_fp16, var_2223_cast_fp16, var_2430_cast_fp16, var_2637_cast_fp16, var_2844_cast_fp16, var_3051_cast_fp16, var_3258_cast_fp16, var_3465_cast_fp16, var_3672_cast_fp16, cache_last_time_cur_cast_fp16))[name = tensor("obj_7_cast_fp16")]; + tensor obj_7_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_7_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor var_3937 = add(x = cache_last_channel_len, y = cache_keep_size)[name = tensor("op_3937")]; + tensor var_3937_promoted_to_fp16_dtype_0 = const()[name = tensor("op_3937_promoted_to_fp16_dtype_0"), val = tensor("fp16")]; + tensor const_293_to_fp16 = const()[name = tensor("const_293_to_fp16"), val = tensor(-inf)]; + tensor var_46_promoted_to_fp16 = const()[name = tensor("op_46_promoted_to_fp16"), val = tensor(0x1.18p+6)]; + tensor var_3937_to_fp16 = cast(dtype = var_3937_promoted_to_fp16_dtype_0, x = var_3937)[name = tensor("cast_184")]; + tensor clip_1_cast_fp16 = clip(alpha = const_293_to_fp16, beta = var_46_promoted_to_fp16, x = var_3937_to_fp16)[name = tensor("clip_1_cast_fp16")]; + tensor var_3964_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("op_3964_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor cast_179_dtype_0 = const()[name = tensor("cast_179_dtype_0"), val = tensor("int32")]; + tensor cast_180_dtype_0 = const()[name = tensor("cast_180_dtype_0"), val = tensor("int32")]; + tensor new_cache_last_channel_len = cast(dtype = cast_180_dtype_0, x = clip_1_cast_fp16)[name = tensor("cast_181")]; + tensor encoded_length = cast(dtype = cast_179_dtype_0, x = clip_0_cast_fp16)[name = tensor("cast_182")]; + tensor new_cache_last_channel = cast(dtype = var_3964_cast_fp16_to_fp32_dtype_0, x = obj_5_cast_fp16)[name = tensor("cast_183")]; + tensor new_cache_last_time = cast(dtype = obj_7_cast_fp16_to_fp32_dtype_0, x = obj_7_cast_fp16)[name = tensor("cast_185")]; + tensor obj_1_cast_fp16 = transpose(perm = obj_1_perm_0, x = audio_signal_cast_fp16)[name = tensor("transpose_85")]; + tensor encoded_output = cast(dtype = obj_1_cast_fp16_to_fp32_dtype_0, x = obj_1_cast_fp16)[name = tensor("cast_186")]; + tensor new_pre_cache = cast(dtype = var_28_cast_fp16_to_fp32_dtype_0, x = var_28_cast_fp16)[name = tensor("cast_198")]; + } -> (encoded_output, encoded_length, new_pre_cache, new_cache_last_channel, new_cache_last_time, new_cache_last_channel_len); +} \ No newline at end of file diff --git a/320ms/streaming_encoder.mlmodelc/weights/weight.bin b/320ms/streaming_encoder.mlmodelc/weights/weight.bin new file mode 100644 index 0000000000000000000000000000000000000000..15654dd70b3be49baf2409d7d99786b2bbfd35da --- /dev/null +++ b/320ms/streaming_encoder.mlmodelc/weights/weight.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff553fcb4f4560cfce1e558e16bb9d856e4b4f6a4c08f800d5c126fce4196822 +size 212865856 diff --git a/320ms/streaming_encoder.mlpackage/Data/com.apple.CoreML/model.mlmodel b/320ms/streaming_encoder.mlpackage/Data/com.apple.CoreML/model.mlmodel new file mode 100644 index 0000000000000000000000000000000000000000..da8d55a342f8863cbab69d42b663e39fc14294aa --- /dev/null +++ b/320ms/streaming_encoder.mlpackage/Data/com.apple.CoreML/model.mlmodel @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2579ec91db1c064c91973e33c01a226d1243eaae3ae204ccc81951ad350b2698 +size 539070 diff --git a/320ms/streaming_encoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin b/320ms/streaming_encoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin new file mode 100644 index 0000000000000000000000000000000000000000..15654dd70b3be49baf2409d7d99786b2bbfd35da --- /dev/null +++ b/320ms/streaming_encoder.mlpackage/Data/com.apple.CoreML/weights/weight.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff553fcb4f4560cfce1e558e16bb9d856e4b4f6a4c08f800d5c126fce4196822 +size 212865856 diff --git a/320ms/streaming_encoder.mlpackage/Manifest.json b/320ms/streaming_encoder.mlpackage/Manifest.json new file mode 100644 index 0000000000000000000000000000000000000000..365b912e230f6f5a69bcb16eda6dd90e0f3d59e4 --- /dev/null +++ b/320ms/streaming_encoder.mlpackage/Manifest.json @@ -0,0 +1,18 @@ +{ + "fileFormatVersion": "1.0.0", + "itemInfoEntries": { + "446DB82B-CCCF-45A7-8BC1-6A0B513ADDAA": { + "author": "com.apple.CoreML", + "description": "CoreML Model Specification", + "name": "model.mlmodel", + "path": "com.apple.CoreML/model.mlmodel" + }, + "56C7CF0D-51B6-4E0A-ACAE-6A71D089B1B7": { + "author": "com.apple.CoreML", + "description": "CoreML Model Weights", + "name": "weights", + "path": "com.apple.CoreML/weights" + } + }, + "rootModelIdentifier": "446DB82B-CCCF-45A7-8BC1-6A0B513ADDAA" +} diff --git a/320ms/streaming_encoder_metadata.json b/320ms/streaming_encoder_metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..87b2d329d7e571863559397e2c855250dc5019b0 --- /dev/null +++ b/320ms/streaming_encoder_metadata.json @@ -0,0 +1,12 @@ +{ + "model_id": "nvidia/parakeet_realtime_eou_120m-v1", + "chunk_ms": 320, + "mel_frames": 64, + "pre_cache_size": 9, + "valid_out_len": 4, + "samples_per_chunk": 10240, + "hidden_dim": 512, + "num_layers": 17, + "cache_channel_size": 70, + "cache_time_size": 8 +} \ No newline at end of file diff --git a/320ms/vocab.json b/320ms/vocab.json new file mode 100644 index 0000000000000000000000000000000000000000..462081190dbb331a0dde13e99d5cfa99b302f0a7 --- /dev/null +++ b/320ms/vocab.json @@ -0,0 +1,1028 @@ +{ + "0": "", + "1": "▁t", + "2": "▁th", + "3": "▁a", + "4": "▁i", + "5": "▁the", + "6": "▁s", + "7": "re", + "8": "▁w", + "9": "▁o", + "10": "in", + "11": "at", + "12": "er", + "13": "nd", + "14": "ou", + "15": "▁c", + "16": "▁b", + "17": "▁h", + "18": "en", + "19": "on", + "20": "▁m", + "21": "▁f", + "22": "ing", + "23": "▁p", + "24": "▁to", + "25": "▁and", + "26": "▁d", + "27": "an", + "28": "or", + "29": "es", + "30": "▁y", + "31": "▁l", + "32": "▁of", + "33": "ll", + "34": "▁in", + "35": "ed", + "36": "it", + "37": "▁g", + "38": "is", + "39": "▁you", + "40": "▁n", + "41": "ar", + "42": "om", + "43": "as", + "44": "ve", + "45": "▁e", + "46": "ic", + "47": "▁it", + "48": "al", + "49": "us", + "50": "▁wh", + "51": "▁we", + "52": "▁be", + "53": "ion", + "54": "ow", + "55": "le", + "56": "▁is", + "57": "et", + "58": "ent", + "59": "ot", + "60": "ut", + "61": "▁re", + "62": "▁on", + "63": "ay", + "64": "▁ha", + "65": "ig", + "66": "▁so", + "67": "ct", + "68": "▁he", + "69": "▁for", + "70": "ver", + "71": "ke", + "72": "ro", + "73": "▁st", + "74": "id", + "75": "▁go", + "76": "all", + "77": "se", + "78": "ly", + "79": "▁u", + "80": "ch", + "81": "st", + "82": "ld", + "83": "▁k", + "84": "ce", + "85": "ur", + "86": "▁li", + "87": "am", + "88": "▁r", + "89": "ht", + "90": "▁j", + "91": "ith", + "92": "▁se", + "93": "ir", + "94": "▁as", + "95": "▁an", + "96": "im", + "97": "▁do", + "98": "ad", + "99": "▁was", + "100": "ight", + "101": "th", + "102": "▁are", + "103": "▁but", + "104": "▁sh", + "105": "ust", + "106": "ally", + "107": "▁not", + "108": "▁or", + "109": "▁com", + "110": "▁can", + "111": "▁me", + "112": "op", + "113": "▁mo", + "114": "▁at", + "115": "ill", + "116": "▁ch", + "117": "▁ne", + "118": "ant", + "119": "▁de", + "120": "▁kn", + "121": "▁one", + "122": "il", + "123": "ol", + "124": "▁con", + "125": "ter", + "126": "▁ab", + "127": "▁fr", + "128": "ere", + "129": "ck", + "130": "▁al", + "131": "▁all", + "132": "qu", + "133": "▁pro", + "134": "▁som", + "135": "ould", + "136": "▁tw", + "137": "ul", + "138": "ra", + "139": "od", + "140": "ers", + "141": "▁su", + "142": "ive", + "143": "▁v", + "144": "use", + "145": "ate", + "146": "ge", + "147": "if", + "148": "▁ex", + "149": "ess", + "150": "pp", + "151": "▁lo", + "152": "out", + "153": "▁if", + "154": "est", + "155": "ain", + "156": "ist", + "157": "and", + "158": "ea", + "159": "very", + "160": "art", + "161": "▁wor", + "162": "▁my", + "163": "ab", + "164": "ment", + "165": "▁bec", + "166": "un", + "167": "ity", + "168": "ri", + "169": "pe", + "170": "ions", + "171": "▁by", + "172": "ok", + "173": "our", + "174": "ort", + "175": "ind", + "176": "ink", + "177": "nt", + "178": "▁up", + "179": "um", + "180": "▁don", + "181": "▁get", + "182": "red", + "183": "▁out", + "184": "el", + "185": "ause", + "186": "res", + "187": "▁ma", + "188": "ich", + "189": "▁us", + "190": "rou", + "191": "▁int", + "192": "em", + "193": "os", + "194": "ies", + "195": "ie", + "196": "▁pl", + "197": "▁tr", + "198": "ven", + "199": "ous", + "200": "▁le", + "201": "▁two", + "202": "ard", + "203": "ine", + "204": "▁co", + "205": "een", + "206": "▁now", + "207": "ty", + "208": "her", + "209": "ack", + "210": "▁pe", + "211": "ame", + "212": "▁how", + "213": "▁who", + "214": "▁see", + "215": "▁tim", + "216": "ect", + "217": "ast", + "218": "▁our", + "219": "ci", + "220": "ree", + "221": "ople", + "222": "gh", + "223": "▁no", + "224": "▁had", + "225": "▁man", + "226": "▁qu", + "227": "▁en", + "228": "ide", + "229": "ure", + "230": "ud", + "231": "so", + "232": "▁his", + "233": "▁sa", + "234": "▁sp", + "235": "▁say", + "236": "ose", + "237": "ther", + "238": "▁act", + "239": "▁ta", + "240": "▁cl", + "241": "ings", + "242": "pt", + "243": "king", + "244": "▁any", + "245": "▁has", + "246": "▁un", + "247": "iv", + "248": "▁im", + "249": "▁ag", + "250": "▁te", + "251": "▁fe", + "252": "one", + "253": "per", + "254": "ong", + "255": "▁po", + "256": "▁ad", + "257": "ff", + "258": "ore", + "259": "itt", + "260": "ans", + "261": "iz", + "262": "eah", + "263": "reat", + "264": "act", + "265": "own", + "266": "hing", + "267": "enty", + "268": "age", + "269": "ber", + "270": "ice", + "271": "▁am", + "272": "ple", + "273": "are", + "274": "▁per", + "275": "und", + "276": "ite", + "277": "ix", + "278": "pl", + "279": "▁way", + "280": "▁did", + "281": "▁pr", + "282": "▁got", + "283": "ars", + "284": "▁she", + "285": "▁let", + "286": "ag", + "287": "▁ac", + "288": "int", + "289": "▁ar", + "290": "ry", + "291": "ign", + "292": "ish", + "293": "▁fir", + "294": "ace", + "295": "ble", + "296": "og", + "297": "ue", + "298": "▁ye", + "299": "ap", + "300": "iff", + "301": "▁ro", + "302": "▁her", + "303": "nder", + "304": "▁ok", + "305": "▁res", + "306": "▁gu", + "307": "ence", + "308": "▁may", + "309": "ated", + "310": "ip", + "311": "▁bo", + "312": "▁him", + "313": "way", + "314": "ac", + "315": "ical", + "316": "ass", + "317": "ase", + "318": "▁dis", + "319": "able", + "320": "ick", + "321": "▁app", + "322": "ance", + "323": "▁pre", + "324": "▁six", + "325": "▁off", + "326": "▁new", + "327": "ia", + "328": "orm", + "329": "ank", + "330": "▁lot", + "331": "ach", + "332": "▁fo", + "333": "inet", + "334": "ire", + "335": "ary", + "336": "ult", + "337": "▁tal", + "338": "▁mu", + "339": "▁bl", + "340": "ount", + "341": "sel", + "342": "vel", + "343": "▁br", + "344": "▁imp", + "345": "ep", + "346": "cess", + "347": "ord", + "348": "▁sc", + "349": "▁inc", + "350": "ound", + "351": "ang", + "352": "be", + "353": "ress", + "354": "uct", + "355": "▁ind", + "356": "▁af", + "357": "ving", + "358": "▁oh", + "359": "▁bet", + "360": "▁use", + "361": "ome", + "362": "ens", + "363": "ys", + "364": "▁bu", + "365": "co", + "366": "ory", + "367": "ater", + "368": "ild", + "369": "ght", + "370": "ial", + "371": "▁day", + "372": "ning", + "373": "na", + "374": "ile", + "375": "▁spe", + "376": "▁mar", + "377": "ody", + "378": "ough", + "379": "ade", + "380": "vers", + "381": "xt", + "382": "▁fl", + "383": "▁ke", + "384": "ian", + "385": "▁sy", + "386": "▁put", + "387": "fore", + "388": "ub", + "389": "▁ph", + "390": "fe", + "391": "▁em", + "392": "▁ser", + "393": "form", + "394": "ting", + "395": "te", + "396": "av", + "397": "ious", + "398": "▁rec", + "399": "ks", + "400": "▁gr", + "401": "ces", + "402": "wn", + "403": "ors", + "404": "▁jo", + "405": "ents", + "406": "▁des", + "407": "▁try", + "408": "▁equ", + "409": "▁z", + "410": "▁rem", + "411": "▁str", + "412": "self", + "413": "▁bit", + "414": "ph", + "415": "ved", + "416": "▁why", + "417": "▁bas", + "418": "▁hel", + "419": "▁rel", + "420": "ath", + "421": "ject", + "422": "ail", + "423": "▁la", + "424": "ual", + "425": "▁god", + "426": "▁nat", + "427": "erm", + "428": "day", + "429": "▁id", + "430": "ft", + "431": "▁wr", + "432": "▁min", + "433": "ates", + "434": "▁gen", + "435": "tain", + "436": "▁ob", + "437": "ull", + "438": "ict", + "439": "▁tra", + "440": "▁end", + "441": "▁hig", + "442": "▁fif", + "443": "oth", + "444": "tern", + "445": "▁its", + "446": "vent", + "447": "▁sm", + "448": "ons", + "449": "▁add", + "450": "iss", + "451": "▁bel", + "452": "ful", + "453": "get", + "454": "▁ele", + "455": "▁rep", + "456": "ak", + "457": "▁ho", + "458": "▁pos", + "459": "▁num", + "460": "ange", + "461": "ves", + "462": "ific", + "463": "urn", + "464": "ise", + "465": "▁cr", + "466": "▁um", + "467": "ward", + "468": "▁reg", + "469": "ady", + "470": "ower", + "471": "uc", + "472": "▁dec", + "473": "lic", + "474": "▁set", + "475": "▁gon", + "476": "▁op", + "477": "▁ear", + "478": "▁sub", + "479": "▁sl", + "480": "les", + "481": "stem", + "482": "cial", + "483": "olog", + "484": "atch", + "485": "ily", + "486": "body", + "487": "nds", + "488": "ular", + "489": "ren", + "490": "▁own", + "491": "▁too", + "492": "cent", + "493": "ible", + "494": "pect", + "495": "ered", + "496": "ways", + "497": "teen", + "498": "▁uh", + "499": "▁big", + "500": "▁mod", + "501": "▁att", + "502": "▁car", + "503": "gr", + "504": "▁acc", + "505": "ied", + "506": "mun", + "507": "ib", + "508": "▁mon", + "509": "▁sch", + "510": "▁pol", + "511": "▁dat", + "512": "▁fin", + "513": "▁sim", + "514": "▁inv", + "515": "▁def", + "516": "ked", + "517": "▁ent", + "518": "▁yes", + "519": "ows", + "520": "ics", + "521": "ited", + "522": "ute", + "523": "ism", + "524": "ps", + "525": "▁ed", + "526": "▁el", + "527": "ably", + "528": "ppen", + "529": "als", + "530": "▁ten", + "531": "ract", + "532": "ss", + "533": "▁ass", + "534": "▁met", + "535": "gan", + "536": "▁eng", + "537": "▁stu", + "538": "ween", + "539": "arch", + "540": "▁gl", + "541": "▁cor", + "542": "▁dr", + "543": "vern", + "544": "▁ty", + "545": "▁run", + "546": "hip", + "547": "cus", + "548": "cond", + "549": "▁ins", + "550": "irty", + "551": "▁pub", + "552": "lud", + "553": "llow", + "554": "▁cou", + "555": "ew", + "556": "iew", + "557": "▁sur", + "558": "ero", + "559": "ood", + "560": "ness", + "561": "▁fun", + "562": "▁eff", + "563": "cept", + "564": "▁ca", + "565": "▁exp", + "566": "duct", + "567": "▁sw", + "568": "ize", + "569": "ope", + "570": "▁par", + "571": "kes", + "572": "cy", + "573": "▁ev", + "574": "▁ref", + "575": "ell", + "576": "▁bus", + "577": "ug", + "578": "rib", + "579": "▁cur", + "580": "mo", + "581": "ock", + "582": "ures", + "583": "air", + "584": "▁war", + "585": "str", + "586": "▁med", + "587": "▁wa", + "588": "▁val", + "589": "▁sin", + "590": "blem", + "591": "▁fam", + "592": "li", + "593": "▁far", + "594": "▁cle", + "595": "▁col", + "596": "mon", + "597": "▁gra", + "598": "led", + "599": "ense", + "600": "tin", + "601": "ues", + "602": "its", + "603": "▁mem", + "604": "▁inf", + "605": "▁eas", + "606": "ideo", + "607": "▁top", + "608": "io", + "609": "pan", + "610": "▁hum", + "611": "▁old", + "612": "ead", + "613": "▁ord", + "614": "ric", + "615": "ants", + "616": "oy", + "617": "esn", + "618": "uck", + "619": "ason", + "620": "ced", + "621": "ool", + "622": "rat", + "623": "ouse", + "624": "▁lar", + "625": "▁art", + "626": "▁wee", + "627": "▁cer", + "628": "ized", + "629": "▁mat", + "630": "con", + "631": "erg", + "632": "land", + "633": "ines", + "634": "▁chr", + "635": "▁aut", + "636": "▁lea", + "637": "▁sou", + "638": "oney", + "639": "tty", + "640": "▁ple", + "641": "ulat", + "642": "oks", + "643": "▁few", + "644": "▁sol", + "645": "▁che", + "646": "chn", + "647": "ird", + "648": "▁bre", + "649": "▁dur", + "650": "▁wom", + "651": "me", + "652": "izat", + "653": "eric", + "654": "ote", + "655": "▁uni", + "656": "eren", + "657": "arn", + "658": "ross", + "659": "ices", + "660": "ten", + "661": "eral", + "662": "ever", + "663": "ieve", + "664": "lish", + "665": "ash", + "666": "▁opp", + "667": "alth", + "668": "ger", + "669": "▁sk", + "670": "▁red", + "671": "peri", + "672": "▁det", + "673": "▁ext", + "674": "ner", + "675": "ah", + "676": "▁var", + "677": "▁loc", + "678": "gram", + "679": "ists", + "680": "ives", + "681": "▁es", + "682": "▁nor", + "683": "tro", + "684": "ale", + "685": "▁iss", + "686": "▁pri", + "687": "gin", + "688": "az", + "689": "oc", + "690": "▁pop", + "691": "ern", + "692": "▁sit", + "693": "ket", + "694": "▁pa", + "695": "▁law", + "696": "ages", + "697": "br", + "698": "▁cam", + "699": "▁mom", + "700": "osed", + "701": "▁bro", + "702": "ne", + "703": "bs", + "704": "▁cre", + "705": "erat", + "706": "▁sec", + "707": "▁cap", + "708": "▁vis", + "709": "▁pat", + "710": "ield", + "711": "iet", + "712": "▁tri", + "713": "up", + "714": "▁bra", + "715": "ts", + "716": "▁mot", + "717": "▁unt", + "718": "put", + "719": "bo", + "720": "ork", + "721": "mer", + "722": "ital", + "723": "▁air", + "724": "ined", + "725": "▁beh", + "726": "▁adv", + "727": "▁ret", + "728": "imes", + "729": "▁tea", + "730": "ural", + "731": "sid", + "732": "ters", + "733": "▁pur", + "734": "▁sci", + "735": "bers", + "736": "ient", + "737": "ier", + "738": "cc", + "739": "sw", + "740": "▁av", + "741": "reen", + "742": "ode", + "743": "ont", + "744": "▁dra", + "745": "ann", + "746": "nect", + "747": "▁x", + "748": "▁eu", + "749": "ton", + "750": "inat", + "751": "ene", + "752": "ared", + "753": "els", + "754": "▁mor", + "755": "▁rat", + "756": "cri", + "757": "▁men", + "758": "▁ah", + "759": "ames", + "760": "▁arm", + "761": "eak", + "762": "▁pay", + "763": "▁hal", + "764": "ins", + "765": "ilit", + "766": "stit", + "767": "▁ra", + "768": "▁leg", + "769": "cl", + "770": "pr", + "771": "▁wal", + "772": "▁bad", + "773": "▁ge", + "774": "roup", + "775": "▁mus", + "776": "man", + "777": "▁gi", + "778": "eds", + "779": "▁aw", + "780": "po", + "781": "ark", + "782": "row", + "783": "▁dep", + "784": "ully", + "785": "ral", + "786": "lect", + "787": "pend", + "788": "▁sev", + "789": "ime", + "790": "gest", + "791": "here", + "792": "▁yet", + "793": "ted", + "794": "▁rev", + "795": "ds", + "796": "▁ask", + "797": "less", + "798": "▁di", + "799": "ets", + "800": "line", + "801": "▁aff", + "802": "ired", + "803": "▁est", + "804": "ken", + "805": "vid", + "806": "most", + "807": "ivid", + "808": "unch", + "809": "par", + "810": "med", + "811": "rop", + "812": "ased", + "813": "eone", + "814": "▁ve", + "815": "▁abs", + "816": "ergy", + "817": "ret", + "818": "▁saw", + "819": "▁ey", + "820": "▁cal", + "821": "uat", + "822": "▁mid", + "823": "vat", + "824": "ream", + "825": "vice", + "826": "ians", + "827": "rent", + "828": "ctor", + "829": "err", + "830": "ush", + "831": "ases", + "832": "▁suc", + "833": "erms", + "834": "ave", + "835": "angu", + "836": "ries", + "837": "▁wo", + "838": "arts", + "839": "▁fil", + "840": "▁fat", + "841": "▁cho", + "842": "orts", + "843": "▁fre", + "844": "ee", + "845": "ught", + "846": "eng", + "847": "ump", + "848": "▁bar", + "849": "ying", + "850": "ane", + "851": "▁tem", + "852": "anks", + "853": "ury", + "854": "iat", + "855": "mit", + "856": "trol", + "857": "▁net", + "858": "▁maj", + "859": "▁cra", + "860": "ling", + "861": "▁fig", + "862": "orn", + "863": "icat", + "864": "pany", + "865": "▁occ", + "866": "ott", + "867": "ands", + "868": "▁exc", + "869": "▁mr", + "870": "ency", + "871": "rope", + "872": "itch", + "873": "▁lit", + "874": "abil", + "875": "not", + "876": "ma", + "877": "▁typ", + "878": "▁opt", + "879": "ob", + "880": "ser", + "881": "ety", + "882": "ms", + "883": "peci", + "884": "aces", + "885": "aut", + "886": "▁hon", + "887": "cuss", + "888": "▁sal", + "889": "▁sor", + "890": "att", + "891": "▁lab", + "892": "▁har", + "893": "urch", + "894": "nded", + "895": "uce", + "896": "ids", + "897": "▁hy", + "898": "▁fut", + "899": "▁ste", + "900": "ours", + "901": "ems", + "902": "utes", + "903": "ng", + "904": "ta", + "905": "▁won", + "906": "▁fa", + "907": "▁env", + "908": "ards", + "909": "▁job", + "910": "ium", + "911": "▁dot", + "912": "▁obv", + "913": "ina", + "914": "side", + "915": "elve", + "916": "cu", + "917": "▁jes", + "918": "▁pot", + "919": "▁pie", + "920": "▁tre", + "921": "▁hey", + "922": "▁mag", + "923": "ron", + "924": "▁key", + "925": "swer", + "926": "▁win", + "927": "ucat", + "928": "work", + "929": "ides", + "930": "▁low", + "931": "▁vol", + "932": "▁oth", + "933": "atic", + "934": "lf", + "935": "ads", + "936": "inds", + "937": "com", + "938": "ths", + "939": "▁ver", + "940": "ised", + "941": "lo", + "942": "▁squ", + "943": "▁cut", + "944": "oked", + "945": "irit", + "946": "ateg", + "947": "ppy", + "948": "mitt", + "949": "come", + "950": "hn", + "951": "igin", + "952": "mand", + "953": "▁dam", + "954": "ho", + "955": "▁da", + "956": "▁fur", + "957": "iron", + "958": "ilar", + "959": "▁fac", + "960": "▁neg", + "961": "▁ago", + "962": "ged", + "963": "miss", + "964": "enth", + "965": "▁dou", + "966": "▁hit", + "967": "▁guy", + "968": "▁bi", + "969": "ove", + "970": "fess", + "971": "ples", + "972": "owed", + "973": "ured", + "974": "▁ris", + "975": "ints", + "976": "rew", + "977": "▁sum", + "978": "▁hu", + "979": "ploy", + "980": "ude", + "981": "ried", + "982": "▁cir", + "983": "▁dev", + "984": "ear", + "985": "▁tot", + "986": "▁ann", + "987": "duc", + "988": "ik", + "989": "pon", + "990": "sted", + "991": "▁ide", + "992": "▁'", + "993": "ipp", + "994": "▁eat", + "995": "▁dom", + "996": "▁", + "997": "e", + "998": "t", + "999": "o", + "1000": "a", + "1001": "i", + "1002": "n", + "1003": "s", + "1004": "r", + "1005": "h", + "1006": "l", + "1007": "d", + "1008": "u", + "1009": "c", + "1010": "m", + "1011": "y", + "1012": "g", + "1013": "w", + "1014": "f", + "1015": "p", + "1016": "b", + "1017": "v", + "1018": "k", + "1019": "'", + "1020": "j", + "1021": "x", + "1022": "q", + "1023": "z", + "1024": "", + "1025": "" +} \ No newline at end of file diff --git a/README.md b/README.md new file mode 100644 index 0000000000000000000000000000000000000000..c581f781ee4e6d4379c815b0aecb326ffabd858d --- /dev/null +++ b/README.md @@ -0,0 +1,117 @@ +--- +license: other +license_name: nvidia-open-model-license +license_link: >- + https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ +language: +- en +metrics: +- wer +library_name: nemo +tags: +- speech-recognition +- FastConformer +- end-of-utterance +- voice agent +pipeline_tag: automatic-speech-recognition +base_model: +- nvidia/parakeet_realtime_eou_120m-v1 +base_model_relation: finetune +--- + + +# Parakeet Realtime EOU 120M — CoreML + +CoreML conversion of [nvidia/parakeet-realtime-eou-120m-v1](https://huggingface.co/nvidia/parakee +t-realtime-eou-120m-v1) for streaming speech recognition with end-of-utterance detection on Apple + Silicon. + +Used by [FluidAudio](https://github.com/FluidInference/FluidAudio) for real-time transcription. + +## Models + +The RNNT pipeline is split into three CoreML models, exported at two chunk sizes: + +| Model | Description | +|-------|-------------| +| `streaming_encoder.mlmodelc` | FastConformer encoder with loopback state caching | +| `decoder.mlmodelc` | 1-layer LSTM decoder (640 hidden units) | +| `joint_decision.mlmodelc` | Joint network for token prediction + EOU detection | + +### Chunk Size Variants + +| Variant | Latency | WER (test-clean) | RTFx (M2) | +|---------|---------|-------------------|------------| +| `160ms/` | 160ms | 8.29% | 4.78x | +| `320ms/` | 320ms | 4.87% | 12.48x | + +Benchmarked on LibriSpeech test-clean (2620 files, 5.40h audio) on Apple M2. + +## Usage with FluidAudio + +```swift +import FluidAudio + +let manager = StreamingEouAsrManager() +await manager.initialize() + +// Transcribe with EOU detection +await manager.startStreaming( + eouCallback: { transcript in + print("Utterance complete: \(transcript)") + }, + partialCallback: { partial in + print("Partial: \(partial)") + } +) + +// Feed audio chunks as they arrive +await manager.feedAudio(samples) +``` + +CLI + +# Transcribe a file +swift run fluidaudio parakeet-eou --input audio.wav + +# Benchmark +swift run -c release fluidaudio parakeet-eou --benchmark --chunk-size 320 + +Architecture + +120M parameter RNNT (Recurrent Neural Network Transducer) with: +- Encoder: 17-layer FastConformer with cache-aware streaming +- Decoder: 1-layer LSTM, 640 hidden size +- Joint: Linear projection with 1027 output classes (1024 tokens + EOU token + SOS + blank) +- EOU token: ID 1024 signals end-of-utterance + +Streaming State + +The encoder maintains loopback state between chunks: +┌─────────────────────┬──────────────────┬───────────────────────┐ +│ State │ Shape │ Description │ +├─────────────────────┼──────────────────┼───────────────────────┤ +│ preCache │ [1, 128, N] │ Mel-level context │ +├─────────────────────┼──────────────────┼───────────────────────┤ +│ cacheLastChannel │ [17, 1, 70, 512] │ Conformer layer cache │ +├─────────────────────┼──────────────────┼───────────────────────┤ +│ cacheLastTime │ [17, 1, 512, 8] │ Temporal cache │ +├─────────────────────┼──────────────────┼───────────────────────┤ +│ cacheLastChannelLen │ [1] │ Cache length tracking │ +└─────────────────────┴──────────────────┴───────────────────────┘ +Export + +Converted from PyTorch using coremltools. To re-export: + +python3 Scripts/ParakeetEOU/Conversion/convert_split_encoder.py \ + --output-dir Models/ParakeetEOU \ + --model-id nvidia/parakeet-realtime-eou-120m-v1 + +License + +NVIDIA Open Model License — see +https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/. + +Original model: https://huggingface.co/nvidia/parakeet-realtime-eou-120m-v1 + +--- \ No newline at end of file diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..9e26dfeeb6e641a33dae4961196235bdb965b21b --- /dev/null +++ b/config.json @@ -0,0 +1 @@ +{} \ No newline at end of file diff --git a/convert_parakeet_eou.py b/convert_parakeet_eou.py new file mode 100644 index 0000000000000000000000000000000000000000..9c865f90c8132840b5755dcd62fcfac15e3f0352 --- /dev/null +++ b/convert_parakeet_eou.py @@ -0,0 +1,740 @@ +#!/usr/bin/env python3 +"""CLI for exporting Parakeet Realtime EOU 120M components to CoreML. + +This model is a cache-aware streaming FastConformer-RNNT model optimized for +low-latency speech recognition with end-of-utterance detection. + +Key differences from Parakeet TDT v3: +- Smaller model (120M vs 600M params) +- No duration outputs (standard RNNT, not TDT) +- Cache-aware streaming encoder (17 layers, attention context [70,1]) +- Special token for end-of-utterance detection +- Optimized for 80-160ms latency + +Reference: https://huggingface.co/nvidia/parakeet_realtime_eou_120m-v1 +""" +from __future__ import annotations + +import json +from dataclasses import asdict +from pathlib import Path +from typing import Dict, Optional, Tuple + +import coremltools as ct +import numpy as np +import soundfile as sf +import torch +import typer + +import nemo.collections.asr as nemo_asr + +from individual_components import ( + DecoderWrapper, + EncoderWrapper, + ExportSettings, + JointWrapper, + JointDecisionWrapper, + JointDecisionSingleStep, + PreprocessorWrapper, + MelEncoderWrapper, + _coreml_convert, +) + +def apply_stft_patch(): + # Monkey patch coremltools.stft to handle extra arguments from newer torch versions + try: + import coremltools.converters.mil.frontend.torch.ops as torch_ops + _original_stft = torch_ops.stft + + def patched_stft(context, node): + if len(node.inputs) > 8: + node.inputs = node.inputs[:8] + return _original_stft(context, node) + + torch_ops.stft = patched_stft + if "stft" in torch_ops._TORCH_OPS_REGISTRY: + torch_ops._TORCH_OPS_REGISTRY["stft"] = patched_stft + print("Monkey patched coremltools.stft for compatibility.") + except Exception as e: + print(f"Warning: Could not monkey patch stft: {e}") + +DEFAULT_MODEL_ID = "nvidia/parakeet_realtime_eou_120m-v1" +AUTHOR = "Fluid Inference" + + +def _compute_length(seconds: float, sample_rate: int) -> int: + return int(round(seconds * sample_rate)) + + +def _prepare_audio( + validation_audio: Optional[Path], + sample_rate: int, + max_samples: int, + seed: Optional[int], +) -> torch.Tensor: + if validation_audio is None: + if seed is not None: + torch.manual_seed(seed) + audio = torch.randn(1, max_samples, dtype=torch.float32) + return audio + + data, sr = sf.read(str(validation_audio), dtype="float32") + if sr != sample_rate: + raise typer.BadParameter( + f"Validation audio sample rate {sr} does not match model rate {sample_rate}" + ) + + if data.ndim > 1: + data = data[:, 0] + + if data.size == 0: + raise typer.BadParameter("Validation audio is empty") + + if data.size < max_samples: + pad_width = max_samples - data.size + data = np.pad(data, (0, pad_width)) + elif data.size > max_samples: + data = data[:max_samples] + + audio = torch.from_numpy(data).unsqueeze(0).to(dtype=torch.float32) + return audio + + +def _save_mlpackage(model: ct.models.MLModel, path: Path, description: str) -> None: + try: + model.minimum_deployment_target = ct.target.iOS17 + except Exception: + pass + model.short_description = description + model.author = AUTHOR + path.parent.mkdir(parents=True, exist_ok=True) + model.save(str(path)) + + +def _tensor_shape(tensor: torch.Tensor) -> Tuple[int, ...]: + return tuple(int(dim) for dim in tensor.shape) + + +def _parse_compute_units(name: str) -> ct.ComputeUnit: + """Parse a human-friendly compute units string into ct.ComputeUnit.""" + normalized = str(name).strip().upper() + mapping = { + "ALL": ct.ComputeUnit.ALL, + "CPU_ONLY": ct.ComputeUnit.CPU_ONLY, + "CPU_AND_GPU": ct.ComputeUnit.CPU_AND_GPU, + "CPU_AND_NE": ct.ComputeUnit.CPU_AND_NE, + "CPU_AND_NEURALENGINE": ct.ComputeUnit.CPU_AND_NE, + } + if normalized not in mapping: + raise typer.BadParameter( + f"Unknown compute units '{name}'. Choose from: " + ", ".join(mapping.keys()) + ) + return mapping[normalized] + + +def _parse_compute_precision(name: Optional[str]) -> Optional[ct.precision]: + """Parse compute precision string into ct.precision or None.""" + if name is None: + return None + normalized = str(name).strip().upper() + if normalized == "": + return None + mapping = { + "FLOAT32": ct.precision.FLOAT32, + "FLOAT16": ct.precision.FLOAT16, + } + if normalized not in mapping: + raise typer.BadParameter( + f"Unknown compute precision '{name}'. Choose from: " + + ", ".join(mapping.keys()) + ) + return mapping[normalized] + + +app = typer.Typer(add_completion=False, pretty_exceptions_show_locals=False) + + +@app.command() +def convert( + nemo_path: Optional[Path] = typer.Option( + None, + "--nemo-path", + exists=True, + resolve_path=True, + help="Path to parakeet_realtime_eou_120m-v1.nemo checkpoint (skip to auto-download)", + ), + model_id: str = typer.Option( + DEFAULT_MODEL_ID, + "--model-id", + help="Model identifier to download when --nemo-path is omitted", + ), + output_dir: Path = typer.Option( + Path("parakeet_eou_coreml"), + help="Directory where mlpackages and metadata will be written", + ), + preprocessor_cu: str = typer.Option( + "CPU_ONLY", + "--preprocessor-cu", + help="Compute units for preprocessor (default CPU_ONLY)", + ), + mel_encoder_cu: str = typer.Option( + "CPU_ONLY", + "--mel-encoder-cu", + help="Compute units for fused mel+encoder (default CPU_ONLY)", + ), + compute_precision: Optional[str] = typer.Option( + None, + "--compute-precision", + help="Export precision: FLOAT32 (default) or FLOAT16 to shrink non-quantized weights.", + ), + max_audio_seconds: float = typer.Option( + 15.0, + "--max-audio-seconds", + help="Maximum audio duration in seconds for the fixed window export", + ), + validation_audio: Optional[Path] = typer.Option( + None, + "--validation-audio", + exists=True, + resolve_path=True, + help="Path to a 16kHz WAV file for tracing (uses random if not provided)", + ), +) -> None: + """Export all Parakeet Realtime EOU sub-modules to CoreML. + + This exports the cache-aware streaming FastConformer-RNNT model for + low-latency speech recognition with end-of-utterance detection. + """ + export_settings = ExportSettings( + output_dir=output_dir, + compute_units=ct.ComputeUnit.CPU_ONLY, + deployment_target=ct.target.iOS17, + compute_precision=_parse_compute_precision(compute_precision), + max_audio_seconds=max_audio_seconds, + max_symbol_steps=1, + ) + + typer.echo("Export configuration:") + typer.echo(asdict(export_settings)) + + output_dir.mkdir(parents=True, exist_ok=True) + pre_cu = _parse_compute_units(preprocessor_cu) + melenc_cu = _parse_compute_units(mel_encoder_cu) + + if nemo_path is not None: + typer.echo(f"Loading NeMo model from {nemo_path}…") + # Try loading as generic ASRModel first, then specific class + try: + asr_model = nemo_asr.models.ASRModel.restore_from( + str(nemo_path), map_location="cpu" + ) + except Exception: + # Fallback to EncDecRNNTBPEModel + asr_model = nemo_asr.models.EncDecRNNTBPEModel.restore_from( + str(nemo_path), map_location="cpu" + ) + checkpoint_meta = { + "type": "file", + "path": str(nemo_path), + } + else: + typer.echo(f"Downloading NeMo model via {model_id}…") + # Use ASRModel.from_pretrained as recommended for this model + try: + asr_model = nemo_asr.models.ASRModel.from_pretrained( + model_id, map_location="cpu" + ) + except Exception: + # Fallback to EncDecRNNTBPEModel + asr_model = nemo_asr.models.EncDecRNNTBPEModel.from_pretrained( + model_id, map_location="cpu" + ) + checkpoint_meta = { + "type": "pretrained", + "model_id": model_id, + } + asr_model.eval() + + # Print model info + typer.echo(f"Model class: {type(asr_model).__name__}") + typer.echo(f"Encoder class: {type(asr_model.encoder).__name__}") + + sample_rate = int(asr_model.cfg.preprocessor.sample_rate) + max_samples = _compute_length(export_settings.max_audio_seconds, sample_rate) + + # Prepare audio for tracing + if validation_audio is not None: + typer.echo(f"Using validation audio: {validation_audio}") + audio_tensor = _prepare_audio(validation_audio, sample_rate, max_samples, seed=None) + else: + typer.echo("Using random audio for tracing (seed=42)") + audio_tensor = _prepare_audio(None, sample_rate, max_samples, seed=42) + + audio_length = torch.tensor([max_samples], dtype=torch.int32) + + preprocessor = PreprocessorWrapper(asr_model.preprocessor.eval()) + encoder = EncoderWrapper(asr_model.encoder.eval()) + decoder = DecoderWrapper(asr_model.decoder.eval()) + joint = JointWrapper(asr_model.joint.eval()) + + decoder_export_flag = getattr(asr_model.decoder, "_rnnt_export", False) + asr_model.decoder._rnnt_export = True + + try: + with torch.no_grad(): + mel_ref, mel_length_ref = preprocessor(audio_tensor, audio_length) + mel_length_ref = mel_length_ref.to(dtype=torch.int32) + encoder_ref, encoder_length_ref, frame_times_ref = encoder( + mel_ref, mel_length_ref + ) + encoder_length_ref = encoder_length_ref.to(dtype=torch.int32) + + # Clone tensors to drop inference flags + mel_ref = mel_ref.clone().detach() + mel_length_ref = mel_length_ref.clone().detach() + encoder_ref = encoder_ref.clone().detach() + encoder_length_ref = encoder_length_ref.clone().detach() + frame_times_ref = frame_times_ref.clone().detach() + + vocab_size = int(asr_model.tokenizer.vocab_size) + decoder_hidden = int(asr_model.decoder.pred_hidden) + decoder_layers = int(asr_model.decoder.pred_rnn_layers) + + # Check if model has extra outputs (TDT-style duration) + num_extra = getattr(asr_model.joint, "num_extra_outputs", 0) + typer.echo(f"Vocab size: {vocab_size}, num_extra_outputs: {num_extra}") + + targets = torch.full( + (1, export_settings.max_symbol_steps), + fill_value=asr_model.decoder.blank_idx, + dtype=torch.int32, + ) + target_lengths = torch.tensor( + [export_settings.max_symbol_steps], dtype=torch.int32 + ) + zero_state = torch.zeros( + decoder_layers, + 1, + decoder_hidden, + dtype=torch.float32, + ) + + with torch.no_grad(): + decoder_ref, h_ref, c_ref = decoder( + targets, target_lengths, zero_state, zero_state + ) + joint_ref = joint(encoder_ref, decoder_ref) + + decoder_ref = decoder_ref.clone() + h_ref = h_ref.clone() + c_ref = c_ref.clone() + joint_ref = joint_ref.clone() + + typer.echo(f"Encoder output shape: {encoder_ref.shape}") + typer.echo(f"Decoder output shape: {decoder_ref.shape}") + typer.echo(f"Joint output shape: {joint_ref.shape}") + + # === Export Preprocessor === + typer.echo("Tracing and converting preprocessor…") + preprocessor = preprocessor.cpu() + audio_tensor = audio_tensor.cpu() + audio_length = audio_length.cpu() + traced_preprocessor = torch.jit.trace( + preprocessor, (audio_tensor, audio_length), strict=False + ) + traced_preprocessor.eval() + preprocessor_inputs = [ + ct.TensorType( + name="audio_signal", + shape=(1, ct.RangeDim(1, max_samples)), + dtype=np.float32, + ), + ct.TensorType(name="audio_length", shape=(1,), dtype=np.int32), + ] + preprocessor_outputs = [ + ct.TensorType(name="mel", dtype=np.float32), + ct.TensorType(name="mel_length", dtype=np.int32), + ] + preprocessor_model = _coreml_convert( + traced_preprocessor, + preprocessor_inputs, + preprocessor_outputs, + export_settings, + compute_units_override=pre_cu, + ) + preprocessor_path = output_dir / "parakeet_eou_preprocessor.mlpackage" + _save_mlpackage( + preprocessor_model, + preprocessor_path, + f"Parakeet EOU preprocessor ({max_audio_seconds}s window)", + ) + + # === Export Encoder === + typer.echo("Tracing and converting encoder…") + traced_encoder = torch.jit.trace( + encoder, (mel_ref, mel_length_ref), strict=False + ) + traced_encoder.eval() + encoder_inputs = [ + ct.TensorType( + name="mel", shape=_tensor_shape(mel_ref), dtype=np.float32 + ), + ct.TensorType(name="mel_length", shape=(1,), dtype=np.int32), + ] + encoder_outputs = [ + ct.TensorType(name="encoder", dtype=np.float32), + ct.TensorType(name="encoder_length", dtype=np.int32), + ct.TensorType(name="frame_times", dtype=np.float32), + ] + encoder_model = _coreml_convert( + traced_encoder, + encoder_inputs, + encoder_outputs, + export_settings, + compute_units_override=ct.ComputeUnit.CPU_ONLY, + ) + encoder_path = output_dir / "parakeet_eou_encoder.mlpackage" + _save_mlpackage( + encoder_model, + encoder_path, + f"Parakeet EOU encoder ({max_audio_seconds}s window)", + ) + + # === Export Fused Mel+Encoder === + typer.echo("Tracing and converting fused mel+encoder…") + mel_encoder = MelEncoderWrapper(preprocessor, encoder) + traced_mel_encoder = torch.jit.trace( + mel_encoder, (audio_tensor, audio_length), strict=False + ) + traced_mel_encoder.eval() + mel_encoder_inputs = [ + ct.TensorType( + name="audio_signal", shape=(1, max_samples), dtype=np.float32 + ), + ct.TensorType(name="audio_length", shape=(1,), dtype=np.int32), + ] + mel_encoder_outputs = [ + ct.TensorType(name="encoder", dtype=np.float32), + ct.TensorType(name="encoder_length", dtype=np.int32), + ct.TensorType(name="frame_times", dtype=np.float32), + ] + mel_encoder_model = _coreml_convert( + traced_mel_encoder, + mel_encoder_inputs, + mel_encoder_outputs, + export_settings, + compute_units_override=melenc_cu, + ) + mel_encoder_path = output_dir / "parakeet_eou_mel_encoder.mlpackage" + _save_mlpackage( + mel_encoder_model, + mel_encoder_path, + f"Parakeet EOU fused Mel+Encoder ({max_audio_seconds}s window)", + ) + + # === Export Decoder === + typer.echo("Tracing and converting decoder…") + traced_decoder = torch.jit.trace( + decoder, + (targets, target_lengths, zero_state, zero_state), + strict=False, + ) + traced_decoder.eval() + decoder_inputs = [ + ct.TensorType( + name="targets", shape=_tensor_shape(targets), dtype=np.int32 + ), + ct.TensorType(name="target_length", shape=(1,), dtype=np.int32), + ct.TensorType( + name="h_in", shape=_tensor_shape(zero_state), dtype=np.float32 + ), + ct.TensorType( + name="c_in", shape=_tensor_shape(zero_state), dtype=np.float32 + ), + ] + decoder_outputs = [ + ct.TensorType(name="decoder", dtype=np.float32), + ct.TensorType(name="h_out", dtype=np.float32), + ct.TensorType(name="c_out", dtype=np.float32), + ] + decoder_model = _coreml_convert( + traced_decoder, + decoder_inputs, + decoder_outputs, + export_settings, + compute_units_override=ct.ComputeUnit.CPU_ONLY, + ) + decoder_path = output_dir / "parakeet_eou_decoder.mlpackage" + _save_mlpackage( + decoder_model, + decoder_path, + "Parakeet EOU decoder (RNNT prediction network)", + ) + + # === Export Joint === + typer.echo("Tracing and converting joint…") + traced_joint = torch.jit.trace( + joint, + (encoder_ref, decoder_ref), + strict=False, + ) + traced_joint.eval() + joint_inputs = [ + ct.TensorType( + name="encoder", shape=_tensor_shape(encoder_ref), dtype=np.float32 + ), + ct.TensorType( + name="decoder", shape=_tensor_shape(decoder_ref), dtype=np.float32 + ), + ] + joint_outputs = [ + ct.TensorType(name="logits", dtype=np.float32), + ] + joint_model = _coreml_convert( + traced_joint, + joint_inputs, + joint_outputs, + export_settings, + compute_units_override=ct.ComputeUnit.CPU_ONLY, + ) + joint_path = output_dir / "parakeet_eou_joint.mlpackage" + _save_mlpackage( + joint_model, + joint_path, + "Parakeet EOU joint network (RNNT)", + ) + + # === Export Joint Decision Head === + typer.echo("Tracing and converting joint decision head…") + joint_decision = JointDecisionWrapper(joint, vocab_size=vocab_size) + traced_joint_decision = torch.jit.trace( + joint_decision, + (encoder_ref, decoder_ref), + strict=False, + ) + traced_joint_decision.eval() + joint_decision_inputs = [ + ct.TensorType( + name="encoder", shape=_tensor_shape(encoder_ref), dtype=np.float32 + ), + ct.TensorType( + name="decoder", shape=_tensor_shape(decoder_ref), dtype=np.float32 + ), + ] + joint_decision_outputs = [ + ct.TensorType(name="token_id", dtype=np.int32), + ct.TensorType(name="token_prob", dtype=np.float32), + ] + joint_decision_model = _coreml_convert( + traced_joint_decision, + joint_decision_inputs, + joint_decision_outputs, + export_settings, + compute_units_override=ct.ComputeUnit.CPU_ONLY, + ) + joint_decision_path = output_dir / "parakeet_eou_joint_decision.mlpackage" + _save_mlpackage( + joint_decision_model, + joint_decision_path, + "Parakeet EOU joint + decision head (softmax, argmax)", + ) + + # === Export Single-Step Joint Decision === + typer.echo("Tracing and converting single-step joint decision…") + jd_single = JointDecisionSingleStep(joint, vocab_size=vocab_size) + # Create single-step slices from refs + enc_step = encoder_ref[:, :, :1].contiguous() + dec_step = decoder_ref[:, :, :1].contiguous() + traced_jd_single = torch.jit.trace( + jd_single, + (enc_step, dec_step), + strict=False, + ) + traced_jd_single.eval() + jd_single_inputs = [ + ct.TensorType( + name="encoder_step", + shape=(1, enc_step.shape[1], 1), + dtype=np.float32, + ), + ct.TensorType( + name="decoder_step", + shape=(1, dec_step.shape[1], 1), + dtype=np.float32, + ), + ] + jd_single_outputs = [ + ct.TensorType(name="token_id", dtype=np.int32), + ct.TensorType(name="token_prob", dtype=np.float32), + ct.TensorType(name="top_k_ids", dtype=np.int32), + ct.TensorType(name="top_k_logits", dtype=np.float32), + ] + jd_single_model = _coreml_convert( + traced_jd_single, + jd_single_inputs, + jd_single_outputs, + export_settings, + compute_units_override=ct.ComputeUnit.CPU_ONLY, + ) + jd_single_path = output_dir / "parakeet_eou_joint_decision_single_step.mlpackage" + _save_mlpackage( + jd_single_model, + jd_single_path, + "Parakeet EOU single-step joint decision (current frame)", + ) + + # === Save Metadata === + metadata: Dict[str, object] = { + "model_id": model_id, + "model_name": "parakeet_realtime_eou_120m-v1", + "model_class": type(asr_model).__name__, + "encoder_class": type(asr_model.encoder).__name__, + "sample_rate": sample_rate, + "max_audio_seconds": export_settings.max_audio_seconds, + "max_audio_samples": max_samples, + "max_symbol_steps": export_settings.max_symbol_steps, + "vocab_size": vocab_size, + "vocab_with_blank": vocab_size + 1, + "decoder_hidden": decoder_hidden, + "decoder_layers": decoder_layers, + "num_extra_outputs": num_extra, + "has_eou_token": True, + "checkpoint": checkpoint_meta, + "coreml": { + "compute_units": export_settings.compute_units.name, + "compute_precision": ( + export_settings.compute_precision.name + if export_settings.compute_precision is not None + else "FLOAT32" + ), + }, + "components": { + "preprocessor": { + "inputs": { + "audio_signal": [1, max_samples], + "audio_length": [1], + }, + "outputs": { + "mel": list(_tensor_shape(mel_ref)), + "mel_length": [1], + }, + "path": preprocessor_path.name, + }, + "encoder": { + "inputs": { + "mel": list(_tensor_shape(mel_ref)), + "mel_length": [1], + }, + "outputs": { + "encoder": list(_tensor_shape(encoder_ref)), + "encoder_length": [1], + "frame_times": [1, _tensor_shape(encoder_ref)[2]], + }, + "path": encoder_path.name, + }, + "mel_encoder": { + "inputs": { + "audio_signal": [1, max_samples], + "audio_length": [1], + }, + "outputs": { + "encoder": list(_tensor_shape(encoder_ref)), + "encoder_length": [1], + "frame_times": [1, _tensor_shape(encoder_ref)[2]], + }, + "path": mel_encoder_path.name, + }, + "decoder": { + "inputs": { + "targets": list(_tensor_shape(targets)), + "target_length": [1], + "h_in": list(_tensor_shape(zero_state)), + "c_in": list(_tensor_shape(zero_state)), + }, + "outputs": { + "decoder": list(_tensor_shape(decoder_ref)), + "h_out": list(_tensor_shape(h_ref)), + "c_out": list(_tensor_shape(c_ref)), + }, + "path": decoder_path.name, + }, + "joint": { + "inputs": { + "encoder": list(_tensor_shape(encoder_ref)), + "decoder": list(_tensor_shape(decoder_ref)), + }, + "outputs": { + "logits": list(_tensor_shape(joint_ref)), + }, + "path": joint_path.name, + }, + "joint_decision": { + "inputs": { + "encoder": list(_tensor_shape(encoder_ref)), + "decoder": list(_tensor_shape(decoder_ref)), + }, + "outputs": { + "token_id": [ + _tensor_shape(encoder_ref)[0], + _tensor_shape(encoder_ref)[2], + _tensor_shape(decoder_ref)[2], + ], + "token_prob": [ + _tensor_shape(encoder_ref)[0], + _tensor_shape(encoder_ref)[2], + _tensor_shape(decoder_ref)[2], + ], + }, + "path": joint_decision_path.name, + }, + "joint_decision_single_step": { + "inputs": { + "encoder_step": [1, _tensor_shape(encoder_ref)[1], 1], + "decoder_step": [1, _tensor_shape(decoder_ref)[1], 1], + }, + "outputs": { + "token_id": [1, 1, 1], + "token_prob": [1, 1, 1], + "top_k_ids": [1, 1, 1, 64], + "top_k_logits": [1, 1, 1, 64], + }, + "path": jd_single_path.name, + }, + }, + } + + # Export tokenizer vocab if available + try: + tokenizer = asr_model.tokenizer + vocab = { + "blank_id": int(asr_model.decoder.blank_idx), + "vocab_size": vocab_size, + } + # Try to get special tokens + if hasattr(tokenizer, "tokenizer"): + inner_tokenizer = tokenizer.tokenizer + if hasattr(inner_tokenizer, "get_vocab"): + full_vocab = inner_tokenizer.get_vocab() + # Find EOU token + eou_token = None + for token, idx in full_vocab.items(): + if "" in token.upper() or "eou" in token.lower(): + eou_token = {"token": token, "id": idx} + break + if eou_token: + vocab["eou_token"] = eou_token + metadata["tokenizer"] = vocab + except Exception as e: + typer.echo(f"Warning: Could not export tokenizer info: {e}") + + metadata_path = output_dir / "metadata.json" + metadata_path.write_text(json.dumps(metadata, indent=2)) + typer.echo(f"\nExport complete. Metadata written to {metadata_path}") + typer.echo(f"Output directory: {output_dir}") + + finally: + asr_model.decoder._rnnt_export = decoder_export_flag + + +if __name__ == "__main__": + app() diff --git a/convert_streaming_encoder.py b/convert_streaming_encoder.py new file mode 100644 index 0000000000000000000000000000000000000000..a1c199c8a6c540dab48b973ab4232db2901b8e72 --- /dev/null +++ b/convert_streaming_encoder.py @@ -0,0 +1,193 @@ + +import torch +import torch.nn as nn +import coremltools as ct +import numpy as np +import typer +from pathlib import Path +from typing import Tuple, List, Optional +import json +import shutil + +# Iimport torch +import coremltools as ct +import numpy as np +import argparse +from nemo.collections.asr.models import EncDecRNNTBPEModel + +app = typer.Typer() + +class LoopbackEncoderWrapper(nn.Module): + """ + Wraps the entire Parakeet Encoder (PreEncode + Conformer) for CoreML Loopback Streaming. + + Inputs: + - audio_signal: [B, D, T] (Mel spectrogram chunk) + - audio_length: [B] + - pre_cache: [B, D, pre_cache_size] (Previous audio context) + - cache_last_channel: [layers, B, cache_size, hidden] + - cache_last_time: [layers, B, hidden, time_cache] + - cache_last_channel_len: [B] + + Outputs: + - encoded_output: [B, D_out, T_out] + - encoded_length: [B] + - new_pre_cache: [B, D, pre_cache_size] + - new_cache_last_channel + - new_cache_last_time + - new_cache_last_channel_len + """ + def __init__(self, encoder, pre_cache_size=16): + super().__init__() + self.encoder = encoder + self.pre_cache_size = pre_cache_size + + def forward( + self, + audio_signal: torch.Tensor, + audio_length: torch.Tensor, + pre_cache: torch.Tensor, + cache_last_channel: torch.Tensor, + cache_last_time: torch.Tensor, + cache_last_channel_len: torch.Tensor + ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor]: + + # 1. Prepend pre_cache to audio_signal + # audio_signal: [B, D, T] + # pre_cache: [B, D, T_cache] + full_input = torch.cat([pre_cache, audio_signal], dim=2) + full_length = audio_length + self.pre_cache_size + + # 2. Extract NEW pre_cache (last N frames of full_input) + # Note: We do this BEFORE processing because we want the raw audio context + new_pre_cache = full_input[:, :, -self.pre_cache_size:] + + # 3. Process with Encoder + # Reconstruct NeMo cache object + current_cache = [cache_last_channel, cache_last_time, cache_last_channel_len] + + encoded, encoded_len, new_cache_channel, new_cache_time, new_cache_len = self.encoder.cache_aware_stream_step( + processed_signal=full_input, + processed_signal_length=full_length, + cache_last_channel=cache_last_channel, + cache_last_time=cache_last_time, + cache_last_channel_len=cache_last_channel_len + ) + + # 4. Drop the first few frames corresponding to pre_cache? + # NeMo's cache_aware_stream_step usually handles the "valid" output frames. + # But since we manually prepended, we might get extra output frames. + # However, for streaming, we usually want the model to see the context but only output the new tokens. + # Let's trust NeMo's streaming logic for now, or check if we need to slice. + # Given we are using 'cache_aware_stream_step', it expects the full context window? + # Actually, standard usage is: input IS the new chunk, but internal convolution looks at past. + # But since we are stateless, we MUST provide the past. + # So passing (pre_cache + chunk) is correct. + + # Cast lengths to Int32 for CoreML + encoded_len_32 = encoded_len.to(dtype=torch.int32) + new_channel_len_32 = new_cache_len.to(dtype=torch.int32) + + return encoded, encoded_len_32, new_pre_cache, new_cache_channel, new_cache_time, new_channel_len_32 + +def _coreml_convert( + traced_model, + inputs, + outputs, + compute_units=ct.ComputeUnit.CPU_ONLY +): + return ct.convert( + traced_model, + inputs=inputs, + outputs=outputs, + compute_units=compute_units, + minimum_deployment_target=ct.target.macOS14, + ) + +def main(): + model_id: str = "nvidia/parakeet_realtime_eou_120m-v1" + output_dir: str = "temp_swift_models/StreamingLoopback" + output_path = Path(output_dir) + output_path.mkdir(parents=True, exist_ok=True) + + print(f"Loading model: {model_id}...") + asr_model = EncDecRNNTBPEModel.from_pretrained(model_name=model_id) + asr_model.eval() + + parser = argparse.ArgumentParser() + parser.add_argument("--chunk-frames", type=int, default=17, help="Number of frames in the input chunk (e.g. 17 for 160ms, 129 for 1.28s)") + args = parser.parse_args() + + encoder = asr_model.encoder + + # --- Configuration --- + # 160ms chunk = 16 frames (but preprocessor produces 17 with padding/centering) + # 1.28s chunk = 128 frames (preprocessor produces 129) + chunk_size_in = args.chunk_frames + mel_dim = 128 + hidden_dim = encoder.d_model # 512 + num_layers = len(encoder.layers) # 17 + + # Cache sizes + cache_channel_size = 70 + cache_time_size = 8 + pre_cache_size = 16 + + print(f"Config: Chunk={chunk_size_in}, Mel={mel_dim}, Hidden={hidden_dim}, Layers={num_layers}") + print(f"Cache: Channel={cache_channel_size}, Time={cache_time_size}, Pre={pre_cache_size}") + + # --- Wrapper --- + wrapper = LoopbackEncoderWrapper(encoder, pre_cache_size=pre_cache_size) + wrapper.eval() + + # --- Test Inputs (for Tracing) --- + batch_size = 1 + test_mel = torch.randn(batch_size, mel_dim, chunk_size_in) + test_mel_len = torch.tensor([chunk_size_in], dtype=torch.int32) + test_pre_cache = torch.zeros(batch_size, mel_dim, pre_cache_size) + + # Initial Cache (Zeros) + test_cache_channel = torch.zeros(num_layers, batch_size, cache_channel_size, hidden_dim) + test_cache_time = torch.zeros(num_layers, batch_size, hidden_dim, cache_time_size) + test_cache_len = torch.zeros(batch_size, dtype=torch.int32) + + print("Tracing model...") + traced_model = torch.jit.trace( + wrapper, + (test_mel, test_mel_len, test_pre_cache, test_cache_channel, test_cache_time, test_cache_len), + strict=False + ) + + # --- CoreML Conversion --- + print("Converting to CoreML...") + + inputs = [ + ct.TensorType(name="audio_signal", shape=(1, 128, chunk_size_in), dtype=np.float32), + ct.TensorType(name="audio_length", shape=(1,), dtype=np.int32), + ct.TensorType(name="pre_cache", shape=(1, 128, pre_cache_size), dtype=np.float32), + ct.TensorType(name="cache_last_channel", shape=(num_layers, 1, cache_channel_size, hidden_dim), dtype=np.float32), + ct.TensorType(name="cache_last_time", shape=(num_layers, 1, hidden_dim, cache_time_size), dtype=np.float32), + ct.TensorType(name="cache_last_channel_len", shape=(1,), dtype=np.int32), + ] + + outputs = [ + ct.TensorType(name="encoded_output", dtype=np.float32), + ct.TensorType(name="encoded_length", dtype=np.int32), + ct.TensorType(name="new_pre_cache", dtype=np.float32), + ct.TensorType(name="new_cache_last_channel", dtype=np.float32), + ct.TensorType(name="new_cache_last_time", dtype=np.float32), + ct.TensorType(name="new_cache_last_channel_len", dtype=np.int32), + ] + + mlmodel = _coreml_convert(traced_model, inputs, outputs) + + save_path = output_path / "streaming_encoder.mlpackage" + mlmodel.save(str(save_path)) + print(f"Saved: {save_path}") + + # Also export Preprocessor, Decoder, Joint for completeness? + # For now, let's assume we reuse the existing ones or export them separately if needed. + # But the user asked specifically for the Encoder loopback. + +if __name__ == "__main__": + main() diff --git a/individual_components.py b/individual_components.py new file mode 100644 index 0000000000000000000000000000000000000000..47271397bc8d9d17cc0fabcf4bb63be7e7c2109c --- /dev/null +++ b/individual_components.py @@ -0,0 +1,250 @@ +#!/usr/bin/env python3 +"""Export Parakeet Realtime EOU 120M RNNT components into CoreML.""" +from __future__ import annotations + +from dataclasses import dataclass +from pathlib import Path +from typing import Optional, Tuple + +import coremltools as ct +import torch + + +@dataclass +class ExportSettings: + output_dir: Path + compute_units: ct.ComputeUnit + deployment_target: Optional[ct.target] + compute_precision: Optional[ct.precision] + max_audio_seconds: float + max_symbol_steps: int + + +class PreprocessorWrapper(torch.nn.Module): + """Wrapper for the audio preprocessor (mel spectrogram extraction).""" + + def __init__(self, module: torch.nn.Module) -> None: + super().__init__() + self.module = module + + def forward( + self, audio_signal: torch.Tensor, length: torch.Tensor + ) -> Tuple[torch.Tensor, torch.Tensor]: + mel, mel_length = self.module( + input_signal=audio_signal, length=length.to(dtype=torch.long) + ) + return mel, mel_length + + +class EncoderWrapper(torch.nn.Module): + """Wrapper for the cache-aware FastConformer encoder. + + Note: For the realtime EOU model, the encoder is cache-aware which means + it can operate in a streaming fashion. For CoreML export, we export + without cache state for simplicity (full-context mode). + """ + + def __init__(self, module: torch.nn.Module) -> None: + super().__init__() + self.module = module + + def forward( + self, features: torch.Tensor, length: torch.Tensor + ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: + encoded, encoded_lengths = self.module( + audio_signal=features, length=length.to(dtype=torch.long) + ) + # Synthesize per-frame timestamps (seconds) using the 80 ms encoder stride. + # Shape: [B, T_enc] + frame_times = ( + torch.arange(encoded.shape[-1], device=encoded.device, dtype=torch.float32) + * 0.08 + ) + return encoded, encoded_lengths, frame_times + + +class DecoderWrapper(torch.nn.Module): + """Wrapper for the RNNT prediction network (decoder).""" + + def __init__(self, module: torch.nn.Module) -> None: + super().__init__() + self.module = module + + def forward( + self, + targets: torch.Tensor, + target_lengths: torch.Tensor, + h_in: torch.Tensor, + c_in: torch.Tensor, + ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: + state = [h_in, c_in] + decoder_output, _, new_state = self.module( + targets=targets.to(dtype=torch.long), + target_length=target_lengths.to(dtype=torch.long), + states=state, + ) + return decoder_output, new_state[0], new_state[1] + + +class JointWrapper(torch.nn.Module): + """Wrapper for the RNNT joint network. + + Note: Unlike Parakeet TDT v3, the realtime EOU model does NOT have + duration outputs (num_extra_outputs). The joint network outputs only + token logits over the vocabulary + blank. + """ + + def __init__(self, module: torch.nn.Module) -> None: + super().__init__() + self.module = module + + def forward( + self, encoder_outputs: torch.Tensor, decoder_outputs: torch.Tensor + ) -> torch.Tensor: + # Input: encoder_outputs [B, D, T], decoder_outputs [B, D, U] + # Transpose to match what projection layers expect + encoder_outputs = encoder_outputs.transpose(1, 2) # [B, T, D] + decoder_outputs = decoder_outputs.transpose(1, 2) # [B, U, D] + + # Apply projections + enc_proj = self.module.enc(encoder_outputs) # [B, T, joint_hidden] + dec_proj = self.module.pred(decoder_outputs) # [B, U, joint_hidden] + + # Explicit broadcasting along T and U to avoid converter ambiguity + x = enc_proj.unsqueeze(2) + dec_proj.unsqueeze(1) # [B, T, U, joint_hidden] + x = self.module.joint_net[0](x) # ReLU + x = self.module.joint_net[1](x) # Dropout (no-op in eval) + out = self.module.joint_net[2](x) # Linear -> logits [B, T, U, vocab+blank] + return out + + +class MelEncoderWrapper(torch.nn.Module): + """Fused wrapper: waveform -> mel -> encoder. + + Inputs: + - audio_signal: [B, S] + - audio_length: [B] + + Outputs: + - encoder: [B, D, T_enc] + - encoder_length: [B] + - frame_times: [T_enc] + """ + + def __init__( + self, preprocessor: PreprocessorWrapper, encoder: EncoderWrapper + ) -> None: + super().__init__() + self.preprocessor = preprocessor + self.encoder = encoder + + def forward( + self, audio_signal: torch.Tensor, audio_length: torch.Tensor + ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: + mel, mel_length = self.preprocessor(audio_signal, audio_length) + encoded, enc_len, frame_times = self.encoder(mel, mel_length.to(dtype=torch.int32)) + return encoded, enc_len, frame_times + + +class JointDecisionWrapper(torch.nn.Module): + """Joint + decision head: outputs label id and label prob. + + Unlike Parakeet TDT v3, this model does NOT have duration outputs. + + Inputs: + - encoder_outputs: [B, D, T] + - decoder_outputs: [B, D, U] + + Returns: + - token_id: [B, T, U] int32 + - token_prob: [B, T, U] float32 + """ + + def __init__(self, joint: JointWrapper, vocab_size: int) -> None: + super().__init__() + self.joint = joint + self.vocab_with_blank = int(vocab_size) + 1 + + def forward(self, encoder_outputs: torch.Tensor, decoder_outputs: torch.Tensor): + logits = self.joint(encoder_outputs, decoder_outputs) + + # Token selection + token_ids = torch.argmax(logits, dim=-1).to(dtype=torch.int32) + token_probs_all = torch.softmax(logits, dim=-1) + # gather expects int64 (long) indices; cast only for gather + token_prob = torch.gather( + token_probs_all, dim=-1, index=token_ids.long().unsqueeze(-1) + ).squeeze(-1) + + return token_ids, token_prob + + +class JointDecisionSingleStep(torch.nn.Module): + """Single-step variant for streaming: encoder_step -> token decision. + + Inputs: + - encoder_step: [B=1, D, T=1] + - decoder_step: [B=1, D, U=1] + + Returns: + - token_id: [1, 1, 1] int32 + - token_prob: [1, 1, 1] float32 + - top_k_ids: [1, 1, 1, K] int32 + - top_k_logits: [1, 1, 1, K] float32 + """ + + def __init__(self, joint: JointWrapper, vocab_size: int, top_k: int = 64) -> None: + super().__init__() + self.joint = joint + self.vocab_with_blank = int(vocab_size) + 1 + self.top_k = int(top_k) + + def forward(self, encoder_step: torch.Tensor, decoder_step: torch.Tensor): + # Reuse JointWrapper which expects [B, D, T] and [B, D, U] + logits = self.joint(encoder_step, decoder_step) # [1, 1, 1, V+blank] + + token_ids = torch.argmax(logits, dim=-1, keepdim=False).to(dtype=torch.int32) + token_probs_all = torch.softmax(logits, dim=-1) + token_prob = torch.gather( + token_probs_all, dim=-1, index=token_ids.long().unsqueeze(-1) + ).squeeze(-1) + + # Also expose top-K candidates for host-side processing + topk_logits, topk_ids_long = torch.topk( + logits, k=min(self.top_k, logits.shape[-1]), dim=-1 + ) + topk_ids = topk_ids_long.to(dtype=torch.int32) + return token_ids, token_prob, topk_ids, topk_logits + + +def _coreml_convert( + traced: torch.jit.ScriptModule, + inputs, + outputs, + settings: ExportSettings, + compute_units_override: Optional[ct.ComputeUnit] = None, + compute_precision: Optional[ct.precision] = None, +) -> ct.models.MLModel: + cu = ( + compute_units_override + if compute_units_override is not None + else settings.compute_units + ) + kwargs = { + "convert_to": "mlprogram", + "inputs": inputs, + "outputs": outputs, + "compute_units": cu, + } + print("Converting:", traced.__class__.__name__) + print("Conversion kwargs:", kwargs) + if settings.deployment_target is not None: + kwargs["minimum_deployment_target"] = settings.deployment_target + + # Priority: explicit argument > settings + if compute_precision is not None: + kwargs["compute_precision"] = compute_precision + elif settings.compute_precision is not None: + kwargs["compute_precision"] = settings.compute_precision + + return ct.convert(traced, **kwargs) diff --git a/metadata.json b/metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..25813362b8502d5f97b8b92d03aec535f93ed0a8 --- /dev/null +++ b/metadata.json @@ -0,0 +1,23 @@ +{ + "model_id": "nvidia/parakeet_realtime_eou_120m-v1", + "model_name": "parakeet_realtime_eou_120m-v1-split", + "streaming_mode": "split_encoder", + "sample_rate": 16000, + "mel_dim": 128, + "hidden_dim": 512, + "num_layers": 17, + "mel_frames_per_chunk": 1005, + "vocab_size": 1026, + "blank_id": 1026, + "decoder_hidden": 640, + "decoder_layers": 1, + "cache_channel_size": 70, + "cache_time_size": 8, + "components": { + "preprocessor": "preprocessor.mlpackage", + "pre_encode": "pre_encode.mlpackage", + "conformer": "conformer_streaming.mlpackage", + "decoder": "decoder.mlpackage", + "joint_decision": "joint_decision.mlpackage" + } +} \ No newline at end of file diff --git a/tokenizer.model b/tokenizer.model new file mode 100644 index 0000000000000000000000000000000000000000..ac8cf05bb6c18f530e2bf89d965c62dbd4feda33 --- /dev/null +++ b/tokenizer.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d556e51ba5b89db64a8cb2be6798fb29974edcadb58b0c7b80418eb5d8752303 +size 258183 diff --git a/vocab.json b/vocab.json new file mode 100644 index 0000000000000000000000000000000000000000..462081190dbb331a0dde13e99d5cfa99b302f0a7 --- /dev/null +++ b/vocab.json @@ -0,0 +1,1028 @@ +{ + "0": "", + "1": "▁t", + "2": "▁th", + "3": "▁a", + "4": "▁i", + "5": "▁the", + "6": "▁s", + "7": "re", + "8": "▁w", + "9": "▁o", + "10": "in", + "11": "at", + "12": "er", + "13": "nd", + "14": "ou", + "15": "▁c", + "16": "▁b", + "17": "▁h", + "18": "en", + "19": "on", + "20": "▁m", + "21": "▁f", + "22": "ing", + "23": "▁p", + "24": "▁to", + "25": "▁and", + "26": "▁d", + "27": "an", + "28": "or", + "29": "es", + "30": "▁y", + "31": "▁l", + "32": "▁of", + "33": "ll", + "34": "▁in", + "35": "ed", + "36": "it", + "37": "▁g", + "38": "is", + "39": "▁you", + "40": "▁n", + "41": "ar", + "42": "om", + "43": "as", + "44": "ve", + "45": "▁e", + "46": "ic", + "47": "▁it", + "48": "al", + "49": "us", + "50": "▁wh", + "51": "▁we", + "52": "▁be", + "53": "ion", + "54": "ow", + "55": "le", + "56": "▁is", + "57": "et", + "58": "ent", + "59": "ot", + "60": "ut", + "61": "▁re", + "62": "▁on", + "63": "ay", + "64": "▁ha", + "65": "ig", + "66": "▁so", + "67": "ct", + "68": "▁he", + "69": "▁for", + "70": "ver", + "71": "ke", + "72": "ro", + "73": "▁st", + "74": "id", + "75": "▁go", + "76": "all", + "77": "se", + "78": "ly", + "79": "▁u", + "80": "ch", + "81": "st", + "82": "ld", + "83": "▁k", + "84": "ce", + "85": "ur", + "86": "▁li", + "87": "am", + "88": "▁r", + "89": "ht", + "90": "▁j", + "91": "ith", + "92": "▁se", + "93": "ir", + "94": "▁as", + "95": "▁an", + "96": "im", + "97": "▁do", + "98": "ad", + "99": "▁was", + "100": "ight", + "101": "th", + "102": "▁are", + "103": "▁but", + "104": "▁sh", + "105": "ust", + "106": "ally", + "107": "▁not", + "108": "▁or", + "109": "▁com", + "110": "▁can", + "111": "▁me", + "112": "op", + "113": "▁mo", + "114": "▁at", + "115": "ill", + "116": "▁ch", + "117": "▁ne", + "118": "ant", + "119": "▁de", + "120": "▁kn", + "121": "▁one", + "122": "il", + "123": "ol", + "124": "▁con", + "125": "ter", + "126": "▁ab", + "127": "▁fr", + "128": "ere", + "129": "ck", + "130": "▁al", + "131": "▁all", + "132": "qu", + "133": "▁pro", + "134": "▁som", + "135": "ould", + "136": "▁tw", + "137": "ul", + "138": "ra", + "139": "od", + "140": "ers", + "141": "▁su", + "142": "ive", + "143": "▁v", + "144": "use", + "145": "ate", + "146": "ge", + "147": "if", + "148": "▁ex", + "149": "ess", + "150": "pp", + "151": "▁lo", + "152": "out", + "153": "▁if", + "154": "est", + "155": "ain", + "156": "ist", + "157": "and", + "158": "ea", + "159": "very", + "160": "art", + "161": "▁wor", + "162": "▁my", + "163": "ab", + "164": "ment", + "165": "▁bec", + "166": "un", + "167": "ity", + "168": "ri", + "169": "pe", + "170": "ions", + "171": "▁by", + "172": "ok", + "173": "our", + "174": "ort", + "175": "ind", + "176": "ink", + "177": "nt", + "178": "▁up", + "179": "um", + "180": "▁don", + "181": "▁get", + "182": "red", + "183": "▁out", + "184": "el", + "185": "ause", + "186": "res", + "187": "▁ma", + "188": "ich", + "189": "▁us", + "190": "rou", + "191": "▁int", + "192": "em", + "193": "os", + "194": "ies", + "195": "ie", + "196": "▁pl", + "197": "▁tr", + "198": "ven", + "199": "ous", + "200": "▁le", + "201": "▁two", + "202": "ard", + "203": "ine", + "204": "▁co", + "205": "een", + "206": "▁now", + "207": "ty", + "208": "her", + "209": "ack", + "210": "▁pe", + "211": "ame", + "212": "▁how", + "213": "▁who", + "214": "▁see", + "215": "▁tim", + "216": "ect", + "217": "ast", + "218": "▁our", + "219": "ci", + "220": "ree", + "221": "ople", + "222": "gh", + "223": "▁no", + "224": "▁had", + "225": "▁man", + "226": "▁qu", + "227": "▁en", + "228": "ide", + "229": "ure", + "230": "ud", + "231": "so", + "232": "▁his", + "233": "▁sa", + "234": "▁sp", + "235": "▁say", + "236": "ose", + "237": "ther", + "238": "▁act", + "239": "▁ta", + "240": "▁cl", + "241": "ings", + "242": "pt", + "243": "king", + "244": "▁any", + "245": "▁has", + "246": "▁un", + "247": "iv", + "248": "▁im", + "249": "▁ag", + "250": "▁te", + "251": "▁fe", + "252": "one", + "253": "per", + "254": "ong", + "255": "▁po", + "256": "▁ad", + "257": "ff", + "258": "ore", + "259": "itt", + "260": "ans", + "261": "iz", + "262": "eah", + "263": "reat", + "264": "act", + "265": "own", + "266": "hing", + "267": "enty", + "268": "age", + "269": "ber", + "270": "ice", + "271": "▁am", + "272": "ple", + "273": "are", + "274": "▁per", + "275": "und", + "276": "ite", + "277": "ix", + "278": "pl", + "279": "▁way", + "280": "▁did", + "281": "▁pr", + "282": "▁got", + "283": "ars", + "284": "▁she", + "285": "▁let", + "286": "ag", + "287": "▁ac", + "288": "int", + "289": "▁ar", + "290": "ry", + "291": "ign", + "292": "ish", + "293": "▁fir", + "294": "ace", + "295": "ble", + "296": "og", + "297": "ue", + "298": "▁ye", + "299": "ap", + "300": "iff", + "301": "▁ro", + "302": "▁her", + "303": "nder", + "304": "▁ok", + "305": "▁res", + "306": "▁gu", + "307": "ence", + "308": "▁may", + "309": "ated", + "310": "ip", + "311": "▁bo", + "312": "▁him", + "313": "way", + "314": "ac", + "315": "ical", + "316": "ass", + "317": "ase", + "318": "▁dis", + "319": "able", + "320": "ick", + "321": "▁app", + "322": "ance", + "323": "▁pre", + "324": "▁six", + "325": "▁off", + "326": "▁new", + "327": "ia", + "328": "orm", + "329": "ank", + "330": "▁lot", + "331": "ach", + "332": "▁fo", + "333": "inet", + "334": "ire", + "335": "ary", + "336": "ult", + "337": "▁tal", + "338": "▁mu", + "339": "▁bl", + "340": "ount", + "341": "sel", + "342": "vel", + "343": "▁br", + "344": "▁imp", + "345": "ep", + "346": "cess", + "347": "ord", + "348": "▁sc", + "349": "▁inc", + "350": "ound", + "351": "ang", + "352": "be", + "353": "ress", + "354": "uct", + "355": "▁ind", + "356": "▁af", + "357": "ving", + "358": "▁oh", + "359": "▁bet", + "360": "▁use", + "361": "ome", + "362": "ens", + "363": "ys", + "364": "▁bu", + "365": "co", + "366": "ory", + "367": "ater", + "368": "ild", + "369": "ght", + "370": "ial", + "371": "▁day", + "372": "ning", + "373": "na", + "374": "ile", + "375": "▁spe", + "376": "▁mar", + "377": "ody", + "378": "ough", + "379": "ade", + "380": "vers", + "381": "xt", + "382": "▁fl", + "383": "▁ke", + "384": "ian", + "385": "▁sy", + "386": "▁put", + "387": "fore", + "388": "ub", + "389": "▁ph", + "390": "fe", + "391": "▁em", + "392": "▁ser", + "393": "form", + "394": "ting", + "395": "te", + "396": "av", + "397": "ious", + "398": "▁rec", + "399": "ks", + "400": "▁gr", + "401": "ces", + "402": "wn", + "403": "ors", + "404": "▁jo", + "405": "ents", + "406": "▁des", + "407": "▁try", + "408": "▁equ", + "409": "▁z", + "410": "▁rem", + "411": "▁str", + "412": "self", + "413": "▁bit", + "414": "ph", + "415": "ved", + "416": "▁why", + "417": "▁bas", + "418": "▁hel", + "419": "▁rel", + "420": "ath", + "421": "ject", + "422": "ail", + "423": "▁la", + "424": "ual", + "425": "▁god", + "426": "▁nat", + "427": "erm", + "428": "day", + "429": "▁id", + "430": "ft", + "431": "▁wr", + "432": "▁min", + "433": "ates", + "434": "▁gen", + "435": "tain", + "436": "▁ob", + "437": "ull", + "438": "ict", + "439": "▁tra", + "440": "▁end", + "441": "▁hig", + "442": "▁fif", + "443": "oth", + "444": "tern", + "445": "▁its", + "446": "vent", + "447": "▁sm", + "448": "ons", + "449": "▁add", + "450": "iss", + "451": "▁bel", + "452": "ful", + "453": "get", + "454": "▁ele", + "455": "▁rep", + "456": "ak", + "457": "▁ho", + "458": "▁pos", + "459": "▁num", + "460": "ange", + "461": "ves", + "462": "ific", + "463": "urn", + "464": "ise", + "465": "▁cr", + "466": "▁um", + "467": "ward", + "468": "▁reg", + "469": "ady", + "470": "ower", + "471": "uc", + "472": "▁dec", + "473": "lic", + "474": "▁set", + "475": "▁gon", + "476": "▁op", + "477": "▁ear", + "478": "▁sub", + "479": "▁sl", + "480": "les", + "481": "stem", + "482": "cial", + "483": "olog", + "484": "atch", + "485": "ily", + "486": "body", + "487": "nds", + "488": "ular", + "489": "ren", + "490": "▁own", + "491": "▁too", + "492": "cent", + "493": "ible", + "494": "pect", + "495": "ered", + "496": "ways", + "497": "teen", + "498": "▁uh", + "499": "▁big", + "500": "▁mod", + "501": "▁att", + "502": "▁car", + "503": "gr", + "504": "▁acc", + "505": "ied", + "506": "mun", + "507": "ib", + "508": "▁mon", + "509": "▁sch", + "510": "▁pol", + "511": "▁dat", + "512": "▁fin", + "513": "▁sim", + "514": "▁inv", + "515": "▁def", + "516": "ked", + "517": "▁ent", + "518": "▁yes", + "519": "ows", + "520": "ics", + "521": "ited", + "522": "ute", + "523": "ism", + "524": "ps", + "525": "▁ed", + "526": "▁el", + "527": "ably", + "528": "ppen", + "529": "als", + "530": "▁ten", + "531": "ract", + "532": "ss", + "533": "▁ass", + "534": "▁met", + "535": "gan", + "536": "▁eng", + "537": "▁stu", + "538": "ween", + "539": "arch", + "540": "▁gl", + "541": "▁cor", + "542": "▁dr", + "543": "vern", + "544": "▁ty", + "545": "▁run", + "546": "hip", + "547": "cus", + "548": "cond", + "549": "▁ins", + "550": "irty", + "551": "▁pub", + "552": "lud", + "553": "llow", + "554": "▁cou", + "555": "ew", + "556": "iew", + "557": "▁sur", + "558": "ero", + "559": "ood", + "560": "ness", + "561": "▁fun", + "562": "▁eff", + "563": "cept", + "564": "▁ca", + "565": "▁exp", + "566": "duct", + "567": "▁sw", + "568": "ize", + "569": "ope", + "570": "▁par", + "571": "kes", + "572": "cy", + "573": "▁ev", + "574": "▁ref", + "575": "ell", + "576": "▁bus", + "577": "ug", + "578": "rib", + "579": "▁cur", + "580": "mo", + "581": "ock", + "582": "ures", + "583": "air", + "584": "▁war", + "585": "str", + "586": "▁med", + "587": "▁wa", + "588": "▁val", + "589": "▁sin", + "590": "blem", + "591": "▁fam", + "592": "li", + "593": "▁far", + "594": "▁cle", + "595": "▁col", + "596": "mon", + "597": "▁gra", + "598": "led", + "599": "ense", + "600": "tin", + "601": "ues", + "602": "its", + "603": "▁mem", + "604": "▁inf", + "605": "▁eas", + "606": "ideo", + "607": "▁top", + "608": "io", + "609": "pan", + "610": "▁hum", + "611": "▁old", + "612": "ead", + "613": "▁ord", + "614": "ric", + "615": "ants", + "616": "oy", + "617": "esn", + "618": "uck", + "619": "ason", + "620": "ced", + "621": "ool", + "622": "rat", + "623": "ouse", + "624": "▁lar", + "625": "▁art", + "626": "▁wee", + "627": "▁cer", + "628": "ized", + "629": "▁mat", + "630": "con", + "631": "erg", + "632": "land", + "633": "ines", + "634": "▁chr", + "635": "▁aut", + "636": "▁lea", + "637": "▁sou", + "638": "oney", + "639": "tty", + "640": "▁ple", + "641": "ulat", + "642": "oks", + "643": "▁few", + "644": "▁sol", + "645": "▁che", + "646": "chn", + "647": "ird", + "648": "▁bre", + "649": "▁dur", + "650": "▁wom", + "651": "me", + "652": "izat", + "653": "eric", + "654": "ote", + "655": "▁uni", + "656": "eren", + "657": "arn", + "658": "ross", + "659": "ices", + "660": "ten", + "661": "eral", + "662": "ever", + "663": "ieve", + "664": "lish", + "665": "ash", + "666": "▁opp", + "667": "alth", + "668": "ger", + "669": "▁sk", + "670": "▁red", + "671": "peri", + "672": "▁det", + "673": "▁ext", + "674": "ner", + "675": "ah", + "676": "▁var", + "677": "▁loc", + "678": "gram", + "679": "ists", + "680": "ives", + "681": "▁es", + "682": "▁nor", + "683": "tro", + "684": "ale", + "685": "▁iss", + "686": "▁pri", + "687": "gin", + "688": "az", + "689": "oc", + "690": "▁pop", + "691": "ern", + "692": "▁sit", + "693": "ket", + "694": "▁pa", + "695": "▁law", + "696": "ages", + "697": "br", + "698": "▁cam", + "699": "▁mom", + "700": "osed", + "701": "▁bro", + "702": "ne", + "703": "bs", + "704": "▁cre", + "705": "erat", + "706": "▁sec", + "707": "▁cap", + "708": "▁vis", + "709": "▁pat", + "710": "ield", + "711": "iet", + "712": "▁tri", + "713": "up", + "714": "▁bra", + "715": "ts", + "716": "▁mot", + "717": "▁unt", + "718": "put", + "719": "bo", + "720": "ork", + "721": "mer", + "722": "ital", + "723": "▁air", + "724": "ined", + "725": "▁beh", + "726": "▁adv", + "727": "▁ret", + "728": "imes", + "729": "▁tea", + "730": "ural", + "731": "sid", + "732": "ters", + "733": "▁pur", + "734": "▁sci", + "735": "bers", + "736": "ient", + "737": "ier", + "738": "cc", + "739": "sw", + "740": "▁av", + "741": "reen", + "742": "ode", + "743": "ont", + "744": "▁dra", + "745": "ann", + "746": "nect", + "747": "▁x", + "748": "▁eu", + "749": "ton", + "750": "inat", + "751": "ene", + "752": "ared", + "753": "els", + "754": "▁mor", + "755": "▁rat", + "756": "cri", + "757": "▁men", + "758": "▁ah", + "759": "ames", + "760": "▁arm", + "761": "eak", + "762": "▁pay", + "763": "▁hal", + "764": "ins", + "765": "ilit", + "766": "stit", + "767": "▁ra", + "768": "▁leg", + "769": "cl", + "770": "pr", + "771": "▁wal", + "772": "▁bad", + "773": "▁ge", + "774": "roup", + "775": "▁mus", + "776": "man", + "777": "▁gi", + "778": "eds", + "779": "▁aw", + "780": "po", + "781": "ark", + "782": "row", + "783": "▁dep", + "784": "ully", + "785": "ral", + "786": "lect", + "787": "pend", + "788": "▁sev", + "789": "ime", + "790": "gest", + "791": "here", + "792": "▁yet", + "793": "ted", + "794": "▁rev", + "795": "ds", + "796": "▁ask", + "797": "less", + "798": "▁di", + "799": "ets", + "800": "line", + "801": "▁aff", + "802": "ired", + "803": "▁est", + "804": "ken", + "805": "vid", + "806": "most", + "807": "ivid", + "808": "unch", + "809": "par", + "810": "med", + "811": "rop", + "812": "ased", + "813": "eone", + "814": "▁ve", + "815": "▁abs", + "816": "ergy", + "817": "ret", + "818": "▁saw", + "819": "▁ey", + "820": "▁cal", + "821": "uat", + "822": "▁mid", + "823": "vat", + "824": "ream", + "825": "vice", + "826": "ians", + "827": "rent", + "828": "ctor", + "829": "err", + "830": "ush", + "831": "ases", + "832": "▁suc", + "833": "erms", + "834": "ave", + "835": "angu", + "836": "ries", + "837": "▁wo", + "838": "arts", + "839": "▁fil", + "840": "▁fat", + "841": "▁cho", + "842": "orts", + "843": "▁fre", + "844": "ee", + "845": "ught", + "846": "eng", + "847": "ump", + "848": "▁bar", + "849": "ying", + "850": "ane", + "851": "▁tem", + "852": "anks", + "853": "ury", + "854": "iat", + "855": "mit", + "856": "trol", + "857": "▁net", + "858": "▁maj", + "859": "▁cra", + "860": "ling", + "861": "▁fig", + "862": "orn", + "863": "icat", + "864": "pany", + "865": "▁occ", + "866": "ott", + "867": "ands", + "868": "▁exc", + "869": "▁mr", + "870": "ency", + "871": "rope", + "872": "itch", + "873": "▁lit", + "874": "abil", + "875": "not", + "876": "ma", + "877": "▁typ", + "878": "▁opt", + "879": "ob", + "880": "ser", + "881": "ety", + "882": "ms", + "883": "peci", + "884": "aces", + "885": "aut", + "886": "▁hon", + "887": "cuss", + "888": "▁sal", + "889": "▁sor", + "890": "att", + "891": "▁lab", + "892": "▁har", + "893": "urch", + "894": "nded", + "895": "uce", + "896": "ids", + "897": "▁hy", + "898": "▁fut", + "899": "▁ste", + "900": "ours", + "901": "ems", + "902": "utes", + "903": "ng", + "904": "ta", + "905": "▁won", + "906": "▁fa", + "907": "▁env", + "908": "ards", + "909": "▁job", + "910": "ium", + "911": "▁dot", + "912": "▁obv", + "913": "ina", + "914": "side", + "915": "elve", + "916": "cu", + "917": "▁jes", + "918": "▁pot", + "919": "▁pie", + "920": "▁tre", + "921": "▁hey", + "922": "▁mag", + "923": "ron", + "924": "▁key", + "925": "swer", + "926": "▁win", + "927": "ucat", + "928": "work", + "929": "ides", + "930": "▁low", + "931": "▁vol", + "932": "▁oth", + "933": "atic", + "934": "lf", + "935": "ads", + "936": "inds", + "937": "com", + "938": "ths", + "939": "▁ver", + "940": "ised", + "941": "lo", + "942": "▁squ", + "943": "▁cut", + "944": "oked", + "945": "irit", + "946": "ateg", + "947": "ppy", + "948": "mitt", + "949": "come", + "950": "hn", + "951": "igin", + "952": "mand", + "953": "▁dam", + "954": "ho", + "955": "▁da", + "956": "▁fur", + "957": "iron", + "958": "ilar", + "959": "▁fac", + "960": "▁neg", + "961": "▁ago", + "962": "ged", + "963": "miss", + "964": "enth", + "965": "▁dou", + "966": "▁hit", + "967": "▁guy", + "968": "▁bi", + "969": "ove", + "970": "fess", + "971": "ples", + "972": "owed", + "973": "ured", + "974": "▁ris", + "975": "ints", + "976": "rew", + "977": "▁sum", + "978": "▁hu", + "979": "ploy", + "980": "ude", + "981": "ried", + "982": "▁cir", + "983": "▁dev", + "984": "ear", + "985": "▁tot", + "986": "▁ann", + "987": "duc", + "988": "ik", + "989": "pon", + "990": "sted", + "991": "▁ide", + "992": "▁'", + "993": "ipp", + "994": "▁eat", + "995": "▁dom", + "996": "▁", + "997": "e", + "998": "t", + "999": "o", + "1000": "a", + "1001": "i", + "1002": "n", + "1003": "s", + "1004": "r", + "1005": "h", + "1006": "l", + "1007": "d", + "1008": "u", + "1009": "c", + "1010": "m", + "1011": "y", + "1012": "g", + "1013": "w", + "1014": "f", + "1015": "p", + "1016": "b", + "1017": "v", + "1018": "k", + "1019": "'", + "1020": "j", + "1021": "x", + "1022": "q", + "1023": "z", + "1024": "", + "1025": "" +} \ No newline at end of file