diff --git "a/models/onnx/ailia-models/LAION-CLAP/CLAP_text_text_branch_RobertaModel_roberta-base.onnx.prototxt" "b/models/onnx/ailia-models/LAION-CLAP/CLAP_text_text_branch_RobertaModel_roberta-base.onnx.prototxt" new file mode 100644--- /dev/null +++ "b/models/onnx/ailia-models/LAION-CLAP/CLAP_text_text_branch_RobertaModel_roberta-base.onnx.prototxt" @@ -0,0 +1,13615 @@ +ir_version: 6 +producer_name: "pytorch" +producer_version: "1.13.0" +model_version: 0 +graph { + name: "torch_jit" + node { + output: "/Constant_output_0" + name: "/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "input_ids" + output: "/Shape_output_0" + name: "/Shape" + op_type: "Shape" + } + node { + output: "/Constant_1_output_0" + name: "/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/Shape_output_0" + input: "/Constant_1_output_0" + output: "/Gather_output_0" + name: "/Gather" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "input_ids" + output: "/Shape_1_output_0" + name: "/Shape_1" + op_type: "Shape" + } + node { + output: "/Constant_2_output_0" + name: "/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/Shape_1_output_0" + input: "/Constant_2_output_0" + output: "/Gather_1_output_0" + name: "/Gather_1" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + output: "onnx::Slice_210" + name: "Constant_7" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + dims: 514 + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/Constant_3_output_0" + name: "/Constant_3" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/Gather_1_output_0" + output: "/Unsqueeze_output_0" + name: "/Unsqueeze" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/Constant_4_output_0" + name: "/Constant_4" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/Constant_5_output_0" + name: "/Constant_5" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "onnx::Slice_210" + input: "/Constant_3_output_0" + input: "/Unsqueeze_output_0" + input: "/Constant_4_output_0" + input: "/Constant_5_output_0" + output: "/Slice_output_0" + name: "/Slice" + op_type: "Slice" + } + node { + input: "/Gather_output_0" + output: "/Unsqueeze_1_output_0" + name: "/Unsqueeze_1" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/Gather_1_output_0" + output: "/Unsqueeze_2_output_0" + name: "/Unsqueeze_2" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/Unsqueeze_1_output_0" + input: "/Unsqueeze_2_output_0" + output: "/Concat_output_0" + name: "/Concat" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + output: "/Constant_6_output_0" + name: "/Constant_6" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\377\377\377\377\377\377\377\377" + } + type: TENSOR + } + } + node { + input: "/Concat_output_0" + input: "/Constant_6_output_0" + output: "/Reshape_output_0" + name: "/Reshape" + op_type: "Reshape" + } + node { + input: "/Reshape_output_0" + output: "/Shape_2_output_0" + name: "/Shape_2" + op_type: "Shape" + } + node { + input: "/Shape_2_output_0" + output: "/ConstantOfShape_output_0" + name: "/ConstantOfShape" + op_type: "ConstantOfShape" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/Constant_7_output_0" + name: "/Constant_7" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\377\377\377\377\377\377\377\377" + } + type: TENSOR + } + } + node { + input: "/ConstantOfShape_output_0" + input: "/Constant_7_output_0" + output: "/Mul_output_0" + name: "/Mul" + op_type: "Mul" + } + node { + input: "/Reshape_output_0" + input: "/Mul_output_0" + output: "/Equal_output_0" + name: "/Equal" + op_type: "Equal" + } + node { + input: "/Equal_output_0" + input: "/ConstantOfShape_output_0" + input: "/Reshape_output_0" + output: "/Where_output_0" + name: "/Where" + op_type: "Where" + } + node { + input: "/Slice_output_0" + input: "/Where_output_0" + output: "/Expand_output_0" + name: "/Expand" + op_type: "Expand" + } + node { + input: "attention_mask" + output: "/Unsqueeze_3_output_0" + name: "/Unsqueeze_3" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 1 + type: INTS + } + } + node { + input: "/Unsqueeze_3_output_0" + output: "/Unsqueeze_4_output_0" + name: "/Unsqueeze_4" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 2 + type: INTS + } + } + node { + input: "/Unsqueeze_4_output_0" + output: "/Cast_output_0" + name: "/Cast" + op_type: "Cast" + attribute { + name: "to" + i: 1 + type: INT + } + } + node { + output: "/Constant_8_output_0" + name: "/Constant_8" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\200?" + } + type: TENSOR + } + } + node { + input: "/Constant_8_output_0" + input: "/Cast_output_0" + output: "/Sub_output_0" + name: "/Sub" + op_type: "Sub" + } + node { + output: "/Constant_9_output_0" + name: "/Constant_9" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000@\034\306" + } + type: TENSOR + } + } + node { + input: "/Sub_output_0" + input: "/Constant_9_output_0" + output: "/Mul_1_output_0" + name: "/Mul_1" + op_type: "Mul" + } + node { + output: "/embeddings/Constant_output_0" + name: "/embeddings/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "input_ids" + input: "/embeddings/Constant_output_0" + output: "/embeddings/Equal_output_0" + name: "/embeddings/Equal" + op_type: "Equal" + } + node { + input: "/embeddings/Equal_output_0" + output: "/embeddings/Not_output_0" + name: "/embeddings/Not" + op_type: "Not" + } + node { + input: "/embeddings/Not_output_0" + output: "/embeddings/Cast_output_0" + name: "/embeddings/Cast" + op_type: "Cast" + attribute { + name: "to" + i: 6 + type: INT + } + } + node { + output: "/embeddings/Constant_1_output_0" + name: "/embeddings/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 6 + raw_data: "\001\000\000\000" + } + type: TENSOR + } + } + node { + input: "/embeddings/Cast_output_0" + input: "/embeddings/Constant_1_output_0" + output: "/embeddings/CumSum_output_0" + name: "/embeddings/CumSum" + op_type: "CumSum" + } + node { + input: "/embeddings/CumSum_output_0" + input: "/embeddings/Cast_output_0" + output: "/embeddings/Mul_output_0" + name: "/embeddings/Mul" + op_type: "Mul" + } + node { + input: "/embeddings/Mul_output_0" + output: "/embeddings/Cast_1_output_0" + name: "/embeddings/Cast_1" + op_type: "Cast" + attribute { + name: "to" + i: 7 + type: INT + } + } + node { + output: "/embeddings/Constant_2_output_0" + name: "/embeddings/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/embeddings/Cast_1_output_0" + input: "/embeddings/Constant_2_output_0" + output: "/embeddings/Add_output_0" + name: "/embeddings/Add" + op_type: "Add" + } + node { + input: "embeddings.word_embeddings.weight" + input: "input_ids" + output: "/embeddings/word_embeddings/Gather_output_0" + name: "/embeddings/word_embeddings/Gather" + op_type: "Gather" + } + node { + input: "embeddings.token_type_embeddings.weight" + input: "/Expand_output_0" + output: "/embeddings/token_type_embeddings/Gather_output_0" + name: "/embeddings/token_type_embeddings/Gather" + op_type: "Gather" + } + node { + input: "/embeddings/word_embeddings/Gather_output_0" + input: "/embeddings/token_type_embeddings/Gather_output_0" + output: "/embeddings/Add_1_output_0" + name: "/embeddings/Add_1" + op_type: "Add" + } + node { + input: "embeddings.position_embeddings.weight" + input: "/embeddings/Add_output_0" + output: "/embeddings/position_embeddings/Gather_output_0" + name: "/embeddings/position_embeddings/Gather" + op_type: "Gather" + } + node { + input: "/embeddings/Add_1_output_0" + input: "/embeddings/position_embeddings/Gather_output_0" + output: "/embeddings/Add_2_output_0" + name: "/embeddings/Add_2" + op_type: "Add" + } + node { + input: "/embeddings/Add_2_output_0" + output: "/embeddings/LayerNorm/ReduceMean_output_0" + name: "/embeddings/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/embeddings/Add_2_output_0" + input: "/embeddings/LayerNorm/ReduceMean_output_0" + output: "/embeddings/LayerNorm/Sub_output_0" + name: "/embeddings/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/embeddings/LayerNorm/Constant_output_0" + name: "/embeddings/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/embeddings/LayerNorm/Sub_output_0" + input: "/embeddings/LayerNorm/Constant_output_0" + output: "/embeddings/LayerNorm/Pow_output_0" + name: "/embeddings/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/embeddings/LayerNorm/Pow_output_0" + output: "/embeddings/LayerNorm/ReduceMean_1_output_0" + name: "/embeddings/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/embeddings/LayerNorm/Constant_1_output_0" + name: "/embeddings/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/embeddings/LayerNorm/ReduceMean_1_output_0" + input: "/embeddings/LayerNorm/Constant_1_output_0" + output: "/embeddings/LayerNorm/Add_output_0" + name: "/embeddings/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/embeddings/LayerNorm/Add_output_0" + output: "/embeddings/LayerNorm/Sqrt_output_0" + name: "/embeddings/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/embeddings/LayerNorm/Sub_output_0" + input: "/embeddings/LayerNorm/Sqrt_output_0" + output: "/embeddings/LayerNorm/Div_output_0" + name: "/embeddings/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/embeddings/LayerNorm/Div_output_0" + input: "embeddings.LayerNorm.weight" + output: "/embeddings/LayerNorm/Mul_output_0" + name: "/embeddings/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/embeddings/LayerNorm/Mul_output_0" + input: "embeddings.LayerNorm.bias" + output: "/embeddings/LayerNorm/Add_1_output_0" + name: "/embeddings/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/embeddings/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1553" + output: "/encoder/layer.0/attention/self/query/MatMul_output_0" + name: "/encoder/layer.0/attention/self/query/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.0.attention.self.query.bias" + input: "/encoder/layer.0/attention/self/query/MatMul_output_0" + output: "/encoder/layer.0/attention/self/query/Add_output_0" + name: "/encoder/layer.0/attention/self/query/Add" + op_type: "Add" + } + node { + input: "/embeddings/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1554" + output: "/encoder/layer.0/attention/self/key/MatMul_output_0" + name: "/encoder/layer.0/attention/self/key/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.0.attention.self.key.bias" + input: "/encoder/layer.0/attention/self/key/MatMul_output_0" + output: "/encoder/layer.0/attention/self/key/Add_output_0" + name: "/encoder/layer.0/attention/self/key/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.0/attention/self/key/Add_output_0" + output: "/encoder/layer.0/attention/self/Shape_output_0" + name: "/encoder/layer.0/attention/self/Shape" + op_type: "Shape" + } + node { + output: "/encoder/layer.0/attention/self/Constant_output_0" + name: "/encoder/layer.0/attention/self/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.0/attention/self/Shape_output_0" + input: "/encoder/layer.0/attention/self/Constant_output_0" + output: "/encoder/layer.0/attention/self/Gather_output_0" + name: "/encoder/layer.0/attention/self/Gather" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.0/attention/self/key/Add_output_0" + output: "/encoder/layer.0/attention/self/Shape_1_output_0" + name: "/encoder/layer.0/attention/self/Shape_1" + op_type: "Shape" + } + node { + output: "/encoder/layer.0/attention/self/Constant_1_output_0" + name: "/encoder/layer.0/attention/self/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.0/attention/self/Shape_1_output_0" + input: "/encoder/layer.0/attention/self/Constant_1_output_0" + output: "/encoder/layer.0/attention/self/Gather_1_output_0" + name: "/encoder/layer.0/attention/self/Gather_1" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.0/attention/self/Gather_output_0" + output: "/encoder/layer.0/attention/self/Unsqueeze_output_0" + name: "/encoder/layer.0/attention/self/Unsqueeze" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.0/attention/self/Gather_1_output_0" + output: "/encoder/layer.0/attention/self/Unsqueeze_1_output_0" + name: "/encoder/layer.0/attention/self/Unsqueeze_1" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.0/attention/self/Constant_2_output_0" + name: "/encoder/layer.0/attention/self/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.0/attention/self/Constant_3_output_0" + name: "/encoder/layer.0/attention/self/Constant_3" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.0/attention/self/Unsqueeze_output_0" + input: "/encoder/layer.0/attention/self/Unsqueeze_1_output_0" + input: "/encoder/layer.0/attention/self/Constant_2_output_0" + input: "/encoder/layer.0/attention/self/Constant_3_output_0" + output: "/encoder/layer.0/attention/self/Concat_output_0" + name: "/encoder/layer.0/attention/self/Concat" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.0/attention/self/key/Add_output_0" + input: "/encoder/layer.0/attention/self/Concat_output_0" + output: "/encoder/layer.0/attention/self/Reshape_output_0" + name: "/encoder/layer.0/attention/self/Reshape" + op_type: "Reshape" + } + node { + input: "/embeddings/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1557" + output: "/encoder/layer.0/attention/self/value/MatMul_output_0" + name: "/encoder/layer.0/attention/self/value/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.0.attention.self.value.bias" + input: "/encoder/layer.0/attention/self/value/MatMul_output_0" + output: "/encoder/layer.0/attention/self/value/Add_output_0" + name: "/encoder/layer.0/attention/self/value/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.0/attention/self/value/Add_output_0" + output: "/encoder/layer.0/attention/self/Shape_2_output_0" + name: "/encoder/layer.0/attention/self/Shape_2" + op_type: "Shape" + } + node { + output: "/encoder/layer.0/attention/self/Constant_4_output_0" + name: "/encoder/layer.0/attention/self/Constant_4" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.0/attention/self/Shape_2_output_0" + input: "/encoder/layer.0/attention/self/Constant_4_output_0" + output: "/encoder/layer.0/attention/self/Gather_2_output_0" + name: "/encoder/layer.0/attention/self/Gather_2" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.0/attention/self/value/Add_output_0" + output: "/encoder/layer.0/attention/self/Shape_3_output_0" + name: "/encoder/layer.0/attention/self/Shape_3" + op_type: "Shape" + } + node { + output: "/encoder/layer.0/attention/self/Constant_5_output_0" + name: "/encoder/layer.0/attention/self/Constant_5" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.0/attention/self/Shape_3_output_0" + input: "/encoder/layer.0/attention/self/Constant_5_output_0" + output: "/encoder/layer.0/attention/self/Gather_3_output_0" + name: "/encoder/layer.0/attention/self/Gather_3" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.0/attention/self/Gather_2_output_0" + output: "/encoder/layer.0/attention/self/Unsqueeze_2_output_0" + name: "/encoder/layer.0/attention/self/Unsqueeze_2" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.0/attention/self/Gather_3_output_0" + output: "/encoder/layer.0/attention/self/Unsqueeze_3_output_0" + name: "/encoder/layer.0/attention/self/Unsqueeze_3" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.0/attention/self/Constant_6_output_0" + name: "/encoder/layer.0/attention/self/Constant_6" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.0/attention/self/Constant_7_output_0" + name: "/encoder/layer.0/attention/self/Constant_7" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.0/attention/self/Unsqueeze_2_output_0" + input: "/encoder/layer.0/attention/self/Unsqueeze_3_output_0" + input: "/encoder/layer.0/attention/self/Constant_6_output_0" + input: "/encoder/layer.0/attention/self/Constant_7_output_0" + output: "/encoder/layer.0/attention/self/Concat_1_output_0" + name: "/encoder/layer.0/attention/self/Concat_1" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.0/attention/self/value/Add_output_0" + input: "/encoder/layer.0/attention/self/Concat_1_output_0" + output: "/encoder/layer.0/attention/self/Reshape_1_output_0" + name: "/encoder/layer.0/attention/self/Reshape_1" + op_type: "Reshape" + } + node { + input: "/encoder/layer.0/attention/self/Reshape_1_output_0" + output: "/encoder/layer.0/attention/self/Transpose_output_0" + name: "/encoder/layer.0/attention/self/Transpose" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.0/attention/self/query/Add_output_0" + output: "/encoder/layer.0/attention/self/Shape_4_output_0" + name: "/encoder/layer.0/attention/self/Shape_4" + op_type: "Shape" + } + node { + output: "/encoder/layer.0/attention/self/Constant_8_output_0" + name: "/encoder/layer.0/attention/self/Constant_8" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.0/attention/self/Shape_4_output_0" + input: "/encoder/layer.0/attention/self/Constant_8_output_0" + output: "/encoder/layer.0/attention/self/Gather_4_output_0" + name: "/encoder/layer.0/attention/self/Gather_4" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.0/attention/self/query/Add_output_0" + output: "/encoder/layer.0/attention/self/Shape_5_output_0" + name: "/encoder/layer.0/attention/self/Shape_5" + op_type: "Shape" + } + node { + output: "/encoder/layer.0/attention/self/Constant_9_output_0" + name: "/encoder/layer.0/attention/self/Constant_9" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.0/attention/self/Shape_5_output_0" + input: "/encoder/layer.0/attention/self/Constant_9_output_0" + output: "/encoder/layer.0/attention/self/Gather_5_output_0" + name: "/encoder/layer.0/attention/self/Gather_5" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.0/attention/self/Gather_4_output_0" + output: "/encoder/layer.0/attention/self/Unsqueeze_4_output_0" + name: "/encoder/layer.0/attention/self/Unsqueeze_4" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.0/attention/self/Gather_5_output_0" + output: "/encoder/layer.0/attention/self/Unsqueeze_5_output_0" + name: "/encoder/layer.0/attention/self/Unsqueeze_5" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.0/attention/self/Constant_10_output_0" + name: "/encoder/layer.0/attention/self/Constant_10" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.0/attention/self/Constant_11_output_0" + name: "/encoder/layer.0/attention/self/Constant_11" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.0/attention/self/Unsqueeze_4_output_0" + input: "/encoder/layer.0/attention/self/Unsqueeze_5_output_0" + input: "/encoder/layer.0/attention/self/Constant_10_output_0" + input: "/encoder/layer.0/attention/self/Constant_11_output_0" + output: "/encoder/layer.0/attention/self/Concat_2_output_0" + name: "/encoder/layer.0/attention/self/Concat_2" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.0/attention/self/query/Add_output_0" + input: "/encoder/layer.0/attention/self/Concat_2_output_0" + output: "/encoder/layer.0/attention/self/Reshape_2_output_0" + name: "/encoder/layer.0/attention/self/Reshape_2" + op_type: "Reshape" + } + node { + input: "/encoder/layer.0/attention/self/Reshape_2_output_0" + output: "/encoder/layer.0/attention/self/Transpose_1_output_0" + name: "/encoder/layer.0/attention/self/Transpose_1" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.0/attention/self/Reshape_output_0" + output: "/encoder/layer.0/attention/self/Transpose_2_output_0" + name: "/encoder/layer.0/attention/self/Transpose_2" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 3 + ints: 1 + type: INTS + } + } + node { + input: "/encoder/layer.0/attention/self/Transpose_1_output_0" + input: "/encoder/layer.0/attention/self/Transpose_2_output_0" + output: "/encoder/layer.0/attention/self/MatMul_output_0" + name: "/encoder/layer.0/attention/self/MatMul" + op_type: "MatMul" + } + node { + output: "/encoder/layer.0/attention/self/Constant_12_output_0" + name: "/encoder/layer.0/attention/self/Constant_12" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000A" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.0/attention/self/MatMul_output_0" + input: "/encoder/layer.0/attention/self/Constant_12_output_0" + output: "/encoder/layer.0/attention/self/Div_output_0" + name: "/encoder/layer.0/attention/self/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.0/attention/self/Div_output_0" + input: "/Mul_1_output_0" + output: "/encoder/layer.0/attention/self/Add_output_0" + name: "/encoder/layer.0/attention/self/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.0/attention/self/Add_output_0" + output: "/encoder/layer.0/attention/self/Softmax_output_0" + name: "/encoder/layer.0/attention/self/Softmax" + op_type: "Softmax" + attribute { + name: "axis" + i: 3 + type: INT + } + } + node { + input: "/encoder/layer.0/attention/self/Softmax_output_0" + input: "/encoder/layer.0/attention/self/Transpose_output_0" + output: "/encoder/layer.0/attention/self/MatMul_1_output_0" + name: "/encoder/layer.0/attention/self/MatMul_1" + op_type: "MatMul" + } + node { + input: "/encoder/layer.0/attention/self/MatMul_1_output_0" + output: "/encoder/layer.0/attention/self/Transpose_3_output_0" + name: "/encoder/layer.0/attention/self/Transpose_3" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.0/attention/self/Transpose_3_output_0" + output: "/encoder/layer.0/attention/self/Shape_6_output_0" + name: "/encoder/layer.0/attention/self/Shape_6" + op_type: "Shape" + } + node { + output: "/encoder/layer.0/attention/self/Constant_13_output_0" + name: "/encoder/layer.0/attention/self/Constant_13" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.0/attention/self/Shape_6_output_0" + input: "/encoder/layer.0/attention/self/Constant_13_output_0" + output: "/encoder/layer.0/attention/self/Gather_6_output_0" + name: "/encoder/layer.0/attention/self/Gather_6" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.0/attention/self/Transpose_3_output_0" + output: "/encoder/layer.0/attention/self/Shape_7_output_0" + name: "/encoder/layer.0/attention/self/Shape_7" + op_type: "Shape" + } + node { + output: "/encoder/layer.0/attention/self/Constant_14_output_0" + name: "/encoder/layer.0/attention/self/Constant_14" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.0/attention/self/Shape_7_output_0" + input: "/encoder/layer.0/attention/self/Constant_14_output_0" + output: "/encoder/layer.0/attention/self/Gather_7_output_0" + name: "/encoder/layer.0/attention/self/Gather_7" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.0/attention/self/Gather_6_output_0" + output: "/encoder/layer.0/attention/self/Unsqueeze_6_output_0" + name: "/encoder/layer.0/attention/self/Unsqueeze_6" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.0/attention/self/Gather_7_output_0" + output: "/encoder/layer.0/attention/self/Unsqueeze_7_output_0" + name: "/encoder/layer.0/attention/self/Unsqueeze_7" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.0/attention/self/Constant_15_output_0" + name: "/encoder/layer.0/attention/self/Constant_15" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\000\003\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.0/attention/self/Unsqueeze_6_output_0" + input: "/encoder/layer.0/attention/self/Unsqueeze_7_output_0" + input: "/encoder/layer.0/attention/self/Constant_15_output_0" + output: "/encoder/layer.0/attention/self/Concat_3_output_0" + name: "/encoder/layer.0/attention/self/Concat_3" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.0/attention/self/Transpose_3_output_0" + input: "/encoder/layer.0/attention/self/Concat_3_output_0" + output: "/encoder/layer.0/attention/self/Reshape_3_output_0" + name: "/encoder/layer.0/attention/self/Reshape_3" + op_type: "Reshape" + } + node { + input: "/encoder/layer.0/attention/self/Reshape_3_output_0" + input: "onnx::MatMul_1563" + output: "/encoder/layer.0/attention/output/dense/MatMul_output_0" + name: "/encoder/layer.0/attention/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.0.attention.output.dense.bias" + input: "/encoder/layer.0/attention/output/dense/MatMul_output_0" + output: "/encoder/layer.0/attention/output/dense/Add_output_0" + name: "/encoder/layer.0/attention/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.0/attention/output/dense/Add_output_0" + input: "/embeddings/LayerNorm/Add_1_output_0" + output: "/encoder/layer.0/attention/output/Add_output_0" + name: "/encoder/layer.0/attention/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.0/attention/output/Add_output_0" + output: "/encoder/layer.0/attention/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.0/attention/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.0/attention/output/Add_output_0" + input: "/encoder/layer.0/attention/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.0/attention/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.0/attention/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.0/attention/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.0/attention/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.0/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.0/attention/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.0/attention/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.0/attention/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.0/attention/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.0/attention/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.0/attention/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.0/attention/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.0/attention/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.0/attention/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.0/attention/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.0/attention/output/LayerNorm/Add_output_0" + name: "/encoder/layer.0/attention/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.0/attention/output/LayerNorm/Add_output_0" + output: "/encoder/layer.0/attention/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.0/attention/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.0/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.0/attention/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.0/attention/output/LayerNorm/Div_output_0" + name: "/encoder/layer.0/attention/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.0/attention/output/LayerNorm/Div_output_0" + input: "encoder.layer.0.attention.output.LayerNorm.weight" + output: "/encoder/layer.0/attention/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.0/attention/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.0/attention/output/LayerNorm/Mul_output_0" + input: "encoder.layer.0.attention.output.LayerNorm.bias" + output: "/encoder/layer.0/attention/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.0/attention/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.0/attention/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1564" + output: "/encoder/layer.0/intermediate/dense/MatMul_output_0" + name: "/encoder/layer.0/intermediate/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.0.intermediate.dense.bias" + input: "/encoder/layer.0/intermediate/dense/MatMul_output_0" + output: "/encoder/layer.0/intermediate/dense/Add_output_0" + name: "/encoder/layer.0/intermediate/dense/Add" + op_type: "Add" + } + node { + output: "/encoder/layer.0/intermediate/intermediate_act_fn/Constant_output_0" + name: "/encoder/layer.0/intermediate/intermediate_act_fn/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\363\004\265?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.0/intermediate/dense/Add_output_0" + input: "/encoder/layer.0/intermediate/intermediate_act_fn/Constant_output_0" + output: "/encoder/layer.0/intermediate/intermediate_act_fn/Div_output_0" + name: "/encoder/layer.0/intermediate/intermediate_act_fn/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.0/intermediate/intermediate_act_fn/Div_output_0" + output: "/encoder/layer.0/intermediate/intermediate_act_fn/Erf_output_0" + name: "/encoder/layer.0/intermediate/intermediate_act_fn/Erf" + op_type: "Erf" + } + node { + output: "/encoder/layer.0/intermediate/intermediate_act_fn/Constant_1_output_0" + name: "/encoder/layer.0/intermediate/intermediate_act_fn/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\200?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.0/intermediate/intermediate_act_fn/Erf_output_0" + input: "/encoder/layer.0/intermediate/intermediate_act_fn/Constant_1_output_0" + output: "/encoder/layer.0/intermediate/intermediate_act_fn/Add_output_0" + name: "/encoder/layer.0/intermediate/intermediate_act_fn/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.0/intermediate/dense/Add_output_0" + input: "/encoder/layer.0/intermediate/intermediate_act_fn/Add_output_0" + output: "/encoder/layer.0/intermediate/intermediate_act_fn/Mul_output_0" + name: "/encoder/layer.0/intermediate/intermediate_act_fn/Mul" + op_type: "Mul" + } + node { + output: "/encoder/layer.0/intermediate/intermediate_act_fn/Constant_2_output_0" + name: "/encoder/layer.0/intermediate/intermediate_act_fn/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.0/intermediate/intermediate_act_fn/Mul_output_0" + input: "/encoder/layer.0/intermediate/intermediate_act_fn/Constant_2_output_0" + output: "/encoder/layer.0/intermediate/intermediate_act_fn/Mul_1_output_0" + name: "/encoder/layer.0/intermediate/intermediate_act_fn/Mul_1" + op_type: "Mul" + } + node { + input: "/encoder/layer.0/intermediate/intermediate_act_fn/Mul_1_output_0" + input: "onnx::MatMul_1565" + output: "/encoder/layer.0/output/dense/MatMul_output_0" + name: "/encoder/layer.0/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.0.output.dense.bias" + input: "/encoder/layer.0/output/dense/MatMul_output_0" + output: "/encoder/layer.0/output/dense/Add_output_0" + name: "/encoder/layer.0/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.0/output/dense/Add_output_0" + input: "/encoder/layer.0/attention/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.0/output/Add_output_0" + name: "/encoder/layer.0/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.0/output/Add_output_0" + output: "/encoder/layer.0/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.0/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.0/output/Add_output_0" + input: "/encoder/layer.0/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.0/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.0/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.0/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.0/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.0/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.0/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.0/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.0/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.0/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.0/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.0/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.0/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.0/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.0/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.0/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.0/output/LayerNorm/Add_output_0" + name: "/encoder/layer.0/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.0/output/LayerNorm/Add_output_0" + output: "/encoder/layer.0/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.0/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.0/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.0/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.0/output/LayerNorm/Div_output_0" + name: "/encoder/layer.0/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.0/output/LayerNorm/Div_output_0" + input: "encoder.layer.0.output.LayerNorm.weight" + output: "/encoder/layer.0/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.0/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.0/output/LayerNorm/Mul_output_0" + input: "encoder.layer.0.output.LayerNorm.bias" + output: "/encoder/layer.0/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.0/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.0/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1566" + output: "/encoder/layer.1/attention/self/query/MatMul_output_0" + name: "/encoder/layer.1/attention/self/query/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.1.attention.self.query.bias" + input: "/encoder/layer.1/attention/self/query/MatMul_output_0" + output: "/encoder/layer.1/attention/self/query/Add_output_0" + name: "/encoder/layer.1/attention/self/query/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.0/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1567" + output: "/encoder/layer.1/attention/self/key/MatMul_output_0" + name: "/encoder/layer.1/attention/self/key/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.1.attention.self.key.bias" + input: "/encoder/layer.1/attention/self/key/MatMul_output_0" + output: "/encoder/layer.1/attention/self/key/Add_output_0" + name: "/encoder/layer.1/attention/self/key/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.1/attention/self/key/Add_output_0" + output: "/encoder/layer.1/attention/self/Shape_output_0" + name: "/encoder/layer.1/attention/self/Shape" + op_type: "Shape" + } + node { + output: "/encoder/layer.1/attention/self/Constant_output_0" + name: "/encoder/layer.1/attention/self/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.1/attention/self/Shape_output_0" + input: "/encoder/layer.1/attention/self/Constant_output_0" + output: "/encoder/layer.1/attention/self/Gather_output_0" + name: "/encoder/layer.1/attention/self/Gather" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.1/attention/self/key/Add_output_0" + output: "/encoder/layer.1/attention/self/Shape_1_output_0" + name: "/encoder/layer.1/attention/self/Shape_1" + op_type: "Shape" + } + node { + output: "/encoder/layer.1/attention/self/Constant_1_output_0" + name: "/encoder/layer.1/attention/self/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.1/attention/self/Shape_1_output_0" + input: "/encoder/layer.1/attention/self/Constant_1_output_0" + output: "/encoder/layer.1/attention/self/Gather_1_output_0" + name: "/encoder/layer.1/attention/self/Gather_1" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.1/attention/self/Gather_output_0" + output: "/encoder/layer.1/attention/self/Unsqueeze_output_0" + name: "/encoder/layer.1/attention/self/Unsqueeze" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.1/attention/self/Gather_1_output_0" + output: "/encoder/layer.1/attention/self/Unsqueeze_1_output_0" + name: "/encoder/layer.1/attention/self/Unsqueeze_1" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.1/attention/self/Constant_2_output_0" + name: "/encoder/layer.1/attention/self/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.1/attention/self/Constant_3_output_0" + name: "/encoder/layer.1/attention/self/Constant_3" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.1/attention/self/Unsqueeze_output_0" + input: "/encoder/layer.1/attention/self/Unsqueeze_1_output_0" + input: "/encoder/layer.1/attention/self/Constant_2_output_0" + input: "/encoder/layer.1/attention/self/Constant_3_output_0" + output: "/encoder/layer.1/attention/self/Concat_output_0" + name: "/encoder/layer.1/attention/self/Concat" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.1/attention/self/key/Add_output_0" + input: "/encoder/layer.1/attention/self/Concat_output_0" + output: "/encoder/layer.1/attention/self/Reshape_output_0" + name: "/encoder/layer.1/attention/self/Reshape" + op_type: "Reshape" + } + node { + input: "/encoder/layer.0/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1570" + output: "/encoder/layer.1/attention/self/value/MatMul_output_0" + name: "/encoder/layer.1/attention/self/value/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.1.attention.self.value.bias" + input: "/encoder/layer.1/attention/self/value/MatMul_output_0" + output: "/encoder/layer.1/attention/self/value/Add_output_0" + name: "/encoder/layer.1/attention/self/value/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.1/attention/self/value/Add_output_0" + output: "/encoder/layer.1/attention/self/Shape_2_output_0" + name: "/encoder/layer.1/attention/self/Shape_2" + op_type: "Shape" + } + node { + output: "/encoder/layer.1/attention/self/Constant_4_output_0" + name: "/encoder/layer.1/attention/self/Constant_4" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.1/attention/self/Shape_2_output_0" + input: "/encoder/layer.1/attention/self/Constant_4_output_0" + output: "/encoder/layer.1/attention/self/Gather_2_output_0" + name: "/encoder/layer.1/attention/self/Gather_2" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.1/attention/self/value/Add_output_0" + output: "/encoder/layer.1/attention/self/Shape_3_output_0" + name: "/encoder/layer.1/attention/self/Shape_3" + op_type: "Shape" + } + node { + output: "/encoder/layer.1/attention/self/Constant_5_output_0" + name: "/encoder/layer.1/attention/self/Constant_5" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.1/attention/self/Shape_3_output_0" + input: "/encoder/layer.1/attention/self/Constant_5_output_0" + output: "/encoder/layer.1/attention/self/Gather_3_output_0" + name: "/encoder/layer.1/attention/self/Gather_3" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.1/attention/self/Gather_2_output_0" + output: "/encoder/layer.1/attention/self/Unsqueeze_2_output_0" + name: "/encoder/layer.1/attention/self/Unsqueeze_2" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.1/attention/self/Gather_3_output_0" + output: "/encoder/layer.1/attention/self/Unsqueeze_3_output_0" + name: "/encoder/layer.1/attention/self/Unsqueeze_3" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.1/attention/self/Constant_6_output_0" + name: "/encoder/layer.1/attention/self/Constant_6" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.1/attention/self/Constant_7_output_0" + name: "/encoder/layer.1/attention/self/Constant_7" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.1/attention/self/Unsqueeze_2_output_0" + input: "/encoder/layer.1/attention/self/Unsqueeze_3_output_0" + input: "/encoder/layer.1/attention/self/Constant_6_output_0" + input: "/encoder/layer.1/attention/self/Constant_7_output_0" + output: "/encoder/layer.1/attention/self/Concat_1_output_0" + name: "/encoder/layer.1/attention/self/Concat_1" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.1/attention/self/value/Add_output_0" + input: "/encoder/layer.1/attention/self/Concat_1_output_0" + output: "/encoder/layer.1/attention/self/Reshape_1_output_0" + name: "/encoder/layer.1/attention/self/Reshape_1" + op_type: "Reshape" + } + node { + input: "/encoder/layer.1/attention/self/Reshape_1_output_0" + output: "/encoder/layer.1/attention/self/Transpose_output_0" + name: "/encoder/layer.1/attention/self/Transpose" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.1/attention/self/query/Add_output_0" + output: "/encoder/layer.1/attention/self/Shape_4_output_0" + name: "/encoder/layer.1/attention/self/Shape_4" + op_type: "Shape" + } + node { + output: "/encoder/layer.1/attention/self/Constant_8_output_0" + name: "/encoder/layer.1/attention/self/Constant_8" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.1/attention/self/Shape_4_output_0" + input: "/encoder/layer.1/attention/self/Constant_8_output_0" + output: "/encoder/layer.1/attention/self/Gather_4_output_0" + name: "/encoder/layer.1/attention/self/Gather_4" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.1/attention/self/query/Add_output_0" + output: "/encoder/layer.1/attention/self/Shape_5_output_0" + name: "/encoder/layer.1/attention/self/Shape_5" + op_type: "Shape" + } + node { + output: "/encoder/layer.1/attention/self/Constant_9_output_0" + name: "/encoder/layer.1/attention/self/Constant_9" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.1/attention/self/Shape_5_output_0" + input: "/encoder/layer.1/attention/self/Constant_9_output_0" + output: "/encoder/layer.1/attention/self/Gather_5_output_0" + name: "/encoder/layer.1/attention/self/Gather_5" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.1/attention/self/Gather_4_output_0" + output: "/encoder/layer.1/attention/self/Unsqueeze_4_output_0" + name: "/encoder/layer.1/attention/self/Unsqueeze_4" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.1/attention/self/Gather_5_output_0" + output: "/encoder/layer.1/attention/self/Unsqueeze_5_output_0" + name: "/encoder/layer.1/attention/self/Unsqueeze_5" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.1/attention/self/Constant_10_output_0" + name: "/encoder/layer.1/attention/self/Constant_10" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.1/attention/self/Constant_11_output_0" + name: "/encoder/layer.1/attention/self/Constant_11" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.1/attention/self/Unsqueeze_4_output_0" + input: "/encoder/layer.1/attention/self/Unsqueeze_5_output_0" + input: "/encoder/layer.1/attention/self/Constant_10_output_0" + input: "/encoder/layer.1/attention/self/Constant_11_output_0" + output: "/encoder/layer.1/attention/self/Concat_2_output_0" + name: "/encoder/layer.1/attention/self/Concat_2" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.1/attention/self/query/Add_output_0" + input: "/encoder/layer.1/attention/self/Concat_2_output_0" + output: "/encoder/layer.1/attention/self/Reshape_2_output_0" + name: "/encoder/layer.1/attention/self/Reshape_2" + op_type: "Reshape" + } + node { + input: "/encoder/layer.1/attention/self/Reshape_2_output_0" + output: "/encoder/layer.1/attention/self/Transpose_1_output_0" + name: "/encoder/layer.1/attention/self/Transpose_1" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.1/attention/self/Reshape_output_0" + output: "/encoder/layer.1/attention/self/Transpose_2_output_0" + name: "/encoder/layer.1/attention/self/Transpose_2" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 3 + ints: 1 + type: INTS + } + } + node { + input: "/encoder/layer.1/attention/self/Transpose_1_output_0" + input: "/encoder/layer.1/attention/self/Transpose_2_output_0" + output: "/encoder/layer.1/attention/self/MatMul_output_0" + name: "/encoder/layer.1/attention/self/MatMul" + op_type: "MatMul" + } + node { + output: "/encoder/layer.1/attention/self/Constant_12_output_0" + name: "/encoder/layer.1/attention/self/Constant_12" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000A" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.1/attention/self/MatMul_output_0" + input: "/encoder/layer.1/attention/self/Constant_12_output_0" + output: "/encoder/layer.1/attention/self/Div_output_0" + name: "/encoder/layer.1/attention/self/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.1/attention/self/Div_output_0" + input: "/Mul_1_output_0" + output: "/encoder/layer.1/attention/self/Add_output_0" + name: "/encoder/layer.1/attention/self/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.1/attention/self/Add_output_0" + output: "/encoder/layer.1/attention/self/Softmax_output_0" + name: "/encoder/layer.1/attention/self/Softmax" + op_type: "Softmax" + attribute { + name: "axis" + i: 3 + type: INT + } + } + node { + input: "/encoder/layer.1/attention/self/Softmax_output_0" + input: "/encoder/layer.1/attention/self/Transpose_output_0" + output: "/encoder/layer.1/attention/self/MatMul_1_output_0" + name: "/encoder/layer.1/attention/self/MatMul_1" + op_type: "MatMul" + } + node { + input: "/encoder/layer.1/attention/self/MatMul_1_output_0" + output: "/encoder/layer.1/attention/self/Transpose_3_output_0" + name: "/encoder/layer.1/attention/self/Transpose_3" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.1/attention/self/Transpose_3_output_0" + output: "/encoder/layer.1/attention/self/Shape_6_output_0" + name: "/encoder/layer.1/attention/self/Shape_6" + op_type: "Shape" + } + node { + output: "/encoder/layer.1/attention/self/Constant_13_output_0" + name: "/encoder/layer.1/attention/self/Constant_13" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.1/attention/self/Shape_6_output_0" + input: "/encoder/layer.1/attention/self/Constant_13_output_0" + output: "/encoder/layer.1/attention/self/Gather_6_output_0" + name: "/encoder/layer.1/attention/self/Gather_6" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.1/attention/self/Transpose_3_output_0" + output: "/encoder/layer.1/attention/self/Shape_7_output_0" + name: "/encoder/layer.1/attention/self/Shape_7" + op_type: "Shape" + } + node { + output: "/encoder/layer.1/attention/self/Constant_14_output_0" + name: "/encoder/layer.1/attention/self/Constant_14" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.1/attention/self/Shape_7_output_0" + input: "/encoder/layer.1/attention/self/Constant_14_output_0" + output: "/encoder/layer.1/attention/self/Gather_7_output_0" + name: "/encoder/layer.1/attention/self/Gather_7" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.1/attention/self/Gather_6_output_0" + output: "/encoder/layer.1/attention/self/Unsqueeze_6_output_0" + name: "/encoder/layer.1/attention/self/Unsqueeze_6" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.1/attention/self/Gather_7_output_0" + output: "/encoder/layer.1/attention/self/Unsqueeze_7_output_0" + name: "/encoder/layer.1/attention/self/Unsqueeze_7" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.1/attention/self/Constant_15_output_0" + name: "/encoder/layer.1/attention/self/Constant_15" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\000\003\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.1/attention/self/Unsqueeze_6_output_0" + input: "/encoder/layer.1/attention/self/Unsqueeze_7_output_0" + input: "/encoder/layer.1/attention/self/Constant_15_output_0" + output: "/encoder/layer.1/attention/self/Concat_3_output_0" + name: "/encoder/layer.1/attention/self/Concat_3" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.1/attention/self/Transpose_3_output_0" + input: "/encoder/layer.1/attention/self/Concat_3_output_0" + output: "/encoder/layer.1/attention/self/Reshape_3_output_0" + name: "/encoder/layer.1/attention/self/Reshape_3" + op_type: "Reshape" + } + node { + input: "/encoder/layer.1/attention/self/Reshape_3_output_0" + input: "onnx::MatMul_1576" + output: "/encoder/layer.1/attention/output/dense/MatMul_output_0" + name: "/encoder/layer.1/attention/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.1.attention.output.dense.bias" + input: "/encoder/layer.1/attention/output/dense/MatMul_output_0" + output: "/encoder/layer.1/attention/output/dense/Add_output_0" + name: "/encoder/layer.1/attention/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.1/attention/output/dense/Add_output_0" + input: "/encoder/layer.0/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.1/attention/output/Add_output_0" + name: "/encoder/layer.1/attention/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.1/attention/output/Add_output_0" + output: "/encoder/layer.1/attention/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.1/attention/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.1/attention/output/Add_output_0" + input: "/encoder/layer.1/attention/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.1/attention/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.1/attention/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.1/attention/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.1/attention/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.1/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.1/attention/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.1/attention/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.1/attention/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.1/attention/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.1/attention/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.1/attention/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.1/attention/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.1/attention/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.1/attention/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.1/attention/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.1/attention/output/LayerNorm/Add_output_0" + name: "/encoder/layer.1/attention/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.1/attention/output/LayerNorm/Add_output_0" + output: "/encoder/layer.1/attention/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.1/attention/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.1/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.1/attention/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.1/attention/output/LayerNorm/Div_output_0" + name: "/encoder/layer.1/attention/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.1/attention/output/LayerNorm/Div_output_0" + input: "encoder.layer.1.attention.output.LayerNorm.weight" + output: "/encoder/layer.1/attention/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.1/attention/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.1/attention/output/LayerNorm/Mul_output_0" + input: "encoder.layer.1.attention.output.LayerNorm.bias" + output: "/encoder/layer.1/attention/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.1/attention/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.1/attention/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1577" + output: "/encoder/layer.1/intermediate/dense/MatMul_output_0" + name: "/encoder/layer.1/intermediate/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.1.intermediate.dense.bias" + input: "/encoder/layer.1/intermediate/dense/MatMul_output_0" + output: "/encoder/layer.1/intermediate/dense/Add_output_0" + name: "/encoder/layer.1/intermediate/dense/Add" + op_type: "Add" + } + node { + output: "/encoder/layer.1/intermediate/intermediate_act_fn/Constant_output_0" + name: "/encoder/layer.1/intermediate/intermediate_act_fn/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\363\004\265?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.1/intermediate/dense/Add_output_0" + input: "/encoder/layer.1/intermediate/intermediate_act_fn/Constant_output_0" + output: "/encoder/layer.1/intermediate/intermediate_act_fn/Div_output_0" + name: "/encoder/layer.1/intermediate/intermediate_act_fn/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.1/intermediate/intermediate_act_fn/Div_output_0" + output: "/encoder/layer.1/intermediate/intermediate_act_fn/Erf_output_0" + name: "/encoder/layer.1/intermediate/intermediate_act_fn/Erf" + op_type: "Erf" + } + node { + output: "/encoder/layer.1/intermediate/intermediate_act_fn/Constant_1_output_0" + name: "/encoder/layer.1/intermediate/intermediate_act_fn/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\200?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.1/intermediate/intermediate_act_fn/Erf_output_0" + input: "/encoder/layer.1/intermediate/intermediate_act_fn/Constant_1_output_0" + output: "/encoder/layer.1/intermediate/intermediate_act_fn/Add_output_0" + name: "/encoder/layer.1/intermediate/intermediate_act_fn/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.1/intermediate/dense/Add_output_0" + input: "/encoder/layer.1/intermediate/intermediate_act_fn/Add_output_0" + output: "/encoder/layer.1/intermediate/intermediate_act_fn/Mul_output_0" + name: "/encoder/layer.1/intermediate/intermediate_act_fn/Mul" + op_type: "Mul" + } + node { + output: "/encoder/layer.1/intermediate/intermediate_act_fn/Constant_2_output_0" + name: "/encoder/layer.1/intermediate/intermediate_act_fn/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.1/intermediate/intermediate_act_fn/Mul_output_0" + input: "/encoder/layer.1/intermediate/intermediate_act_fn/Constant_2_output_0" + output: "/encoder/layer.1/intermediate/intermediate_act_fn/Mul_1_output_0" + name: "/encoder/layer.1/intermediate/intermediate_act_fn/Mul_1" + op_type: "Mul" + } + node { + input: "/encoder/layer.1/intermediate/intermediate_act_fn/Mul_1_output_0" + input: "onnx::MatMul_1578" + output: "/encoder/layer.1/output/dense/MatMul_output_0" + name: "/encoder/layer.1/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.1.output.dense.bias" + input: "/encoder/layer.1/output/dense/MatMul_output_0" + output: "/encoder/layer.1/output/dense/Add_output_0" + name: "/encoder/layer.1/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.1/output/dense/Add_output_0" + input: "/encoder/layer.1/attention/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.1/output/Add_output_0" + name: "/encoder/layer.1/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.1/output/Add_output_0" + output: "/encoder/layer.1/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.1/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.1/output/Add_output_0" + input: "/encoder/layer.1/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.1/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.1/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.1/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.1/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.1/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.1/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.1/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.1/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.1/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.1/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.1/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.1/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.1/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.1/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.1/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.1/output/LayerNorm/Add_output_0" + name: "/encoder/layer.1/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.1/output/LayerNorm/Add_output_0" + output: "/encoder/layer.1/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.1/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.1/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.1/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.1/output/LayerNorm/Div_output_0" + name: "/encoder/layer.1/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.1/output/LayerNorm/Div_output_0" + input: "encoder.layer.1.output.LayerNorm.weight" + output: "/encoder/layer.1/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.1/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.1/output/LayerNorm/Mul_output_0" + input: "encoder.layer.1.output.LayerNorm.bias" + output: "/encoder/layer.1/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.1/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.1/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1579" + output: "/encoder/layer.2/attention/self/query/MatMul_output_0" + name: "/encoder/layer.2/attention/self/query/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.2.attention.self.query.bias" + input: "/encoder/layer.2/attention/self/query/MatMul_output_0" + output: "/encoder/layer.2/attention/self/query/Add_output_0" + name: "/encoder/layer.2/attention/self/query/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.1/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1580" + output: "/encoder/layer.2/attention/self/key/MatMul_output_0" + name: "/encoder/layer.2/attention/self/key/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.2.attention.self.key.bias" + input: "/encoder/layer.2/attention/self/key/MatMul_output_0" + output: "/encoder/layer.2/attention/self/key/Add_output_0" + name: "/encoder/layer.2/attention/self/key/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.2/attention/self/key/Add_output_0" + output: "/encoder/layer.2/attention/self/Shape_output_0" + name: "/encoder/layer.2/attention/self/Shape" + op_type: "Shape" + } + node { + output: "/encoder/layer.2/attention/self/Constant_output_0" + name: "/encoder/layer.2/attention/self/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.2/attention/self/Shape_output_0" + input: "/encoder/layer.2/attention/self/Constant_output_0" + output: "/encoder/layer.2/attention/self/Gather_output_0" + name: "/encoder/layer.2/attention/self/Gather" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.2/attention/self/key/Add_output_0" + output: "/encoder/layer.2/attention/self/Shape_1_output_0" + name: "/encoder/layer.2/attention/self/Shape_1" + op_type: "Shape" + } + node { + output: "/encoder/layer.2/attention/self/Constant_1_output_0" + name: "/encoder/layer.2/attention/self/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.2/attention/self/Shape_1_output_0" + input: "/encoder/layer.2/attention/self/Constant_1_output_0" + output: "/encoder/layer.2/attention/self/Gather_1_output_0" + name: "/encoder/layer.2/attention/self/Gather_1" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.2/attention/self/Gather_output_0" + output: "/encoder/layer.2/attention/self/Unsqueeze_output_0" + name: "/encoder/layer.2/attention/self/Unsqueeze" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.2/attention/self/Gather_1_output_0" + output: "/encoder/layer.2/attention/self/Unsqueeze_1_output_0" + name: "/encoder/layer.2/attention/self/Unsqueeze_1" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.2/attention/self/Constant_2_output_0" + name: "/encoder/layer.2/attention/self/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.2/attention/self/Constant_3_output_0" + name: "/encoder/layer.2/attention/self/Constant_3" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.2/attention/self/Unsqueeze_output_0" + input: "/encoder/layer.2/attention/self/Unsqueeze_1_output_0" + input: "/encoder/layer.2/attention/self/Constant_2_output_0" + input: "/encoder/layer.2/attention/self/Constant_3_output_0" + output: "/encoder/layer.2/attention/self/Concat_output_0" + name: "/encoder/layer.2/attention/self/Concat" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.2/attention/self/key/Add_output_0" + input: "/encoder/layer.2/attention/self/Concat_output_0" + output: "/encoder/layer.2/attention/self/Reshape_output_0" + name: "/encoder/layer.2/attention/self/Reshape" + op_type: "Reshape" + } + node { + input: "/encoder/layer.1/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1583" + output: "/encoder/layer.2/attention/self/value/MatMul_output_0" + name: "/encoder/layer.2/attention/self/value/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.2.attention.self.value.bias" + input: "/encoder/layer.2/attention/self/value/MatMul_output_0" + output: "/encoder/layer.2/attention/self/value/Add_output_0" + name: "/encoder/layer.2/attention/self/value/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.2/attention/self/value/Add_output_0" + output: "/encoder/layer.2/attention/self/Shape_2_output_0" + name: "/encoder/layer.2/attention/self/Shape_2" + op_type: "Shape" + } + node { + output: "/encoder/layer.2/attention/self/Constant_4_output_0" + name: "/encoder/layer.2/attention/self/Constant_4" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.2/attention/self/Shape_2_output_0" + input: "/encoder/layer.2/attention/self/Constant_4_output_0" + output: "/encoder/layer.2/attention/self/Gather_2_output_0" + name: "/encoder/layer.2/attention/self/Gather_2" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.2/attention/self/value/Add_output_0" + output: "/encoder/layer.2/attention/self/Shape_3_output_0" + name: "/encoder/layer.2/attention/self/Shape_3" + op_type: "Shape" + } + node { + output: "/encoder/layer.2/attention/self/Constant_5_output_0" + name: "/encoder/layer.2/attention/self/Constant_5" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.2/attention/self/Shape_3_output_0" + input: "/encoder/layer.2/attention/self/Constant_5_output_0" + output: "/encoder/layer.2/attention/self/Gather_3_output_0" + name: "/encoder/layer.2/attention/self/Gather_3" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.2/attention/self/Gather_2_output_0" + output: "/encoder/layer.2/attention/self/Unsqueeze_2_output_0" + name: "/encoder/layer.2/attention/self/Unsqueeze_2" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.2/attention/self/Gather_3_output_0" + output: "/encoder/layer.2/attention/self/Unsqueeze_3_output_0" + name: "/encoder/layer.2/attention/self/Unsqueeze_3" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.2/attention/self/Constant_6_output_0" + name: "/encoder/layer.2/attention/self/Constant_6" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.2/attention/self/Constant_7_output_0" + name: "/encoder/layer.2/attention/self/Constant_7" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.2/attention/self/Unsqueeze_2_output_0" + input: "/encoder/layer.2/attention/self/Unsqueeze_3_output_0" + input: "/encoder/layer.2/attention/self/Constant_6_output_0" + input: "/encoder/layer.2/attention/self/Constant_7_output_0" + output: "/encoder/layer.2/attention/self/Concat_1_output_0" + name: "/encoder/layer.2/attention/self/Concat_1" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.2/attention/self/value/Add_output_0" + input: "/encoder/layer.2/attention/self/Concat_1_output_0" + output: "/encoder/layer.2/attention/self/Reshape_1_output_0" + name: "/encoder/layer.2/attention/self/Reshape_1" + op_type: "Reshape" + } + node { + input: "/encoder/layer.2/attention/self/Reshape_1_output_0" + output: "/encoder/layer.2/attention/self/Transpose_output_0" + name: "/encoder/layer.2/attention/self/Transpose" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.2/attention/self/query/Add_output_0" + output: "/encoder/layer.2/attention/self/Shape_4_output_0" + name: "/encoder/layer.2/attention/self/Shape_4" + op_type: "Shape" + } + node { + output: "/encoder/layer.2/attention/self/Constant_8_output_0" + name: "/encoder/layer.2/attention/self/Constant_8" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.2/attention/self/Shape_4_output_0" + input: "/encoder/layer.2/attention/self/Constant_8_output_0" + output: "/encoder/layer.2/attention/self/Gather_4_output_0" + name: "/encoder/layer.2/attention/self/Gather_4" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.2/attention/self/query/Add_output_0" + output: "/encoder/layer.2/attention/self/Shape_5_output_0" + name: "/encoder/layer.2/attention/self/Shape_5" + op_type: "Shape" + } + node { + output: "/encoder/layer.2/attention/self/Constant_9_output_0" + name: "/encoder/layer.2/attention/self/Constant_9" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.2/attention/self/Shape_5_output_0" + input: "/encoder/layer.2/attention/self/Constant_9_output_0" + output: "/encoder/layer.2/attention/self/Gather_5_output_0" + name: "/encoder/layer.2/attention/self/Gather_5" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.2/attention/self/Gather_4_output_0" + output: "/encoder/layer.2/attention/self/Unsqueeze_4_output_0" + name: "/encoder/layer.2/attention/self/Unsqueeze_4" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.2/attention/self/Gather_5_output_0" + output: "/encoder/layer.2/attention/self/Unsqueeze_5_output_0" + name: "/encoder/layer.2/attention/self/Unsqueeze_5" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.2/attention/self/Constant_10_output_0" + name: "/encoder/layer.2/attention/self/Constant_10" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.2/attention/self/Constant_11_output_0" + name: "/encoder/layer.2/attention/self/Constant_11" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.2/attention/self/Unsqueeze_4_output_0" + input: "/encoder/layer.2/attention/self/Unsqueeze_5_output_0" + input: "/encoder/layer.2/attention/self/Constant_10_output_0" + input: "/encoder/layer.2/attention/self/Constant_11_output_0" + output: "/encoder/layer.2/attention/self/Concat_2_output_0" + name: "/encoder/layer.2/attention/self/Concat_2" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.2/attention/self/query/Add_output_0" + input: "/encoder/layer.2/attention/self/Concat_2_output_0" + output: "/encoder/layer.2/attention/self/Reshape_2_output_0" + name: "/encoder/layer.2/attention/self/Reshape_2" + op_type: "Reshape" + } + node { + input: "/encoder/layer.2/attention/self/Reshape_2_output_0" + output: "/encoder/layer.2/attention/self/Transpose_1_output_0" + name: "/encoder/layer.2/attention/self/Transpose_1" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.2/attention/self/Reshape_output_0" + output: "/encoder/layer.2/attention/self/Transpose_2_output_0" + name: "/encoder/layer.2/attention/self/Transpose_2" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 3 + ints: 1 + type: INTS + } + } + node { + input: "/encoder/layer.2/attention/self/Transpose_1_output_0" + input: "/encoder/layer.2/attention/self/Transpose_2_output_0" + output: "/encoder/layer.2/attention/self/MatMul_output_0" + name: "/encoder/layer.2/attention/self/MatMul" + op_type: "MatMul" + } + node { + output: "/encoder/layer.2/attention/self/Constant_12_output_0" + name: "/encoder/layer.2/attention/self/Constant_12" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000A" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.2/attention/self/MatMul_output_0" + input: "/encoder/layer.2/attention/self/Constant_12_output_0" + output: "/encoder/layer.2/attention/self/Div_output_0" + name: "/encoder/layer.2/attention/self/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.2/attention/self/Div_output_0" + input: "/Mul_1_output_0" + output: "/encoder/layer.2/attention/self/Add_output_0" + name: "/encoder/layer.2/attention/self/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.2/attention/self/Add_output_0" + output: "/encoder/layer.2/attention/self/Softmax_output_0" + name: "/encoder/layer.2/attention/self/Softmax" + op_type: "Softmax" + attribute { + name: "axis" + i: 3 + type: INT + } + } + node { + input: "/encoder/layer.2/attention/self/Softmax_output_0" + input: "/encoder/layer.2/attention/self/Transpose_output_0" + output: "/encoder/layer.2/attention/self/MatMul_1_output_0" + name: "/encoder/layer.2/attention/self/MatMul_1" + op_type: "MatMul" + } + node { + input: "/encoder/layer.2/attention/self/MatMul_1_output_0" + output: "/encoder/layer.2/attention/self/Transpose_3_output_0" + name: "/encoder/layer.2/attention/self/Transpose_3" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.2/attention/self/Transpose_3_output_0" + output: "/encoder/layer.2/attention/self/Shape_6_output_0" + name: "/encoder/layer.2/attention/self/Shape_6" + op_type: "Shape" + } + node { + output: "/encoder/layer.2/attention/self/Constant_13_output_0" + name: "/encoder/layer.2/attention/self/Constant_13" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.2/attention/self/Shape_6_output_0" + input: "/encoder/layer.2/attention/self/Constant_13_output_0" + output: "/encoder/layer.2/attention/self/Gather_6_output_0" + name: "/encoder/layer.2/attention/self/Gather_6" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.2/attention/self/Transpose_3_output_0" + output: "/encoder/layer.2/attention/self/Shape_7_output_0" + name: "/encoder/layer.2/attention/self/Shape_7" + op_type: "Shape" + } + node { + output: "/encoder/layer.2/attention/self/Constant_14_output_0" + name: "/encoder/layer.2/attention/self/Constant_14" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.2/attention/self/Shape_7_output_0" + input: "/encoder/layer.2/attention/self/Constant_14_output_0" + output: "/encoder/layer.2/attention/self/Gather_7_output_0" + name: "/encoder/layer.2/attention/self/Gather_7" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.2/attention/self/Gather_6_output_0" + output: "/encoder/layer.2/attention/self/Unsqueeze_6_output_0" + name: "/encoder/layer.2/attention/self/Unsqueeze_6" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.2/attention/self/Gather_7_output_0" + output: "/encoder/layer.2/attention/self/Unsqueeze_7_output_0" + name: "/encoder/layer.2/attention/self/Unsqueeze_7" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.2/attention/self/Constant_15_output_0" + name: "/encoder/layer.2/attention/self/Constant_15" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\000\003\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.2/attention/self/Unsqueeze_6_output_0" + input: "/encoder/layer.2/attention/self/Unsqueeze_7_output_0" + input: "/encoder/layer.2/attention/self/Constant_15_output_0" + output: "/encoder/layer.2/attention/self/Concat_3_output_0" + name: "/encoder/layer.2/attention/self/Concat_3" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.2/attention/self/Transpose_3_output_0" + input: "/encoder/layer.2/attention/self/Concat_3_output_0" + output: "/encoder/layer.2/attention/self/Reshape_3_output_0" + name: "/encoder/layer.2/attention/self/Reshape_3" + op_type: "Reshape" + } + node { + input: "/encoder/layer.2/attention/self/Reshape_3_output_0" + input: "onnx::MatMul_1589" + output: "/encoder/layer.2/attention/output/dense/MatMul_output_0" + name: "/encoder/layer.2/attention/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.2.attention.output.dense.bias" + input: "/encoder/layer.2/attention/output/dense/MatMul_output_0" + output: "/encoder/layer.2/attention/output/dense/Add_output_0" + name: "/encoder/layer.2/attention/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.2/attention/output/dense/Add_output_0" + input: "/encoder/layer.1/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.2/attention/output/Add_output_0" + name: "/encoder/layer.2/attention/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.2/attention/output/Add_output_0" + output: "/encoder/layer.2/attention/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.2/attention/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.2/attention/output/Add_output_0" + input: "/encoder/layer.2/attention/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.2/attention/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.2/attention/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.2/attention/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.2/attention/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.2/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.2/attention/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.2/attention/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.2/attention/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.2/attention/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.2/attention/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.2/attention/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.2/attention/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.2/attention/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.2/attention/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.2/attention/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.2/attention/output/LayerNorm/Add_output_0" + name: "/encoder/layer.2/attention/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.2/attention/output/LayerNorm/Add_output_0" + output: "/encoder/layer.2/attention/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.2/attention/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.2/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.2/attention/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.2/attention/output/LayerNorm/Div_output_0" + name: "/encoder/layer.2/attention/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.2/attention/output/LayerNorm/Div_output_0" + input: "encoder.layer.2.attention.output.LayerNorm.weight" + output: "/encoder/layer.2/attention/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.2/attention/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.2/attention/output/LayerNorm/Mul_output_0" + input: "encoder.layer.2.attention.output.LayerNorm.bias" + output: "/encoder/layer.2/attention/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.2/attention/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.2/attention/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1590" + output: "/encoder/layer.2/intermediate/dense/MatMul_output_0" + name: "/encoder/layer.2/intermediate/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.2.intermediate.dense.bias" + input: "/encoder/layer.2/intermediate/dense/MatMul_output_0" + output: "/encoder/layer.2/intermediate/dense/Add_output_0" + name: "/encoder/layer.2/intermediate/dense/Add" + op_type: "Add" + } + node { + output: "/encoder/layer.2/intermediate/intermediate_act_fn/Constant_output_0" + name: "/encoder/layer.2/intermediate/intermediate_act_fn/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\363\004\265?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.2/intermediate/dense/Add_output_0" + input: "/encoder/layer.2/intermediate/intermediate_act_fn/Constant_output_0" + output: "/encoder/layer.2/intermediate/intermediate_act_fn/Div_output_0" + name: "/encoder/layer.2/intermediate/intermediate_act_fn/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.2/intermediate/intermediate_act_fn/Div_output_0" + output: "/encoder/layer.2/intermediate/intermediate_act_fn/Erf_output_0" + name: "/encoder/layer.2/intermediate/intermediate_act_fn/Erf" + op_type: "Erf" + } + node { + output: "/encoder/layer.2/intermediate/intermediate_act_fn/Constant_1_output_0" + name: "/encoder/layer.2/intermediate/intermediate_act_fn/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\200?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.2/intermediate/intermediate_act_fn/Erf_output_0" + input: "/encoder/layer.2/intermediate/intermediate_act_fn/Constant_1_output_0" + output: "/encoder/layer.2/intermediate/intermediate_act_fn/Add_output_0" + name: "/encoder/layer.2/intermediate/intermediate_act_fn/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.2/intermediate/dense/Add_output_0" + input: "/encoder/layer.2/intermediate/intermediate_act_fn/Add_output_0" + output: "/encoder/layer.2/intermediate/intermediate_act_fn/Mul_output_0" + name: "/encoder/layer.2/intermediate/intermediate_act_fn/Mul" + op_type: "Mul" + } + node { + output: "/encoder/layer.2/intermediate/intermediate_act_fn/Constant_2_output_0" + name: "/encoder/layer.2/intermediate/intermediate_act_fn/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.2/intermediate/intermediate_act_fn/Mul_output_0" + input: "/encoder/layer.2/intermediate/intermediate_act_fn/Constant_2_output_0" + output: "/encoder/layer.2/intermediate/intermediate_act_fn/Mul_1_output_0" + name: "/encoder/layer.2/intermediate/intermediate_act_fn/Mul_1" + op_type: "Mul" + } + node { + input: "/encoder/layer.2/intermediate/intermediate_act_fn/Mul_1_output_0" + input: "onnx::MatMul_1591" + output: "/encoder/layer.2/output/dense/MatMul_output_0" + name: "/encoder/layer.2/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.2.output.dense.bias" + input: "/encoder/layer.2/output/dense/MatMul_output_0" + output: "/encoder/layer.2/output/dense/Add_output_0" + name: "/encoder/layer.2/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.2/output/dense/Add_output_0" + input: "/encoder/layer.2/attention/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.2/output/Add_output_0" + name: "/encoder/layer.2/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.2/output/Add_output_0" + output: "/encoder/layer.2/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.2/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.2/output/Add_output_0" + input: "/encoder/layer.2/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.2/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.2/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.2/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.2/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.2/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.2/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.2/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.2/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.2/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.2/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.2/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.2/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.2/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.2/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.2/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.2/output/LayerNorm/Add_output_0" + name: "/encoder/layer.2/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.2/output/LayerNorm/Add_output_0" + output: "/encoder/layer.2/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.2/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.2/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.2/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.2/output/LayerNorm/Div_output_0" + name: "/encoder/layer.2/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.2/output/LayerNorm/Div_output_0" + input: "encoder.layer.2.output.LayerNorm.weight" + output: "/encoder/layer.2/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.2/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.2/output/LayerNorm/Mul_output_0" + input: "encoder.layer.2.output.LayerNorm.bias" + output: "/encoder/layer.2/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.2/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.2/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1592" + output: "/encoder/layer.3/attention/self/query/MatMul_output_0" + name: "/encoder/layer.3/attention/self/query/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.3.attention.self.query.bias" + input: "/encoder/layer.3/attention/self/query/MatMul_output_0" + output: "/encoder/layer.3/attention/self/query/Add_output_0" + name: "/encoder/layer.3/attention/self/query/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.2/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1593" + output: "/encoder/layer.3/attention/self/key/MatMul_output_0" + name: "/encoder/layer.3/attention/self/key/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.3.attention.self.key.bias" + input: "/encoder/layer.3/attention/self/key/MatMul_output_0" + output: "/encoder/layer.3/attention/self/key/Add_output_0" + name: "/encoder/layer.3/attention/self/key/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.3/attention/self/key/Add_output_0" + output: "/encoder/layer.3/attention/self/Shape_output_0" + name: "/encoder/layer.3/attention/self/Shape" + op_type: "Shape" + } + node { + output: "/encoder/layer.3/attention/self/Constant_output_0" + name: "/encoder/layer.3/attention/self/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.3/attention/self/Shape_output_0" + input: "/encoder/layer.3/attention/self/Constant_output_0" + output: "/encoder/layer.3/attention/self/Gather_output_0" + name: "/encoder/layer.3/attention/self/Gather" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.3/attention/self/key/Add_output_0" + output: "/encoder/layer.3/attention/self/Shape_1_output_0" + name: "/encoder/layer.3/attention/self/Shape_1" + op_type: "Shape" + } + node { + output: "/encoder/layer.3/attention/self/Constant_1_output_0" + name: "/encoder/layer.3/attention/self/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.3/attention/self/Shape_1_output_0" + input: "/encoder/layer.3/attention/self/Constant_1_output_0" + output: "/encoder/layer.3/attention/self/Gather_1_output_0" + name: "/encoder/layer.3/attention/self/Gather_1" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.3/attention/self/Gather_output_0" + output: "/encoder/layer.3/attention/self/Unsqueeze_output_0" + name: "/encoder/layer.3/attention/self/Unsqueeze" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.3/attention/self/Gather_1_output_0" + output: "/encoder/layer.3/attention/self/Unsqueeze_1_output_0" + name: "/encoder/layer.3/attention/self/Unsqueeze_1" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.3/attention/self/Constant_2_output_0" + name: "/encoder/layer.3/attention/self/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.3/attention/self/Constant_3_output_0" + name: "/encoder/layer.3/attention/self/Constant_3" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.3/attention/self/Unsqueeze_output_0" + input: "/encoder/layer.3/attention/self/Unsqueeze_1_output_0" + input: "/encoder/layer.3/attention/self/Constant_2_output_0" + input: "/encoder/layer.3/attention/self/Constant_3_output_0" + output: "/encoder/layer.3/attention/self/Concat_output_0" + name: "/encoder/layer.3/attention/self/Concat" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.3/attention/self/key/Add_output_0" + input: "/encoder/layer.3/attention/self/Concat_output_0" + output: "/encoder/layer.3/attention/self/Reshape_output_0" + name: "/encoder/layer.3/attention/self/Reshape" + op_type: "Reshape" + } + node { + input: "/encoder/layer.2/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1596" + output: "/encoder/layer.3/attention/self/value/MatMul_output_0" + name: "/encoder/layer.3/attention/self/value/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.3.attention.self.value.bias" + input: "/encoder/layer.3/attention/self/value/MatMul_output_0" + output: "/encoder/layer.3/attention/self/value/Add_output_0" + name: "/encoder/layer.3/attention/self/value/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.3/attention/self/value/Add_output_0" + output: "/encoder/layer.3/attention/self/Shape_2_output_0" + name: "/encoder/layer.3/attention/self/Shape_2" + op_type: "Shape" + } + node { + output: "/encoder/layer.3/attention/self/Constant_4_output_0" + name: "/encoder/layer.3/attention/self/Constant_4" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.3/attention/self/Shape_2_output_0" + input: "/encoder/layer.3/attention/self/Constant_4_output_0" + output: "/encoder/layer.3/attention/self/Gather_2_output_0" + name: "/encoder/layer.3/attention/self/Gather_2" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.3/attention/self/value/Add_output_0" + output: "/encoder/layer.3/attention/self/Shape_3_output_0" + name: "/encoder/layer.3/attention/self/Shape_3" + op_type: "Shape" + } + node { + output: "/encoder/layer.3/attention/self/Constant_5_output_0" + name: "/encoder/layer.3/attention/self/Constant_5" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.3/attention/self/Shape_3_output_0" + input: "/encoder/layer.3/attention/self/Constant_5_output_0" + output: "/encoder/layer.3/attention/self/Gather_3_output_0" + name: "/encoder/layer.3/attention/self/Gather_3" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.3/attention/self/Gather_2_output_0" + output: "/encoder/layer.3/attention/self/Unsqueeze_2_output_0" + name: "/encoder/layer.3/attention/self/Unsqueeze_2" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.3/attention/self/Gather_3_output_0" + output: "/encoder/layer.3/attention/self/Unsqueeze_3_output_0" + name: "/encoder/layer.3/attention/self/Unsqueeze_3" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.3/attention/self/Constant_6_output_0" + name: "/encoder/layer.3/attention/self/Constant_6" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.3/attention/self/Constant_7_output_0" + name: "/encoder/layer.3/attention/self/Constant_7" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.3/attention/self/Unsqueeze_2_output_0" + input: "/encoder/layer.3/attention/self/Unsqueeze_3_output_0" + input: "/encoder/layer.3/attention/self/Constant_6_output_0" + input: "/encoder/layer.3/attention/self/Constant_7_output_0" + output: "/encoder/layer.3/attention/self/Concat_1_output_0" + name: "/encoder/layer.3/attention/self/Concat_1" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.3/attention/self/value/Add_output_0" + input: "/encoder/layer.3/attention/self/Concat_1_output_0" + output: "/encoder/layer.3/attention/self/Reshape_1_output_0" + name: "/encoder/layer.3/attention/self/Reshape_1" + op_type: "Reshape" + } + node { + input: "/encoder/layer.3/attention/self/Reshape_1_output_0" + output: "/encoder/layer.3/attention/self/Transpose_output_0" + name: "/encoder/layer.3/attention/self/Transpose" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.3/attention/self/query/Add_output_0" + output: "/encoder/layer.3/attention/self/Shape_4_output_0" + name: "/encoder/layer.3/attention/self/Shape_4" + op_type: "Shape" + } + node { + output: "/encoder/layer.3/attention/self/Constant_8_output_0" + name: "/encoder/layer.3/attention/self/Constant_8" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.3/attention/self/Shape_4_output_0" + input: "/encoder/layer.3/attention/self/Constant_8_output_0" + output: "/encoder/layer.3/attention/self/Gather_4_output_0" + name: "/encoder/layer.3/attention/self/Gather_4" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.3/attention/self/query/Add_output_0" + output: "/encoder/layer.3/attention/self/Shape_5_output_0" + name: "/encoder/layer.3/attention/self/Shape_5" + op_type: "Shape" + } + node { + output: "/encoder/layer.3/attention/self/Constant_9_output_0" + name: "/encoder/layer.3/attention/self/Constant_9" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.3/attention/self/Shape_5_output_0" + input: "/encoder/layer.3/attention/self/Constant_9_output_0" + output: "/encoder/layer.3/attention/self/Gather_5_output_0" + name: "/encoder/layer.3/attention/self/Gather_5" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.3/attention/self/Gather_4_output_0" + output: "/encoder/layer.3/attention/self/Unsqueeze_4_output_0" + name: "/encoder/layer.3/attention/self/Unsqueeze_4" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.3/attention/self/Gather_5_output_0" + output: "/encoder/layer.3/attention/self/Unsqueeze_5_output_0" + name: "/encoder/layer.3/attention/self/Unsqueeze_5" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.3/attention/self/Constant_10_output_0" + name: "/encoder/layer.3/attention/self/Constant_10" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.3/attention/self/Constant_11_output_0" + name: "/encoder/layer.3/attention/self/Constant_11" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.3/attention/self/Unsqueeze_4_output_0" + input: "/encoder/layer.3/attention/self/Unsqueeze_5_output_0" + input: "/encoder/layer.3/attention/self/Constant_10_output_0" + input: "/encoder/layer.3/attention/self/Constant_11_output_0" + output: "/encoder/layer.3/attention/self/Concat_2_output_0" + name: "/encoder/layer.3/attention/self/Concat_2" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.3/attention/self/query/Add_output_0" + input: "/encoder/layer.3/attention/self/Concat_2_output_0" + output: "/encoder/layer.3/attention/self/Reshape_2_output_0" + name: "/encoder/layer.3/attention/self/Reshape_2" + op_type: "Reshape" + } + node { + input: "/encoder/layer.3/attention/self/Reshape_2_output_0" + output: "/encoder/layer.3/attention/self/Transpose_1_output_0" + name: "/encoder/layer.3/attention/self/Transpose_1" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.3/attention/self/Reshape_output_0" + output: "/encoder/layer.3/attention/self/Transpose_2_output_0" + name: "/encoder/layer.3/attention/self/Transpose_2" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 3 + ints: 1 + type: INTS + } + } + node { + input: "/encoder/layer.3/attention/self/Transpose_1_output_0" + input: "/encoder/layer.3/attention/self/Transpose_2_output_0" + output: "/encoder/layer.3/attention/self/MatMul_output_0" + name: "/encoder/layer.3/attention/self/MatMul" + op_type: "MatMul" + } + node { + output: "/encoder/layer.3/attention/self/Constant_12_output_0" + name: "/encoder/layer.3/attention/self/Constant_12" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000A" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.3/attention/self/MatMul_output_0" + input: "/encoder/layer.3/attention/self/Constant_12_output_0" + output: "/encoder/layer.3/attention/self/Div_output_0" + name: "/encoder/layer.3/attention/self/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.3/attention/self/Div_output_0" + input: "/Mul_1_output_0" + output: "/encoder/layer.3/attention/self/Add_output_0" + name: "/encoder/layer.3/attention/self/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.3/attention/self/Add_output_0" + output: "/encoder/layer.3/attention/self/Softmax_output_0" + name: "/encoder/layer.3/attention/self/Softmax" + op_type: "Softmax" + attribute { + name: "axis" + i: 3 + type: INT + } + } + node { + input: "/encoder/layer.3/attention/self/Softmax_output_0" + input: "/encoder/layer.3/attention/self/Transpose_output_0" + output: "/encoder/layer.3/attention/self/MatMul_1_output_0" + name: "/encoder/layer.3/attention/self/MatMul_1" + op_type: "MatMul" + } + node { + input: "/encoder/layer.3/attention/self/MatMul_1_output_0" + output: "/encoder/layer.3/attention/self/Transpose_3_output_0" + name: "/encoder/layer.3/attention/self/Transpose_3" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.3/attention/self/Transpose_3_output_0" + output: "/encoder/layer.3/attention/self/Shape_6_output_0" + name: "/encoder/layer.3/attention/self/Shape_6" + op_type: "Shape" + } + node { + output: "/encoder/layer.3/attention/self/Constant_13_output_0" + name: "/encoder/layer.3/attention/self/Constant_13" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.3/attention/self/Shape_6_output_0" + input: "/encoder/layer.3/attention/self/Constant_13_output_0" + output: "/encoder/layer.3/attention/self/Gather_6_output_0" + name: "/encoder/layer.3/attention/self/Gather_6" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.3/attention/self/Transpose_3_output_0" + output: "/encoder/layer.3/attention/self/Shape_7_output_0" + name: "/encoder/layer.3/attention/self/Shape_7" + op_type: "Shape" + } + node { + output: "/encoder/layer.3/attention/self/Constant_14_output_0" + name: "/encoder/layer.3/attention/self/Constant_14" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.3/attention/self/Shape_7_output_0" + input: "/encoder/layer.3/attention/self/Constant_14_output_0" + output: "/encoder/layer.3/attention/self/Gather_7_output_0" + name: "/encoder/layer.3/attention/self/Gather_7" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.3/attention/self/Gather_6_output_0" + output: "/encoder/layer.3/attention/self/Unsqueeze_6_output_0" + name: "/encoder/layer.3/attention/self/Unsqueeze_6" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.3/attention/self/Gather_7_output_0" + output: "/encoder/layer.3/attention/self/Unsqueeze_7_output_0" + name: "/encoder/layer.3/attention/self/Unsqueeze_7" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.3/attention/self/Constant_15_output_0" + name: "/encoder/layer.3/attention/self/Constant_15" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\000\003\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.3/attention/self/Unsqueeze_6_output_0" + input: "/encoder/layer.3/attention/self/Unsqueeze_7_output_0" + input: "/encoder/layer.3/attention/self/Constant_15_output_0" + output: "/encoder/layer.3/attention/self/Concat_3_output_0" + name: "/encoder/layer.3/attention/self/Concat_3" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.3/attention/self/Transpose_3_output_0" + input: "/encoder/layer.3/attention/self/Concat_3_output_0" + output: "/encoder/layer.3/attention/self/Reshape_3_output_0" + name: "/encoder/layer.3/attention/self/Reshape_3" + op_type: "Reshape" + } + node { + input: "/encoder/layer.3/attention/self/Reshape_3_output_0" + input: "onnx::MatMul_1602" + output: "/encoder/layer.3/attention/output/dense/MatMul_output_0" + name: "/encoder/layer.3/attention/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.3.attention.output.dense.bias" + input: "/encoder/layer.3/attention/output/dense/MatMul_output_0" + output: "/encoder/layer.3/attention/output/dense/Add_output_0" + name: "/encoder/layer.3/attention/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.3/attention/output/dense/Add_output_0" + input: "/encoder/layer.2/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.3/attention/output/Add_output_0" + name: "/encoder/layer.3/attention/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.3/attention/output/Add_output_0" + output: "/encoder/layer.3/attention/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.3/attention/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.3/attention/output/Add_output_0" + input: "/encoder/layer.3/attention/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.3/attention/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.3/attention/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.3/attention/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.3/attention/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.3/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.3/attention/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.3/attention/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.3/attention/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.3/attention/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.3/attention/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.3/attention/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.3/attention/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.3/attention/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.3/attention/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.3/attention/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.3/attention/output/LayerNorm/Add_output_0" + name: "/encoder/layer.3/attention/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.3/attention/output/LayerNorm/Add_output_0" + output: "/encoder/layer.3/attention/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.3/attention/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.3/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.3/attention/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.3/attention/output/LayerNorm/Div_output_0" + name: "/encoder/layer.3/attention/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.3/attention/output/LayerNorm/Div_output_0" + input: "encoder.layer.3.attention.output.LayerNorm.weight" + output: "/encoder/layer.3/attention/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.3/attention/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.3/attention/output/LayerNorm/Mul_output_0" + input: "encoder.layer.3.attention.output.LayerNorm.bias" + output: "/encoder/layer.3/attention/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.3/attention/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.3/attention/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1603" + output: "/encoder/layer.3/intermediate/dense/MatMul_output_0" + name: "/encoder/layer.3/intermediate/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.3.intermediate.dense.bias" + input: "/encoder/layer.3/intermediate/dense/MatMul_output_0" + output: "/encoder/layer.3/intermediate/dense/Add_output_0" + name: "/encoder/layer.3/intermediate/dense/Add" + op_type: "Add" + } + node { + output: "/encoder/layer.3/intermediate/intermediate_act_fn/Constant_output_0" + name: "/encoder/layer.3/intermediate/intermediate_act_fn/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\363\004\265?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.3/intermediate/dense/Add_output_0" + input: "/encoder/layer.3/intermediate/intermediate_act_fn/Constant_output_0" + output: "/encoder/layer.3/intermediate/intermediate_act_fn/Div_output_0" + name: "/encoder/layer.3/intermediate/intermediate_act_fn/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.3/intermediate/intermediate_act_fn/Div_output_0" + output: "/encoder/layer.3/intermediate/intermediate_act_fn/Erf_output_0" + name: "/encoder/layer.3/intermediate/intermediate_act_fn/Erf" + op_type: "Erf" + } + node { + output: "/encoder/layer.3/intermediate/intermediate_act_fn/Constant_1_output_0" + name: "/encoder/layer.3/intermediate/intermediate_act_fn/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\200?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.3/intermediate/intermediate_act_fn/Erf_output_0" + input: "/encoder/layer.3/intermediate/intermediate_act_fn/Constant_1_output_0" + output: "/encoder/layer.3/intermediate/intermediate_act_fn/Add_output_0" + name: "/encoder/layer.3/intermediate/intermediate_act_fn/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.3/intermediate/dense/Add_output_0" + input: "/encoder/layer.3/intermediate/intermediate_act_fn/Add_output_0" + output: "/encoder/layer.3/intermediate/intermediate_act_fn/Mul_output_0" + name: "/encoder/layer.3/intermediate/intermediate_act_fn/Mul" + op_type: "Mul" + } + node { + output: "/encoder/layer.3/intermediate/intermediate_act_fn/Constant_2_output_0" + name: "/encoder/layer.3/intermediate/intermediate_act_fn/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.3/intermediate/intermediate_act_fn/Mul_output_0" + input: "/encoder/layer.3/intermediate/intermediate_act_fn/Constant_2_output_0" + output: "/encoder/layer.3/intermediate/intermediate_act_fn/Mul_1_output_0" + name: "/encoder/layer.3/intermediate/intermediate_act_fn/Mul_1" + op_type: "Mul" + } + node { + input: "/encoder/layer.3/intermediate/intermediate_act_fn/Mul_1_output_0" + input: "onnx::MatMul_1604" + output: "/encoder/layer.3/output/dense/MatMul_output_0" + name: "/encoder/layer.3/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.3.output.dense.bias" + input: "/encoder/layer.3/output/dense/MatMul_output_0" + output: "/encoder/layer.3/output/dense/Add_output_0" + name: "/encoder/layer.3/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.3/output/dense/Add_output_0" + input: "/encoder/layer.3/attention/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.3/output/Add_output_0" + name: "/encoder/layer.3/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.3/output/Add_output_0" + output: "/encoder/layer.3/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.3/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.3/output/Add_output_0" + input: "/encoder/layer.3/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.3/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.3/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.3/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.3/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.3/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.3/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.3/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.3/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.3/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.3/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.3/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.3/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.3/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.3/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.3/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.3/output/LayerNorm/Add_output_0" + name: "/encoder/layer.3/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.3/output/LayerNorm/Add_output_0" + output: "/encoder/layer.3/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.3/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.3/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.3/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.3/output/LayerNorm/Div_output_0" + name: "/encoder/layer.3/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.3/output/LayerNorm/Div_output_0" + input: "encoder.layer.3.output.LayerNorm.weight" + output: "/encoder/layer.3/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.3/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.3/output/LayerNorm/Mul_output_0" + input: "encoder.layer.3.output.LayerNorm.bias" + output: "/encoder/layer.3/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.3/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.3/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1605" + output: "/encoder/layer.4/attention/self/query/MatMul_output_0" + name: "/encoder/layer.4/attention/self/query/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.4.attention.self.query.bias" + input: "/encoder/layer.4/attention/self/query/MatMul_output_0" + output: "/encoder/layer.4/attention/self/query/Add_output_0" + name: "/encoder/layer.4/attention/self/query/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.3/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1606" + output: "/encoder/layer.4/attention/self/key/MatMul_output_0" + name: "/encoder/layer.4/attention/self/key/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.4.attention.self.key.bias" + input: "/encoder/layer.4/attention/self/key/MatMul_output_0" + output: "/encoder/layer.4/attention/self/key/Add_output_0" + name: "/encoder/layer.4/attention/self/key/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.4/attention/self/key/Add_output_0" + output: "/encoder/layer.4/attention/self/Shape_output_0" + name: "/encoder/layer.4/attention/self/Shape" + op_type: "Shape" + } + node { + output: "/encoder/layer.4/attention/self/Constant_output_0" + name: "/encoder/layer.4/attention/self/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.4/attention/self/Shape_output_0" + input: "/encoder/layer.4/attention/self/Constant_output_0" + output: "/encoder/layer.4/attention/self/Gather_output_0" + name: "/encoder/layer.4/attention/self/Gather" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.4/attention/self/key/Add_output_0" + output: "/encoder/layer.4/attention/self/Shape_1_output_0" + name: "/encoder/layer.4/attention/self/Shape_1" + op_type: "Shape" + } + node { + output: "/encoder/layer.4/attention/self/Constant_1_output_0" + name: "/encoder/layer.4/attention/self/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.4/attention/self/Shape_1_output_0" + input: "/encoder/layer.4/attention/self/Constant_1_output_0" + output: "/encoder/layer.4/attention/self/Gather_1_output_0" + name: "/encoder/layer.4/attention/self/Gather_1" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.4/attention/self/Gather_output_0" + output: "/encoder/layer.4/attention/self/Unsqueeze_output_0" + name: "/encoder/layer.4/attention/self/Unsqueeze" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.4/attention/self/Gather_1_output_0" + output: "/encoder/layer.4/attention/self/Unsqueeze_1_output_0" + name: "/encoder/layer.4/attention/self/Unsqueeze_1" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.4/attention/self/Constant_2_output_0" + name: "/encoder/layer.4/attention/self/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.4/attention/self/Constant_3_output_0" + name: "/encoder/layer.4/attention/self/Constant_3" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.4/attention/self/Unsqueeze_output_0" + input: "/encoder/layer.4/attention/self/Unsqueeze_1_output_0" + input: "/encoder/layer.4/attention/self/Constant_2_output_0" + input: "/encoder/layer.4/attention/self/Constant_3_output_0" + output: "/encoder/layer.4/attention/self/Concat_output_0" + name: "/encoder/layer.4/attention/self/Concat" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.4/attention/self/key/Add_output_0" + input: "/encoder/layer.4/attention/self/Concat_output_0" + output: "/encoder/layer.4/attention/self/Reshape_output_0" + name: "/encoder/layer.4/attention/self/Reshape" + op_type: "Reshape" + } + node { + input: "/encoder/layer.3/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1609" + output: "/encoder/layer.4/attention/self/value/MatMul_output_0" + name: "/encoder/layer.4/attention/self/value/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.4.attention.self.value.bias" + input: "/encoder/layer.4/attention/self/value/MatMul_output_0" + output: "/encoder/layer.4/attention/self/value/Add_output_0" + name: "/encoder/layer.4/attention/self/value/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.4/attention/self/value/Add_output_0" + output: "/encoder/layer.4/attention/self/Shape_2_output_0" + name: "/encoder/layer.4/attention/self/Shape_2" + op_type: "Shape" + } + node { + output: "/encoder/layer.4/attention/self/Constant_4_output_0" + name: "/encoder/layer.4/attention/self/Constant_4" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.4/attention/self/Shape_2_output_0" + input: "/encoder/layer.4/attention/self/Constant_4_output_0" + output: "/encoder/layer.4/attention/self/Gather_2_output_0" + name: "/encoder/layer.4/attention/self/Gather_2" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.4/attention/self/value/Add_output_0" + output: "/encoder/layer.4/attention/self/Shape_3_output_0" + name: "/encoder/layer.4/attention/self/Shape_3" + op_type: "Shape" + } + node { + output: "/encoder/layer.4/attention/self/Constant_5_output_0" + name: "/encoder/layer.4/attention/self/Constant_5" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.4/attention/self/Shape_3_output_0" + input: "/encoder/layer.4/attention/self/Constant_5_output_0" + output: "/encoder/layer.4/attention/self/Gather_3_output_0" + name: "/encoder/layer.4/attention/self/Gather_3" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.4/attention/self/Gather_2_output_0" + output: "/encoder/layer.4/attention/self/Unsqueeze_2_output_0" + name: "/encoder/layer.4/attention/self/Unsqueeze_2" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.4/attention/self/Gather_3_output_0" + output: "/encoder/layer.4/attention/self/Unsqueeze_3_output_0" + name: "/encoder/layer.4/attention/self/Unsqueeze_3" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.4/attention/self/Constant_6_output_0" + name: "/encoder/layer.4/attention/self/Constant_6" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.4/attention/self/Constant_7_output_0" + name: "/encoder/layer.4/attention/self/Constant_7" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.4/attention/self/Unsqueeze_2_output_0" + input: "/encoder/layer.4/attention/self/Unsqueeze_3_output_0" + input: "/encoder/layer.4/attention/self/Constant_6_output_0" + input: "/encoder/layer.4/attention/self/Constant_7_output_0" + output: "/encoder/layer.4/attention/self/Concat_1_output_0" + name: "/encoder/layer.4/attention/self/Concat_1" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.4/attention/self/value/Add_output_0" + input: "/encoder/layer.4/attention/self/Concat_1_output_0" + output: "/encoder/layer.4/attention/self/Reshape_1_output_0" + name: "/encoder/layer.4/attention/self/Reshape_1" + op_type: "Reshape" + } + node { + input: "/encoder/layer.4/attention/self/Reshape_1_output_0" + output: "/encoder/layer.4/attention/self/Transpose_output_0" + name: "/encoder/layer.4/attention/self/Transpose" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.4/attention/self/query/Add_output_0" + output: "/encoder/layer.4/attention/self/Shape_4_output_0" + name: "/encoder/layer.4/attention/self/Shape_4" + op_type: "Shape" + } + node { + output: "/encoder/layer.4/attention/self/Constant_8_output_0" + name: "/encoder/layer.4/attention/self/Constant_8" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.4/attention/self/Shape_4_output_0" + input: "/encoder/layer.4/attention/self/Constant_8_output_0" + output: "/encoder/layer.4/attention/self/Gather_4_output_0" + name: "/encoder/layer.4/attention/self/Gather_4" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.4/attention/self/query/Add_output_0" + output: "/encoder/layer.4/attention/self/Shape_5_output_0" + name: "/encoder/layer.4/attention/self/Shape_5" + op_type: "Shape" + } + node { + output: "/encoder/layer.4/attention/self/Constant_9_output_0" + name: "/encoder/layer.4/attention/self/Constant_9" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.4/attention/self/Shape_5_output_0" + input: "/encoder/layer.4/attention/self/Constant_9_output_0" + output: "/encoder/layer.4/attention/self/Gather_5_output_0" + name: "/encoder/layer.4/attention/self/Gather_5" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.4/attention/self/Gather_4_output_0" + output: "/encoder/layer.4/attention/self/Unsqueeze_4_output_0" + name: "/encoder/layer.4/attention/self/Unsqueeze_4" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.4/attention/self/Gather_5_output_0" + output: "/encoder/layer.4/attention/self/Unsqueeze_5_output_0" + name: "/encoder/layer.4/attention/self/Unsqueeze_5" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.4/attention/self/Constant_10_output_0" + name: "/encoder/layer.4/attention/self/Constant_10" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.4/attention/self/Constant_11_output_0" + name: "/encoder/layer.4/attention/self/Constant_11" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.4/attention/self/Unsqueeze_4_output_0" + input: "/encoder/layer.4/attention/self/Unsqueeze_5_output_0" + input: "/encoder/layer.4/attention/self/Constant_10_output_0" + input: "/encoder/layer.4/attention/self/Constant_11_output_0" + output: "/encoder/layer.4/attention/self/Concat_2_output_0" + name: "/encoder/layer.4/attention/self/Concat_2" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.4/attention/self/query/Add_output_0" + input: "/encoder/layer.4/attention/self/Concat_2_output_0" + output: "/encoder/layer.4/attention/self/Reshape_2_output_0" + name: "/encoder/layer.4/attention/self/Reshape_2" + op_type: "Reshape" + } + node { + input: "/encoder/layer.4/attention/self/Reshape_2_output_0" + output: "/encoder/layer.4/attention/self/Transpose_1_output_0" + name: "/encoder/layer.4/attention/self/Transpose_1" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.4/attention/self/Reshape_output_0" + output: "/encoder/layer.4/attention/self/Transpose_2_output_0" + name: "/encoder/layer.4/attention/self/Transpose_2" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 3 + ints: 1 + type: INTS + } + } + node { + input: "/encoder/layer.4/attention/self/Transpose_1_output_0" + input: "/encoder/layer.4/attention/self/Transpose_2_output_0" + output: "/encoder/layer.4/attention/self/MatMul_output_0" + name: "/encoder/layer.4/attention/self/MatMul" + op_type: "MatMul" + } + node { + output: "/encoder/layer.4/attention/self/Constant_12_output_0" + name: "/encoder/layer.4/attention/self/Constant_12" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000A" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.4/attention/self/MatMul_output_0" + input: "/encoder/layer.4/attention/self/Constant_12_output_0" + output: "/encoder/layer.4/attention/self/Div_output_0" + name: "/encoder/layer.4/attention/self/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.4/attention/self/Div_output_0" + input: "/Mul_1_output_0" + output: "/encoder/layer.4/attention/self/Add_output_0" + name: "/encoder/layer.4/attention/self/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.4/attention/self/Add_output_0" + output: "/encoder/layer.4/attention/self/Softmax_output_0" + name: "/encoder/layer.4/attention/self/Softmax" + op_type: "Softmax" + attribute { + name: "axis" + i: 3 + type: INT + } + } + node { + input: "/encoder/layer.4/attention/self/Softmax_output_0" + input: "/encoder/layer.4/attention/self/Transpose_output_0" + output: "/encoder/layer.4/attention/self/MatMul_1_output_0" + name: "/encoder/layer.4/attention/self/MatMul_1" + op_type: "MatMul" + } + node { + input: "/encoder/layer.4/attention/self/MatMul_1_output_0" + output: "/encoder/layer.4/attention/self/Transpose_3_output_0" + name: "/encoder/layer.4/attention/self/Transpose_3" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.4/attention/self/Transpose_3_output_0" + output: "/encoder/layer.4/attention/self/Shape_6_output_0" + name: "/encoder/layer.4/attention/self/Shape_6" + op_type: "Shape" + } + node { + output: "/encoder/layer.4/attention/self/Constant_13_output_0" + name: "/encoder/layer.4/attention/self/Constant_13" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.4/attention/self/Shape_6_output_0" + input: "/encoder/layer.4/attention/self/Constant_13_output_0" + output: "/encoder/layer.4/attention/self/Gather_6_output_0" + name: "/encoder/layer.4/attention/self/Gather_6" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.4/attention/self/Transpose_3_output_0" + output: "/encoder/layer.4/attention/self/Shape_7_output_0" + name: "/encoder/layer.4/attention/self/Shape_7" + op_type: "Shape" + } + node { + output: "/encoder/layer.4/attention/self/Constant_14_output_0" + name: "/encoder/layer.4/attention/self/Constant_14" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.4/attention/self/Shape_7_output_0" + input: "/encoder/layer.4/attention/self/Constant_14_output_0" + output: "/encoder/layer.4/attention/self/Gather_7_output_0" + name: "/encoder/layer.4/attention/self/Gather_7" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.4/attention/self/Gather_6_output_0" + output: "/encoder/layer.4/attention/self/Unsqueeze_6_output_0" + name: "/encoder/layer.4/attention/self/Unsqueeze_6" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.4/attention/self/Gather_7_output_0" + output: "/encoder/layer.4/attention/self/Unsqueeze_7_output_0" + name: "/encoder/layer.4/attention/self/Unsqueeze_7" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.4/attention/self/Constant_15_output_0" + name: "/encoder/layer.4/attention/self/Constant_15" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\000\003\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.4/attention/self/Unsqueeze_6_output_0" + input: "/encoder/layer.4/attention/self/Unsqueeze_7_output_0" + input: "/encoder/layer.4/attention/self/Constant_15_output_0" + output: "/encoder/layer.4/attention/self/Concat_3_output_0" + name: "/encoder/layer.4/attention/self/Concat_3" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.4/attention/self/Transpose_3_output_0" + input: "/encoder/layer.4/attention/self/Concat_3_output_0" + output: "/encoder/layer.4/attention/self/Reshape_3_output_0" + name: "/encoder/layer.4/attention/self/Reshape_3" + op_type: "Reshape" + } + node { + input: "/encoder/layer.4/attention/self/Reshape_3_output_0" + input: "onnx::MatMul_1615" + output: "/encoder/layer.4/attention/output/dense/MatMul_output_0" + name: "/encoder/layer.4/attention/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.4.attention.output.dense.bias" + input: "/encoder/layer.4/attention/output/dense/MatMul_output_0" + output: "/encoder/layer.4/attention/output/dense/Add_output_0" + name: "/encoder/layer.4/attention/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.4/attention/output/dense/Add_output_0" + input: "/encoder/layer.3/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.4/attention/output/Add_output_0" + name: "/encoder/layer.4/attention/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.4/attention/output/Add_output_0" + output: "/encoder/layer.4/attention/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.4/attention/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.4/attention/output/Add_output_0" + input: "/encoder/layer.4/attention/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.4/attention/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.4/attention/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.4/attention/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.4/attention/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.4/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.4/attention/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.4/attention/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.4/attention/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.4/attention/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.4/attention/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.4/attention/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.4/attention/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.4/attention/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.4/attention/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.4/attention/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.4/attention/output/LayerNorm/Add_output_0" + name: "/encoder/layer.4/attention/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.4/attention/output/LayerNorm/Add_output_0" + output: "/encoder/layer.4/attention/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.4/attention/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.4/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.4/attention/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.4/attention/output/LayerNorm/Div_output_0" + name: "/encoder/layer.4/attention/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.4/attention/output/LayerNorm/Div_output_0" + input: "encoder.layer.4.attention.output.LayerNorm.weight" + output: "/encoder/layer.4/attention/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.4/attention/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.4/attention/output/LayerNorm/Mul_output_0" + input: "encoder.layer.4.attention.output.LayerNorm.bias" + output: "/encoder/layer.4/attention/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.4/attention/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.4/attention/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1616" + output: "/encoder/layer.4/intermediate/dense/MatMul_output_0" + name: "/encoder/layer.4/intermediate/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.4.intermediate.dense.bias" + input: "/encoder/layer.4/intermediate/dense/MatMul_output_0" + output: "/encoder/layer.4/intermediate/dense/Add_output_0" + name: "/encoder/layer.4/intermediate/dense/Add" + op_type: "Add" + } + node { + output: "/encoder/layer.4/intermediate/intermediate_act_fn/Constant_output_0" + name: "/encoder/layer.4/intermediate/intermediate_act_fn/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\363\004\265?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.4/intermediate/dense/Add_output_0" + input: "/encoder/layer.4/intermediate/intermediate_act_fn/Constant_output_0" + output: "/encoder/layer.4/intermediate/intermediate_act_fn/Div_output_0" + name: "/encoder/layer.4/intermediate/intermediate_act_fn/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.4/intermediate/intermediate_act_fn/Div_output_0" + output: "/encoder/layer.4/intermediate/intermediate_act_fn/Erf_output_0" + name: "/encoder/layer.4/intermediate/intermediate_act_fn/Erf" + op_type: "Erf" + } + node { + output: "/encoder/layer.4/intermediate/intermediate_act_fn/Constant_1_output_0" + name: "/encoder/layer.4/intermediate/intermediate_act_fn/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\200?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.4/intermediate/intermediate_act_fn/Erf_output_0" + input: "/encoder/layer.4/intermediate/intermediate_act_fn/Constant_1_output_0" + output: "/encoder/layer.4/intermediate/intermediate_act_fn/Add_output_0" + name: "/encoder/layer.4/intermediate/intermediate_act_fn/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.4/intermediate/dense/Add_output_0" + input: "/encoder/layer.4/intermediate/intermediate_act_fn/Add_output_0" + output: "/encoder/layer.4/intermediate/intermediate_act_fn/Mul_output_0" + name: "/encoder/layer.4/intermediate/intermediate_act_fn/Mul" + op_type: "Mul" + } + node { + output: "/encoder/layer.4/intermediate/intermediate_act_fn/Constant_2_output_0" + name: "/encoder/layer.4/intermediate/intermediate_act_fn/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.4/intermediate/intermediate_act_fn/Mul_output_0" + input: "/encoder/layer.4/intermediate/intermediate_act_fn/Constant_2_output_0" + output: "/encoder/layer.4/intermediate/intermediate_act_fn/Mul_1_output_0" + name: "/encoder/layer.4/intermediate/intermediate_act_fn/Mul_1" + op_type: "Mul" + } + node { + input: "/encoder/layer.4/intermediate/intermediate_act_fn/Mul_1_output_0" + input: "onnx::MatMul_1617" + output: "/encoder/layer.4/output/dense/MatMul_output_0" + name: "/encoder/layer.4/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.4.output.dense.bias" + input: "/encoder/layer.4/output/dense/MatMul_output_0" + output: "/encoder/layer.4/output/dense/Add_output_0" + name: "/encoder/layer.4/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.4/output/dense/Add_output_0" + input: "/encoder/layer.4/attention/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.4/output/Add_output_0" + name: "/encoder/layer.4/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.4/output/Add_output_0" + output: "/encoder/layer.4/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.4/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.4/output/Add_output_0" + input: "/encoder/layer.4/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.4/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.4/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.4/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.4/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.4/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.4/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.4/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.4/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.4/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.4/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.4/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.4/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.4/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.4/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.4/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.4/output/LayerNorm/Add_output_0" + name: "/encoder/layer.4/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.4/output/LayerNorm/Add_output_0" + output: "/encoder/layer.4/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.4/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.4/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.4/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.4/output/LayerNorm/Div_output_0" + name: "/encoder/layer.4/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.4/output/LayerNorm/Div_output_0" + input: "encoder.layer.4.output.LayerNorm.weight" + output: "/encoder/layer.4/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.4/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.4/output/LayerNorm/Mul_output_0" + input: "encoder.layer.4.output.LayerNorm.bias" + output: "/encoder/layer.4/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.4/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.4/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1618" + output: "/encoder/layer.5/attention/self/query/MatMul_output_0" + name: "/encoder/layer.5/attention/self/query/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.5.attention.self.query.bias" + input: "/encoder/layer.5/attention/self/query/MatMul_output_0" + output: "/encoder/layer.5/attention/self/query/Add_output_0" + name: "/encoder/layer.5/attention/self/query/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.4/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1619" + output: "/encoder/layer.5/attention/self/key/MatMul_output_0" + name: "/encoder/layer.5/attention/self/key/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.5.attention.self.key.bias" + input: "/encoder/layer.5/attention/self/key/MatMul_output_0" + output: "/encoder/layer.5/attention/self/key/Add_output_0" + name: "/encoder/layer.5/attention/self/key/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.5/attention/self/key/Add_output_0" + output: "/encoder/layer.5/attention/self/Shape_output_0" + name: "/encoder/layer.5/attention/self/Shape" + op_type: "Shape" + } + node { + output: "/encoder/layer.5/attention/self/Constant_output_0" + name: "/encoder/layer.5/attention/self/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.5/attention/self/Shape_output_0" + input: "/encoder/layer.5/attention/self/Constant_output_0" + output: "/encoder/layer.5/attention/self/Gather_output_0" + name: "/encoder/layer.5/attention/self/Gather" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.5/attention/self/key/Add_output_0" + output: "/encoder/layer.5/attention/self/Shape_1_output_0" + name: "/encoder/layer.5/attention/self/Shape_1" + op_type: "Shape" + } + node { + output: "/encoder/layer.5/attention/self/Constant_1_output_0" + name: "/encoder/layer.5/attention/self/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.5/attention/self/Shape_1_output_0" + input: "/encoder/layer.5/attention/self/Constant_1_output_0" + output: "/encoder/layer.5/attention/self/Gather_1_output_0" + name: "/encoder/layer.5/attention/self/Gather_1" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.5/attention/self/Gather_output_0" + output: "/encoder/layer.5/attention/self/Unsqueeze_output_0" + name: "/encoder/layer.5/attention/self/Unsqueeze" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.5/attention/self/Gather_1_output_0" + output: "/encoder/layer.5/attention/self/Unsqueeze_1_output_0" + name: "/encoder/layer.5/attention/self/Unsqueeze_1" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.5/attention/self/Constant_2_output_0" + name: "/encoder/layer.5/attention/self/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.5/attention/self/Constant_3_output_0" + name: "/encoder/layer.5/attention/self/Constant_3" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.5/attention/self/Unsqueeze_output_0" + input: "/encoder/layer.5/attention/self/Unsqueeze_1_output_0" + input: "/encoder/layer.5/attention/self/Constant_2_output_0" + input: "/encoder/layer.5/attention/self/Constant_3_output_0" + output: "/encoder/layer.5/attention/self/Concat_output_0" + name: "/encoder/layer.5/attention/self/Concat" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.5/attention/self/key/Add_output_0" + input: "/encoder/layer.5/attention/self/Concat_output_0" + output: "/encoder/layer.5/attention/self/Reshape_output_0" + name: "/encoder/layer.5/attention/self/Reshape" + op_type: "Reshape" + } + node { + input: "/encoder/layer.4/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1622" + output: "/encoder/layer.5/attention/self/value/MatMul_output_0" + name: "/encoder/layer.5/attention/self/value/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.5.attention.self.value.bias" + input: "/encoder/layer.5/attention/self/value/MatMul_output_0" + output: "/encoder/layer.5/attention/self/value/Add_output_0" + name: "/encoder/layer.5/attention/self/value/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.5/attention/self/value/Add_output_0" + output: "/encoder/layer.5/attention/self/Shape_2_output_0" + name: "/encoder/layer.5/attention/self/Shape_2" + op_type: "Shape" + } + node { + output: "/encoder/layer.5/attention/self/Constant_4_output_0" + name: "/encoder/layer.5/attention/self/Constant_4" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.5/attention/self/Shape_2_output_0" + input: "/encoder/layer.5/attention/self/Constant_4_output_0" + output: "/encoder/layer.5/attention/self/Gather_2_output_0" + name: "/encoder/layer.5/attention/self/Gather_2" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.5/attention/self/value/Add_output_0" + output: "/encoder/layer.5/attention/self/Shape_3_output_0" + name: "/encoder/layer.5/attention/self/Shape_3" + op_type: "Shape" + } + node { + output: "/encoder/layer.5/attention/self/Constant_5_output_0" + name: "/encoder/layer.5/attention/self/Constant_5" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.5/attention/self/Shape_3_output_0" + input: "/encoder/layer.5/attention/self/Constant_5_output_0" + output: "/encoder/layer.5/attention/self/Gather_3_output_0" + name: "/encoder/layer.5/attention/self/Gather_3" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.5/attention/self/Gather_2_output_0" + output: "/encoder/layer.5/attention/self/Unsqueeze_2_output_0" + name: "/encoder/layer.5/attention/self/Unsqueeze_2" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.5/attention/self/Gather_3_output_0" + output: "/encoder/layer.5/attention/self/Unsqueeze_3_output_0" + name: "/encoder/layer.5/attention/self/Unsqueeze_3" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.5/attention/self/Constant_6_output_0" + name: "/encoder/layer.5/attention/self/Constant_6" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.5/attention/self/Constant_7_output_0" + name: "/encoder/layer.5/attention/self/Constant_7" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.5/attention/self/Unsqueeze_2_output_0" + input: "/encoder/layer.5/attention/self/Unsqueeze_3_output_0" + input: "/encoder/layer.5/attention/self/Constant_6_output_0" + input: "/encoder/layer.5/attention/self/Constant_7_output_0" + output: "/encoder/layer.5/attention/self/Concat_1_output_0" + name: "/encoder/layer.5/attention/self/Concat_1" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.5/attention/self/value/Add_output_0" + input: "/encoder/layer.5/attention/self/Concat_1_output_0" + output: "/encoder/layer.5/attention/self/Reshape_1_output_0" + name: "/encoder/layer.5/attention/self/Reshape_1" + op_type: "Reshape" + } + node { + input: "/encoder/layer.5/attention/self/Reshape_1_output_0" + output: "/encoder/layer.5/attention/self/Transpose_output_0" + name: "/encoder/layer.5/attention/self/Transpose" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.5/attention/self/query/Add_output_0" + output: "/encoder/layer.5/attention/self/Shape_4_output_0" + name: "/encoder/layer.5/attention/self/Shape_4" + op_type: "Shape" + } + node { + output: "/encoder/layer.5/attention/self/Constant_8_output_0" + name: "/encoder/layer.5/attention/self/Constant_8" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.5/attention/self/Shape_4_output_0" + input: "/encoder/layer.5/attention/self/Constant_8_output_0" + output: "/encoder/layer.5/attention/self/Gather_4_output_0" + name: "/encoder/layer.5/attention/self/Gather_4" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.5/attention/self/query/Add_output_0" + output: "/encoder/layer.5/attention/self/Shape_5_output_0" + name: "/encoder/layer.5/attention/self/Shape_5" + op_type: "Shape" + } + node { + output: "/encoder/layer.5/attention/self/Constant_9_output_0" + name: "/encoder/layer.5/attention/self/Constant_9" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.5/attention/self/Shape_5_output_0" + input: "/encoder/layer.5/attention/self/Constant_9_output_0" + output: "/encoder/layer.5/attention/self/Gather_5_output_0" + name: "/encoder/layer.5/attention/self/Gather_5" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.5/attention/self/Gather_4_output_0" + output: "/encoder/layer.5/attention/self/Unsqueeze_4_output_0" + name: "/encoder/layer.5/attention/self/Unsqueeze_4" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.5/attention/self/Gather_5_output_0" + output: "/encoder/layer.5/attention/self/Unsqueeze_5_output_0" + name: "/encoder/layer.5/attention/self/Unsqueeze_5" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.5/attention/self/Constant_10_output_0" + name: "/encoder/layer.5/attention/self/Constant_10" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.5/attention/self/Constant_11_output_0" + name: "/encoder/layer.5/attention/self/Constant_11" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.5/attention/self/Unsqueeze_4_output_0" + input: "/encoder/layer.5/attention/self/Unsqueeze_5_output_0" + input: "/encoder/layer.5/attention/self/Constant_10_output_0" + input: "/encoder/layer.5/attention/self/Constant_11_output_0" + output: "/encoder/layer.5/attention/self/Concat_2_output_0" + name: "/encoder/layer.5/attention/self/Concat_2" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.5/attention/self/query/Add_output_0" + input: "/encoder/layer.5/attention/self/Concat_2_output_0" + output: "/encoder/layer.5/attention/self/Reshape_2_output_0" + name: "/encoder/layer.5/attention/self/Reshape_2" + op_type: "Reshape" + } + node { + input: "/encoder/layer.5/attention/self/Reshape_2_output_0" + output: "/encoder/layer.5/attention/self/Transpose_1_output_0" + name: "/encoder/layer.5/attention/self/Transpose_1" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.5/attention/self/Reshape_output_0" + output: "/encoder/layer.5/attention/self/Transpose_2_output_0" + name: "/encoder/layer.5/attention/self/Transpose_2" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 3 + ints: 1 + type: INTS + } + } + node { + input: "/encoder/layer.5/attention/self/Transpose_1_output_0" + input: "/encoder/layer.5/attention/self/Transpose_2_output_0" + output: "/encoder/layer.5/attention/self/MatMul_output_0" + name: "/encoder/layer.5/attention/self/MatMul" + op_type: "MatMul" + } + node { + output: "/encoder/layer.5/attention/self/Constant_12_output_0" + name: "/encoder/layer.5/attention/self/Constant_12" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000A" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.5/attention/self/MatMul_output_0" + input: "/encoder/layer.5/attention/self/Constant_12_output_0" + output: "/encoder/layer.5/attention/self/Div_output_0" + name: "/encoder/layer.5/attention/self/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.5/attention/self/Div_output_0" + input: "/Mul_1_output_0" + output: "/encoder/layer.5/attention/self/Add_output_0" + name: "/encoder/layer.5/attention/self/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.5/attention/self/Add_output_0" + output: "/encoder/layer.5/attention/self/Softmax_output_0" + name: "/encoder/layer.5/attention/self/Softmax" + op_type: "Softmax" + attribute { + name: "axis" + i: 3 + type: INT + } + } + node { + input: "/encoder/layer.5/attention/self/Softmax_output_0" + input: "/encoder/layer.5/attention/self/Transpose_output_0" + output: "/encoder/layer.5/attention/self/MatMul_1_output_0" + name: "/encoder/layer.5/attention/self/MatMul_1" + op_type: "MatMul" + } + node { + input: "/encoder/layer.5/attention/self/MatMul_1_output_0" + output: "/encoder/layer.5/attention/self/Transpose_3_output_0" + name: "/encoder/layer.5/attention/self/Transpose_3" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.5/attention/self/Transpose_3_output_0" + output: "/encoder/layer.5/attention/self/Shape_6_output_0" + name: "/encoder/layer.5/attention/self/Shape_6" + op_type: "Shape" + } + node { + output: "/encoder/layer.5/attention/self/Constant_13_output_0" + name: "/encoder/layer.5/attention/self/Constant_13" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.5/attention/self/Shape_6_output_0" + input: "/encoder/layer.5/attention/self/Constant_13_output_0" + output: "/encoder/layer.5/attention/self/Gather_6_output_0" + name: "/encoder/layer.5/attention/self/Gather_6" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.5/attention/self/Transpose_3_output_0" + output: "/encoder/layer.5/attention/self/Shape_7_output_0" + name: "/encoder/layer.5/attention/self/Shape_7" + op_type: "Shape" + } + node { + output: "/encoder/layer.5/attention/self/Constant_14_output_0" + name: "/encoder/layer.5/attention/self/Constant_14" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.5/attention/self/Shape_7_output_0" + input: "/encoder/layer.5/attention/self/Constant_14_output_0" + output: "/encoder/layer.5/attention/self/Gather_7_output_0" + name: "/encoder/layer.5/attention/self/Gather_7" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.5/attention/self/Gather_6_output_0" + output: "/encoder/layer.5/attention/self/Unsqueeze_6_output_0" + name: "/encoder/layer.5/attention/self/Unsqueeze_6" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.5/attention/self/Gather_7_output_0" + output: "/encoder/layer.5/attention/self/Unsqueeze_7_output_0" + name: "/encoder/layer.5/attention/self/Unsqueeze_7" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.5/attention/self/Constant_15_output_0" + name: "/encoder/layer.5/attention/self/Constant_15" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\000\003\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.5/attention/self/Unsqueeze_6_output_0" + input: "/encoder/layer.5/attention/self/Unsqueeze_7_output_0" + input: "/encoder/layer.5/attention/self/Constant_15_output_0" + output: "/encoder/layer.5/attention/self/Concat_3_output_0" + name: "/encoder/layer.5/attention/self/Concat_3" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.5/attention/self/Transpose_3_output_0" + input: "/encoder/layer.5/attention/self/Concat_3_output_0" + output: "/encoder/layer.5/attention/self/Reshape_3_output_0" + name: "/encoder/layer.5/attention/self/Reshape_3" + op_type: "Reshape" + } + node { + input: "/encoder/layer.5/attention/self/Reshape_3_output_0" + input: "onnx::MatMul_1628" + output: "/encoder/layer.5/attention/output/dense/MatMul_output_0" + name: "/encoder/layer.5/attention/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.5.attention.output.dense.bias" + input: "/encoder/layer.5/attention/output/dense/MatMul_output_0" + output: "/encoder/layer.5/attention/output/dense/Add_output_0" + name: "/encoder/layer.5/attention/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.5/attention/output/dense/Add_output_0" + input: "/encoder/layer.4/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.5/attention/output/Add_output_0" + name: "/encoder/layer.5/attention/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.5/attention/output/Add_output_0" + output: "/encoder/layer.5/attention/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.5/attention/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.5/attention/output/Add_output_0" + input: "/encoder/layer.5/attention/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.5/attention/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.5/attention/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.5/attention/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.5/attention/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.5/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.5/attention/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.5/attention/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.5/attention/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.5/attention/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.5/attention/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.5/attention/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.5/attention/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.5/attention/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.5/attention/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.5/attention/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.5/attention/output/LayerNorm/Add_output_0" + name: "/encoder/layer.5/attention/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.5/attention/output/LayerNorm/Add_output_0" + output: "/encoder/layer.5/attention/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.5/attention/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.5/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.5/attention/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.5/attention/output/LayerNorm/Div_output_0" + name: "/encoder/layer.5/attention/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.5/attention/output/LayerNorm/Div_output_0" + input: "encoder.layer.5.attention.output.LayerNorm.weight" + output: "/encoder/layer.5/attention/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.5/attention/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.5/attention/output/LayerNorm/Mul_output_0" + input: "encoder.layer.5.attention.output.LayerNorm.bias" + output: "/encoder/layer.5/attention/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.5/attention/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.5/attention/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1629" + output: "/encoder/layer.5/intermediate/dense/MatMul_output_0" + name: "/encoder/layer.5/intermediate/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.5.intermediate.dense.bias" + input: "/encoder/layer.5/intermediate/dense/MatMul_output_0" + output: "/encoder/layer.5/intermediate/dense/Add_output_0" + name: "/encoder/layer.5/intermediate/dense/Add" + op_type: "Add" + } + node { + output: "/encoder/layer.5/intermediate/intermediate_act_fn/Constant_output_0" + name: "/encoder/layer.5/intermediate/intermediate_act_fn/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\363\004\265?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.5/intermediate/dense/Add_output_0" + input: "/encoder/layer.5/intermediate/intermediate_act_fn/Constant_output_0" + output: "/encoder/layer.5/intermediate/intermediate_act_fn/Div_output_0" + name: "/encoder/layer.5/intermediate/intermediate_act_fn/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.5/intermediate/intermediate_act_fn/Div_output_0" + output: "/encoder/layer.5/intermediate/intermediate_act_fn/Erf_output_0" + name: "/encoder/layer.5/intermediate/intermediate_act_fn/Erf" + op_type: "Erf" + } + node { + output: "/encoder/layer.5/intermediate/intermediate_act_fn/Constant_1_output_0" + name: "/encoder/layer.5/intermediate/intermediate_act_fn/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\200?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.5/intermediate/intermediate_act_fn/Erf_output_0" + input: "/encoder/layer.5/intermediate/intermediate_act_fn/Constant_1_output_0" + output: "/encoder/layer.5/intermediate/intermediate_act_fn/Add_output_0" + name: "/encoder/layer.5/intermediate/intermediate_act_fn/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.5/intermediate/dense/Add_output_0" + input: "/encoder/layer.5/intermediate/intermediate_act_fn/Add_output_0" + output: "/encoder/layer.5/intermediate/intermediate_act_fn/Mul_output_0" + name: "/encoder/layer.5/intermediate/intermediate_act_fn/Mul" + op_type: "Mul" + } + node { + output: "/encoder/layer.5/intermediate/intermediate_act_fn/Constant_2_output_0" + name: "/encoder/layer.5/intermediate/intermediate_act_fn/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.5/intermediate/intermediate_act_fn/Mul_output_0" + input: "/encoder/layer.5/intermediate/intermediate_act_fn/Constant_2_output_0" + output: "/encoder/layer.5/intermediate/intermediate_act_fn/Mul_1_output_0" + name: "/encoder/layer.5/intermediate/intermediate_act_fn/Mul_1" + op_type: "Mul" + } + node { + input: "/encoder/layer.5/intermediate/intermediate_act_fn/Mul_1_output_0" + input: "onnx::MatMul_1630" + output: "/encoder/layer.5/output/dense/MatMul_output_0" + name: "/encoder/layer.5/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.5.output.dense.bias" + input: "/encoder/layer.5/output/dense/MatMul_output_0" + output: "/encoder/layer.5/output/dense/Add_output_0" + name: "/encoder/layer.5/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.5/output/dense/Add_output_0" + input: "/encoder/layer.5/attention/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.5/output/Add_output_0" + name: "/encoder/layer.5/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.5/output/Add_output_0" + output: "/encoder/layer.5/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.5/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.5/output/Add_output_0" + input: "/encoder/layer.5/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.5/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.5/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.5/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.5/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.5/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.5/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.5/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.5/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.5/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.5/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.5/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.5/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.5/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.5/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.5/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.5/output/LayerNorm/Add_output_0" + name: "/encoder/layer.5/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.5/output/LayerNorm/Add_output_0" + output: "/encoder/layer.5/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.5/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.5/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.5/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.5/output/LayerNorm/Div_output_0" + name: "/encoder/layer.5/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.5/output/LayerNorm/Div_output_0" + input: "encoder.layer.5.output.LayerNorm.weight" + output: "/encoder/layer.5/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.5/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.5/output/LayerNorm/Mul_output_0" + input: "encoder.layer.5.output.LayerNorm.bias" + output: "/encoder/layer.5/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.5/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.5/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1631" + output: "/encoder/layer.6/attention/self/query/MatMul_output_0" + name: "/encoder/layer.6/attention/self/query/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.6.attention.self.query.bias" + input: "/encoder/layer.6/attention/self/query/MatMul_output_0" + output: "/encoder/layer.6/attention/self/query/Add_output_0" + name: "/encoder/layer.6/attention/self/query/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.5/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1632" + output: "/encoder/layer.6/attention/self/key/MatMul_output_0" + name: "/encoder/layer.6/attention/self/key/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.6.attention.self.key.bias" + input: "/encoder/layer.6/attention/self/key/MatMul_output_0" + output: "/encoder/layer.6/attention/self/key/Add_output_0" + name: "/encoder/layer.6/attention/self/key/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.6/attention/self/key/Add_output_0" + output: "/encoder/layer.6/attention/self/Shape_output_0" + name: "/encoder/layer.6/attention/self/Shape" + op_type: "Shape" + } + node { + output: "/encoder/layer.6/attention/self/Constant_output_0" + name: "/encoder/layer.6/attention/self/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.6/attention/self/Shape_output_0" + input: "/encoder/layer.6/attention/self/Constant_output_0" + output: "/encoder/layer.6/attention/self/Gather_output_0" + name: "/encoder/layer.6/attention/self/Gather" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.6/attention/self/key/Add_output_0" + output: "/encoder/layer.6/attention/self/Shape_1_output_0" + name: "/encoder/layer.6/attention/self/Shape_1" + op_type: "Shape" + } + node { + output: "/encoder/layer.6/attention/self/Constant_1_output_0" + name: "/encoder/layer.6/attention/self/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.6/attention/self/Shape_1_output_0" + input: "/encoder/layer.6/attention/self/Constant_1_output_0" + output: "/encoder/layer.6/attention/self/Gather_1_output_0" + name: "/encoder/layer.6/attention/self/Gather_1" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.6/attention/self/Gather_output_0" + output: "/encoder/layer.6/attention/self/Unsqueeze_output_0" + name: "/encoder/layer.6/attention/self/Unsqueeze" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.6/attention/self/Gather_1_output_0" + output: "/encoder/layer.6/attention/self/Unsqueeze_1_output_0" + name: "/encoder/layer.6/attention/self/Unsqueeze_1" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.6/attention/self/Constant_2_output_0" + name: "/encoder/layer.6/attention/self/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.6/attention/self/Constant_3_output_0" + name: "/encoder/layer.6/attention/self/Constant_3" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.6/attention/self/Unsqueeze_output_0" + input: "/encoder/layer.6/attention/self/Unsqueeze_1_output_0" + input: "/encoder/layer.6/attention/self/Constant_2_output_0" + input: "/encoder/layer.6/attention/self/Constant_3_output_0" + output: "/encoder/layer.6/attention/self/Concat_output_0" + name: "/encoder/layer.6/attention/self/Concat" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.6/attention/self/key/Add_output_0" + input: "/encoder/layer.6/attention/self/Concat_output_0" + output: "/encoder/layer.6/attention/self/Reshape_output_0" + name: "/encoder/layer.6/attention/self/Reshape" + op_type: "Reshape" + } + node { + input: "/encoder/layer.5/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1635" + output: "/encoder/layer.6/attention/self/value/MatMul_output_0" + name: "/encoder/layer.6/attention/self/value/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.6.attention.self.value.bias" + input: "/encoder/layer.6/attention/self/value/MatMul_output_0" + output: "/encoder/layer.6/attention/self/value/Add_output_0" + name: "/encoder/layer.6/attention/self/value/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.6/attention/self/value/Add_output_0" + output: "/encoder/layer.6/attention/self/Shape_2_output_0" + name: "/encoder/layer.6/attention/self/Shape_2" + op_type: "Shape" + } + node { + output: "/encoder/layer.6/attention/self/Constant_4_output_0" + name: "/encoder/layer.6/attention/self/Constant_4" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.6/attention/self/Shape_2_output_0" + input: "/encoder/layer.6/attention/self/Constant_4_output_0" + output: "/encoder/layer.6/attention/self/Gather_2_output_0" + name: "/encoder/layer.6/attention/self/Gather_2" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.6/attention/self/value/Add_output_0" + output: "/encoder/layer.6/attention/self/Shape_3_output_0" + name: "/encoder/layer.6/attention/self/Shape_3" + op_type: "Shape" + } + node { + output: "/encoder/layer.6/attention/self/Constant_5_output_0" + name: "/encoder/layer.6/attention/self/Constant_5" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.6/attention/self/Shape_3_output_0" + input: "/encoder/layer.6/attention/self/Constant_5_output_0" + output: "/encoder/layer.6/attention/self/Gather_3_output_0" + name: "/encoder/layer.6/attention/self/Gather_3" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.6/attention/self/Gather_2_output_0" + output: "/encoder/layer.6/attention/self/Unsqueeze_2_output_0" + name: "/encoder/layer.6/attention/self/Unsqueeze_2" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.6/attention/self/Gather_3_output_0" + output: "/encoder/layer.6/attention/self/Unsqueeze_3_output_0" + name: "/encoder/layer.6/attention/self/Unsqueeze_3" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.6/attention/self/Constant_6_output_0" + name: "/encoder/layer.6/attention/self/Constant_6" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.6/attention/self/Constant_7_output_0" + name: "/encoder/layer.6/attention/self/Constant_7" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.6/attention/self/Unsqueeze_2_output_0" + input: "/encoder/layer.6/attention/self/Unsqueeze_3_output_0" + input: "/encoder/layer.6/attention/self/Constant_6_output_0" + input: "/encoder/layer.6/attention/self/Constant_7_output_0" + output: "/encoder/layer.6/attention/self/Concat_1_output_0" + name: "/encoder/layer.6/attention/self/Concat_1" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.6/attention/self/value/Add_output_0" + input: "/encoder/layer.6/attention/self/Concat_1_output_0" + output: "/encoder/layer.6/attention/self/Reshape_1_output_0" + name: "/encoder/layer.6/attention/self/Reshape_1" + op_type: "Reshape" + } + node { + input: "/encoder/layer.6/attention/self/Reshape_1_output_0" + output: "/encoder/layer.6/attention/self/Transpose_output_0" + name: "/encoder/layer.6/attention/self/Transpose" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.6/attention/self/query/Add_output_0" + output: "/encoder/layer.6/attention/self/Shape_4_output_0" + name: "/encoder/layer.6/attention/self/Shape_4" + op_type: "Shape" + } + node { + output: "/encoder/layer.6/attention/self/Constant_8_output_0" + name: "/encoder/layer.6/attention/self/Constant_8" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.6/attention/self/Shape_4_output_0" + input: "/encoder/layer.6/attention/self/Constant_8_output_0" + output: "/encoder/layer.6/attention/self/Gather_4_output_0" + name: "/encoder/layer.6/attention/self/Gather_4" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.6/attention/self/query/Add_output_0" + output: "/encoder/layer.6/attention/self/Shape_5_output_0" + name: "/encoder/layer.6/attention/self/Shape_5" + op_type: "Shape" + } + node { + output: "/encoder/layer.6/attention/self/Constant_9_output_0" + name: "/encoder/layer.6/attention/self/Constant_9" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.6/attention/self/Shape_5_output_0" + input: "/encoder/layer.6/attention/self/Constant_9_output_0" + output: "/encoder/layer.6/attention/self/Gather_5_output_0" + name: "/encoder/layer.6/attention/self/Gather_5" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.6/attention/self/Gather_4_output_0" + output: "/encoder/layer.6/attention/self/Unsqueeze_4_output_0" + name: "/encoder/layer.6/attention/self/Unsqueeze_4" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.6/attention/self/Gather_5_output_0" + output: "/encoder/layer.6/attention/self/Unsqueeze_5_output_0" + name: "/encoder/layer.6/attention/self/Unsqueeze_5" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.6/attention/self/Constant_10_output_0" + name: "/encoder/layer.6/attention/self/Constant_10" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.6/attention/self/Constant_11_output_0" + name: "/encoder/layer.6/attention/self/Constant_11" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.6/attention/self/Unsqueeze_4_output_0" + input: "/encoder/layer.6/attention/self/Unsqueeze_5_output_0" + input: "/encoder/layer.6/attention/self/Constant_10_output_0" + input: "/encoder/layer.6/attention/self/Constant_11_output_0" + output: "/encoder/layer.6/attention/self/Concat_2_output_0" + name: "/encoder/layer.6/attention/self/Concat_2" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.6/attention/self/query/Add_output_0" + input: "/encoder/layer.6/attention/self/Concat_2_output_0" + output: "/encoder/layer.6/attention/self/Reshape_2_output_0" + name: "/encoder/layer.6/attention/self/Reshape_2" + op_type: "Reshape" + } + node { + input: "/encoder/layer.6/attention/self/Reshape_2_output_0" + output: "/encoder/layer.6/attention/self/Transpose_1_output_0" + name: "/encoder/layer.6/attention/self/Transpose_1" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.6/attention/self/Reshape_output_0" + output: "/encoder/layer.6/attention/self/Transpose_2_output_0" + name: "/encoder/layer.6/attention/self/Transpose_2" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 3 + ints: 1 + type: INTS + } + } + node { + input: "/encoder/layer.6/attention/self/Transpose_1_output_0" + input: "/encoder/layer.6/attention/self/Transpose_2_output_0" + output: "/encoder/layer.6/attention/self/MatMul_output_0" + name: "/encoder/layer.6/attention/self/MatMul" + op_type: "MatMul" + } + node { + output: "/encoder/layer.6/attention/self/Constant_12_output_0" + name: "/encoder/layer.6/attention/self/Constant_12" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000A" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.6/attention/self/MatMul_output_0" + input: "/encoder/layer.6/attention/self/Constant_12_output_0" + output: "/encoder/layer.6/attention/self/Div_output_0" + name: "/encoder/layer.6/attention/self/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.6/attention/self/Div_output_0" + input: "/Mul_1_output_0" + output: "/encoder/layer.6/attention/self/Add_output_0" + name: "/encoder/layer.6/attention/self/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.6/attention/self/Add_output_0" + output: "/encoder/layer.6/attention/self/Softmax_output_0" + name: "/encoder/layer.6/attention/self/Softmax" + op_type: "Softmax" + attribute { + name: "axis" + i: 3 + type: INT + } + } + node { + input: "/encoder/layer.6/attention/self/Softmax_output_0" + input: "/encoder/layer.6/attention/self/Transpose_output_0" + output: "/encoder/layer.6/attention/self/MatMul_1_output_0" + name: "/encoder/layer.6/attention/self/MatMul_1" + op_type: "MatMul" + } + node { + input: "/encoder/layer.6/attention/self/MatMul_1_output_0" + output: "/encoder/layer.6/attention/self/Transpose_3_output_0" + name: "/encoder/layer.6/attention/self/Transpose_3" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.6/attention/self/Transpose_3_output_0" + output: "/encoder/layer.6/attention/self/Shape_6_output_0" + name: "/encoder/layer.6/attention/self/Shape_6" + op_type: "Shape" + } + node { + output: "/encoder/layer.6/attention/self/Constant_13_output_0" + name: "/encoder/layer.6/attention/self/Constant_13" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.6/attention/self/Shape_6_output_0" + input: "/encoder/layer.6/attention/self/Constant_13_output_0" + output: "/encoder/layer.6/attention/self/Gather_6_output_0" + name: "/encoder/layer.6/attention/self/Gather_6" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.6/attention/self/Transpose_3_output_0" + output: "/encoder/layer.6/attention/self/Shape_7_output_0" + name: "/encoder/layer.6/attention/self/Shape_7" + op_type: "Shape" + } + node { + output: "/encoder/layer.6/attention/self/Constant_14_output_0" + name: "/encoder/layer.6/attention/self/Constant_14" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.6/attention/self/Shape_7_output_0" + input: "/encoder/layer.6/attention/self/Constant_14_output_0" + output: "/encoder/layer.6/attention/self/Gather_7_output_0" + name: "/encoder/layer.6/attention/self/Gather_7" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.6/attention/self/Gather_6_output_0" + output: "/encoder/layer.6/attention/self/Unsqueeze_6_output_0" + name: "/encoder/layer.6/attention/self/Unsqueeze_6" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.6/attention/self/Gather_7_output_0" + output: "/encoder/layer.6/attention/self/Unsqueeze_7_output_0" + name: "/encoder/layer.6/attention/self/Unsqueeze_7" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.6/attention/self/Constant_15_output_0" + name: "/encoder/layer.6/attention/self/Constant_15" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\000\003\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.6/attention/self/Unsqueeze_6_output_0" + input: "/encoder/layer.6/attention/self/Unsqueeze_7_output_0" + input: "/encoder/layer.6/attention/self/Constant_15_output_0" + output: "/encoder/layer.6/attention/self/Concat_3_output_0" + name: "/encoder/layer.6/attention/self/Concat_3" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.6/attention/self/Transpose_3_output_0" + input: "/encoder/layer.6/attention/self/Concat_3_output_0" + output: "/encoder/layer.6/attention/self/Reshape_3_output_0" + name: "/encoder/layer.6/attention/self/Reshape_3" + op_type: "Reshape" + } + node { + input: "/encoder/layer.6/attention/self/Reshape_3_output_0" + input: "onnx::MatMul_1641" + output: "/encoder/layer.6/attention/output/dense/MatMul_output_0" + name: "/encoder/layer.6/attention/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.6.attention.output.dense.bias" + input: "/encoder/layer.6/attention/output/dense/MatMul_output_0" + output: "/encoder/layer.6/attention/output/dense/Add_output_0" + name: "/encoder/layer.6/attention/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.6/attention/output/dense/Add_output_0" + input: "/encoder/layer.5/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.6/attention/output/Add_output_0" + name: "/encoder/layer.6/attention/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.6/attention/output/Add_output_0" + output: "/encoder/layer.6/attention/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.6/attention/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.6/attention/output/Add_output_0" + input: "/encoder/layer.6/attention/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.6/attention/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.6/attention/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.6/attention/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.6/attention/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.6/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.6/attention/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.6/attention/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.6/attention/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.6/attention/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.6/attention/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.6/attention/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.6/attention/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.6/attention/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.6/attention/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.6/attention/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.6/attention/output/LayerNorm/Add_output_0" + name: "/encoder/layer.6/attention/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.6/attention/output/LayerNorm/Add_output_0" + output: "/encoder/layer.6/attention/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.6/attention/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.6/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.6/attention/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.6/attention/output/LayerNorm/Div_output_0" + name: "/encoder/layer.6/attention/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.6/attention/output/LayerNorm/Div_output_0" + input: "encoder.layer.6.attention.output.LayerNorm.weight" + output: "/encoder/layer.6/attention/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.6/attention/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.6/attention/output/LayerNorm/Mul_output_0" + input: "encoder.layer.6.attention.output.LayerNorm.bias" + output: "/encoder/layer.6/attention/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.6/attention/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.6/attention/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1642" + output: "/encoder/layer.6/intermediate/dense/MatMul_output_0" + name: "/encoder/layer.6/intermediate/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.6.intermediate.dense.bias" + input: "/encoder/layer.6/intermediate/dense/MatMul_output_0" + output: "/encoder/layer.6/intermediate/dense/Add_output_0" + name: "/encoder/layer.6/intermediate/dense/Add" + op_type: "Add" + } + node { + output: "/encoder/layer.6/intermediate/intermediate_act_fn/Constant_output_0" + name: "/encoder/layer.6/intermediate/intermediate_act_fn/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\363\004\265?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.6/intermediate/dense/Add_output_0" + input: "/encoder/layer.6/intermediate/intermediate_act_fn/Constant_output_0" + output: "/encoder/layer.6/intermediate/intermediate_act_fn/Div_output_0" + name: "/encoder/layer.6/intermediate/intermediate_act_fn/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.6/intermediate/intermediate_act_fn/Div_output_0" + output: "/encoder/layer.6/intermediate/intermediate_act_fn/Erf_output_0" + name: "/encoder/layer.6/intermediate/intermediate_act_fn/Erf" + op_type: "Erf" + } + node { + output: "/encoder/layer.6/intermediate/intermediate_act_fn/Constant_1_output_0" + name: "/encoder/layer.6/intermediate/intermediate_act_fn/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\200?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.6/intermediate/intermediate_act_fn/Erf_output_0" + input: "/encoder/layer.6/intermediate/intermediate_act_fn/Constant_1_output_0" + output: "/encoder/layer.6/intermediate/intermediate_act_fn/Add_output_0" + name: "/encoder/layer.6/intermediate/intermediate_act_fn/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.6/intermediate/dense/Add_output_0" + input: "/encoder/layer.6/intermediate/intermediate_act_fn/Add_output_0" + output: "/encoder/layer.6/intermediate/intermediate_act_fn/Mul_output_0" + name: "/encoder/layer.6/intermediate/intermediate_act_fn/Mul" + op_type: "Mul" + } + node { + output: "/encoder/layer.6/intermediate/intermediate_act_fn/Constant_2_output_0" + name: "/encoder/layer.6/intermediate/intermediate_act_fn/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.6/intermediate/intermediate_act_fn/Mul_output_0" + input: "/encoder/layer.6/intermediate/intermediate_act_fn/Constant_2_output_0" + output: "/encoder/layer.6/intermediate/intermediate_act_fn/Mul_1_output_0" + name: "/encoder/layer.6/intermediate/intermediate_act_fn/Mul_1" + op_type: "Mul" + } + node { + input: "/encoder/layer.6/intermediate/intermediate_act_fn/Mul_1_output_0" + input: "onnx::MatMul_1643" + output: "/encoder/layer.6/output/dense/MatMul_output_0" + name: "/encoder/layer.6/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.6.output.dense.bias" + input: "/encoder/layer.6/output/dense/MatMul_output_0" + output: "/encoder/layer.6/output/dense/Add_output_0" + name: "/encoder/layer.6/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.6/output/dense/Add_output_0" + input: "/encoder/layer.6/attention/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.6/output/Add_output_0" + name: "/encoder/layer.6/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.6/output/Add_output_0" + output: "/encoder/layer.6/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.6/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.6/output/Add_output_0" + input: "/encoder/layer.6/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.6/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.6/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.6/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.6/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.6/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.6/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.6/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.6/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.6/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.6/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.6/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.6/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.6/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.6/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.6/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.6/output/LayerNorm/Add_output_0" + name: "/encoder/layer.6/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.6/output/LayerNorm/Add_output_0" + output: "/encoder/layer.6/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.6/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.6/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.6/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.6/output/LayerNorm/Div_output_0" + name: "/encoder/layer.6/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.6/output/LayerNorm/Div_output_0" + input: "encoder.layer.6.output.LayerNorm.weight" + output: "/encoder/layer.6/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.6/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.6/output/LayerNorm/Mul_output_0" + input: "encoder.layer.6.output.LayerNorm.bias" + output: "/encoder/layer.6/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.6/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.6/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1644" + output: "/encoder/layer.7/attention/self/query/MatMul_output_0" + name: "/encoder/layer.7/attention/self/query/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.7.attention.self.query.bias" + input: "/encoder/layer.7/attention/self/query/MatMul_output_0" + output: "/encoder/layer.7/attention/self/query/Add_output_0" + name: "/encoder/layer.7/attention/self/query/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.6/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1645" + output: "/encoder/layer.7/attention/self/key/MatMul_output_0" + name: "/encoder/layer.7/attention/self/key/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.7.attention.self.key.bias" + input: "/encoder/layer.7/attention/self/key/MatMul_output_0" + output: "/encoder/layer.7/attention/self/key/Add_output_0" + name: "/encoder/layer.7/attention/self/key/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.7/attention/self/key/Add_output_0" + output: "/encoder/layer.7/attention/self/Shape_output_0" + name: "/encoder/layer.7/attention/self/Shape" + op_type: "Shape" + } + node { + output: "/encoder/layer.7/attention/self/Constant_output_0" + name: "/encoder/layer.7/attention/self/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.7/attention/self/Shape_output_0" + input: "/encoder/layer.7/attention/self/Constant_output_0" + output: "/encoder/layer.7/attention/self/Gather_output_0" + name: "/encoder/layer.7/attention/self/Gather" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.7/attention/self/key/Add_output_0" + output: "/encoder/layer.7/attention/self/Shape_1_output_0" + name: "/encoder/layer.7/attention/self/Shape_1" + op_type: "Shape" + } + node { + output: "/encoder/layer.7/attention/self/Constant_1_output_0" + name: "/encoder/layer.7/attention/self/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.7/attention/self/Shape_1_output_0" + input: "/encoder/layer.7/attention/self/Constant_1_output_0" + output: "/encoder/layer.7/attention/self/Gather_1_output_0" + name: "/encoder/layer.7/attention/self/Gather_1" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.7/attention/self/Gather_output_0" + output: "/encoder/layer.7/attention/self/Unsqueeze_output_0" + name: "/encoder/layer.7/attention/self/Unsqueeze" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.7/attention/self/Gather_1_output_0" + output: "/encoder/layer.7/attention/self/Unsqueeze_1_output_0" + name: "/encoder/layer.7/attention/self/Unsqueeze_1" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.7/attention/self/Constant_2_output_0" + name: "/encoder/layer.7/attention/self/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.7/attention/self/Constant_3_output_0" + name: "/encoder/layer.7/attention/self/Constant_3" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.7/attention/self/Unsqueeze_output_0" + input: "/encoder/layer.7/attention/self/Unsqueeze_1_output_0" + input: "/encoder/layer.7/attention/self/Constant_2_output_0" + input: "/encoder/layer.7/attention/self/Constant_3_output_0" + output: "/encoder/layer.7/attention/self/Concat_output_0" + name: "/encoder/layer.7/attention/self/Concat" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.7/attention/self/key/Add_output_0" + input: "/encoder/layer.7/attention/self/Concat_output_0" + output: "/encoder/layer.7/attention/self/Reshape_output_0" + name: "/encoder/layer.7/attention/self/Reshape" + op_type: "Reshape" + } + node { + input: "/encoder/layer.6/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1648" + output: "/encoder/layer.7/attention/self/value/MatMul_output_0" + name: "/encoder/layer.7/attention/self/value/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.7.attention.self.value.bias" + input: "/encoder/layer.7/attention/self/value/MatMul_output_0" + output: "/encoder/layer.7/attention/self/value/Add_output_0" + name: "/encoder/layer.7/attention/self/value/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.7/attention/self/value/Add_output_0" + output: "/encoder/layer.7/attention/self/Shape_2_output_0" + name: "/encoder/layer.7/attention/self/Shape_2" + op_type: "Shape" + } + node { + output: "/encoder/layer.7/attention/self/Constant_4_output_0" + name: "/encoder/layer.7/attention/self/Constant_4" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.7/attention/self/Shape_2_output_0" + input: "/encoder/layer.7/attention/self/Constant_4_output_0" + output: "/encoder/layer.7/attention/self/Gather_2_output_0" + name: "/encoder/layer.7/attention/self/Gather_2" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.7/attention/self/value/Add_output_0" + output: "/encoder/layer.7/attention/self/Shape_3_output_0" + name: "/encoder/layer.7/attention/self/Shape_3" + op_type: "Shape" + } + node { + output: "/encoder/layer.7/attention/self/Constant_5_output_0" + name: "/encoder/layer.7/attention/self/Constant_5" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.7/attention/self/Shape_3_output_0" + input: "/encoder/layer.7/attention/self/Constant_5_output_0" + output: "/encoder/layer.7/attention/self/Gather_3_output_0" + name: "/encoder/layer.7/attention/self/Gather_3" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.7/attention/self/Gather_2_output_0" + output: "/encoder/layer.7/attention/self/Unsqueeze_2_output_0" + name: "/encoder/layer.7/attention/self/Unsqueeze_2" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.7/attention/self/Gather_3_output_0" + output: "/encoder/layer.7/attention/self/Unsqueeze_3_output_0" + name: "/encoder/layer.7/attention/self/Unsqueeze_3" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.7/attention/self/Constant_6_output_0" + name: "/encoder/layer.7/attention/self/Constant_6" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.7/attention/self/Constant_7_output_0" + name: "/encoder/layer.7/attention/self/Constant_7" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.7/attention/self/Unsqueeze_2_output_0" + input: "/encoder/layer.7/attention/self/Unsqueeze_3_output_0" + input: "/encoder/layer.7/attention/self/Constant_6_output_0" + input: "/encoder/layer.7/attention/self/Constant_7_output_0" + output: "/encoder/layer.7/attention/self/Concat_1_output_0" + name: "/encoder/layer.7/attention/self/Concat_1" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.7/attention/self/value/Add_output_0" + input: "/encoder/layer.7/attention/self/Concat_1_output_0" + output: "/encoder/layer.7/attention/self/Reshape_1_output_0" + name: "/encoder/layer.7/attention/self/Reshape_1" + op_type: "Reshape" + } + node { + input: "/encoder/layer.7/attention/self/Reshape_1_output_0" + output: "/encoder/layer.7/attention/self/Transpose_output_0" + name: "/encoder/layer.7/attention/self/Transpose" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.7/attention/self/query/Add_output_0" + output: "/encoder/layer.7/attention/self/Shape_4_output_0" + name: "/encoder/layer.7/attention/self/Shape_4" + op_type: "Shape" + } + node { + output: "/encoder/layer.7/attention/self/Constant_8_output_0" + name: "/encoder/layer.7/attention/self/Constant_8" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.7/attention/self/Shape_4_output_0" + input: "/encoder/layer.7/attention/self/Constant_8_output_0" + output: "/encoder/layer.7/attention/self/Gather_4_output_0" + name: "/encoder/layer.7/attention/self/Gather_4" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.7/attention/self/query/Add_output_0" + output: "/encoder/layer.7/attention/self/Shape_5_output_0" + name: "/encoder/layer.7/attention/self/Shape_5" + op_type: "Shape" + } + node { + output: "/encoder/layer.7/attention/self/Constant_9_output_0" + name: "/encoder/layer.7/attention/self/Constant_9" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.7/attention/self/Shape_5_output_0" + input: "/encoder/layer.7/attention/self/Constant_9_output_0" + output: "/encoder/layer.7/attention/self/Gather_5_output_0" + name: "/encoder/layer.7/attention/self/Gather_5" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.7/attention/self/Gather_4_output_0" + output: "/encoder/layer.7/attention/self/Unsqueeze_4_output_0" + name: "/encoder/layer.7/attention/self/Unsqueeze_4" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.7/attention/self/Gather_5_output_0" + output: "/encoder/layer.7/attention/self/Unsqueeze_5_output_0" + name: "/encoder/layer.7/attention/self/Unsqueeze_5" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.7/attention/self/Constant_10_output_0" + name: "/encoder/layer.7/attention/self/Constant_10" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.7/attention/self/Constant_11_output_0" + name: "/encoder/layer.7/attention/self/Constant_11" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.7/attention/self/Unsqueeze_4_output_0" + input: "/encoder/layer.7/attention/self/Unsqueeze_5_output_0" + input: "/encoder/layer.7/attention/self/Constant_10_output_0" + input: "/encoder/layer.7/attention/self/Constant_11_output_0" + output: "/encoder/layer.7/attention/self/Concat_2_output_0" + name: "/encoder/layer.7/attention/self/Concat_2" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.7/attention/self/query/Add_output_0" + input: "/encoder/layer.7/attention/self/Concat_2_output_0" + output: "/encoder/layer.7/attention/self/Reshape_2_output_0" + name: "/encoder/layer.7/attention/self/Reshape_2" + op_type: "Reshape" + } + node { + input: "/encoder/layer.7/attention/self/Reshape_2_output_0" + output: "/encoder/layer.7/attention/self/Transpose_1_output_0" + name: "/encoder/layer.7/attention/self/Transpose_1" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.7/attention/self/Reshape_output_0" + output: "/encoder/layer.7/attention/self/Transpose_2_output_0" + name: "/encoder/layer.7/attention/self/Transpose_2" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 3 + ints: 1 + type: INTS + } + } + node { + input: "/encoder/layer.7/attention/self/Transpose_1_output_0" + input: "/encoder/layer.7/attention/self/Transpose_2_output_0" + output: "/encoder/layer.7/attention/self/MatMul_output_0" + name: "/encoder/layer.7/attention/self/MatMul" + op_type: "MatMul" + } + node { + output: "/encoder/layer.7/attention/self/Constant_12_output_0" + name: "/encoder/layer.7/attention/self/Constant_12" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000A" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.7/attention/self/MatMul_output_0" + input: "/encoder/layer.7/attention/self/Constant_12_output_0" + output: "/encoder/layer.7/attention/self/Div_output_0" + name: "/encoder/layer.7/attention/self/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.7/attention/self/Div_output_0" + input: "/Mul_1_output_0" + output: "/encoder/layer.7/attention/self/Add_output_0" + name: "/encoder/layer.7/attention/self/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.7/attention/self/Add_output_0" + output: "/encoder/layer.7/attention/self/Softmax_output_0" + name: "/encoder/layer.7/attention/self/Softmax" + op_type: "Softmax" + attribute { + name: "axis" + i: 3 + type: INT + } + } + node { + input: "/encoder/layer.7/attention/self/Softmax_output_0" + input: "/encoder/layer.7/attention/self/Transpose_output_0" + output: "/encoder/layer.7/attention/self/MatMul_1_output_0" + name: "/encoder/layer.7/attention/self/MatMul_1" + op_type: "MatMul" + } + node { + input: "/encoder/layer.7/attention/self/MatMul_1_output_0" + output: "/encoder/layer.7/attention/self/Transpose_3_output_0" + name: "/encoder/layer.7/attention/self/Transpose_3" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.7/attention/self/Transpose_3_output_0" + output: "/encoder/layer.7/attention/self/Shape_6_output_0" + name: "/encoder/layer.7/attention/self/Shape_6" + op_type: "Shape" + } + node { + output: "/encoder/layer.7/attention/self/Constant_13_output_0" + name: "/encoder/layer.7/attention/self/Constant_13" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.7/attention/self/Shape_6_output_0" + input: "/encoder/layer.7/attention/self/Constant_13_output_0" + output: "/encoder/layer.7/attention/self/Gather_6_output_0" + name: "/encoder/layer.7/attention/self/Gather_6" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.7/attention/self/Transpose_3_output_0" + output: "/encoder/layer.7/attention/self/Shape_7_output_0" + name: "/encoder/layer.7/attention/self/Shape_7" + op_type: "Shape" + } + node { + output: "/encoder/layer.7/attention/self/Constant_14_output_0" + name: "/encoder/layer.7/attention/self/Constant_14" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.7/attention/self/Shape_7_output_0" + input: "/encoder/layer.7/attention/self/Constant_14_output_0" + output: "/encoder/layer.7/attention/self/Gather_7_output_0" + name: "/encoder/layer.7/attention/self/Gather_7" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.7/attention/self/Gather_6_output_0" + output: "/encoder/layer.7/attention/self/Unsqueeze_6_output_0" + name: "/encoder/layer.7/attention/self/Unsqueeze_6" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.7/attention/self/Gather_7_output_0" + output: "/encoder/layer.7/attention/self/Unsqueeze_7_output_0" + name: "/encoder/layer.7/attention/self/Unsqueeze_7" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.7/attention/self/Constant_15_output_0" + name: "/encoder/layer.7/attention/self/Constant_15" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\000\003\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.7/attention/self/Unsqueeze_6_output_0" + input: "/encoder/layer.7/attention/self/Unsqueeze_7_output_0" + input: "/encoder/layer.7/attention/self/Constant_15_output_0" + output: "/encoder/layer.7/attention/self/Concat_3_output_0" + name: "/encoder/layer.7/attention/self/Concat_3" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.7/attention/self/Transpose_3_output_0" + input: "/encoder/layer.7/attention/self/Concat_3_output_0" + output: "/encoder/layer.7/attention/self/Reshape_3_output_0" + name: "/encoder/layer.7/attention/self/Reshape_3" + op_type: "Reshape" + } + node { + input: "/encoder/layer.7/attention/self/Reshape_3_output_0" + input: "onnx::MatMul_1654" + output: "/encoder/layer.7/attention/output/dense/MatMul_output_0" + name: "/encoder/layer.7/attention/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.7.attention.output.dense.bias" + input: "/encoder/layer.7/attention/output/dense/MatMul_output_0" + output: "/encoder/layer.7/attention/output/dense/Add_output_0" + name: "/encoder/layer.7/attention/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.7/attention/output/dense/Add_output_0" + input: "/encoder/layer.6/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.7/attention/output/Add_output_0" + name: "/encoder/layer.7/attention/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.7/attention/output/Add_output_0" + output: "/encoder/layer.7/attention/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.7/attention/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.7/attention/output/Add_output_0" + input: "/encoder/layer.7/attention/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.7/attention/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.7/attention/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.7/attention/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.7/attention/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.7/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.7/attention/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.7/attention/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.7/attention/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.7/attention/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.7/attention/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.7/attention/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.7/attention/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.7/attention/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.7/attention/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.7/attention/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.7/attention/output/LayerNorm/Add_output_0" + name: "/encoder/layer.7/attention/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.7/attention/output/LayerNorm/Add_output_0" + output: "/encoder/layer.7/attention/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.7/attention/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.7/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.7/attention/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.7/attention/output/LayerNorm/Div_output_0" + name: "/encoder/layer.7/attention/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.7/attention/output/LayerNorm/Div_output_0" + input: "encoder.layer.7.attention.output.LayerNorm.weight" + output: "/encoder/layer.7/attention/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.7/attention/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.7/attention/output/LayerNorm/Mul_output_0" + input: "encoder.layer.7.attention.output.LayerNorm.bias" + output: "/encoder/layer.7/attention/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.7/attention/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.7/attention/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1655" + output: "/encoder/layer.7/intermediate/dense/MatMul_output_0" + name: "/encoder/layer.7/intermediate/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.7.intermediate.dense.bias" + input: "/encoder/layer.7/intermediate/dense/MatMul_output_0" + output: "/encoder/layer.7/intermediate/dense/Add_output_0" + name: "/encoder/layer.7/intermediate/dense/Add" + op_type: "Add" + } + node { + output: "/encoder/layer.7/intermediate/intermediate_act_fn/Constant_output_0" + name: "/encoder/layer.7/intermediate/intermediate_act_fn/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\363\004\265?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.7/intermediate/dense/Add_output_0" + input: "/encoder/layer.7/intermediate/intermediate_act_fn/Constant_output_0" + output: "/encoder/layer.7/intermediate/intermediate_act_fn/Div_output_0" + name: "/encoder/layer.7/intermediate/intermediate_act_fn/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.7/intermediate/intermediate_act_fn/Div_output_0" + output: "/encoder/layer.7/intermediate/intermediate_act_fn/Erf_output_0" + name: "/encoder/layer.7/intermediate/intermediate_act_fn/Erf" + op_type: "Erf" + } + node { + output: "/encoder/layer.7/intermediate/intermediate_act_fn/Constant_1_output_0" + name: "/encoder/layer.7/intermediate/intermediate_act_fn/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\200?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.7/intermediate/intermediate_act_fn/Erf_output_0" + input: "/encoder/layer.7/intermediate/intermediate_act_fn/Constant_1_output_0" + output: "/encoder/layer.7/intermediate/intermediate_act_fn/Add_output_0" + name: "/encoder/layer.7/intermediate/intermediate_act_fn/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.7/intermediate/dense/Add_output_0" + input: "/encoder/layer.7/intermediate/intermediate_act_fn/Add_output_0" + output: "/encoder/layer.7/intermediate/intermediate_act_fn/Mul_output_0" + name: "/encoder/layer.7/intermediate/intermediate_act_fn/Mul" + op_type: "Mul" + } + node { + output: "/encoder/layer.7/intermediate/intermediate_act_fn/Constant_2_output_0" + name: "/encoder/layer.7/intermediate/intermediate_act_fn/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.7/intermediate/intermediate_act_fn/Mul_output_0" + input: "/encoder/layer.7/intermediate/intermediate_act_fn/Constant_2_output_0" + output: "/encoder/layer.7/intermediate/intermediate_act_fn/Mul_1_output_0" + name: "/encoder/layer.7/intermediate/intermediate_act_fn/Mul_1" + op_type: "Mul" + } + node { + input: "/encoder/layer.7/intermediate/intermediate_act_fn/Mul_1_output_0" + input: "onnx::MatMul_1656" + output: "/encoder/layer.7/output/dense/MatMul_output_0" + name: "/encoder/layer.7/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.7.output.dense.bias" + input: "/encoder/layer.7/output/dense/MatMul_output_0" + output: "/encoder/layer.7/output/dense/Add_output_0" + name: "/encoder/layer.7/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.7/output/dense/Add_output_0" + input: "/encoder/layer.7/attention/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.7/output/Add_output_0" + name: "/encoder/layer.7/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.7/output/Add_output_0" + output: "/encoder/layer.7/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.7/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.7/output/Add_output_0" + input: "/encoder/layer.7/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.7/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.7/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.7/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.7/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.7/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.7/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.7/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.7/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.7/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.7/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.7/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.7/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.7/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.7/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.7/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.7/output/LayerNorm/Add_output_0" + name: "/encoder/layer.7/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.7/output/LayerNorm/Add_output_0" + output: "/encoder/layer.7/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.7/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.7/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.7/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.7/output/LayerNorm/Div_output_0" + name: "/encoder/layer.7/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.7/output/LayerNorm/Div_output_0" + input: "encoder.layer.7.output.LayerNorm.weight" + output: "/encoder/layer.7/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.7/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.7/output/LayerNorm/Mul_output_0" + input: "encoder.layer.7.output.LayerNorm.bias" + output: "/encoder/layer.7/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.7/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.7/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1657" + output: "/encoder/layer.8/attention/self/query/MatMul_output_0" + name: "/encoder/layer.8/attention/self/query/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.8.attention.self.query.bias" + input: "/encoder/layer.8/attention/self/query/MatMul_output_0" + output: "/encoder/layer.8/attention/self/query/Add_output_0" + name: "/encoder/layer.8/attention/self/query/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.7/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1658" + output: "/encoder/layer.8/attention/self/key/MatMul_output_0" + name: "/encoder/layer.8/attention/self/key/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.8.attention.self.key.bias" + input: "/encoder/layer.8/attention/self/key/MatMul_output_0" + output: "/encoder/layer.8/attention/self/key/Add_output_0" + name: "/encoder/layer.8/attention/self/key/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.8/attention/self/key/Add_output_0" + output: "/encoder/layer.8/attention/self/Shape_output_0" + name: "/encoder/layer.8/attention/self/Shape" + op_type: "Shape" + } + node { + output: "/encoder/layer.8/attention/self/Constant_output_0" + name: "/encoder/layer.8/attention/self/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.8/attention/self/Shape_output_0" + input: "/encoder/layer.8/attention/self/Constant_output_0" + output: "/encoder/layer.8/attention/self/Gather_output_0" + name: "/encoder/layer.8/attention/self/Gather" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.8/attention/self/key/Add_output_0" + output: "/encoder/layer.8/attention/self/Shape_1_output_0" + name: "/encoder/layer.8/attention/self/Shape_1" + op_type: "Shape" + } + node { + output: "/encoder/layer.8/attention/self/Constant_1_output_0" + name: "/encoder/layer.8/attention/self/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.8/attention/self/Shape_1_output_0" + input: "/encoder/layer.8/attention/self/Constant_1_output_0" + output: "/encoder/layer.8/attention/self/Gather_1_output_0" + name: "/encoder/layer.8/attention/self/Gather_1" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.8/attention/self/Gather_output_0" + output: "/encoder/layer.8/attention/self/Unsqueeze_output_0" + name: "/encoder/layer.8/attention/self/Unsqueeze" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.8/attention/self/Gather_1_output_0" + output: "/encoder/layer.8/attention/self/Unsqueeze_1_output_0" + name: "/encoder/layer.8/attention/self/Unsqueeze_1" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.8/attention/self/Constant_2_output_0" + name: "/encoder/layer.8/attention/self/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.8/attention/self/Constant_3_output_0" + name: "/encoder/layer.8/attention/self/Constant_3" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.8/attention/self/Unsqueeze_output_0" + input: "/encoder/layer.8/attention/self/Unsqueeze_1_output_0" + input: "/encoder/layer.8/attention/self/Constant_2_output_0" + input: "/encoder/layer.8/attention/self/Constant_3_output_0" + output: "/encoder/layer.8/attention/self/Concat_output_0" + name: "/encoder/layer.8/attention/self/Concat" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.8/attention/self/key/Add_output_0" + input: "/encoder/layer.8/attention/self/Concat_output_0" + output: "/encoder/layer.8/attention/self/Reshape_output_0" + name: "/encoder/layer.8/attention/self/Reshape" + op_type: "Reshape" + } + node { + input: "/encoder/layer.7/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1661" + output: "/encoder/layer.8/attention/self/value/MatMul_output_0" + name: "/encoder/layer.8/attention/self/value/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.8.attention.self.value.bias" + input: "/encoder/layer.8/attention/self/value/MatMul_output_0" + output: "/encoder/layer.8/attention/self/value/Add_output_0" + name: "/encoder/layer.8/attention/self/value/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.8/attention/self/value/Add_output_0" + output: "/encoder/layer.8/attention/self/Shape_2_output_0" + name: "/encoder/layer.8/attention/self/Shape_2" + op_type: "Shape" + } + node { + output: "/encoder/layer.8/attention/self/Constant_4_output_0" + name: "/encoder/layer.8/attention/self/Constant_4" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.8/attention/self/Shape_2_output_0" + input: "/encoder/layer.8/attention/self/Constant_4_output_0" + output: "/encoder/layer.8/attention/self/Gather_2_output_0" + name: "/encoder/layer.8/attention/self/Gather_2" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.8/attention/self/value/Add_output_0" + output: "/encoder/layer.8/attention/self/Shape_3_output_0" + name: "/encoder/layer.8/attention/self/Shape_3" + op_type: "Shape" + } + node { + output: "/encoder/layer.8/attention/self/Constant_5_output_0" + name: "/encoder/layer.8/attention/self/Constant_5" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.8/attention/self/Shape_3_output_0" + input: "/encoder/layer.8/attention/self/Constant_5_output_0" + output: "/encoder/layer.8/attention/self/Gather_3_output_0" + name: "/encoder/layer.8/attention/self/Gather_3" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.8/attention/self/Gather_2_output_0" + output: "/encoder/layer.8/attention/self/Unsqueeze_2_output_0" + name: "/encoder/layer.8/attention/self/Unsqueeze_2" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.8/attention/self/Gather_3_output_0" + output: "/encoder/layer.8/attention/self/Unsqueeze_3_output_0" + name: "/encoder/layer.8/attention/self/Unsqueeze_3" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.8/attention/self/Constant_6_output_0" + name: "/encoder/layer.8/attention/self/Constant_6" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.8/attention/self/Constant_7_output_0" + name: "/encoder/layer.8/attention/self/Constant_7" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.8/attention/self/Unsqueeze_2_output_0" + input: "/encoder/layer.8/attention/self/Unsqueeze_3_output_0" + input: "/encoder/layer.8/attention/self/Constant_6_output_0" + input: "/encoder/layer.8/attention/self/Constant_7_output_0" + output: "/encoder/layer.8/attention/self/Concat_1_output_0" + name: "/encoder/layer.8/attention/self/Concat_1" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.8/attention/self/value/Add_output_0" + input: "/encoder/layer.8/attention/self/Concat_1_output_0" + output: "/encoder/layer.8/attention/self/Reshape_1_output_0" + name: "/encoder/layer.8/attention/self/Reshape_1" + op_type: "Reshape" + } + node { + input: "/encoder/layer.8/attention/self/Reshape_1_output_0" + output: "/encoder/layer.8/attention/self/Transpose_output_0" + name: "/encoder/layer.8/attention/self/Transpose" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.8/attention/self/query/Add_output_0" + output: "/encoder/layer.8/attention/self/Shape_4_output_0" + name: "/encoder/layer.8/attention/self/Shape_4" + op_type: "Shape" + } + node { + output: "/encoder/layer.8/attention/self/Constant_8_output_0" + name: "/encoder/layer.8/attention/self/Constant_8" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.8/attention/self/Shape_4_output_0" + input: "/encoder/layer.8/attention/self/Constant_8_output_0" + output: "/encoder/layer.8/attention/self/Gather_4_output_0" + name: "/encoder/layer.8/attention/self/Gather_4" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.8/attention/self/query/Add_output_0" + output: "/encoder/layer.8/attention/self/Shape_5_output_0" + name: "/encoder/layer.8/attention/self/Shape_5" + op_type: "Shape" + } + node { + output: "/encoder/layer.8/attention/self/Constant_9_output_0" + name: "/encoder/layer.8/attention/self/Constant_9" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.8/attention/self/Shape_5_output_0" + input: "/encoder/layer.8/attention/self/Constant_9_output_0" + output: "/encoder/layer.8/attention/self/Gather_5_output_0" + name: "/encoder/layer.8/attention/self/Gather_5" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.8/attention/self/Gather_4_output_0" + output: "/encoder/layer.8/attention/self/Unsqueeze_4_output_0" + name: "/encoder/layer.8/attention/self/Unsqueeze_4" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.8/attention/self/Gather_5_output_0" + output: "/encoder/layer.8/attention/self/Unsqueeze_5_output_0" + name: "/encoder/layer.8/attention/self/Unsqueeze_5" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.8/attention/self/Constant_10_output_0" + name: "/encoder/layer.8/attention/self/Constant_10" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.8/attention/self/Constant_11_output_0" + name: "/encoder/layer.8/attention/self/Constant_11" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.8/attention/self/Unsqueeze_4_output_0" + input: "/encoder/layer.8/attention/self/Unsqueeze_5_output_0" + input: "/encoder/layer.8/attention/self/Constant_10_output_0" + input: "/encoder/layer.8/attention/self/Constant_11_output_0" + output: "/encoder/layer.8/attention/self/Concat_2_output_0" + name: "/encoder/layer.8/attention/self/Concat_2" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.8/attention/self/query/Add_output_0" + input: "/encoder/layer.8/attention/self/Concat_2_output_0" + output: "/encoder/layer.8/attention/self/Reshape_2_output_0" + name: "/encoder/layer.8/attention/self/Reshape_2" + op_type: "Reshape" + } + node { + input: "/encoder/layer.8/attention/self/Reshape_2_output_0" + output: "/encoder/layer.8/attention/self/Transpose_1_output_0" + name: "/encoder/layer.8/attention/self/Transpose_1" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.8/attention/self/Reshape_output_0" + output: "/encoder/layer.8/attention/self/Transpose_2_output_0" + name: "/encoder/layer.8/attention/self/Transpose_2" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 3 + ints: 1 + type: INTS + } + } + node { + input: "/encoder/layer.8/attention/self/Transpose_1_output_0" + input: "/encoder/layer.8/attention/self/Transpose_2_output_0" + output: "/encoder/layer.8/attention/self/MatMul_output_0" + name: "/encoder/layer.8/attention/self/MatMul" + op_type: "MatMul" + } + node { + output: "/encoder/layer.8/attention/self/Constant_12_output_0" + name: "/encoder/layer.8/attention/self/Constant_12" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000A" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.8/attention/self/MatMul_output_0" + input: "/encoder/layer.8/attention/self/Constant_12_output_0" + output: "/encoder/layer.8/attention/self/Div_output_0" + name: "/encoder/layer.8/attention/self/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.8/attention/self/Div_output_0" + input: "/Mul_1_output_0" + output: "/encoder/layer.8/attention/self/Add_output_0" + name: "/encoder/layer.8/attention/self/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.8/attention/self/Add_output_0" + output: "/encoder/layer.8/attention/self/Softmax_output_0" + name: "/encoder/layer.8/attention/self/Softmax" + op_type: "Softmax" + attribute { + name: "axis" + i: 3 + type: INT + } + } + node { + input: "/encoder/layer.8/attention/self/Softmax_output_0" + input: "/encoder/layer.8/attention/self/Transpose_output_0" + output: "/encoder/layer.8/attention/self/MatMul_1_output_0" + name: "/encoder/layer.8/attention/self/MatMul_1" + op_type: "MatMul" + } + node { + input: "/encoder/layer.8/attention/self/MatMul_1_output_0" + output: "/encoder/layer.8/attention/self/Transpose_3_output_0" + name: "/encoder/layer.8/attention/self/Transpose_3" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.8/attention/self/Transpose_3_output_0" + output: "/encoder/layer.8/attention/self/Shape_6_output_0" + name: "/encoder/layer.8/attention/self/Shape_6" + op_type: "Shape" + } + node { + output: "/encoder/layer.8/attention/self/Constant_13_output_0" + name: "/encoder/layer.8/attention/self/Constant_13" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.8/attention/self/Shape_6_output_0" + input: "/encoder/layer.8/attention/self/Constant_13_output_0" + output: "/encoder/layer.8/attention/self/Gather_6_output_0" + name: "/encoder/layer.8/attention/self/Gather_6" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.8/attention/self/Transpose_3_output_0" + output: "/encoder/layer.8/attention/self/Shape_7_output_0" + name: "/encoder/layer.8/attention/self/Shape_7" + op_type: "Shape" + } + node { + output: "/encoder/layer.8/attention/self/Constant_14_output_0" + name: "/encoder/layer.8/attention/self/Constant_14" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.8/attention/self/Shape_7_output_0" + input: "/encoder/layer.8/attention/self/Constant_14_output_0" + output: "/encoder/layer.8/attention/self/Gather_7_output_0" + name: "/encoder/layer.8/attention/self/Gather_7" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.8/attention/self/Gather_6_output_0" + output: "/encoder/layer.8/attention/self/Unsqueeze_6_output_0" + name: "/encoder/layer.8/attention/self/Unsqueeze_6" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.8/attention/self/Gather_7_output_0" + output: "/encoder/layer.8/attention/self/Unsqueeze_7_output_0" + name: "/encoder/layer.8/attention/self/Unsqueeze_7" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.8/attention/self/Constant_15_output_0" + name: "/encoder/layer.8/attention/self/Constant_15" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\000\003\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.8/attention/self/Unsqueeze_6_output_0" + input: "/encoder/layer.8/attention/self/Unsqueeze_7_output_0" + input: "/encoder/layer.8/attention/self/Constant_15_output_0" + output: "/encoder/layer.8/attention/self/Concat_3_output_0" + name: "/encoder/layer.8/attention/self/Concat_3" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.8/attention/self/Transpose_3_output_0" + input: "/encoder/layer.8/attention/self/Concat_3_output_0" + output: "/encoder/layer.8/attention/self/Reshape_3_output_0" + name: "/encoder/layer.8/attention/self/Reshape_3" + op_type: "Reshape" + } + node { + input: "/encoder/layer.8/attention/self/Reshape_3_output_0" + input: "onnx::MatMul_1667" + output: "/encoder/layer.8/attention/output/dense/MatMul_output_0" + name: "/encoder/layer.8/attention/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.8.attention.output.dense.bias" + input: "/encoder/layer.8/attention/output/dense/MatMul_output_0" + output: "/encoder/layer.8/attention/output/dense/Add_output_0" + name: "/encoder/layer.8/attention/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.8/attention/output/dense/Add_output_0" + input: "/encoder/layer.7/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.8/attention/output/Add_output_0" + name: "/encoder/layer.8/attention/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.8/attention/output/Add_output_0" + output: "/encoder/layer.8/attention/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.8/attention/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.8/attention/output/Add_output_0" + input: "/encoder/layer.8/attention/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.8/attention/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.8/attention/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.8/attention/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.8/attention/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.8/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.8/attention/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.8/attention/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.8/attention/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.8/attention/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.8/attention/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.8/attention/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.8/attention/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.8/attention/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.8/attention/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.8/attention/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.8/attention/output/LayerNorm/Add_output_0" + name: "/encoder/layer.8/attention/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.8/attention/output/LayerNorm/Add_output_0" + output: "/encoder/layer.8/attention/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.8/attention/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.8/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.8/attention/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.8/attention/output/LayerNorm/Div_output_0" + name: "/encoder/layer.8/attention/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.8/attention/output/LayerNorm/Div_output_0" + input: "encoder.layer.8.attention.output.LayerNorm.weight" + output: "/encoder/layer.8/attention/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.8/attention/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.8/attention/output/LayerNorm/Mul_output_0" + input: "encoder.layer.8.attention.output.LayerNorm.bias" + output: "/encoder/layer.8/attention/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.8/attention/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.8/attention/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1668" + output: "/encoder/layer.8/intermediate/dense/MatMul_output_0" + name: "/encoder/layer.8/intermediate/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.8.intermediate.dense.bias" + input: "/encoder/layer.8/intermediate/dense/MatMul_output_0" + output: "/encoder/layer.8/intermediate/dense/Add_output_0" + name: "/encoder/layer.8/intermediate/dense/Add" + op_type: "Add" + } + node { + output: "/encoder/layer.8/intermediate/intermediate_act_fn/Constant_output_0" + name: "/encoder/layer.8/intermediate/intermediate_act_fn/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\363\004\265?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.8/intermediate/dense/Add_output_0" + input: "/encoder/layer.8/intermediate/intermediate_act_fn/Constant_output_0" + output: "/encoder/layer.8/intermediate/intermediate_act_fn/Div_output_0" + name: "/encoder/layer.8/intermediate/intermediate_act_fn/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.8/intermediate/intermediate_act_fn/Div_output_0" + output: "/encoder/layer.8/intermediate/intermediate_act_fn/Erf_output_0" + name: "/encoder/layer.8/intermediate/intermediate_act_fn/Erf" + op_type: "Erf" + } + node { + output: "/encoder/layer.8/intermediate/intermediate_act_fn/Constant_1_output_0" + name: "/encoder/layer.8/intermediate/intermediate_act_fn/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\200?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.8/intermediate/intermediate_act_fn/Erf_output_0" + input: "/encoder/layer.8/intermediate/intermediate_act_fn/Constant_1_output_0" + output: "/encoder/layer.8/intermediate/intermediate_act_fn/Add_output_0" + name: "/encoder/layer.8/intermediate/intermediate_act_fn/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.8/intermediate/dense/Add_output_0" + input: "/encoder/layer.8/intermediate/intermediate_act_fn/Add_output_0" + output: "/encoder/layer.8/intermediate/intermediate_act_fn/Mul_output_0" + name: "/encoder/layer.8/intermediate/intermediate_act_fn/Mul" + op_type: "Mul" + } + node { + output: "/encoder/layer.8/intermediate/intermediate_act_fn/Constant_2_output_0" + name: "/encoder/layer.8/intermediate/intermediate_act_fn/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.8/intermediate/intermediate_act_fn/Mul_output_0" + input: "/encoder/layer.8/intermediate/intermediate_act_fn/Constant_2_output_0" + output: "/encoder/layer.8/intermediate/intermediate_act_fn/Mul_1_output_0" + name: "/encoder/layer.8/intermediate/intermediate_act_fn/Mul_1" + op_type: "Mul" + } + node { + input: "/encoder/layer.8/intermediate/intermediate_act_fn/Mul_1_output_0" + input: "onnx::MatMul_1669" + output: "/encoder/layer.8/output/dense/MatMul_output_0" + name: "/encoder/layer.8/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.8.output.dense.bias" + input: "/encoder/layer.8/output/dense/MatMul_output_0" + output: "/encoder/layer.8/output/dense/Add_output_0" + name: "/encoder/layer.8/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.8/output/dense/Add_output_0" + input: "/encoder/layer.8/attention/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.8/output/Add_output_0" + name: "/encoder/layer.8/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.8/output/Add_output_0" + output: "/encoder/layer.8/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.8/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.8/output/Add_output_0" + input: "/encoder/layer.8/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.8/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.8/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.8/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.8/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.8/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.8/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.8/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.8/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.8/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.8/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.8/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.8/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.8/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.8/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.8/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.8/output/LayerNorm/Add_output_0" + name: "/encoder/layer.8/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.8/output/LayerNorm/Add_output_0" + output: "/encoder/layer.8/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.8/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.8/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.8/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.8/output/LayerNorm/Div_output_0" + name: "/encoder/layer.8/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.8/output/LayerNorm/Div_output_0" + input: "encoder.layer.8.output.LayerNorm.weight" + output: "/encoder/layer.8/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.8/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.8/output/LayerNorm/Mul_output_0" + input: "encoder.layer.8.output.LayerNorm.bias" + output: "/encoder/layer.8/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.8/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.8/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1670" + output: "/encoder/layer.9/attention/self/query/MatMul_output_0" + name: "/encoder/layer.9/attention/self/query/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.9.attention.self.query.bias" + input: "/encoder/layer.9/attention/self/query/MatMul_output_0" + output: "/encoder/layer.9/attention/self/query/Add_output_0" + name: "/encoder/layer.9/attention/self/query/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.8/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1671" + output: "/encoder/layer.9/attention/self/key/MatMul_output_0" + name: "/encoder/layer.9/attention/self/key/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.9.attention.self.key.bias" + input: "/encoder/layer.9/attention/self/key/MatMul_output_0" + output: "/encoder/layer.9/attention/self/key/Add_output_0" + name: "/encoder/layer.9/attention/self/key/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.9/attention/self/key/Add_output_0" + output: "/encoder/layer.9/attention/self/Shape_output_0" + name: "/encoder/layer.9/attention/self/Shape" + op_type: "Shape" + } + node { + output: "/encoder/layer.9/attention/self/Constant_output_0" + name: "/encoder/layer.9/attention/self/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.9/attention/self/Shape_output_0" + input: "/encoder/layer.9/attention/self/Constant_output_0" + output: "/encoder/layer.9/attention/self/Gather_output_0" + name: "/encoder/layer.9/attention/self/Gather" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.9/attention/self/key/Add_output_0" + output: "/encoder/layer.9/attention/self/Shape_1_output_0" + name: "/encoder/layer.9/attention/self/Shape_1" + op_type: "Shape" + } + node { + output: "/encoder/layer.9/attention/self/Constant_1_output_0" + name: "/encoder/layer.9/attention/self/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.9/attention/self/Shape_1_output_0" + input: "/encoder/layer.9/attention/self/Constant_1_output_0" + output: "/encoder/layer.9/attention/self/Gather_1_output_0" + name: "/encoder/layer.9/attention/self/Gather_1" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.9/attention/self/Gather_output_0" + output: "/encoder/layer.9/attention/self/Unsqueeze_output_0" + name: "/encoder/layer.9/attention/self/Unsqueeze" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.9/attention/self/Gather_1_output_0" + output: "/encoder/layer.9/attention/self/Unsqueeze_1_output_0" + name: "/encoder/layer.9/attention/self/Unsqueeze_1" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.9/attention/self/Constant_2_output_0" + name: "/encoder/layer.9/attention/self/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.9/attention/self/Constant_3_output_0" + name: "/encoder/layer.9/attention/self/Constant_3" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.9/attention/self/Unsqueeze_output_0" + input: "/encoder/layer.9/attention/self/Unsqueeze_1_output_0" + input: "/encoder/layer.9/attention/self/Constant_2_output_0" + input: "/encoder/layer.9/attention/self/Constant_3_output_0" + output: "/encoder/layer.9/attention/self/Concat_output_0" + name: "/encoder/layer.9/attention/self/Concat" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.9/attention/self/key/Add_output_0" + input: "/encoder/layer.9/attention/self/Concat_output_0" + output: "/encoder/layer.9/attention/self/Reshape_output_0" + name: "/encoder/layer.9/attention/self/Reshape" + op_type: "Reshape" + } + node { + input: "/encoder/layer.8/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1674" + output: "/encoder/layer.9/attention/self/value/MatMul_output_0" + name: "/encoder/layer.9/attention/self/value/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.9.attention.self.value.bias" + input: "/encoder/layer.9/attention/self/value/MatMul_output_0" + output: "/encoder/layer.9/attention/self/value/Add_output_0" + name: "/encoder/layer.9/attention/self/value/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.9/attention/self/value/Add_output_0" + output: "/encoder/layer.9/attention/self/Shape_2_output_0" + name: "/encoder/layer.9/attention/self/Shape_2" + op_type: "Shape" + } + node { + output: "/encoder/layer.9/attention/self/Constant_4_output_0" + name: "/encoder/layer.9/attention/self/Constant_4" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.9/attention/self/Shape_2_output_0" + input: "/encoder/layer.9/attention/self/Constant_4_output_0" + output: "/encoder/layer.9/attention/self/Gather_2_output_0" + name: "/encoder/layer.9/attention/self/Gather_2" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.9/attention/self/value/Add_output_0" + output: "/encoder/layer.9/attention/self/Shape_3_output_0" + name: "/encoder/layer.9/attention/self/Shape_3" + op_type: "Shape" + } + node { + output: "/encoder/layer.9/attention/self/Constant_5_output_0" + name: "/encoder/layer.9/attention/self/Constant_5" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.9/attention/self/Shape_3_output_0" + input: "/encoder/layer.9/attention/self/Constant_5_output_0" + output: "/encoder/layer.9/attention/self/Gather_3_output_0" + name: "/encoder/layer.9/attention/self/Gather_3" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.9/attention/self/Gather_2_output_0" + output: "/encoder/layer.9/attention/self/Unsqueeze_2_output_0" + name: "/encoder/layer.9/attention/self/Unsqueeze_2" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.9/attention/self/Gather_3_output_0" + output: "/encoder/layer.9/attention/self/Unsqueeze_3_output_0" + name: "/encoder/layer.9/attention/self/Unsqueeze_3" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.9/attention/self/Constant_6_output_0" + name: "/encoder/layer.9/attention/self/Constant_6" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.9/attention/self/Constant_7_output_0" + name: "/encoder/layer.9/attention/self/Constant_7" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.9/attention/self/Unsqueeze_2_output_0" + input: "/encoder/layer.9/attention/self/Unsqueeze_3_output_0" + input: "/encoder/layer.9/attention/self/Constant_6_output_0" + input: "/encoder/layer.9/attention/self/Constant_7_output_0" + output: "/encoder/layer.9/attention/self/Concat_1_output_0" + name: "/encoder/layer.9/attention/self/Concat_1" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.9/attention/self/value/Add_output_0" + input: "/encoder/layer.9/attention/self/Concat_1_output_0" + output: "/encoder/layer.9/attention/self/Reshape_1_output_0" + name: "/encoder/layer.9/attention/self/Reshape_1" + op_type: "Reshape" + } + node { + input: "/encoder/layer.9/attention/self/Reshape_1_output_0" + output: "/encoder/layer.9/attention/self/Transpose_output_0" + name: "/encoder/layer.9/attention/self/Transpose" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.9/attention/self/query/Add_output_0" + output: "/encoder/layer.9/attention/self/Shape_4_output_0" + name: "/encoder/layer.9/attention/self/Shape_4" + op_type: "Shape" + } + node { + output: "/encoder/layer.9/attention/self/Constant_8_output_0" + name: "/encoder/layer.9/attention/self/Constant_8" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.9/attention/self/Shape_4_output_0" + input: "/encoder/layer.9/attention/self/Constant_8_output_0" + output: "/encoder/layer.9/attention/self/Gather_4_output_0" + name: "/encoder/layer.9/attention/self/Gather_4" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.9/attention/self/query/Add_output_0" + output: "/encoder/layer.9/attention/self/Shape_5_output_0" + name: "/encoder/layer.9/attention/self/Shape_5" + op_type: "Shape" + } + node { + output: "/encoder/layer.9/attention/self/Constant_9_output_0" + name: "/encoder/layer.9/attention/self/Constant_9" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.9/attention/self/Shape_5_output_0" + input: "/encoder/layer.9/attention/self/Constant_9_output_0" + output: "/encoder/layer.9/attention/self/Gather_5_output_0" + name: "/encoder/layer.9/attention/self/Gather_5" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.9/attention/self/Gather_4_output_0" + output: "/encoder/layer.9/attention/self/Unsqueeze_4_output_0" + name: "/encoder/layer.9/attention/self/Unsqueeze_4" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.9/attention/self/Gather_5_output_0" + output: "/encoder/layer.9/attention/self/Unsqueeze_5_output_0" + name: "/encoder/layer.9/attention/self/Unsqueeze_5" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.9/attention/self/Constant_10_output_0" + name: "/encoder/layer.9/attention/self/Constant_10" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.9/attention/self/Constant_11_output_0" + name: "/encoder/layer.9/attention/self/Constant_11" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.9/attention/self/Unsqueeze_4_output_0" + input: "/encoder/layer.9/attention/self/Unsqueeze_5_output_0" + input: "/encoder/layer.9/attention/self/Constant_10_output_0" + input: "/encoder/layer.9/attention/self/Constant_11_output_0" + output: "/encoder/layer.9/attention/self/Concat_2_output_0" + name: "/encoder/layer.9/attention/self/Concat_2" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.9/attention/self/query/Add_output_0" + input: "/encoder/layer.9/attention/self/Concat_2_output_0" + output: "/encoder/layer.9/attention/self/Reshape_2_output_0" + name: "/encoder/layer.9/attention/self/Reshape_2" + op_type: "Reshape" + } + node { + input: "/encoder/layer.9/attention/self/Reshape_2_output_0" + output: "/encoder/layer.9/attention/self/Transpose_1_output_0" + name: "/encoder/layer.9/attention/self/Transpose_1" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.9/attention/self/Reshape_output_0" + output: "/encoder/layer.9/attention/self/Transpose_2_output_0" + name: "/encoder/layer.9/attention/self/Transpose_2" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 3 + ints: 1 + type: INTS + } + } + node { + input: "/encoder/layer.9/attention/self/Transpose_1_output_0" + input: "/encoder/layer.9/attention/self/Transpose_2_output_0" + output: "/encoder/layer.9/attention/self/MatMul_output_0" + name: "/encoder/layer.9/attention/self/MatMul" + op_type: "MatMul" + } + node { + output: "/encoder/layer.9/attention/self/Constant_12_output_0" + name: "/encoder/layer.9/attention/self/Constant_12" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000A" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.9/attention/self/MatMul_output_0" + input: "/encoder/layer.9/attention/self/Constant_12_output_0" + output: "/encoder/layer.9/attention/self/Div_output_0" + name: "/encoder/layer.9/attention/self/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.9/attention/self/Div_output_0" + input: "/Mul_1_output_0" + output: "/encoder/layer.9/attention/self/Add_output_0" + name: "/encoder/layer.9/attention/self/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.9/attention/self/Add_output_0" + output: "/encoder/layer.9/attention/self/Softmax_output_0" + name: "/encoder/layer.9/attention/self/Softmax" + op_type: "Softmax" + attribute { + name: "axis" + i: 3 + type: INT + } + } + node { + input: "/encoder/layer.9/attention/self/Softmax_output_0" + input: "/encoder/layer.9/attention/self/Transpose_output_0" + output: "/encoder/layer.9/attention/self/MatMul_1_output_0" + name: "/encoder/layer.9/attention/self/MatMul_1" + op_type: "MatMul" + } + node { + input: "/encoder/layer.9/attention/self/MatMul_1_output_0" + output: "/encoder/layer.9/attention/self/Transpose_3_output_0" + name: "/encoder/layer.9/attention/self/Transpose_3" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.9/attention/self/Transpose_3_output_0" + output: "/encoder/layer.9/attention/self/Shape_6_output_0" + name: "/encoder/layer.9/attention/self/Shape_6" + op_type: "Shape" + } + node { + output: "/encoder/layer.9/attention/self/Constant_13_output_0" + name: "/encoder/layer.9/attention/self/Constant_13" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.9/attention/self/Shape_6_output_0" + input: "/encoder/layer.9/attention/self/Constant_13_output_0" + output: "/encoder/layer.9/attention/self/Gather_6_output_0" + name: "/encoder/layer.9/attention/self/Gather_6" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.9/attention/self/Transpose_3_output_0" + output: "/encoder/layer.9/attention/self/Shape_7_output_0" + name: "/encoder/layer.9/attention/self/Shape_7" + op_type: "Shape" + } + node { + output: "/encoder/layer.9/attention/self/Constant_14_output_0" + name: "/encoder/layer.9/attention/self/Constant_14" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.9/attention/self/Shape_7_output_0" + input: "/encoder/layer.9/attention/self/Constant_14_output_0" + output: "/encoder/layer.9/attention/self/Gather_7_output_0" + name: "/encoder/layer.9/attention/self/Gather_7" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.9/attention/self/Gather_6_output_0" + output: "/encoder/layer.9/attention/self/Unsqueeze_6_output_0" + name: "/encoder/layer.9/attention/self/Unsqueeze_6" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.9/attention/self/Gather_7_output_0" + output: "/encoder/layer.9/attention/self/Unsqueeze_7_output_0" + name: "/encoder/layer.9/attention/self/Unsqueeze_7" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.9/attention/self/Constant_15_output_0" + name: "/encoder/layer.9/attention/self/Constant_15" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\000\003\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.9/attention/self/Unsqueeze_6_output_0" + input: "/encoder/layer.9/attention/self/Unsqueeze_7_output_0" + input: "/encoder/layer.9/attention/self/Constant_15_output_0" + output: "/encoder/layer.9/attention/self/Concat_3_output_0" + name: "/encoder/layer.9/attention/self/Concat_3" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.9/attention/self/Transpose_3_output_0" + input: "/encoder/layer.9/attention/self/Concat_3_output_0" + output: "/encoder/layer.9/attention/self/Reshape_3_output_0" + name: "/encoder/layer.9/attention/self/Reshape_3" + op_type: "Reshape" + } + node { + input: "/encoder/layer.9/attention/self/Reshape_3_output_0" + input: "onnx::MatMul_1680" + output: "/encoder/layer.9/attention/output/dense/MatMul_output_0" + name: "/encoder/layer.9/attention/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.9.attention.output.dense.bias" + input: "/encoder/layer.9/attention/output/dense/MatMul_output_0" + output: "/encoder/layer.9/attention/output/dense/Add_output_0" + name: "/encoder/layer.9/attention/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.9/attention/output/dense/Add_output_0" + input: "/encoder/layer.8/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.9/attention/output/Add_output_0" + name: "/encoder/layer.9/attention/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.9/attention/output/Add_output_0" + output: "/encoder/layer.9/attention/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.9/attention/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.9/attention/output/Add_output_0" + input: "/encoder/layer.9/attention/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.9/attention/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.9/attention/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.9/attention/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.9/attention/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.9/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.9/attention/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.9/attention/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.9/attention/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.9/attention/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.9/attention/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.9/attention/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.9/attention/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.9/attention/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.9/attention/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.9/attention/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.9/attention/output/LayerNorm/Add_output_0" + name: "/encoder/layer.9/attention/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.9/attention/output/LayerNorm/Add_output_0" + output: "/encoder/layer.9/attention/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.9/attention/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.9/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.9/attention/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.9/attention/output/LayerNorm/Div_output_0" + name: "/encoder/layer.9/attention/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.9/attention/output/LayerNorm/Div_output_0" + input: "encoder.layer.9.attention.output.LayerNorm.weight" + output: "/encoder/layer.9/attention/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.9/attention/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.9/attention/output/LayerNorm/Mul_output_0" + input: "encoder.layer.9.attention.output.LayerNorm.bias" + output: "/encoder/layer.9/attention/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.9/attention/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.9/attention/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1681" + output: "/encoder/layer.9/intermediate/dense/MatMul_output_0" + name: "/encoder/layer.9/intermediate/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.9.intermediate.dense.bias" + input: "/encoder/layer.9/intermediate/dense/MatMul_output_0" + output: "/encoder/layer.9/intermediate/dense/Add_output_0" + name: "/encoder/layer.9/intermediate/dense/Add" + op_type: "Add" + } + node { + output: "/encoder/layer.9/intermediate/intermediate_act_fn/Constant_output_0" + name: "/encoder/layer.9/intermediate/intermediate_act_fn/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\363\004\265?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.9/intermediate/dense/Add_output_0" + input: "/encoder/layer.9/intermediate/intermediate_act_fn/Constant_output_0" + output: "/encoder/layer.9/intermediate/intermediate_act_fn/Div_output_0" + name: "/encoder/layer.9/intermediate/intermediate_act_fn/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.9/intermediate/intermediate_act_fn/Div_output_0" + output: "/encoder/layer.9/intermediate/intermediate_act_fn/Erf_output_0" + name: "/encoder/layer.9/intermediate/intermediate_act_fn/Erf" + op_type: "Erf" + } + node { + output: "/encoder/layer.9/intermediate/intermediate_act_fn/Constant_1_output_0" + name: "/encoder/layer.9/intermediate/intermediate_act_fn/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\200?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.9/intermediate/intermediate_act_fn/Erf_output_0" + input: "/encoder/layer.9/intermediate/intermediate_act_fn/Constant_1_output_0" + output: "/encoder/layer.9/intermediate/intermediate_act_fn/Add_output_0" + name: "/encoder/layer.9/intermediate/intermediate_act_fn/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.9/intermediate/dense/Add_output_0" + input: "/encoder/layer.9/intermediate/intermediate_act_fn/Add_output_0" + output: "/encoder/layer.9/intermediate/intermediate_act_fn/Mul_output_0" + name: "/encoder/layer.9/intermediate/intermediate_act_fn/Mul" + op_type: "Mul" + } + node { + output: "/encoder/layer.9/intermediate/intermediate_act_fn/Constant_2_output_0" + name: "/encoder/layer.9/intermediate/intermediate_act_fn/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.9/intermediate/intermediate_act_fn/Mul_output_0" + input: "/encoder/layer.9/intermediate/intermediate_act_fn/Constant_2_output_0" + output: "/encoder/layer.9/intermediate/intermediate_act_fn/Mul_1_output_0" + name: "/encoder/layer.9/intermediate/intermediate_act_fn/Mul_1" + op_type: "Mul" + } + node { + input: "/encoder/layer.9/intermediate/intermediate_act_fn/Mul_1_output_0" + input: "onnx::MatMul_1682" + output: "/encoder/layer.9/output/dense/MatMul_output_0" + name: "/encoder/layer.9/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.9.output.dense.bias" + input: "/encoder/layer.9/output/dense/MatMul_output_0" + output: "/encoder/layer.9/output/dense/Add_output_0" + name: "/encoder/layer.9/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.9/output/dense/Add_output_0" + input: "/encoder/layer.9/attention/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.9/output/Add_output_0" + name: "/encoder/layer.9/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.9/output/Add_output_0" + output: "/encoder/layer.9/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.9/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.9/output/Add_output_0" + input: "/encoder/layer.9/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.9/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.9/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.9/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.9/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.9/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.9/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.9/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.9/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.9/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.9/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.9/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.9/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.9/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.9/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.9/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.9/output/LayerNorm/Add_output_0" + name: "/encoder/layer.9/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.9/output/LayerNorm/Add_output_0" + output: "/encoder/layer.9/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.9/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.9/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.9/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.9/output/LayerNorm/Div_output_0" + name: "/encoder/layer.9/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.9/output/LayerNorm/Div_output_0" + input: "encoder.layer.9.output.LayerNorm.weight" + output: "/encoder/layer.9/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.9/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.9/output/LayerNorm/Mul_output_0" + input: "encoder.layer.9.output.LayerNorm.bias" + output: "/encoder/layer.9/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.9/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.9/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1683" + output: "/encoder/layer.10/attention/self/query/MatMul_output_0" + name: "/encoder/layer.10/attention/self/query/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.10.attention.self.query.bias" + input: "/encoder/layer.10/attention/self/query/MatMul_output_0" + output: "/encoder/layer.10/attention/self/query/Add_output_0" + name: "/encoder/layer.10/attention/self/query/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.9/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1684" + output: "/encoder/layer.10/attention/self/key/MatMul_output_0" + name: "/encoder/layer.10/attention/self/key/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.10.attention.self.key.bias" + input: "/encoder/layer.10/attention/self/key/MatMul_output_0" + output: "/encoder/layer.10/attention/self/key/Add_output_0" + name: "/encoder/layer.10/attention/self/key/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.10/attention/self/key/Add_output_0" + output: "/encoder/layer.10/attention/self/Shape_output_0" + name: "/encoder/layer.10/attention/self/Shape" + op_type: "Shape" + } + node { + output: "/encoder/layer.10/attention/self/Constant_output_0" + name: "/encoder/layer.10/attention/self/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.10/attention/self/Shape_output_0" + input: "/encoder/layer.10/attention/self/Constant_output_0" + output: "/encoder/layer.10/attention/self/Gather_output_0" + name: "/encoder/layer.10/attention/self/Gather" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.10/attention/self/key/Add_output_0" + output: "/encoder/layer.10/attention/self/Shape_1_output_0" + name: "/encoder/layer.10/attention/self/Shape_1" + op_type: "Shape" + } + node { + output: "/encoder/layer.10/attention/self/Constant_1_output_0" + name: "/encoder/layer.10/attention/self/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.10/attention/self/Shape_1_output_0" + input: "/encoder/layer.10/attention/self/Constant_1_output_0" + output: "/encoder/layer.10/attention/self/Gather_1_output_0" + name: "/encoder/layer.10/attention/self/Gather_1" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.10/attention/self/Gather_output_0" + output: "/encoder/layer.10/attention/self/Unsqueeze_output_0" + name: "/encoder/layer.10/attention/self/Unsqueeze" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.10/attention/self/Gather_1_output_0" + output: "/encoder/layer.10/attention/self/Unsqueeze_1_output_0" + name: "/encoder/layer.10/attention/self/Unsqueeze_1" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.10/attention/self/Constant_2_output_0" + name: "/encoder/layer.10/attention/self/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.10/attention/self/Constant_3_output_0" + name: "/encoder/layer.10/attention/self/Constant_3" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.10/attention/self/Unsqueeze_output_0" + input: "/encoder/layer.10/attention/self/Unsqueeze_1_output_0" + input: "/encoder/layer.10/attention/self/Constant_2_output_0" + input: "/encoder/layer.10/attention/self/Constant_3_output_0" + output: "/encoder/layer.10/attention/self/Concat_output_0" + name: "/encoder/layer.10/attention/self/Concat" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.10/attention/self/key/Add_output_0" + input: "/encoder/layer.10/attention/self/Concat_output_0" + output: "/encoder/layer.10/attention/self/Reshape_output_0" + name: "/encoder/layer.10/attention/self/Reshape" + op_type: "Reshape" + } + node { + input: "/encoder/layer.9/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1687" + output: "/encoder/layer.10/attention/self/value/MatMul_output_0" + name: "/encoder/layer.10/attention/self/value/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.10.attention.self.value.bias" + input: "/encoder/layer.10/attention/self/value/MatMul_output_0" + output: "/encoder/layer.10/attention/self/value/Add_output_0" + name: "/encoder/layer.10/attention/self/value/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.10/attention/self/value/Add_output_0" + output: "/encoder/layer.10/attention/self/Shape_2_output_0" + name: "/encoder/layer.10/attention/self/Shape_2" + op_type: "Shape" + } + node { + output: "/encoder/layer.10/attention/self/Constant_4_output_0" + name: "/encoder/layer.10/attention/self/Constant_4" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.10/attention/self/Shape_2_output_0" + input: "/encoder/layer.10/attention/self/Constant_4_output_0" + output: "/encoder/layer.10/attention/self/Gather_2_output_0" + name: "/encoder/layer.10/attention/self/Gather_2" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.10/attention/self/value/Add_output_0" + output: "/encoder/layer.10/attention/self/Shape_3_output_0" + name: "/encoder/layer.10/attention/self/Shape_3" + op_type: "Shape" + } + node { + output: "/encoder/layer.10/attention/self/Constant_5_output_0" + name: "/encoder/layer.10/attention/self/Constant_5" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.10/attention/self/Shape_3_output_0" + input: "/encoder/layer.10/attention/self/Constant_5_output_0" + output: "/encoder/layer.10/attention/self/Gather_3_output_0" + name: "/encoder/layer.10/attention/self/Gather_3" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.10/attention/self/Gather_2_output_0" + output: "/encoder/layer.10/attention/self/Unsqueeze_2_output_0" + name: "/encoder/layer.10/attention/self/Unsqueeze_2" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.10/attention/self/Gather_3_output_0" + output: "/encoder/layer.10/attention/self/Unsqueeze_3_output_0" + name: "/encoder/layer.10/attention/self/Unsqueeze_3" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.10/attention/self/Constant_6_output_0" + name: "/encoder/layer.10/attention/self/Constant_6" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.10/attention/self/Constant_7_output_0" + name: "/encoder/layer.10/attention/self/Constant_7" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.10/attention/self/Unsqueeze_2_output_0" + input: "/encoder/layer.10/attention/self/Unsqueeze_3_output_0" + input: "/encoder/layer.10/attention/self/Constant_6_output_0" + input: "/encoder/layer.10/attention/self/Constant_7_output_0" + output: "/encoder/layer.10/attention/self/Concat_1_output_0" + name: "/encoder/layer.10/attention/self/Concat_1" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.10/attention/self/value/Add_output_0" + input: "/encoder/layer.10/attention/self/Concat_1_output_0" + output: "/encoder/layer.10/attention/self/Reshape_1_output_0" + name: "/encoder/layer.10/attention/self/Reshape_1" + op_type: "Reshape" + } + node { + input: "/encoder/layer.10/attention/self/Reshape_1_output_0" + output: "/encoder/layer.10/attention/self/Transpose_output_0" + name: "/encoder/layer.10/attention/self/Transpose" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.10/attention/self/query/Add_output_0" + output: "/encoder/layer.10/attention/self/Shape_4_output_0" + name: "/encoder/layer.10/attention/self/Shape_4" + op_type: "Shape" + } + node { + output: "/encoder/layer.10/attention/self/Constant_8_output_0" + name: "/encoder/layer.10/attention/self/Constant_8" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.10/attention/self/Shape_4_output_0" + input: "/encoder/layer.10/attention/self/Constant_8_output_0" + output: "/encoder/layer.10/attention/self/Gather_4_output_0" + name: "/encoder/layer.10/attention/self/Gather_4" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.10/attention/self/query/Add_output_0" + output: "/encoder/layer.10/attention/self/Shape_5_output_0" + name: "/encoder/layer.10/attention/self/Shape_5" + op_type: "Shape" + } + node { + output: "/encoder/layer.10/attention/self/Constant_9_output_0" + name: "/encoder/layer.10/attention/self/Constant_9" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.10/attention/self/Shape_5_output_0" + input: "/encoder/layer.10/attention/self/Constant_9_output_0" + output: "/encoder/layer.10/attention/self/Gather_5_output_0" + name: "/encoder/layer.10/attention/self/Gather_5" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.10/attention/self/Gather_4_output_0" + output: "/encoder/layer.10/attention/self/Unsqueeze_4_output_0" + name: "/encoder/layer.10/attention/self/Unsqueeze_4" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.10/attention/self/Gather_5_output_0" + output: "/encoder/layer.10/attention/self/Unsqueeze_5_output_0" + name: "/encoder/layer.10/attention/self/Unsqueeze_5" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.10/attention/self/Constant_10_output_0" + name: "/encoder/layer.10/attention/self/Constant_10" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.10/attention/self/Constant_11_output_0" + name: "/encoder/layer.10/attention/self/Constant_11" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.10/attention/self/Unsqueeze_4_output_0" + input: "/encoder/layer.10/attention/self/Unsqueeze_5_output_0" + input: "/encoder/layer.10/attention/self/Constant_10_output_0" + input: "/encoder/layer.10/attention/self/Constant_11_output_0" + output: "/encoder/layer.10/attention/self/Concat_2_output_0" + name: "/encoder/layer.10/attention/self/Concat_2" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.10/attention/self/query/Add_output_0" + input: "/encoder/layer.10/attention/self/Concat_2_output_0" + output: "/encoder/layer.10/attention/self/Reshape_2_output_0" + name: "/encoder/layer.10/attention/self/Reshape_2" + op_type: "Reshape" + } + node { + input: "/encoder/layer.10/attention/self/Reshape_2_output_0" + output: "/encoder/layer.10/attention/self/Transpose_1_output_0" + name: "/encoder/layer.10/attention/self/Transpose_1" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.10/attention/self/Reshape_output_0" + output: "/encoder/layer.10/attention/self/Transpose_2_output_0" + name: "/encoder/layer.10/attention/self/Transpose_2" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 3 + ints: 1 + type: INTS + } + } + node { + input: "/encoder/layer.10/attention/self/Transpose_1_output_0" + input: "/encoder/layer.10/attention/self/Transpose_2_output_0" + output: "/encoder/layer.10/attention/self/MatMul_output_0" + name: "/encoder/layer.10/attention/self/MatMul" + op_type: "MatMul" + } + node { + output: "/encoder/layer.10/attention/self/Constant_12_output_0" + name: "/encoder/layer.10/attention/self/Constant_12" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000A" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.10/attention/self/MatMul_output_0" + input: "/encoder/layer.10/attention/self/Constant_12_output_0" + output: "/encoder/layer.10/attention/self/Div_output_0" + name: "/encoder/layer.10/attention/self/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.10/attention/self/Div_output_0" + input: "/Mul_1_output_0" + output: "/encoder/layer.10/attention/self/Add_output_0" + name: "/encoder/layer.10/attention/self/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.10/attention/self/Add_output_0" + output: "/encoder/layer.10/attention/self/Softmax_output_0" + name: "/encoder/layer.10/attention/self/Softmax" + op_type: "Softmax" + attribute { + name: "axis" + i: 3 + type: INT + } + } + node { + input: "/encoder/layer.10/attention/self/Softmax_output_0" + input: "/encoder/layer.10/attention/self/Transpose_output_0" + output: "/encoder/layer.10/attention/self/MatMul_1_output_0" + name: "/encoder/layer.10/attention/self/MatMul_1" + op_type: "MatMul" + } + node { + input: "/encoder/layer.10/attention/self/MatMul_1_output_0" + output: "/encoder/layer.10/attention/self/Transpose_3_output_0" + name: "/encoder/layer.10/attention/self/Transpose_3" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.10/attention/self/Transpose_3_output_0" + output: "/encoder/layer.10/attention/self/Shape_6_output_0" + name: "/encoder/layer.10/attention/self/Shape_6" + op_type: "Shape" + } + node { + output: "/encoder/layer.10/attention/self/Constant_13_output_0" + name: "/encoder/layer.10/attention/self/Constant_13" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.10/attention/self/Shape_6_output_0" + input: "/encoder/layer.10/attention/self/Constant_13_output_0" + output: "/encoder/layer.10/attention/self/Gather_6_output_0" + name: "/encoder/layer.10/attention/self/Gather_6" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.10/attention/self/Transpose_3_output_0" + output: "/encoder/layer.10/attention/self/Shape_7_output_0" + name: "/encoder/layer.10/attention/self/Shape_7" + op_type: "Shape" + } + node { + output: "/encoder/layer.10/attention/self/Constant_14_output_0" + name: "/encoder/layer.10/attention/self/Constant_14" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.10/attention/self/Shape_7_output_0" + input: "/encoder/layer.10/attention/self/Constant_14_output_0" + output: "/encoder/layer.10/attention/self/Gather_7_output_0" + name: "/encoder/layer.10/attention/self/Gather_7" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.10/attention/self/Gather_6_output_0" + output: "/encoder/layer.10/attention/self/Unsqueeze_6_output_0" + name: "/encoder/layer.10/attention/self/Unsqueeze_6" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.10/attention/self/Gather_7_output_0" + output: "/encoder/layer.10/attention/self/Unsqueeze_7_output_0" + name: "/encoder/layer.10/attention/self/Unsqueeze_7" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.10/attention/self/Constant_15_output_0" + name: "/encoder/layer.10/attention/self/Constant_15" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\000\003\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.10/attention/self/Unsqueeze_6_output_0" + input: "/encoder/layer.10/attention/self/Unsqueeze_7_output_0" + input: "/encoder/layer.10/attention/self/Constant_15_output_0" + output: "/encoder/layer.10/attention/self/Concat_3_output_0" + name: "/encoder/layer.10/attention/self/Concat_3" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.10/attention/self/Transpose_3_output_0" + input: "/encoder/layer.10/attention/self/Concat_3_output_0" + output: "/encoder/layer.10/attention/self/Reshape_3_output_0" + name: "/encoder/layer.10/attention/self/Reshape_3" + op_type: "Reshape" + } + node { + input: "/encoder/layer.10/attention/self/Reshape_3_output_0" + input: "onnx::MatMul_1693" + output: "/encoder/layer.10/attention/output/dense/MatMul_output_0" + name: "/encoder/layer.10/attention/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.10.attention.output.dense.bias" + input: "/encoder/layer.10/attention/output/dense/MatMul_output_0" + output: "/encoder/layer.10/attention/output/dense/Add_output_0" + name: "/encoder/layer.10/attention/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.10/attention/output/dense/Add_output_0" + input: "/encoder/layer.9/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.10/attention/output/Add_output_0" + name: "/encoder/layer.10/attention/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.10/attention/output/Add_output_0" + output: "/encoder/layer.10/attention/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.10/attention/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.10/attention/output/Add_output_0" + input: "/encoder/layer.10/attention/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.10/attention/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.10/attention/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.10/attention/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.10/attention/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.10/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.10/attention/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.10/attention/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.10/attention/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.10/attention/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.10/attention/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.10/attention/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.10/attention/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.10/attention/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.10/attention/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.10/attention/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.10/attention/output/LayerNorm/Add_output_0" + name: "/encoder/layer.10/attention/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.10/attention/output/LayerNorm/Add_output_0" + output: "/encoder/layer.10/attention/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.10/attention/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.10/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.10/attention/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.10/attention/output/LayerNorm/Div_output_0" + name: "/encoder/layer.10/attention/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.10/attention/output/LayerNorm/Div_output_0" + input: "encoder.layer.10.attention.output.LayerNorm.weight" + output: "/encoder/layer.10/attention/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.10/attention/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.10/attention/output/LayerNorm/Mul_output_0" + input: "encoder.layer.10.attention.output.LayerNorm.bias" + output: "/encoder/layer.10/attention/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.10/attention/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.10/attention/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1694" + output: "/encoder/layer.10/intermediate/dense/MatMul_output_0" + name: "/encoder/layer.10/intermediate/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.10.intermediate.dense.bias" + input: "/encoder/layer.10/intermediate/dense/MatMul_output_0" + output: "/encoder/layer.10/intermediate/dense/Add_output_0" + name: "/encoder/layer.10/intermediate/dense/Add" + op_type: "Add" + } + node { + output: "/encoder/layer.10/intermediate/intermediate_act_fn/Constant_output_0" + name: "/encoder/layer.10/intermediate/intermediate_act_fn/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\363\004\265?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.10/intermediate/dense/Add_output_0" + input: "/encoder/layer.10/intermediate/intermediate_act_fn/Constant_output_0" + output: "/encoder/layer.10/intermediate/intermediate_act_fn/Div_output_0" + name: "/encoder/layer.10/intermediate/intermediate_act_fn/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.10/intermediate/intermediate_act_fn/Div_output_0" + output: "/encoder/layer.10/intermediate/intermediate_act_fn/Erf_output_0" + name: "/encoder/layer.10/intermediate/intermediate_act_fn/Erf" + op_type: "Erf" + } + node { + output: "/encoder/layer.10/intermediate/intermediate_act_fn/Constant_1_output_0" + name: "/encoder/layer.10/intermediate/intermediate_act_fn/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\200?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.10/intermediate/intermediate_act_fn/Erf_output_0" + input: "/encoder/layer.10/intermediate/intermediate_act_fn/Constant_1_output_0" + output: "/encoder/layer.10/intermediate/intermediate_act_fn/Add_output_0" + name: "/encoder/layer.10/intermediate/intermediate_act_fn/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.10/intermediate/dense/Add_output_0" + input: "/encoder/layer.10/intermediate/intermediate_act_fn/Add_output_0" + output: "/encoder/layer.10/intermediate/intermediate_act_fn/Mul_output_0" + name: "/encoder/layer.10/intermediate/intermediate_act_fn/Mul" + op_type: "Mul" + } + node { + output: "/encoder/layer.10/intermediate/intermediate_act_fn/Constant_2_output_0" + name: "/encoder/layer.10/intermediate/intermediate_act_fn/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.10/intermediate/intermediate_act_fn/Mul_output_0" + input: "/encoder/layer.10/intermediate/intermediate_act_fn/Constant_2_output_0" + output: "/encoder/layer.10/intermediate/intermediate_act_fn/Mul_1_output_0" + name: "/encoder/layer.10/intermediate/intermediate_act_fn/Mul_1" + op_type: "Mul" + } + node { + input: "/encoder/layer.10/intermediate/intermediate_act_fn/Mul_1_output_0" + input: "onnx::MatMul_1695" + output: "/encoder/layer.10/output/dense/MatMul_output_0" + name: "/encoder/layer.10/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.10.output.dense.bias" + input: "/encoder/layer.10/output/dense/MatMul_output_0" + output: "/encoder/layer.10/output/dense/Add_output_0" + name: "/encoder/layer.10/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.10/output/dense/Add_output_0" + input: "/encoder/layer.10/attention/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.10/output/Add_output_0" + name: "/encoder/layer.10/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.10/output/Add_output_0" + output: "/encoder/layer.10/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.10/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.10/output/Add_output_0" + input: "/encoder/layer.10/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.10/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.10/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.10/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.10/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.10/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.10/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.10/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.10/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.10/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.10/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.10/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.10/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.10/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.10/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.10/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.10/output/LayerNorm/Add_output_0" + name: "/encoder/layer.10/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.10/output/LayerNorm/Add_output_0" + output: "/encoder/layer.10/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.10/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.10/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.10/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.10/output/LayerNorm/Div_output_0" + name: "/encoder/layer.10/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.10/output/LayerNorm/Div_output_0" + input: "encoder.layer.10.output.LayerNorm.weight" + output: "/encoder/layer.10/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.10/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.10/output/LayerNorm/Mul_output_0" + input: "encoder.layer.10.output.LayerNorm.bias" + output: "/encoder/layer.10/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.10/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.10/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1696" + output: "/encoder/layer.11/attention/self/query/MatMul_output_0" + name: "/encoder/layer.11/attention/self/query/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.11.attention.self.query.bias" + input: "/encoder/layer.11/attention/self/query/MatMul_output_0" + output: "/encoder/layer.11/attention/self/query/Add_output_0" + name: "/encoder/layer.11/attention/self/query/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.10/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1697" + output: "/encoder/layer.11/attention/self/key/MatMul_output_0" + name: "/encoder/layer.11/attention/self/key/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.11.attention.self.key.bias" + input: "/encoder/layer.11/attention/self/key/MatMul_output_0" + output: "/encoder/layer.11/attention/self/key/Add_output_0" + name: "/encoder/layer.11/attention/self/key/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.11/attention/self/key/Add_output_0" + output: "/encoder/layer.11/attention/self/Shape_output_0" + name: "/encoder/layer.11/attention/self/Shape" + op_type: "Shape" + } + node { + output: "/encoder/layer.11/attention/self/Constant_output_0" + name: "/encoder/layer.11/attention/self/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.11/attention/self/Shape_output_0" + input: "/encoder/layer.11/attention/self/Constant_output_0" + output: "/encoder/layer.11/attention/self/Gather_output_0" + name: "/encoder/layer.11/attention/self/Gather" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.11/attention/self/key/Add_output_0" + output: "/encoder/layer.11/attention/self/Shape_1_output_0" + name: "/encoder/layer.11/attention/self/Shape_1" + op_type: "Shape" + } + node { + output: "/encoder/layer.11/attention/self/Constant_1_output_0" + name: "/encoder/layer.11/attention/self/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.11/attention/self/Shape_1_output_0" + input: "/encoder/layer.11/attention/self/Constant_1_output_0" + output: "/encoder/layer.11/attention/self/Gather_1_output_0" + name: "/encoder/layer.11/attention/self/Gather_1" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.11/attention/self/Gather_output_0" + output: "/encoder/layer.11/attention/self/Unsqueeze_output_0" + name: "/encoder/layer.11/attention/self/Unsqueeze" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.11/attention/self/Gather_1_output_0" + output: "/encoder/layer.11/attention/self/Unsqueeze_1_output_0" + name: "/encoder/layer.11/attention/self/Unsqueeze_1" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.11/attention/self/Constant_2_output_0" + name: "/encoder/layer.11/attention/self/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.11/attention/self/Constant_3_output_0" + name: "/encoder/layer.11/attention/self/Constant_3" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.11/attention/self/Unsqueeze_output_0" + input: "/encoder/layer.11/attention/self/Unsqueeze_1_output_0" + input: "/encoder/layer.11/attention/self/Constant_2_output_0" + input: "/encoder/layer.11/attention/self/Constant_3_output_0" + output: "/encoder/layer.11/attention/self/Concat_output_0" + name: "/encoder/layer.11/attention/self/Concat" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.11/attention/self/key/Add_output_0" + input: "/encoder/layer.11/attention/self/Concat_output_0" + output: "/encoder/layer.11/attention/self/Reshape_output_0" + name: "/encoder/layer.11/attention/self/Reshape" + op_type: "Reshape" + } + node { + input: "/encoder/layer.10/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1700" + output: "/encoder/layer.11/attention/self/value/MatMul_output_0" + name: "/encoder/layer.11/attention/self/value/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.11.attention.self.value.bias" + input: "/encoder/layer.11/attention/self/value/MatMul_output_0" + output: "/encoder/layer.11/attention/self/value/Add_output_0" + name: "/encoder/layer.11/attention/self/value/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.11/attention/self/value/Add_output_0" + output: "/encoder/layer.11/attention/self/Shape_2_output_0" + name: "/encoder/layer.11/attention/self/Shape_2" + op_type: "Shape" + } + node { + output: "/encoder/layer.11/attention/self/Constant_4_output_0" + name: "/encoder/layer.11/attention/self/Constant_4" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.11/attention/self/Shape_2_output_0" + input: "/encoder/layer.11/attention/self/Constant_4_output_0" + output: "/encoder/layer.11/attention/self/Gather_2_output_0" + name: "/encoder/layer.11/attention/self/Gather_2" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.11/attention/self/value/Add_output_0" + output: "/encoder/layer.11/attention/self/Shape_3_output_0" + name: "/encoder/layer.11/attention/self/Shape_3" + op_type: "Shape" + } + node { + output: "/encoder/layer.11/attention/self/Constant_5_output_0" + name: "/encoder/layer.11/attention/self/Constant_5" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.11/attention/self/Shape_3_output_0" + input: "/encoder/layer.11/attention/self/Constant_5_output_0" + output: "/encoder/layer.11/attention/self/Gather_3_output_0" + name: "/encoder/layer.11/attention/self/Gather_3" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.11/attention/self/Gather_2_output_0" + output: "/encoder/layer.11/attention/self/Unsqueeze_2_output_0" + name: "/encoder/layer.11/attention/self/Unsqueeze_2" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.11/attention/self/Gather_3_output_0" + output: "/encoder/layer.11/attention/self/Unsqueeze_3_output_0" + name: "/encoder/layer.11/attention/self/Unsqueeze_3" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.11/attention/self/Constant_6_output_0" + name: "/encoder/layer.11/attention/self/Constant_6" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.11/attention/self/Constant_7_output_0" + name: "/encoder/layer.11/attention/self/Constant_7" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.11/attention/self/Unsqueeze_2_output_0" + input: "/encoder/layer.11/attention/self/Unsqueeze_3_output_0" + input: "/encoder/layer.11/attention/self/Constant_6_output_0" + input: "/encoder/layer.11/attention/self/Constant_7_output_0" + output: "/encoder/layer.11/attention/self/Concat_1_output_0" + name: "/encoder/layer.11/attention/self/Concat_1" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.11/attention/self/value/Add_output_0" + input: "/encoder/layer.11/attention/self/Concat_1_output_0" + output: "/encoder/layer.11/attention/self/Reshape_1_output_0" + name: "/encoder/layer.11/attention/self/Reshape_1" + op_type: "Reshape" + } + node { + input: "/encoder/layer.11/attention/self/Reshape_1_output_0" + output: "/encoder/layer.11/attention/self/Transpose_output_0" + name: "/encoder/layer.11/attention/self/Transpose" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.11/attention/self/query/Add_output_0" + output: "/encoder/layer.11/attention/self/Shape_4_output_0" + name: "/encoder/layer.11/attention/self/Shape_4" + op_type: "Shape" + } + node { + output: "/encoder/layer.11/attention/self/Constant_8_output_0" + name: "/encoder/layer.11/attention/self/Constant_8" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.11/attention/self/Shape_4_output_0" + input: "/encoder/layer.11/attention/self/Constant_8_output_0" + output: "/encoder/layer.11/attention/self/Gather_4_output_0" + name: "/encoder/layer.11/attention/self/Gather_4" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.11/attention/self/query/Add_output_0" + output: "/encoder/layer.11/attention/self/Shape_5_output_0" + name: "/encoder/layer.11/attention/self/Shape_5" + op_type: "Shape" + } + node { + output: "/encoder/layer.11/attention/self/Constant_9_output_0" + name: "/encoder/layer.11/attention/self/Constant_9" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.11/attention/self/Shape_5_output_0" + input: "/encoder/layer.11/attention/self/Constant_9_output_0" + output: "/encoder/layer.11/attention/self/Gather_5_output_0" + name: "/encoder/layer.11/attention/self/Gather_5" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.11/attention/self/Gather_4_output_0" + output: "/encoder/layer.11/attention/self/Unsqueeze_4_output_0" + name: "/encoder/layer.11/attention/self/Unsqueeze_4" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.11/attention/self/Gather_5_output_0" + output: "/encoder/layer.11/attention/self/Unsqueeze_5_output_0" + name: "/encoder/layer.11/attention/self/Unsqueeze_5" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.11/attention/self/Constant_10_output_0" + name: "/encoder/layer.11/attention/self/Constant_10" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\014\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + output: "/encoder/layer.11/attention/self/Constant_11_output_0" + name: "/encoder/layer.11/attention/self/Constant_11" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "@\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.11/attention/self/Unsqueeze_4_output_0" + input: "/encoder/layer.11/attention/self/Unsqueeze_5_output_0" + input: "/encoder/layer.11/attention/self/Constant_10_output_0" + input: "/encoder/layer.11/attention/self/Constant_11_output_0" + output: "/encoder/layer.11/attention/self/Concat_2_output_0" + name: "/encoder/layer.11/attention/self/Concat_2" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.11/attention/self/query/Add_output_0" + input: "/encoder/layer.11/attention/self/Concat_2_output_0" + output: "/encoder/layer.11/attention/self/Reshape_2_output_0" + name: "/encoder/layer.11/attention/self/Reshape_2" + op_type: "Reshape" + } + node { + input: "/encoder/layer.11/attention/self/Reshape_2_output_0" + output: "/encoder/layer.11/attention/self/Transpose_1_output_0" + name: "/encoder/layer.11/attention/self/Transpose_1" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.11/attention/self/Reshape_output_0" + output: "/encoder/layer.11/attention/self/Transpose_2_output_0" + name: "/encoder/layer.11/attention/self/Transpose_2" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 3 + ints: 1 + type: INTS + } + } + node { + input: "/encoder/layer.11/attention/self/Transpose_1_output_0" + input: "/encoder/layer.11/attention/self/Transpose_2_output_0" + output: "/encoder/layer.11/attention/self/MatMul_output_0" + name: "/encoder/layer.11/attention/self/MatMul" + op_type: "MatMul" + } + node { + output: "/encoder/layer.11/attention/self/Constant_12_output_0" + name: "/encoder/layer.11/attention/self/Constant_12" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000A" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.11/attention/self/MatMul_output_0" + input: "/encoder/layer.11/attention/self/Constant_12_output_0" + output: "/encoder/layer.11/attention/self/Div_output_0" + name: "/encoder/layer.11/attention/self/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.11/attention/self/Div_output_0" + input: "/Mul_1_output_0" + output: "/encoder/layer.11/attention/self/Add_output_0" + name: "/encoder/layer.11/attention/self/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.11/attention/self/Add_output_0" + output: "/encoder/layer.11/attention/self/Softmax_output_0" + name: "/encoder/layer.11/attention/self/Softmax" + op_type: "Softmax" + attribute { + name: "axis" + i: 3 + type: INT + } + } + node { + input: "/encoder/layer.11/attention/self/Softmax_output_0" + input: "/encoder/layer.11/attention/self/Transpose_output_0" + output: "/encoder/layer.11/attention/self/MatMul_1_output_0" + name: "/encoder/layer.11/attention/self/MatMul_1" + op_type: "MatMul" + } + node { + input: "/encoder/layer.11/attention/self/MatMul_1_output_0" + output: "/encoder/layer.11/attention/self/Transpose_3_output_0" + name: "/encoder/layer.11/attention/self/Transpose_3" + op_type: "Transpose" + attribute { + name: "perm" + ints: 0 + ints: 2 + ints: 1 + ints: 3 + type: INTS + } + } + node { + input: "/encoder/layer.11/attention/self/Transpose_3_output_0" + output: "/encoder/layer.11/attention/self/Shape_6_output_0" + name: "/encoder/layer.11/attention/self/Shape_6" + op_type: "Shape" + } + node { + output: "/encoder/layer.11/attention/self/Constant_13_output_0" + name: "/encoder/layer.11/attention/self/Constant_13" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\000\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.11/attention/self/Shape_6_output_0" + input: "/encoder/layer.11/attention/self/Constant_13_output_0" + output: "/encoder/layer.11/attention/self/Gather_6_output_0" + name: "/encoder/layer.11/attention/self/Gather_6" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.11/attention/self/Transpose_3_output_0" + output: "/encoder/layer.11/attention/self/Shape_7_output_0" + name: "/encoder/layer.11/attention/self/Shape_7" + op_type: "Shape" + } + node { + output: "/encoder/layer.11/attention/self/Constant_14_output_0" + name: "/encoder/layer.11/attention/self/Constant_14" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 7 + raw_data: "\001\000\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.11/attention/self/Shape_7_output_0" + input: "/encoder/layer.11/attention/self/Constant_14_output_0" + output: "/encoder/layer.11/attention/self/Gather_7_output_0" + name: "/encoder/layer.11/attention/self/Gather_7" + op_type: "Gather" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.11/attention/self/Gather_6_output_0" + output: "/encoder/layer.11/attention/self/Unsqueeze_6_output_0" + name: "/encoder/layer.11/attention/self/Unsqueeze_6" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + input: "/encoder/layer.11/attention/self/Gather_7_output_0" + output: "/encoder/layer.11/attention/self/Unsqueeze_7_output_0" + name: "/encoder/layer.11/attention/self/Unsqueeze_7" + op_type: "Unsqueeze" + attribute { + name: "axes" + ints: 0 + type: INTS + } + } + node { + output: "/encoder/layer.11/attention/self/Constant_15_output_0" + name: "/encoder/layer.11/attention/self/Constant_15" + op_type: "Constant" + attribute { + name: "value" + t { + dims: 1 + data_type: 7 + raw_data: "\000\003\000\000\000\000\000\000" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.11/attention/self/Unsqueeze_6_output_0" + input: "/encoder/layer.11/attention/self/Unsqueeze_7_output_0" + input: "/encoder/layer.11/attention/self/Constant_15_output_0" + output: "/encoder/layer.11/attention/self/Concat_3_output_0" + name: "/encoder/layer.11/attention/self/Concat_3" + op_type: "Concat" + attribute { + name: "axis" + i: 0 + type: INT + } + } + node { + input: "/encoder/layer.11/attention/self/Transpose_3_output_0" + input: "/encoder/layer.11/attention/self/Concat_3_output_0" + output: "/encoder/layer.11/attention/self/Reshape_3_output_0" + name: "/encoder/layer.11/attention/self/Reshape_3" + op_type: "Reshape" + } + node { + input: "/encoder/layer.11/attention/self/Reshape_3_output_0" + input: "onnx::MatMul_1706" + output: "/encoder/layer.11/attention/output/dense/MatMul_output_0" + name: "/encoder/layer.11/attention/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.11.attention.output.dense.bias" + input: "/encoder/layer.11/attention/output/dense/MatMul_output_0" + output: "/encoder/layer.11/attention/output/dense/Add_output_0" + name: "/encoder/layer.11/attention/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.11/attention/output/dense/Add_output_0" + input: "/encoder/layer.10/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.11/attention/output/Add_output_0" + name: "/encoder/layer.11/attention/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.11/attention/output/Add_output_0" + output: "/encoder/layer.11/attention/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.11/attention/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.11/attention/output/Add_output_0" + input: "/encoder/layer.11/attention/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.11/attention/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.11/attention/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.11/attention/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.11/attention/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.11/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.11/attention/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.11/attention/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.11/attention/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.11/attention/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.11/attention/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.11/attention/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.11/attention/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.11/attention/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.11/attention/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.11/attention/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.11/attention/output/LayerNorm/Add_output_0" + name: "/encoder/layer.11/attention/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.11/attention/output/LayerNorm/Add_output_0" + output: "/encoder/layer.11/attention/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.11/attention/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.11/attention/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.11/attention/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.11/attention/output/LayerNorm/Div_output_0" + name: "/encoder/layer.11/attention/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.11/attention/output/LayerNorm/Div_output_0" + input: "encoder.layer.11.attention.output.LayerNorm.weight" + output: "/encoder/layer.11/attention/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.11/attention/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.11/attention/output/LayerNorm/Mul_output_0" + input: "encoder.layer.11.attention.output.LayerNorm.bias" + output: "/encoder/layer.11/attention/output/LayerNorm/Add_1_output_0" + name: "/encoder/layer.11/attention/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "/encoder/layer.11/attention/output/LayerNorm/Add_1_output_0" + input: "onnx::MatMul_1707" + output: "/encoder/layer.11/intermediate/dense/MatMul_output_0" + name: "/encoder/layer.11/intermediate/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.11.intermediate.dense.bias" + input: "/encoder/layer.11/intermediate/dense/MatMul_output_0" + output: "/encoder/layer.11/intermediate/dense/Add_output_0" + name: "/encoder/layer.11/intermediate/dense/Add" + op_type: "Add" + } + node { + output: "/encoder/layer.11/intermediate/intermediate_act_fn/Constant_output_0" + name: "/encoder/layer.11/intermediate/intermediate_act_fn/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\363\004\265?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.11/intermediate/dense/Add_output_0" + input: "/encoder/layer.11/intermediate/intermediate_act_fn/Constant_output_0" + output: "/encoder/layer.11/intermediate/intermediate_act_fn/Div_output_0" + name: "/encoder/layer.11/intermediate/intermediate_act_fn/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.11/intermediate/intermediate_act_fn/Div_output_0" + output: "/encoder/layer.11/intermediate/intermediate_act_fn/Erf_output_0" + name: "/encoder/layer.11/intermediate/intermediate_act_fn/Erf" + op_type: "Erf" + } + node { + output: "/encoder/layer.11/intermediate/intermediate_act_fn/Constant_1_output_0" + name: "/encoder/layer.11/intermediate/intermediate_act_fn/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\200?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.11/intermediate/intermediate_act_fn/Erf_output_0" + input: "/encoder/layer.11/intermediate/intermediate_act_fn/Constant_1_output_0" + output: "/encoder/layer.11/intermediate/intermediate_act_fn/Add_output_0" + name: "/encoder/layer.11/intermediate/intermediate_act_fn/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.11/intermediate/dense/Add_output_0" + input: "/encoder/layer.11/intermediate/intermediate_act_fn/Add_output_0" + output: "/encoder/layer.11/intermediate/intermediate_act_fn/Mul_output_0" + name: "/encoder/layer.11/intermediate/intermediate_act_fn/Mul" + op_type: "Mul" + } + node { + output: "/encoder/layer.11/intermediate/intermediate_act_fn/Constant_2_output_0" + name: "/encoder/layer.11/intermediate/intermediate_act_fn/Constant_2" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000?" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.11/intermediate/intermediate_act_fn/Mul_output_0" + input: "/encoder/layer.11/intermediate/intermediate_act_fn/Constant_2_output_0" + output: "/encoder/layer.11/intermediate/intermediate_act_fn/Mul_1_output_0" + name: "/encoder/layer.11/intermediate/intermediate_act_fn/Mul_1" + op_type: "Mul" + } + node { + input: "/encoder/layer.11/intermediate/intermediate_act_fn/Mul_1_output_0" + input: "onnx::MatMul_1708" + output: "/encoder/layer.11/output/dense/MatMul_output_0" + name: "/encoder/layer.11/output/dense/MatMul" + op_type: "MatMul" + } + node { + input: "encoder.layer.11.output.dense.bias" + input: "/encoder/layer.11/output/dense/MatMul_output_0" + output: "/encoder/layer.11/output/dense/Add_output_0" + name: "/encoder/layer.11/output/dense/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.11/output/dense/Add_output_0" + input: "/encoder/layer.11/attention/output/LayerNorm/Add_1_output_0" + output: "/encoder/layer.11/output/Add_output_0" + name: "/encoder/layer.11/output/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.11/output/Add_output_0" + output: "/encoder/layer.11/output/LayerNorm/ReduceMean_output_0" + name: "/encoder/layer.11/output/LayerNorm/ReduceMean" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + input: "/encoder/layer.11/output/Add_output_0" + input: "/encoder/layer.11/output/LayerNorm/ReduceMean_output_0" + output: "/encoder/layer.11/output/LayerNorm/Sub_output_0" + name: "/encoder/layer.11/output/LayerNorm/Sub" + op_type: "Sub" + } + node { + output: "/encoder/layer.11/output/LayerNorm/Constant_output_0" + name: "/encoder/layer.11/output/LayerNorm/Constant" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\000\000\000@" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.11/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.11/output/LayerNorm/Constant_output_0" + output: "/encoder/layer.11/output/LayerNorm/Pow_output_0" + name: "/encoder/layer.11/output/LayerNorm/Pow" + op_type: "Pow" + } + node { + input: "/encoder/layer.11/output/LayerNorm/Pow_output_0" + output: "/encoder/layer.11/output/LayerNorm/ReduceMean_1_output_0" + name: "/encoder/layer.11/output/LayerNorm/ReduceMean_1" + op_type: "ReduceMean" + attribute { + name: "axes" + ints: -1 + type: INTS + } + } + node { + output: "/encoder/layer.11/output/LayerNorm/Constant_1_output_0" + name: "/encoder/layer.11/output/LayerNorm/Constant_1" + op_type: "Constant" + attribute { + name: "value" + t { + data_type: 1 + raw_data: "\254\305\'7" + } + type: TENSOR + } + } + node { + input: "/encoder/layer.11/output/LayerNorm/ReduceMean_1_output_0" + input: "/encoder/layer.11/output/LayerNorm/Constant_1_output_0" + output: "/encoder/layer.11/output/LayerNorm/Add_output_0" + name: "/encoder/layer.11/output/LayerNorm/Add" + op_type: "Add" + } + node { + input: "/encoder/layer.11/output/LayerNorm/Add_output_0" + output: "/encoder/layer.11/output/LayerNorm/Sqrt_output_0" + name: "/encoder/layer.11/output/LayerNorm/Sqrt" + op_type: "Sqrt" + } + node { + input: "/encoder/layer.11/output/LayerNorm/Sub_output_0" + input: "/encoder/layer.11/output/LayerNorm/Sqrt_output_0" + output: "/encoder/layer.11/output/LayerNorm/Div_output_0" + name: "/encoder/layer.11/output/LayerNorm/Div" + op_type: "Div" + } + node { + input: "/encoder/layer.11/output/LayerNorm/Div_output_0" + input: "encoder.layer.11.output.LayerNorm.weight" + output: "/encoder/layer.11/output/LayerNorm/Mul_output_0" + name: "/encoder/layer.11/output/LayerNorm/Mul" + op_type: "Mul" + } + node { + input: "/encoder/layer.11/output/LayerNorm/Mul_output_0" + input: "encoder.layer.11.output.LayerNorm.bias" + output: "onnx::Gather_1547" + name: "/encoder/layer.11/output/LayerNorm/Add_1" + op_type: "Add" + } + node { + input: "onnx::Gather_1547" + input: "/Constant_output_0" + output: "/pooler/Gather_output_0" + name: "/pooler/Gather" + op_type: "Gather" + attribute { + name: "axis" + i: 1 + type: INT + } + } + node { + input: "/pooler/Gather_output_0" + input: "pooler.dense.weight" + input: "pooler.dense.bias" + output: "/pooler/dense/Gemm_output_0" + name: "/pooler/dense/Gemm" + op_type: "Gemm" + attribute { + name: "alpha" + f: 1.0 + type: FLOAT + } + attribute { + name: "beta" + f: 1.0 + type: FLOAT + } + attribute { + name: "transB" + i: 1 + type: INT + } + } + node { + input: "/pooler/dense/Gemm_output_0" + output: "1550" + name: "/pooler/activation/Tanh" + op_type: "Tanh" + } + initializer { + dims: 50265 + dims: 768 + data_type: 1 + name: "embeddings.word_embeddings.weight" + } + initializer { + dims: 514 + dims: 768 + data_type: 1 + name: "embeddings.position_embeddings.weight" + } + initializer { + dims: 1 + dims: 768 + data_type: 1 + name: "embeddings.token_type_embeddings.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "embeddings.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "embeddings.LayerNorm.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.0.attention.self.query.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.0.attention.self.key.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.0.attention.self.value.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.0.attention.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.0.attention.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.0.attention.output.LayerNorm.bias" + } + initializer { + dims: 3072 + data_type: 1 + name: "encoder.layer.0.intermediate.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.0.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.0.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.0.output.LayerNorm.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.1.attention.self.query.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.1.attention.self.key.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.1.attention.self.value.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.1.attention.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.1.attention.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.1.attention.output.LayerNorm.bias" + } + initializer { + dims: 3072 + data_type: 1 + name: "encoder.layer.1.intermediate.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.1.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.1.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.1.output.LayerNorm.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.2.attention.self.query.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.2.attention.self.key.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.2.attention.self.value.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.2.attention.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.2.attention.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.2.attention.output.LayerNorm.bias" + } + initializer { + dims: 3072 + data_type: 1 + name: "encoder.layer.2.intermediate.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.2.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.2.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.2.output.LayerNorm.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.3.attention.self.query.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.3.attention.self.key.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.3.attention.self.value.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.3.attention.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.3.attention.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.3.attention.output.LayerNorm.bias" + } + initializer { + dims: 3072 + data_type: 1 + name: "encoder.layer.3.intermediate.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.3.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.3.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.3.output.LayerNorm.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.4.attention.self.query.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.4.attention.self.key.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.4.attention.self.value.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.4.attention.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.4.attention.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.4.attention.output.LayerNorm.bias" + } + initializer { + dims: 3072 + data_type: 1 + name: "encoder.layer.4.intermediate.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.4.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.4.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.4.output.LayerNorm.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.5.attention.self.query.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.5.attention.self.key.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.5.attention.self.value.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.5.attention.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.5.attention.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.5.attention.output.LayerNorm.bias" + } + initializer { + dims: 3072 + data_type: 1 + name: "encoder.layer.5.intermediate.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.5.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.5.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.5.output.LayerNorm.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.6.attention.self.query.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.6.attention.self.key.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.6.attention.self.value.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.6.attention.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.6.attention.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.6.attention.output.LayerNorm.bias" + } + initializer { + dims: 3072 + data_type: 1 + name: "encoder.layer.6.intermediate.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.6.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.6.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.6.output.LayerNorm.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.7.attention.self.query.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.7.attention.self.key.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.7.attention.self.value.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.7.attention.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.7.attention.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.7.attention.output.LayerNorm.bias" + } + initializer { + dims: 3072 + data_type: 1 + name: "encoder.layer.7.intermediate.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.7.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.7.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.7.output.LayerNorm.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.8.attention.self.query.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.8.attention.self.key.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.8.attention.self.value.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.8.attention.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.8.attention.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.8.attention.output.LayerNorm.bias" + } + initializer { + dims: 3072 + data_type: 1 + name: "encoder.layer.8.intermediate.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.8.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.8.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.8.output.LayerNorm.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.9.attention.self.query.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.9.attention.self.key.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.9.attention.self.value.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.9.attention.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.9.attention.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.9.attention.output.LayerNorm.bias" + } + initializer { + dims: 3072 + data_type: 1 + name: "encoder.layer.9.intermediate.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.9.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.9.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.9.output.LayerNorm.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.10.attention.self.query.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.10.attention.self.key.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.10.attention.self.value.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.10.attention.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.10.attention.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.10.attention.output.LayerNorm.bias" + } + initializer { + dims: 3072 + data_type: 1 + name: "encoder.layer.10.intermediate.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.10.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.10.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.10.output.LayerNorm.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.11.attention.self.query.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.11.attention.self.key.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.11.attention.self.value.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.11.attention.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.11.attention.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.11.attention.output.LayerNorm.bias" + } + initializer { + dims: 3072 + data_type: 1 + name: "encoder.layer.11.intermediate.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.11.output.dense.bias" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.11.output.LayerNorm.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "encoder.layer.11.output.LayerNorm.bias" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "pooler.dense.weight" + } + initializer { + dims: 768 + data_type: 1 + name: "pooler.dense.bias" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1553" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1554" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1557" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1563" + } + initializer { + dims: 768 + dims: 3072 + data_type: 1 + name: "onnx::MatMul_1564" + } + initializer { + dims: 3072 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1565" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1566" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1567" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1570" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1576" + } + initializer { + dims: 768 + dims: 3072 + data_type: 1 + name: "onnx::MatMul_1577" + } + initializer { + dims: 3072 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1578" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1579" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1580" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1583" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1589" + } + initializer { + dims: 768 + dims: 3072 + data_type: 1 + name: "onnx::MatMul_1590" + } + initializer { + dims: 3072 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1591" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1592" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1593" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1596" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1602" + } + initializer { + dims: 768 + dims: 3072 + data_type: 1 + name: "onnx::MatMul_1603" + } + initializer { + dims: 3072 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1604" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1605" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1606" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1609" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1615" + } + initializer { + dims: 768 + dims: 3072 + data_type: 1 + name: "onnx::MatMul_1616" + } + initializer { + dims: 3072 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1617" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1618" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1619" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1622" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1628" + } + initializer { + dims: 768 + dims: 3072 + data_type: 1 + name: "onnx::MatMul_1629" + } + initializer { + dims: 3072 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1630" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1631" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1632" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1635" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1641" + } + initializer { + dims: 768 + dims: 3072 + data_type: 1 + name: "onnx::MatMul_1642" + } + initializer { + dims: 3072 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1643" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1644" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1645" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1648" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1654" + } + initializer { + dims: 768 + dims: 3072 + data_type: 1 + name: "onnx::MatMul_1655" + } + initializer { + dims: 3072 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1656" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1657" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1658" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1661" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1667" + } + initializer { + dims: 768 + dims: 3072 + data_type: 1 + name: "onnx::MatMul_1668" + } + initializer { + dims: 3072 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1669" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1670" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1671" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1674" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1680" + } + initializer { + dims: 768 + dims: 3072 + data_type: 1 + name: "onnx::MatMul_1681" + } + initializer { + dims: 3072 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1682" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1683" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1684" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1687" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1693" + } + initializer { + dims: 768 + dims: 3072 + data_type: 1 + name: "onnx::MatMul_1694" + } + initializer { + dims: 3072 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1695" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1696" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1697" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1700" + } + initializer { + dims: 768 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1706" + } + initializer { + dims: 768 + dims: 3072 + data_type: 1 + name: "onnx::MatMul_1707" + } + initializer { + dims: 3072 + dims: 768 + data_type: 1 + name: "onnx::MatMul_1708" + } + input { + name: "input_ids" + type { + tensor_type { + elem_type: 7 + shape { + dim { + dim_param: "batch_size" + } + dim { + dim_value: 77 + } + } + } + } + } + input { + name: "attention_mask" + type { + tensor_type { + elem_type: 7 + shape { + dim { + dim_param: "batch_size" + } + dim { + dim_value: 77 + } + } + } + } + } + output { + name: "onnx::Gather_1547" + type { + tensor_type { + elem_type: 1 + shape { + dim { + dim_param: "Addonnx::Gather_1547_dim_0" + } + dim { + dim_param: "Addonnx::Gather_1547_dim_1" + } + dim { + dim_value: 768 + } + } + } + } + } + output { + name: "1550" + type { + tensor_type { + elem_type: 1 + shape { + dim { + dim_param: "Addonnx::Gather_1547_dim_0" + } + dim { + dim_value: 768 + } + } + } + } + } +} +opset_import { + version: 11 +}