katuni4ka commited on
Commit
0fd5106
·
verified ·
1 Parent(s): be9c872

Upload 23 files

Browse files
feature_extractor/preprocessor_config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": {
3
+ "height": 224,
4
+ "width": 224
5
+ },
6
+ "do_center_crop": true,
7
+ "do_convert_rgb": true,
8
+ "do_normalize": true,
9
+ "do_rescale": true,
10
+ "do_resize": true,
11
+ "image_mean": [
12
+ 0.48145466,
13
+ 0.4578275,
14
+ 0.40821073
15
+ ],
16
+ "image_processor_type": "CLIPImageProcessor",
17
+ "image_std": [
18
+ 0.26862954,
19
+ 0.26130258,
20
+ 0.27577711
21
+ ],
22
+ "resample": 3,
23
+ "rescale_factor": 0.00392156862745098,
24
+ "size": {
25
+ "shortest_edge": 224
26
+ }
27
+ }
model_index.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "StableDiffusionPipeline",
3
+ "_diffusers_version": "0.32.2",
4
+ "_name_or_path": "hf-internal-testing/tiny-stable-diffusion-torch",
5
+ "feature_extractor": [
6
+ "transformers",
7
+ "CLIPImageProcessor"
8
+ ],
9
+ "image_encoder": [
10
+ null,
11
+ null
12
+ ],
13
+ "requires_safety_checker": true,
14
+ "safety_checker": [
15
+ null,
16
+ null
17
+ ],
18
+ "scheduler": [
19
+ "diffusers",
20
+ "PNDMScheduler"
21
+ ],
22
+ "text_encoder": [
23
+ "transformers",
24
+ "CLIPTextModel"
25
+ ],
26
+ "tokenizer": [
27
+ "transformers",
28
+ "CLIPTokenizer"
29
+ ],
30
+ "unet": [
31
+ "diffusers",
32
+ "UNet2DConditionModel"
33
+ ],
34
+ "vae": [
35
+ "diffusers",
36
+ "AutoencoderKL"
37
+ ]
38
+ }
scheduler/scheduler_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "PNDMScheduler",
3
+ "_diffusers_version": "0.32.2",
4
+ "beta_end": 0.012,
5
+ "beta_schedule": "scaled_linear",
6
+ "beta_start": 0.00085,
7
+ "clip_sample": false,
8
+ "num_train_timesteps": 1000,
9
+ "prediction_type": "epsilon",
10
+ "set_alpha_to_one": false,
11
+ "skip_prk_steps": true,
12
+ "steps_offset": 1,
13
+ "timestep_spacing": "leading",
14
+ "trained_betas": null
15
+ }
text_encoder/config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_attn_implementation_autoset": true,
3
+ "_name_or_path": "/home/ea/.cache/huggingface/hub/models--hf-internal-testing--tiny-stable-diffusion-torch/snapshots/a88cdfbd91f96ec7f61eb7484b652ff0f4ee701d/text_encoder",
4
+ "architectures": [
5
+ "CLIPTextModel"
6
+ ],
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 0,
9
+ "dropout": 0.0,
10
+ "eos_token_id": 2,
11
+ "export_model_type": "clip-text",
12
+ "hidden_act": "quick_gelu",
13
+ "hidden_size": 32,
14
+ "initializer_factor": 1.0,
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 37,
17
+ "layer_norm_eps": 1e-05,
18
+ "max_position_embeddings": 77,
19
+ "model_type": "clip_text_model",
20
+ "num_attention_heads": 4,
21
+ "num_hidden_layers": 5,
22
+ "pad_token_id": 1,
23
+ "projection_dim": 512,
24
+ "torch_dtype": "float32",
25
+ "transformers_version": "4.49.0",
26
+ "vocab_size": 1000
27
+ }
text_encoder/openvino_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a591b39e8622573d849f6b585ef1ad3be809563dd935f35e5f9a830201676368
3
+ size 267916
text_encoder/openvino_model.xml ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer/merges.txt ADDED
@@ -0,0 +1,647 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #version: 0.2
2
+ Ġ t
3
+ Ġt h
4
+ Ġ a
5
+ Ġth e</w>
6
+ i n
7
+ Ġ o
8
+ Ġ ,</w>
9
+ Ġ s
10
+ e d</w>
11
+ Ġ w
12
+ e r
13
+ Ġ .</w>
14
+ Ġ i
15
+ r e
16
+ Ġ c
17
+ n d</w>
18
+ Ġ f
19
+ Ġ b
20
+ a t
21
+ Ġo f</w>
22
+ e r</w>
23
+ e n
24
+ a r
25
+ o r
26
+ i t
27
+ Ġ p
28
+ Ġ h
29
+ Ġa nd</w>
30
+ o n
31
+ in g</w>
32
+ a n
33
+ r o
34
+ Ġ m
35
+ Ġ d
36
+ e s</w>
37
+ Ġi n</w>
38
+ o n</w>
39
+ Ġt o</w>
40
+ o u
41
+ i s
42
+ Ġ a</w>
43
+ i c
44
+ Ġ T
45
+ a l
46
+ Ġ l
47
+ Ġ =</w>
48
+ Ġ re
49
+ Ġ "</w>
50
+ e s
51
+ Ġ S
52
+ a s</w>
53
+ a l</w>
54
+ i l
55
+ e l
56
+ i on</w>
57
+ Ġ A
58
+ Ġ C
59
+ Ġ 1
60
+ Ġ Ċ</w>
61
+ u r
62
+ ĠT h
63
+ Ġ n
64
+ a s
65
+ Ġ @
66
+ e c
67
+ o m
68
+ a c
69
+ Ġ e
70
+ Ġw as</w>
71
+ Ġ M
72
+ o r</w>
73
+ a n</w>
74
+ a m
75
+ e n</w>
76
+ o l
77
+ Ġ in
78
+ Ġ g
79
+ Ġ '</w>
80
+ Ġ B
81
+ l y</w>
82
+ a t</w>
83
+ i v
84
+ t s</w>
85
+ ĠTh e</w>
86
+ u s
87
+ - @</w>
88
+ Ġ@ -@</w>
89
+ i s</w>
90
+ Ġ I
91
+ Ġw h
92
+ i g
93
+ Ġ H
94
+ Ġs t
95
+ o s
96
+ u n
97
+ t h
98
+ Ġ P
99
+ Ġw it
100
+ Ġth at</w>
101
+ i r
102
+ Ġa s</w>
103
+ e m
104
+ Ġo n</w>
105
+ r a
106
+ Ġf or</w>
107
+ Ġ R
108
+ e t
109
+ o w
110
+ Ġ 2
111
+ i d
112
+ Ġ D
113
+ l e</w>
114
+ Ġwit h</w>
115
+ l a
116
+ en t</w>
117
+ i m
118
+ Ġ F
119
+ e a
120
+ i on
121
+ Ġb y</w>
122
+ Ġ )</w>
123
+ Ġ (</w>
124
+ Ġa l
125
+ Ġc on
126
+ en t
127
+ Ġ W
128
+ Ġi s</w>
129
+ er e</w>
130
+ Ġ G
131
+ Ġ N
132
+ Ġ L
133
+ Ġh a
134
+ er s</w>
135
+ r i
136
+ t h</w>
137
+ t ed</w>
138
+ u c
139
+ Ġ J
140
+ Ġ1 9
141
+ e v
142
+ u l
143
+ Ġ v
144
+ c e</w>
145
+ at ion</w>
146
+ ro m</w>
147
+ Ġb e
148
+ Ġ E
149
+ i n</w>
150
+ Ġth e
151
+ Ġf rom</w>
152
+ Ġ O
153
+ t er</w>
154
+ Ġp ro
155
+ Ġa r
156
+ a d
157
+ Ġc om
158
+ i c</w>
159
+ a g
160
+ Ġh is</w>
161
+ Ġs h
162
+ Ġa t</w>
163
+ o v
164
+ i es</w>
165
+ o o
166
+ p p
167
+ s t
168
+ c h
169
+ Ġ r
170
+ Ġ2 0
171
+ a y</w>
172
+ i f
173
+ Ġw ere</w>
174
+ Ġc h
175
+ u t</w>
176
+ s t</w>
177
+ u t
178
+ d s</w>
179
+ o p
180
+ u m
181
+ Ġi t</w>
182
+ o c
183
+ t er
184
+ l e
185
+ ig h
186
+ u d
187
+ Ġe x
188
+ ion s</w>
189
+ at e</w>
190
+ it y</w>
191
+ at ed</w>
192
+ Ġ un
193
+ e p
194
+ q u
195
+ Ġn o
196
+ Ġ K
197
+ iv e</w>
198
+ is t
199
+ Ġo n
200
+ am e</w>
201
+ ou n
202
+ i r</w>
203
+ a b
204
+ Ġ â
205
+ in g
206
+ Ġh e</w>
207
+ l d</w>
208
+ u g
209
+ ic h</w>
210
+ Ġa n</w>
211
+ e d
212
+ Ġ k
213
+ Ġâ Ģ
214
+ Ġha d</w>
215
+ v e</w>
216
+ a in
217
+ Ġs e
218
+ t ion</w>
219
+ or e</w>
220
+ re s
221
+ Ġwh ich</w>
222
+ ĠI n</w>
223
+ o d
224
+ th er</w>
225
+ a k
226
+ Ġs p
227
+ a r</w>
228
+ Ġ y
229
+ ĠC h
230
+ on g</w>
231
+ Ġa c
232
+ es t</w>
233
+ Ġ U
234
+ a p
235
+ f f
236
+ al ly</w>
237
+ r it
238
+ ĠS t
239
+ u b
240
+ g e</w>
241
+ b er</w>
242
+ e t</w>
243
+ Ġb e</w>
244
+ e ar
245
+ Ġre c
246
+ er s
247
+ Ġf ir
248
+ o t
249
+ Ġar e</w>
250
+ Ġa n
251
+ c h</w>
252
+ o g
253
+ i a</w>
254
+ es t
255
+ in e</w>
256
+ il l
257
+ an d
258
+ e l</w>
259
+ ar y</w>
260
+ e w</w>
261
+ i d</w>
262
+ Ġf or
263
+ Ġ ;</w>
264
+ Ġcom p
265
+ Ġ V
266
+ Ġin c
267
+ t r
268
+ Ġ20 0
269
+ Ġthe ir</w>
270
+ u s</w>
271
+ Ġb ut</w>
272
+ r an
273
+ ic al</w>
274
+ Ġfir st</w>
275
+ Ġd e
276
+ Ġin t
277
+ Ġ ro
278
+ s o</w>
279
+ ĠâĢ ĵ</w>
280
+ Ġno t</w>
281
+ d ing</w>
282
+ f ter</w>
283
+ ur e</w>
284
+ Ġp ar
285
+ Ġ :</w>
286
+ i an</w>
287
+ Ġt w
288
+ ou ld</w>
289
+ Ġal so</w>
290
+ Ġi ts</w>
291
+ Ġw or
292
+ u m</w>
293
+ Ġo r</w>
294
+ os t</w>
295
+ 0 0</w>
296
+ ou r
297
+ ar d</w>
298
+ Ġre s
299
+ m p
300
+ u e</w>
301
+ Ġa b
302
+ is h</w>
303
+ Ġcon t
304
+ Ġa d
305
+ ow n</w>
306
+ al l</w>
307
+ ou g
308
+ Ġh er</w>
309
+ as t</w>
310
+ Ġ en
311
+ om e</w>
312
+ al l
313
+ d ed</w>
314
+ o w</w>
315
+ Ġha ve</w>
316
+ Ġ us
317
+ ea r</w>
318
+ ac k</w>
319
+ d uc
320
+ i al</w>
321
+ s s
322
+ en ts</w>
323
+ a in</w>
324
+ t ing</w>
325
+ Ġon e</w>
326
+ es s
327
+ Ġh as</w>
328
+ igh t</w>
329
+ a v
330
+ Ġe v
331
+ ou t</w>
332
+ a y
333
+ en ce</w>
334
+ Ġbe en</w>
335
+ e w
336
+ Ġtw o</w>
337
+ Ġc l
338
+ d er</w>
339
+ im e</w>
340
+ k s</w>
341
+ es s</w>
342
+ is h
343
+ . @</w>
344
+ Ġ@ .@</w>
345
+ Ġp la
346
+ Ġp l
347
+ Ġo r
348
+ u p</w>
349
+ m ent</w>
350
+ ur ing</w>
351
+ ol l
352
+ ĠI n
353
+ Ġth is</w>
354
+ Ġb ec
355
+ Ġcom m
356
+ Ġd is
357
+ at er</w>
358
+ ag e</w>
359
+ Ġa pp
360
+ ou s</w>
361
+ e y</w>
362
+ i l</w>
363
+ p er
364
+ ĠA l
365
+ ion al</w>
366
+ l ud
367
+ el y</w>
368
+ t t
369
+ il e</w>
370
+ i z
371
+ Ġ j
372
+ Ġwh o</w>
373
+ Ġa g
374
+ i b
375
+ Ġthe y</w>
376
+ f or
377
+ Ġo v
378
+ at h
379
+ e g
380
+ Ġs c
381
+ i p
382
+ Ġ20 1
383
+ Ġ 3
384
+ Ġp er
385
+ or y</w>
386
+ Ġd es
387
+ id e</w>
388
+ Ġs er
389
+ s e</w>
390
+ ĠH e</w>
391
+ la nd</w>
392
+ at ions</w>
393
+ r ic
394
+ i t</w>
395
+ re s</w>
396
+ er ed</w>
397
+ Ġp re
398
+ ĠS h
399
+ an ce</w>
400
+ or t</w>
401
+ an t</w>
402
+ , @</w>
403
+ Ġ@ ,@</w>
404
+ el l</w>
405
+ Ġ Y
406
+ n ed</w>
407
+ el l
408
+ it e</w>
409
+ Ġinc lud
410
+ Ġre p
411
+ Ġa fter</w>
412
+ Ġs uc
413
+ re e</w>
414
+ an y</w>
415
+ i m</w>
416
+ or t
417
+ Ġ1 8
418
+ Ġs u
419
+ ad e</w>
420
+ ou r</w>
421
+ ĠU n
422
+ ĠI t</w>
423
+ i k
424
+ ĠM ar
425
+ em ber</w>
426
+ Ġ 1</w>
427
+ e en</w>
428
+ a nd</w>
429
+ Ġs ec
430
+ ic e</w>
431
+ Ġt ime</w>
432
+ ĠA n
433
+ Ġint o</w>
434
+ Ġf in
435
+ Ġo ther</w>
436
+ Ġa tt
437
+ il l</w>
438
+ re n
439
+ ac h
440
+ as s
441
+ er al</w>
442
+ es e</w>
443
+ s h
444
+ al s</w>
445
+ it ion</w>
446
+ oug h</w>
447
+ l es</w>
448
+ am p
449
+ Ġw ould</w>
450
+ Ġm ore</w>
451
+ ro ug
452
+ ri b
453
+ er y</w>
454
+ ac e</w>
455
+ Ġ A</w>
456
+ Ġpla y
457
+ it ed</w>
458
+ k ed</w>
459
+ is t</w>
460
+ i ed</w>
461
+ Ġ 2</w>
462
+ as ed</w>
463
+ ing s</w>
464
+ an g
465
+ a m</w>
466
+ i p</w>
467
+ Ġb o
468
+ ab le</w>
469
+ t y</w>
470
+ Ġch ar
471
+ Ġc ent
472
+ et w
473
+ at es</w>
474
+ ro p
475
+ Ġ I</w>
476
+ u nd</w>
477
+ ĠA m
478
+ c es</w>
479
+ o in
480
+ Ġin ter
481
+ u p
482
+ c t
483
+ on e</w>
484
+ Ġt ra
485
+ an t
486
+ ec t
487
+ Ġal l</w>
488
+ e f
489
+ Ġcon s
490
+ ub l
491
+ n ing</w>
492
+ an s</w>
493
+ Ġf e
494
+ us t</w>
495
+ Ġ 0
496
+ Ġre m
497
+ as e</w>
498
+ on g
499
+ Ġwh en</w>
500
+ e b
501
+ ĠW h
502
+ Ġe ar
503
+ ev er</w>
504
+ Ġov er</w>
505
+ Ġk n
506
+ a us
507
+ Ġp os
508
+ a d</w>
509
+ er m
510
+ Ġsh e</w>
511
+ Ġ ra
512
+ Ġd uring</w>
513
+ as on</w>
514
+ v i
515
+ Ġex p
516
+ Ġl ea
517
+ Ġ el
518
+ Ġ 4
519
+ Ġon ly</w>
520
+ o nd</w>
521
+ Ġd ec
522
+ Ġac c
523
+ Ġo ff
524
+ is s
525
+ Ġf l
526
+ ĠE n
527
+ o t</w>
528
+ en s
529
+ os e</w>
530
+ ak e</w>
531
+ o m</w>
532
+ Ġs ev
533
+ ac h</w>
534
+ etw een</w>
535
+ er n
536
+ Ġ 3</w>
537
+ Ġp r
538
+ Ġg ro
539
+ r uc
540
+ Ġd i
541
+ Ġ19 9
542
+ ĠA r
543
+ Ġg ame</w>
544
+ Ġh im</w>
545
+ oo k</w>
546
+ Ġ up</w>
547
+ Ġab out</w>
548
+ Ġre l
549
+ for m
550
+ Ġth ree</w>
551
+ at t
552
+ ĠC om
553
+ Ġs a
554
+ ear s</w>
555
+ Ġ 5
556
+ r y</w>
557
+ Ġi mp
558
+ Ġm ost</w>
559
+ f er
560
+ Ġp res
561
+ Ġf il
562
+ Ġb etween</w>
563
+ Ġbe g
564
+ p h
565
+ or s</w>
566
+ Ġth an</w>
567
+ Ġrec or
568
+ o b
569
+ er ic
570
+ at ing</w>
571
+ Ġth roug
572
+ k ing</w>
573
+ Ġo ut</w>
574
+ Ġn um
575
+ oo d</w>
576
+ oll ow
577
+ ac t
578
+ u il
579
+ Ġc re
580
+ ol og
581
+ at ional</w>
582
+ Ġpro duc
583
+ Ġwh ile</w>
584
+ Ġl ater</w>
585
+ Ġw rit
586
+ e x
587
+ Ġst ar
588
+ Ġsp ec
589
+ e e
590
+ ish ed</w>
591
+ Ġre g
592
+ is ion</w>
593
+ ou th</w>
594
+ Ġre le
595
+ Ġa ss
596
+ Ġse ason</w>
597
+ Ġm ade</w>
598
+ il y</w>
599
+ r u
600
+ o y
601
+ t ur
602
+ t e</w>
603
+ Ġ qu
604
+ Ġm ov
605
+ ur y</w>
606
+ ĠAm eric
607
+ em ent</w>
608
+ c c
609
+ ou nd</w>
610
+ Ġl ar
611
+ Ġfor m
612
+ ec t</w>
613
+ Ġde f
614
+ Ġm us
615
+ ĠP ar
616
+ Ġm e
617
+ Ġs ub
618
+ w ay</w>
619
+ o p</w>
620
+ o h
621
+ el d</w>
622
+ i e</w>
623
+ em p
624
+ am es</w>
625
+ er n</w>
626
+ Ġn or
627
+ iv ed</w>
628
+ ev el
629
+ Ġsuc h</w>
630
+ ar ds</w>
631
+ Ġin d
632
+ ik e</w>
633
+ Ġg en
634
+ er t
635
+ Ġy ear</w>
636
+ Ġus ed</w>
637
+ Ġn ew</w>
638
+ Ġ 5</w>
639
+ Ġal b
640
+ s p
641
+ y p
642
+ Ġwit h
643
+ Ġwh ere</w>
644
+ ic s</w>
645
+ ĠTh is</w>
646
+ Ġthe m</w>
647
+ w n</w>
tokenizer/openvino_detokenizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a37f92bcd32849b1894e47d0a21c6f8c6283ae52e5de0ce04a067f521a1e43c
3
+ size 8236
tokenizer/openvino_detokenizer.xml ADDED
@@ -0,0 +1,388 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <?xml version="1.0"?>
2
+ <net name="detokenizer" version="11">
3
+ <layers>
4
+ <layer id="0" name="Parameter_54567" type="Parameter" version="opset1">
5
+ <data shape="?,?" element_type="i64" />
6
+ <output>
7
+ <port id="0" precision="I64" names="Parameter_54567">
8
+ <dim>-1</dim>
9
+ <dim>-1</dim>
10
+ </port>
11
+ </output>
12
+ </layer>
13
+ <layer id="1" name="Convert_54593" type="Convert" version="opset1">
14
+ <data destination_type="i32" />
15
+ <input>
16
+ <port id="0" precision="I64">
17
+ <dim>-1</dim>
18
+ <dim>-1</dim>
19
+ </port>
20
+ </input>
21
+ <output>
22
+ <port id="1" precision="I32">
23
+ <dim>-1</dim>
24
+ <dim>-1</dim>
25
+ </port>
26
+ </output>
27
+ </layer>
28
+ <layer id="2" name="Constant_54531" type="Const" version="opset1">
29
+ <data element_type="u8" shape="8200" offset="0" size="8200" />
30
+ <output>
31
+ <port id="0" precision="U8">
32
+ <dim>8200</dim>
33
+ </port>
34
+ </output>
35
+ </layer>
36
+ <layer id="3" name="StringTensorUnpack_54532" type="StringTensorUnpack" version="extension">
37
+ <data mode="begins_ends" />
38
+ <input>
39
+ <port id="0" precision="U8">
40
+ <dim>8200</dim>
41
+ </port>
42
+ </input>
43
+ <output>
44
+ <port id="1" precision="I32">
45
+ <dim>-1</dim>
46
+ </port>
47
+ <port id="2" precision="I32">
48
+ <dim>-1</dim>
49
+ </port>
50
+ <port id="3" precision="U8">
51
+ <dim>-1</dim>
52
+ </port>
53
+ </output>
54
+ </layer>
55
+ <layer id="4" name="Constant_54571" type="Const" version="opset1">
56
+ <data element_type="i32" shape="2" offset="8200" size="8" />
57
+ <output>
58
+ <port id="0" precision="I32">
59
+ <dim>2</dim>
60
+ </port>
61
+ </output>
62
+ </layer>
63
+ <layer id="5" name="Constant_54569" type="Const" version="opset1">
64
+ <data element_type="i32" shape="1" offset="8208" size="4" />
65
+ <output>
66
+ <port id="0" precision="I32">
67
+ <dim>1</dim>
68
+ </port>
69
+ </output>
70
+ </layer>
71
+ <layer id="6" name="Constant_54568" type="Const" version="opset1">
72
+ <data element_type="i32" shape="1" offset="8212" size="4" />
73
+ <output>
74
+ <port id="0" precision="I32">
75
+ <dim>1</dim>
76
+ </port>
77
+ </output>
78
+ </layer>
79
+ <layer id="7" name="Constant_54570" type="Const" version="opset1">
80
+ <data element_type="i32" shape="1" offset="8216" size="4" />
81
+ <output>
82
+ <port id="0" precision="I32">
83
+ <dim>1</dim>
84
+ </port>
85
+ </output>
86
+ </layer>
87
+ <layer id="8" name="Constant_54573" type="Const" version="opset1">
88
+ <data element_type="i64" shape="1" offset="8220" size="8" />
89
+ <output>
90
+ <port id="0" precision="I64">
91
+ <dim>1</dim>
92
+ </port>
93
+ </output>
94
+ </layer>
95
+ <layer id="9" name="Slice_54572" type="Slice" version="opset8">
96
+ <input>
97
+ <port id="0" precision="I32">
98
+ <dim>2</dim>
99
+ </port>
100
+ <port id="1" precision="I32">
101
+ <dim>1</dim>
102
+ </port>
103
+ <port id="2" precision="I32">
104
+ <dim>1</dim>
105
+ </port>
106
+ <port id="3" precision="I32">
107
+ <dim>1</dim>
108
+ </port>
109
+ <port id="4" precision="I64">
110
+ <dim>1</dim>
111
+ </port>
112
+ </input>
113
+ <output>
114
+ <port id="5" precision="I32">
115
+ <dim>2</dim>
116
+ </port>
117
+ </output>
118
+ </layer>
119
+ <layer id="10" name="VocabDecoder_54574" type="VocabDecoder" version="extension">
120
+ <data skip_tokens="" />
121
+ <input>
122
+ <port id="0" precision="I32">
123
+ <dim>-1</dim>
124
+ <dim>-1</dim>
125
+ </port>
126
+ <port id="1" precision="I32">
127
+ <dim>-1</dim>
128
+ </port>
129
+ <port id="2" precision="I32">
130
+ <dim>-1</dim>
131
+ </port>
132
+ <port id="3" precision="U8">
133
+ <dim>-1</dim>
134
+ </port>
135
+ <port id="4" precision="I32">
136
+ <dim>2</dim>
137
+ </port>
138
+ </input>
139
+ <output>
140
+ <port id="5" precision="I32">
141
+ <dim>-1</dim>
142
+ </port>
143
+ <port id="6" precision="I32">
144
+ <dim>-1</dim>
145
+ </port>
146
+ <port id="7" precision="I32">
147
+ <dim>-1</dim>
148
+ </port>
149
+ <port id="8" precision="I32">
150
+ <dim>-1</dim>
151
+ </port>
152
+ <port id="9" precision="U8">
153
+ <dim>-1</dim>
154
+ </port>
155
+ </output>
156
+ </layer>
157
+ <layer id="11" name="FuzeRagged_54575" type="FuzeRagged" version="extension">
158
+ <input>
159
+ <port id="0" precision="I32">
160
+ <dim>-1</dim>
161
+ </port>
162
+ <port id="1" precision="I32">
163
+ <dim>-1</dim>
164
+ </port>
165
+ <port id="2" precision="I32">
166
+ <dim>-1</dim>
167
+ </port>
168
+ <port id="3" precision="I32">
169
+ <dim>-1</dim>
170
+ </port>
171
+ </input>
172
+ <output>
173
+ <port id="4" precision="I32">
174
+ <dim>-1</dim>
175
+ </port>
176
+ <port id="5" precision="I32">
177
+ <dim>-1</dim>
178
+ </port>
179
+ </output>
180
+ </layer>
181
+ <layer id="12" name="UTF8Validate_54576" type="UTF8Validate" version="extension">
182
+ <data replace_mode="true" />
183
+ <input>
184
+ <port id="0" precision="I32">
185
+ <dim>-1</dim>
186
+ </port>
187
+ <port id="1" precision="I32">
188
+ <dim>-1</dim>
189
+ </port>
190
+ <port id="2" precision="U8">
191
+ <dim>-1</dim>
192
+ </port>
193
+ </input>
194
+ <output>
195
+ <port id="3" precision="I32">
196
+ <dim>-1</dim>
197
+ </port>
198
+ <port id="4" precision="I32">
199
+ <dim>-1</dim>
200
+ </port>
201
+ <port id="5" precision="U8">
202
+ <dim>-1</dim>
203
+ </port>
204
+ </output>
205
+ </layer>
206
+ <layer id="13" name="Constant_54578" type="Const" version="opset1">
207
+ <data element_type="u8" shape="4" offset="8228" size="4" />
208
+ <output>
209
+ <port id="0" precision="U8">
210
+ <dim>4</dim>
211
+ </port>
212
+ </output>
213
+ </layer>
214
+ <layer id="14" name="Constant_54580" type="Const" version="opset1">
215
+ <data element_type="u8" shape="1" offset="8232" size="1" />
216
+ <output>
217
+ <port id="0" precision="U8">
218
+ <dim>1</dim>
219
+ </port>
220
+ </output>
221
+ </layer>
222
+ <layer id="15" name="RegexNormalization_54581" type="RegexNormalization" version="extension">
223
+ <data global_replace="true" />
224
+ <input>
225
+ <port id="0" precision="I32">
226
+ <dim>-1</dim>
227
+ </port>
228
+ <port id="1" precision="I32">
229
+ <dim>-1</dim>
230
+ </port>
231
+ <port id="2" precision="U8">
232
+ <dim>-1</dim>
233
+ </port>
234
+ <port id="3" precision="U8">
235
+ <dim>4</dim>
236
+ </port>
237
+ <port id="4" precision="U8">
238
+ <dim>1</dim>
239
+ </port>
240
+ </input>
241
+ <output>
242
+ <port id="5" precision="I32">
243
+ <dim>-1</dim>
244
+ </port>
245
+ <port id="6" precision="I32">
246
+ <dim>-1</dim>
247
+ </port>
248
+ <port id="7" precision="U8">
249
+ <dim>-1</dim>
250
+ </port>
251
+ </output>
252
+ </layer>
253
+ <layer id="16" name="Constant_54583" type="Const" version="opset1">
254
+ <data element_type="u8" shape="2" offset="8233" size="2" />
255
+ <output>
256
+ <port id="0" precision="U8">
257
+ <dim>2</dim>
258
+ </port>
259
+ </output>
260
+ </layer>
261
+ <layer id="17" name="Constant_54585" type="Const" version="opset1">
262
+ <data element_type="u8" shape="0" offset="8235" size="1" />
263
+ <output>
264
+ <port id="0" precision="U8">
265
+ <dim>0</dim>
266
+ </port>
267
+ </output>
268
+ </layer>
269
+ <layer id="18" name="RegexNormalization_54586" type="RegexNormalization" version="extension">
270
+ <data global_replace="true" />
271
+ <input>
272
+ <port id="0" precision="I32">
273
+ <dim>-1</dim>
274
+ </port>
275
+ <port id="1" precision="I32">
276
+ <dim>-1</dim>
277
+ </port>
278
+ <port id="2" precision="U8">
279
+ <dim>-1</dim>
280
+ </port>
281
+ <port id="3" precision="U8">
282
+ <dim>2</dim>
283
+ </port>
284
+ <port id="4" precision="U8">
285
+ <dim>0</dim>
286
+ </port>
287
+ </input>
288
+ <output>
289
+ <port id="5" precision="I32">
290
+ <dim>-1</dim>
291
+ </port>
292
+ <port id="6" precision="I32">
293
+ <dim>-1</dim>
294
+ </port>
295
+ <port id="7" precision="U8">
296
+ <dim>-1</dim>
297
+ </port>
298
+ </output>
299
+ </layer>
300
+ <layer id="19" name="StringTensorPack_54587" type="StringTensorPack" version="extension">
301
+ <data mode="begins_ends" />
302
+ <input>
303
+ <port id="0" precision="I32">
304
+ <dim>-1</dim>
305
+ </port>
306
+ <port id="1" precision="I32">
307
+ <dim>-1</dim>
308
+ </port>
309
+ <port id="2" precision="U8">
310
+ <dim>-1</dim>
311
+ </port>
312
+ </input>
313
+ <output>
314
+ <port id="3" precision="STRING" names="string_output">
315
+ <dim>-1</dim>
316
+ </port>
317
+ </output>
318
+ </layer>
319
+ <layer id="20" name="Result_54588" type="Result" version="opset1">
320
+ <input>
321
+ <port id="0" precision="STRING">
322
+ <dim>-1</dim>
323
+ </port>
324
+ </input>
325
+ </layer>
326
+ </layers>
327
+ <edges>
328
+ <edge from-layer="0" from-port="0" to-layer="1" to-port="0" />
329
+ <edge from-layer="1" from-port="1" to-layer="10" to-port="0" />
330
+ <edge from-layer="2" from-port="0" to-layer="3" to-port="0" />
331
+ <edge from-layer="3" from-port="2" to-layer="10" to-port="2" />
332
+ <edge from-layer="3" from-port="3" to-layer="10" to-port="3" />
333
+ <edge from-layer="3" from-port="1" to-layer="10" to-port="1" />
334
+ <edge from-layer="4" from-port="0" to-layer="9" to-port="0" />
335
+ <edge from-layer="5" from-port="0" to-layer="9" to-port="1" />
336
+ <edge from-layer="6" from-port="0" to-layer="9" to-port="2" />
337
+ <edge from-layer="7" from-port="0" to-layer="9" to-port="3" />
338
+ <edge from-layer="8" from-port="0" to-layer="9" to-port="4" />
339
+ <edge from-layer="9" from-port="5" to-layer="10" to-port="4" />
340
+ <edge from-layer="10" from-port="5" to-layer="11" to-port="0" />
341
+ <edge from-layer="10" from-port="6" to-layer="11" to-port="1" />
342
+ <edge from-layer="10" from-port="7" to-layer="11" to-port="2" />
343
+ <edge from-layer="10" from-port="8" to-layer="11" to-port="3" />
344
+ <edge from-layer="10" from-port="9" to-layer="12" to-port="2" />
345
+ <edge from-layer="11" from-port="4" to-layer="12" to-port="0" />
346
+ <edge from-layer="11" from-port="5" to-layer="12" to-port="1" />
347
+ <edge from-layer="12" from-port="3" to-layer="15" to-port="0" />
348
+ <edge from-layer="12" from-port="4" to-layer="15" to-port="1" />
349
+ <edge from-layer="12" from-port="5" to-layer="15" to-port="2" />
350
+ <edge from-layer="13" from-port="0" to-layer="15" to-port="3" />
351
+ <edge from-layer="14" from-port="0" to-layer="15" to-port="4" />
352
+ <edge from-layer="15" from-port="5" to-layer="18" to-port="0" />
353
+ <edge from-layer="15" from-port="6" to-layer="18" to-port="1" />
354
+ <edge from-layer="15" from-port="7" to-layer="18" to-port="2" />
355
+ <edge from-layer="16" from-port="0" to-layer="18" to-port="3" />
356
+ <edge from-layer="17" from-port="0" to-layer="18" to-port="4" />
357
+ <edge from-layer="18" from-port="5" to-layer="19" to-port="0" />
358
+ <edge from-layer="18" from-port="6" to-layer="19" to-port="1" />
359
+ <edge from-layer="18" from-port="7" to-layer="19" to-port="2" />
360
+ <edge from-layer="19" from-port="3" to-layer="20" to-port="0" />
361
+ </edges>
362
+ <rt_info>
363
+ <add_attention_mask value="True" />
364
+ <add_prefix_space />
365
+ <add_special_tokens value="True" />
366
+ <bos_token_id value="0" />
367
+ <clean_up_tokenization_spaces />
368
+ <detokenizer_input_type value="i64" />
369
+ <eos_token_id value="1" />
370
+ <handle_special_tokens_with_re />
371
+ <number_of_inputs value="1" />
372
+ <openvino_tokenizers_version value="2025.0.0.0" />
373
+ <openvino_version value="2025.0.0" />
374
+ <original_tokenizer_class value="&lt;class 'transformers.models.clip.tokenization_clip_fast.CLIPTokenizerFast'>" />
375
+ <pad_token_id value="1" />
376
+ <sentencepiece_version value="0.2.0" />
377
+ <skip_special_tokens value="True" />
378
+ <streaming_detokenizer value="False" />
379
+ <tiktoken_version value="0.7.0" />
380
+ <tokenizer_output_type value="i64" />
381
+ <tokenizers_version value="0.21.0" />
382
+ <transformers_version value="4.49.0" />
383
+ <use_max_padding value="False" />
384
+ <use_sentencepiece_backend value="False" />
385
+ <utf8_replace_mode value="replace" />
386
+ <with_detokenizer value="True" />
387
+ </rt_info>
388
+ </net>
tokenizer/openvino_tokenizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e25dcae2d1053855189d36bb62591bb1ba1bc7af7d58574b5219f06b9b9cc82
3
+ size 16791
tokenizer/openvino_tokenizer.xml ADDED
@@ -0,0 +1,995 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <?xml version="1.0"?>
2
+ <net name="tokenizer" version="11">
3
+ <layers>
4
+ <layer id="0" name="Parameter_54440" type="Parameter" version="opset1">
5
+ <data shape="?" element_type="string" />
6
+ <output>
7
+ <port id="0" precision="STRING" names="Parameter_54440">
8
+ <dim>-1</dim>
9
+ </port>
10
+ </output>
11
+ </layer>
12
+ <layer id="1" name="Constant_54548" type="Const" version="opset1">
13
+ <data element_type="i32" shape="" offset="0" size="4" />
14
+ <output>
15
+ <port id="0" precision="I32" />
16
+ </output>
17
+ </layer>
18
+ <layer id="2" name="Constant_54549" type="Const" version="opset1">
19
+ <data element_type="i32" shape="" offset="4" size="4" />
20
+ <output>
21
+ <port id="0" precision="I32" />
22
+ </output>
23
+ </layer>
24
+ <layer id="3" name="Constant_54550" type="Const" version="opset1">
25
+ <data element_type="i32" shape="1" offset="0" size="4" />
26
+ <output>
27
+ <port id="0" precision="I32">
28
+ <dim>1</dim>
29
+ </port>
30
+ </output>
31
+ </layer>
32
+ <layer id="4" name="Constant_54446" type="Const" version="opset1">
33
+ <data element_type="i64" shape="" offset="8" size="8" />
34
+ <output>
35
+ <port id="0" precision="I64" />
36
+ </output>
37
+ </layer>
38
+ <layer id="5" name="StringTensorUnpack_54441" type="StringTensorUnpack" version="extension">
39
+ <data mode="begins_ends" />
40
+ <input>
41
+ <port id="0" precision="STRING">
42
+ <dim>-1</dim>
43
+ </port>
44
+ </input>
45
+ <output>
46
+ <port id="1" precision="I32">
47
+ <dim>-1</dim>
48
+ </port>
49
+ <port id="2" precision="I32">
50
+ <dim>-1</dim>
51
+ </port>
52
+ <port id="3" precision="U8">
53
+ <dim>-1</dim>
54
+ </port>
55
+ </output>
56
+ </layer>
57
+ <layer id="6" name="ShapeOf_54442" type="ShapeOf" version="opset3">
58
+ <data output_type="i64" />
59
+ <input>
60
+ <port id="0" precision="I32">
61
+ <dim>-1</dim>
62
+ </port>
63
+ </input>
64
+ <output>
65
+ <port id="1" precision="I64">
66
+ <dim>1</dim>
67
+ </port>
68
+ </output>
69
+ </layer>
70
+ <layer id="7" name="Constant_54443" type="Const" version="opset1">
71
+ <data element_type="i64" shape="" offset="8" size="8" />
72
+ <output>
73
+ <port id="0" precision="I64" />
74
+ </output>
75
+ </layer>
76
+ <layer id="8" name="Constant_54444" type="Const" version="opset1">
77
+ <data element_type="i64" shape="" offset="8" size="8" />
78
+ <output>
79
+ <port id="0" precision="I64" />
80
+ </output>
81
+ </layer>
82
+ <layer id="9" name="Gather_54445" type="Gather" version="opset8">
83
+ <data batch_dims="0" />
84
+ <input>
85
+ <port id="0" precision="I64">
86
+ <dim>1</dim>
87
+ </port>
88
+ <port id="1" precision="I64" />
89
+ <port id="2" precision="I64" />
90
+ </input>
91
+ <output>
92
+ <port id="3" precision="I64" />
93
+ </output>
94
+ </layer>
95
+ <layer id="10" name="Constant_54447" type="Const" version="opset1">
96
+ <data element_type="i64" shape="" offset="16" size="8" />
97
+ <output>
98
+ <port id="0" precision="I64" />
99
+ </output>
100
+ </layer>
101
+ <layer id="11" name="Range_54448" type="Range" version="opset4">
102
+ <data output_type="i32" />
103
+ <input>
104
+ <port id="0" precision="I64" />
105
+ <port id="1" precision="I64" />
106
+ <port id="2" precision="I64" />
107
+ </input>
108
+ <output>
109
+ <port id="3" precision="I32">
110
+ <dim>-1</dim>
111
+ </port>
112
+ </output>
113
+ </layer>
114
+ <layer id="12" name="Constant_54449" type="Const" version="opset1">
115
+ <data element_type="i64" shape="" offset="16" size="8" />
116
+ <output>
117
+ <port id="0" precision="I64" />
118
+ </output>
119
+ </layer>
120
+ <layer id="13" name="Constant_54450" type="Const" version="opset1">
121
+ <data element_type="i64" shape="" offset="16" size="8" />
122
+ <output>
123
+ <port id="0" precision="I64" />
124
+ </output>
125
+ </layer>
126
+ <layer id="14" name="Add_54451" type="Add" version="opset1">
127
+ <data auto_broadcast="numpy" />
128
+ <input>
129
+ <port id="0" precision="I64" />
130
+ <port id="1" precision="I64" />
131
+ </input>
132
+ <output>
133
+ <port id="2" precision="I64" />
134
+ </output>
135
+ </layer>
136
+ <layer id="15" name="Constant_54452" type="Const" version="opset1">
137
+ <data element_type="i64" shape="" offset="16" size="8" />
138
+ <output>
139
+ <port id="0" precision="I64" />
140
+ </output>
141
+ </layer>
142
+ <layer id="16" name="Range_54453" type="Range" version="opset4">
143
+ <data output_type="i32" />
144
+ <input>
145
+ <port id="0" precision="I64" />
146
+ <port id="1" precision="I64" />
147
+ <port id="2" precision="I64" />
148
+ </input>
149
+ <output>
150
+ <port id="3" precision="I32">
151
+ <dim>-1</dim>
152
+ </port>
153
+ </output>
154
+ </layer>
155
+ <layer id="17" name="Constant_54515" type="Const" version="opset1">
156
+ <data element_type="u8" shape="41" offset="24" size="41" />
157
+ <output>
158
+ <port id="0" precision="U8">
159
+ <dim>41</dim>
160
+ </port>
161
+ </output>
162
+ </layer>
163
+ <layer id="18" name="SpecialTokensSplit_54516" type="SpecialTokensSplit" version="extension">
164
+ <input>
165
+ <port id="0" precision="I32">
166
+ <dim>-1</dim>
167
+ </port>
168
+ <port id="1" precision="I32">
169
+ <dim>-1</dim>
170
+ </port>
171
+ <port id="2" precision="I32">
172
+ <dim>-1</dim>
173
+ </port>
174
+ <port id="3" precision="I32">
175
+ <dim>-1</dim>
176
+ </port>
177
+ <port id="4" precision="U8">
178
+ <dim>-1</dim>
179
+ </port>
180
+ <port id="5" precision="U8">
181
+ <dim>41</dim>
182
+ </port>
183
+ </input>
184
+ <output>
185
+ <port id="6" precision="I32">
186
+ <dim>-1</dim>
187
+ </port>
188
+ <port id="7" precision="I32">
189
+ <dim>-1</dim>
190
+ </port>
191
+ <port id="8" precision="I32">
192
+ <dim>-1</dim>
193
+ </port>
194
+ <port id="9" precision="I32">
195
+ <dim>-1</dim>
196
+ </port>
197
+ <port id="10" precision="U8">
198
+ <dim>-1</dim>
199
+ </port>
200
+ <port id="11" precision="BOOL">
201
+ <dim>-1</dim>
202
+ </port>
203
+ </output>
204
+ </layer>
205
+ <layer id="19" name="NormalizeUnicode_54517" type="NormalizeUnicode" version="extension">
206
+ <data normalization_form="NFC" />
207
+ <input>
208
+ <port id="0" precision="I32">
209
+ <dim>-1</dim>
210
+ </port>
211
+ <port id="1" precision="I32">
212
+ <dim>-1</dim>
213
+ </port>
214
+ <port id="2" precision="U8">
215
+ <dim>-1</dim>
216
+ </port>
217
+ <port id="3" precision="BOOL">
218
+ <dim>-1</dim>
219
+ </port>
220
+ </input>
221
+ <output>
222
+ <port id="4" precision="I32">
223
+ <dim>-1</dim>
224
+ </port>
225
+ <port id="5" precision="I32">
226
+ <dim>-1</dim>
227
+ </port>
228
+ <port id="6" precision="U8">
229
+ <dim>-1</dim>
230
+ </port>
231
+ <port id="7" precision="BOOL">
232
+ <dim>-1</dim>
233
+ </port>
234
+ </output>
235
+ </layer>
236
+ <layer id="20" name="Constant_54519" type="Const" version="opset1">
237
+ <data element_type="u8" shape="3" offset="65" size="3" />
238
+ <output>
239
+ <port id="0" precision="U8">
240
+ <dim>3</dim>
241
+ </port>
242
+ </output>
243
+ </layer>
244
+ <layer id="21" name="Constant_54521" type="Const" version="opset1">
245
+ <data element_type="u8" shape="1" offset="68" size="1" />
246
+ <output>
247
+ <port id="0" precision="U8">
248
+ <dim>1</dim>
249
+ </port>
250
+ </output>
251
+ </layer>
252
+ <layer id="22" name="RegexNormalization_54522" type="RegexNormalization" version="extension">
253
+ <data global_replace="true" />
254
+ <input>
255
+ <port id="0" precision="I32">
256
+ <dim>-1</dim>
257
+ </port>
258
+ <port id="1" precision="I32">
259
+ <dim>-1</dim>
260
+ </port>
261
+ <port id="2" precision="U8">
262
+ <dim>-1</dim>
263
+ </port>
264
+ <port id="3" precision="BOOL">
265
+ <dim>-1</dim>
266
+ </port>
267
+ <port id="4" precision="U8">
268
+ <dim>3</dim>
269
+ </port>
270
+ <port id="5" precision="U8">
271
+ <dim>1</dim>
272
+ </port>
273
+ </input>
274
+ <output>
275
+ <port id="6" precision="I32">
276
+ <dim>-1</dim>
277
+ </port>
278
+ <port id="7" precision="I32">
279
+ <dim>-1</dim>
280
+ </port>
281
+ <port id="8" precision="U8">
282
+ <dim>-1</dim>
283
+ </port>
284
+ <port id="9" precision="BOOL">
285
+ <dim>-1</dim>
286
+ </port>
287
+ </output>
288
+ </layer>
289
+ <layer id="23" name="CaseFold_54523" type="CaseFold" version="extension">
290
+ <data encoding="utf-8" />
291
+ <input>
292
+ <port id="0" precision="I32">
293
+ <dim>-1</dim>
294
+ </port>
295
+ <port id="1" precision="I32">
296
+ <dim>-1</dim>
297
+ </port>
298
+ <port id="2" precision="U8">
299
+ <dim>-1</dim>
300
+ </port>
301
+ <port id="3" precision="BOOL">
302
+ <dim>-1</dim>
303
+ </port>
304
+ </input>
305
+ <output>
306
+ <port id="4" precision="I32">
307
+ <dim>-1</dim>
308
+ </port>
309
+ <port id="5" precision="I32">
310
+ <dim>-1</dim>
311
+ </port>
312
+ <port id="6" precision="U8">
313
+ <dim>-1</dim>
314
+ </port>
315
+ <port id="7" precision="BOOL">
316
+ <dim>-1</dim>
317
+ </port>
318
+ </output>
319
+ </layer>
320
+ <layer id="24" name="Constant_54525" type="Const" version="opset1">
321
+ <data element_type="u8" shape="57" offset="69" size="57" />
322
+ <output>
323
+ <port id="0" precision="U8">
324
+ <dim>57</dim>
325
+ </port>
326
+ </output>
327
+ </layer>
328
+ <layer id="25" name="RegexSplit_54526" type="RegexSplit" version="extension">
329
+ <data behaviour="remove" invert="true" max_splits="-1" />
330
+ <input>
331
+ <port id="0" precision="I32">
332
+ <dim>-1</dim>
333
+ </port>
334
+ <port id="1" precision="I32">
335
+ <dim>-1</dim>
336
+ </port>
337
+ <port id="2" precision="I32">
338
+ <dim>-1</dim>
339
+ </port>
340
+ <port id="3" precision="I32">
341
+ <dim>-1</dim>
342
+ </port>
343
+ <port id="4" precision="U8">
344
+ <dim>-1</dim>
345
+ </port>
346
+ <port id="5" precision="BOOL">
347
+ <dim>-1</dim>
348
+ </port>
349
+ <port id="6" precision="U8">
350
+ <dim>57</dim>
351
+ </port>
352
+ </input>
353
+ <output>
354
+ <port id="7" precision="I32">
355
+ <dim>-1</dim>
356
+ </port>
357
+ <port id="8" precision="I32">
358
+ <dim>-1</dim>
359
+ </port>
360
+ <port id="9" precision="I32">
361
+ <dim>-1</dim>
362
+ </port>
363
+ <port id="10" precision="I32">
364
+ <dim>-1</dim>
365
+ </port>
366
+ <port id="11" precision="U8">
367
+ <dim>-1</dim>
368
+ </port>
369
+ <port id="12" precision="BOOL">
370
+ <dim>-1</dim>
371
+ </port>
372
+ </output>
373
+ </layer>
374
+ <layer id="26" name="Constant_54528" type="Const" version="opset1">
375
+ <data element_type="u8" shape="64" offset="126" size="64" />
376
+ <output>
377
+ <port id="0" precision="U8">
378
+ <dim>64</dim>
379
+ </port>
380
+ </output>
381
+ </layer>
382
+ <layer id="27" name="RegexSplit_54529" type="RegexSplit" version="extension">
383
+ <data behaviour="isolate" invert="false" max_splits="-1" />
384
+ <input>
385
+ <port id="0" precision="I32">
386
+ <dim>-1</dim>
387
+ </port>
388
+ <port id="1" precision="I32">
389
+ <dim>-1</dim>
390
+ </port>
391
+ <port id="2" precision="I32">
392
+ <dim>-1</dim>
393
+ </port>
394
+ <port id="3" precision="I32">
395
+ <dim>-1</dim>
396
+ </port>
397
+ <port id="4" precision="U8">
398
+ <dim>-1</dim>
399
+ </port>
400
+ <port id="5" precision="BOOL">
401
+ <dim>-1</dim>
402
+ </port>
403
+ <port id="6" precision="U8">
404
+ <dim>64</dim>
405
+ </port>
406
+ </input>
407
+ <output>
408
+ <port id="7" precision="I32">
409
+ <dim>-1</dim>
410
+ </port>
411
+ <port id="8" precision="I32">
412
+ <dim>-1</dim>
413
+ </port>
414
+ <port id="9" precision="I32">
415
+ <dim>-1</dim>
416
+ </port>
417
+ <port id="10" precision="I32">
418
+ <dim>-1</dim>
419
+ </port>
420
+ <port id="11" precision="U8">
421
+ <dim>-1</dim>
422
+ </port>
423
+ <port id="12" precision="BOOL">
424
+ <dim>-1</dim>
425
+ </port>
426
+ </output>
427
+ </layer>
428
+ <layer id="28" name="Constant_54531" type="Const" version="opset1">
429
+ <data element_type="u8" shape="8200" offset="190" size="8200" />
430
+ <output>
431
+ <port id="0" precision="U8">
432
+ <dim>8200</dim>
433
+ </port>
434
+ </output>
435
+ </layer>
436
+ <layer id="29" name="StringTensorUnpack_54532" type="StringTensorUnpack" version="extension">
437
+ <data mode="begins_ends" />
438
+ <input>
439
+ <port id="0" precision="U8">
440
+ <dim>8200</dim>
441
+ </port>
442
+ </input>
443
+ <output>
444
+ <port id="1" precision="I32">
445
+ <dim>-1</dim>
446
+ </port>
447
+ <port id="2" precision="I32">
448
+ <dim>-1</dim>
449
+ </port>
450
+ <port id="3" precision="U8">
451
+ <dim>-1</dim>
452
+ </port>
453
+ </output>
454
+ </layer>
455
+ <layer id="30" name="Constant_54537" type="Const" version="opset1">
456
+ <data element_type="u8" shape="3698" offset="8390" size="3698" />
457
+ <output>
458
+ <port id="0" precision="U8">
459
+ <dim>3698</dim>
460
+ </port>
461
+ </output>
462
+ </layer>
463
+ <layer id="31" name="StringTensorUnpack_54538" type="StringTensorUnpack" version="extension">
464
+ <data mode="begins_ends" />
465
+ <input>
466
+ <port id="0" precision="U8">
467
+ <dim>3698</dim>
468
+ </port>
469
+ </input>
470
+ <output>
471
+ <port id="1" precision="I32">
472
+ <dim>-1</dim>
473
+ </port>
474
+ <port id="2" precision="I32">
475
+ <dim>-1</dim>
476
+ </port>
477
+ <port id="3" precision="U8">
478
+ <dim>-1</dim>
479
+ </port>
480
+ </output>
481
+ </layer>
482
+ <layer id="32" name="Constant_54540" type="Const" version="opset1">
483
+ <data element_type="u8" shape="4662" offset="12088" size="4662" />
484
+ <output>
485
+ <port id="0" precision="U8">
486
+ <dim>4662</dim>
487
+ </port>
488
+ </output>
489
+ </layer>
490
+ <layer id="33" name="StringTensorUnpack_54541" type="StringTensorUnpack" version="extension">
491
+ <data mode="begins_ends" />
492
+ <input>
493
+ <port id="0" precision="U8">
494
+ <dim>4662</dim>
495
+ </port>
496
+ </input>
497
+ <output>
498
+ <port id="1" precision="I32">
499
+ <dim>-1</dim>
500
+ </port>
501
+ <port id="2" precision="I32">
502
+ <dim>-1</dim>
503
+ </port>
504
+ <port id="3" precision="U8">
505
+ <dim>-1</dim>
506
+ </port>
507
+ </output>
508
+ </layer>
509
+ <layer id="34" name="Constant_54534" type="Const" version="opset1">
510
+ <data element_type="u8" shape="25" offset="16750" size="25" />
511
+ <output>
512
+ <port id="0" precision="U8">
513
+ <dim>25</dim>
514
+ </port>
515
+ </output>
516
+ </layer>
517
+ <layer id="35" name="StringTensorUnpack_54535" type="StringTensorUnpack" version="extension">
518
+ <data mode="begins_ends" />
519
+ <input>
520
+ <port id="0" precision="U8">
521
+ <dim>25</dim>
522
+ </port>
523
+ </input>
524
+ <output>
525
+ <port id="1" precision="I32">
526
+ <dim>-1</dim>
527
+ </port>
528
+ <port id="2" precision="I32">
529
+ <dim>-1</dim>
530
+ </port>
531
+ <port id="3" precision="U8">
532
+ <dim>-1</dim>
533
+ </port>
534
+ </output>
535
+ </layer>
536
+ <layer id="36" name="Constant_54542" type="Const" version="opset1">
537
+ <data element_type="i32" shape="1" offset="4" size="4" />
538
+ <output>
539
+ <port id="0" precision="I32">
540
+ <dim>1</dim>
541
+ </port>
542
+ </output>
543
+ </layer>
544
+ <layer id="37" name="BPETokenizer_54543" type="BPETokenizer" version="extension">
545
+ <data unk_token="&lt;|endoftext|>" fuse_unk="false" suffix_indicator="" end_suffix="&lt;/w>" byte_fallback="false" cache_capacity="20000" />
546
+ <input>
547
+ <port id="0" precision="I32">
548
+ <dim>-1</dim>
549
+ </port>
550
+ <port id="1" precision="I32">
551
+ <dim>-1</dim>
552
+ </port>
553
+ <port id="2" precision="I32">
554
+ <dim>-1</dim>
555
+ </port>
556
+ <port id="3" precision="I32">
557
+ <dim>-1</dim>
558
+ </port>
559
+ <port id="4" precision="U8">
560
+ <dim>-1</dim>
561
+ </port>
562
+ <port id="5" precision="I32">
563
+ <dim>-1</dim>
564
+ </port>
565
+ <port id="6" precision="I32">
566
+ <dim>-1</dim>
567
+ </port>
568
+ <port id="7" precision="U8">
569
+ <dim>-1</dim>
570
+ </port>
571
+ <port id="8" precision="I32">
572
+ <dim>-1</dim>
573
+ </port>
574
+ <port id="9" precision="I32">
575
+ <dim>-1</dim>
576
+ </port>
577
+ <port id="10" precision="U8">
578
+ <dim>-1</dim>
579
+ </port>
580
+ <port id="11" precision="I32">
581
+ <dim>-1</dim>
582
+ </port>
583
+ <port id="12" precision="I32">
584
+ <dim>-1</dim>
585
+ </port>
586
+ <port id="13" precision="U8">
587
+ <dim>-1</dim>
588
+ </port>
589
+ <port id="14" precision="I32">
590
+ <dim>-1</dim>
591
+ </port>
592
+ <port id="15" precision="I32">
593
+ <dim>-1</dim>
594
+ </port>
595
+ <port id="16" precision="U8">
596
+ <dim>-1</dim>
597
+ </port>
598
+ <port id="17" precision="I32">
599
+ <dim>1</dim>
600
+ </port>
601
+ </input>
602
+ <output>
603
+ <port id="18" precision="I32">
604
+ <dim>-1</dim>
605
+ </port>
606
+ <port id="19" precision="I32">
607
+ <dim>-1</dim>
608
+ </port>
609
+ <port id="20" precision="I32">
610
+ <dim>-1</dim>
611
+ </port>
612
+ </output>
613
+ </layer>
614
+ <layer id="38" name="Subtract_54544" type="Subtract" version="opset1">
615
+ <data auto_broadcast="numpy" />
616
+ <input>
617
+ <port id="0" precision="I32">
618
+ <dim>-1</dim>
619
+ </port>
620
+ <port id="1" precision="I32">
621
+ <dim>-1</dim>
622
+ </port>
623
+ </input>
624
+ <output>
625
+ <port id="2" precision="I32">
626
+ <dim>-1</dim>
627
+ </port>
628
+ </output>
629
+ </layer>
630
+ <layer id="39" name="Constant_54545" type="Const" version="opset1">
631
+ <data element_type="i32" shape="" offset="16775" size="4" />
632
+ <output>
633
+ <port id="0" precision="I32" />
634
+ </output>
635
+ </layer>
636
+ <layer id="40" name="Minimum_54546" type="Minimum" version="opset1">
637
+ <data auto_broadcast="numpy" />
638
+ <input>
639
+ <port id="0" precision="I32">
640
+ <dim>-1</dim>
641
+ </port>
642
+ <port id="1" precision="I32" />
643
+ </input>
644
+ <output>
645
+ <port id="2" precision="I32">
646
+ <dim>-1</dim>
647
+ </port>
648
+ </output>
649
+ </layer>
650
+ <layer id="41" name="Add_54547" type="Add" version="opset1">
651
+ <data auto_broadcast="numpy" />
652
+ <input>
653
+ <port id="0" precision="I32">
654
+ <dim>-1</dim>
655
+ </port>
656
+ <port id="1" precision="I32">
657
+ <dim>-1</dim>
658
+ </port>
659
+ </input>
660
+ <output>
661
+ <port id="2" precision="I32">
662
+ <dim>-1</dim>
663
+ </port>
664
+ </output>
665
+ </layer>
666
+ <layer id="42" name="Constant_54551" type="Const" version="opset1">
667
+ <data element_type="i32" shape="" offset="0" size="4" />
668
+ <output>
669
+ <port id="0" precision="I32" />
670
+ </output>
671
+ </layer>
672
+ <layer id="43" name="Constant_54552" type="Const" version="opset1">
673
+ <data element_type="i32" shape="" offset="4" size="4" />
674
+ <output>
675
+ <port id="0" precision="I32" />
676
+ </output>
677
+ </layer>
678
+ <layer id="44" name="Constant_54553" type="Const" version="opset1">
679
+ <data element_type="i32" shape="1" offset="4" size="4" />
680
+ <output>
681
+ <port id="0" precision="I32">
682
+ <dim>1</dim>
683
+ </port>
684
+ </output>
685
+ </layer>
686
+ <layer id="45" name="Constant_54554" type="Const" version="opset1">
687
+ <data element_type="i32" shape="3" offset="16779" size="12" />
688
+ <output>
689
+ <port id="0" precision="I32">
690
+ <dim>3</dim>
691
+ </port>
692
+ </output>
693
+ </layer>
694
+ <layer id="46" name="CombineSegments_54555" type="CombineSegments" version="extension">
695
+ <input>
696
+ <port id="0" precision="I32" />
697
+ <port id="1" precision="I32" />
698
+ <port id="2" precision="I32">
699
+ <dim>1</dim>
700
+ </port>
701
+ <port id="3" precision="I32">
702
+ <dim>-1</dim>
703
+ </port>
704
+ <port id="4" precision="I32">
705
+ <dim>-1</dim>
706
+ </port>
707
+ <port id="5" precision="I32">
708
+ <dim>-1</dim>
709
+ </port>
710
+ <port id="6" precision="I32" />
711
+ <port id="7" precision="I32" />
712
+ <port id="8" precision="I32">
713
+ <dim>1</dim>
714
+ </port>
715
+ <port id="9" precision="I32">
716
+ <dim>3</dim>
717
+ </port>
718
+ </input>
719
+ <output>
720
+ <port id="10" precision="I32">
721
+ <dim>-1</dim>
722
+ </port>
723
+ <port id="11" precision="I32">
724
+ <dim>-1</dim>
725
+ </port>
726
+ <port id="12" precision="I32">
727
+ <dim>-1</dim>
728
+ </port>
729
+ <port id="13" precision="I32">
730
+ <dim>-1</dim>
731
+ </port>
732
+ <port id="14" precision="I32">
733
+ <dim>-1</dim>
734
+ </port>
735
+ <port id="15" precision="I32">
736
+ <dim>-1</dim>
737
+ </port>
738
+ </output>
739
+ </layer>
740
+ <layer id="47" name="Subtract_54556" type="Subtract" version="opset1">
741
+ <data auto_broadcast="numpy" />
742
+ <input>
743
+ <port id="0" precision="I32">
744
+ <dim>-1</dim>
745
+ </port>
746
+ <port id="1" precision="I32">
747
+ <dim>-1</dim>
748
+ </port>
749
+ </input>
750
+ <output>
751
+ <port id="2" precision="I32">
752
+ <dim>-1</dim>
753
+ </port>
754
+ </output>
755
+ </layer>
756
+ <layer id="48" name="Constant_54557" type="Const" version="opset1">
757
+ <data element_type="i32" shape="" offset="0" size="4" />
758
+ <output>
759
+ <port id="0" precision="I32" />
760
+ </output>
761
+ </layer>
762
+ <layer id="49" name="ReduceMax_54558" type="ReduceMax" version="opset1">
763
+ <data keep_dims="false" />
764
+ <input>
765
+ <port id="0" precision="I32">
766
+ <dim>-1</dim>
767
+ </port>
768
+ <port id="1" precision="I32" />
769
+ </input>
770
+ <output>
771
+ <port id="2" precision="I32" />
772
+ </output>
773
+ </layer>
774
+ <layer id="50" name="Constant_54559" type="Const" version="opset1">
775
+ <data element_type="i32" shape="" offset="4" size="4" />
776
+ <output>
777
+ <port id="0" precision="I32" />
778
+ </output>
779
+ </layer>
780
+ <layer id="51" name="RaggedToDense_54560" type="RaggedToDense" version="extension">
781
+ <data pad_right="true" />
782
+ <input>
783
+ <port id="0" precision="I32">
784
+ <dim>-1</dim>
785
+ </port>
786
+ <port id="1" precision="I32">
787
+ <dim>-1</dim>
788
+ </port>
789
+ <port id="2" precision="I32">
790
+ <dim>-1</dim>
791
+ </port>
792
+ <port id="3" precision="I32" />
793
+ <port id="4" precision="I32" />
794
+ </input>
795
+ <output>
796
+ <port id="5" precision="I32">
797
+ <dim>-1</dim>
798
+ <dim>-1</dim>
799
+ </port>
800
+ <port id="6" precision="BOOL">
801
+ <dim>-1</dim>
802
+ <dim>-1</dim>
803
+ </port>
804
+ </output>
805
+ </layer>
806
+ <layer id="52" name="Convert_54561" type="Convert" version="opset1">
807
+ <data destination_type="i32" />
808
+ <input>
809
+ <port id="0" precision="BOOL">
810
+ <dim>-1</dim>
811
+ <dim>-1</dim>
812
+ </port>
813
+ </input>
814
+ <output>
815
+ <port id="1" precision="I32">
816
+ <dim>-1</dim>
817
+ <dim>-1</dim>
818
+ </port>
819
+ </output>
820
+ </layer>
821
+ <layer id="53" name="Convert_54561.0" type="Convert" version="opset1">
822
+ <data destination_type="i64" />
823
+ <input>
824
+ <port id="0" precision="I32">
825
+ <dim>-1</dim>
826
+ <dim>-1</dim>
827
+ </port>
828
+ </input>
829
+ <output>
830
+ <port id="1" precision="I64" names="attention_mask">
831
+ <dim>-1</dim>
832
+ <dim>-1</dim>
833
+ </port>
834
+ </output>
835
+ </layer>
836
+ <layer id="55" name="RaggedToDense_54560.0" type="Convert" version="opset1">
837
+ <data destination_type="i64" />
838
+ <input>
839
+ <port id="0" precision="I32">
840
+ <dim>-1</dim>
841
+ <dim>-1</dim>
842
+ </port>
843
+ </input>
844
+ <output>
845
+ <port id="1" precision="I64" names="input_ids">
846
+ <dim>-1</dim>
847
+ <dim>-1</dim>
848
+ </port>
849
+ </output>
850
+ </layer>
851
+ <layer id="56" name="Result_54564" type="Result" version="opset1">
852
+ <input>
853
+ <port id="0" precision="I64">
854
+ <dim>-1</dim>
855
+ <dim>-1</dim>
856
+ </port>
857
+ </input>
858
+ </layer>
859
+ <layer id="54" name="Result_54566" type="Result" version="opset1">
860
+ <input>
861
+ <port id="0" precision="I64">
862
+ <dim>-1</dim>
863
+ <dim>-1</dim>
864
+ </port>
865
+ </input>
866
+ </layer>
867
+ </layers>
868
+ <edges>
869
+ <edge from-layer="0" from-port="0" to-layer="5" to-port="0" />
870
+ <edge from-layer="1" from-port="0" to-layer="46" to-port="0" />
871
+ <edge from-layer="2" from-port="0" to-layer="46" to-port="1" />
872
+ <edge from-layer="3" from-port="0" to-layer="46" to-port="2" />
873
+ <edge from-layer="4" from-port="0" to-layer="11" to-port="0" />
874
+ <edge from-layer="5" from-port="1" to-layer="6" to-port="0" />
875
+ <edge from-layer="5" from-port="3" to-layer="18" to-port="4" />
876
+ <edge from-layer="5" from-port="2" to-layer="18" to-port="3" />
877
+ <edge from-layer="5" from-port="1" to-layer="18" to-port="2" />
878
+ <edge from-layer="6" from-port="1" to-layer="9" to-port="0" />
879
+ <edge from-layer="7" from-port="0" to-layer="9" to-port="1" />
880
+ <edge from-layer="8" from-port="0" to-layer="9" to-port="2" />
881
+ <edge from-layer="9" from-port="3" to-layer="11" to-port="1" />
882
+ <edge from-layer="9" from-port="3" to-layer="14" to-port="0" />
883
+ <edge from-layer="10" from-port="0" to-layer="11" to-port="2" />
884
+ <edge from-layer="11" from-port="3" to-layer="18" to-port="0" />
885
+ <edge from-layer="12" from-port="0" to-layer="16" to-port="0" />
886
+ <edge from-layer="13" from-port="0" to-layer="14" to-port="1" />
887
+ <edge from-layer="14" from-port="2" to-layer="16" to-port="1" />
888
+ <edge from-layer="15" from-port="0" to-layer="16" to-port="2" />
889
+ <edge from-layer="16" from-port="3" to-layer="18" to-port="1" />
890
+ <edge from-layer="17" from-port="0" to-layer="18" to-port="5" />
891
+ <edge from-layer="18" from-port="6" to-layer="25" to-port="0" />
892
+ <edge from-layer="18" from-port="7" to-layer="25" to-port="1" />
893
+ <edge from-layer="18" from-port="11" to-layer="19" to-port="3" />
894
+ <edge from-layer="18" from-port="10" to-layer="19" to-port="2" />
895
+ <edge from-layer="18" from-port="9" to-layer="19" to-port="1" />
896
+ <edge from-layer="18" from-port="8" to-layer="19" to-port="0" />
897
+ <edge from-layer="19" from-port="5" to-layer="22" to-port="1" />
898
+ <edge from-layer="19" from-port="6" to-layer="22" to-port="2" />
899
+ <edge from-layer="19" from-port="7" to-layer="22" to-port="3" />
900
+ <edge from-layer="19" from-port="4" to-layer="22" to-port="0" />
901
+ <edge from-layer="20" from-port="0" to-layer="22" to-port="4" />
902
+ <edge from-layer="21" from-port="0" to-layer="22" to-port="5" />
903
+ <edge from-layer="22" from-port="9" to-layer="23" to-port="3" />
904
+ <edge from-layer="22" from-port="8" to-layer="23" to-port="2" />
905
+ <edge from-layer="22" from-port="7" to-layer="23" to-port="1" />
906
+ <edge from-layer="22" from-port="6" to-layer="23" to-port="0" />
907
+ <edge from-layer="23" from-port="5" to-layer="25" to-port="3" />
908
+ <edge from-layer="23" from-port="4" to-layer="25" to-port="2" />
909
+ <edge from-layer="23" from-port="7" to-layer="25" to-port="5" />
910
+ <edge from-layer="23" from-port="6" to-layer="25" to-port="4" />
911
+ <edge from-layer="24" from-port="0" to-layer="25" to-port="6" />
912
+ <edge from-layer="25" from-port="7" to-layer="27" to-port="0" />
913
+ <edge from-layer="25" from-port="8" to-layer="27" to-port="1" />
914
+ <edge from-layer="25" from-port="9" to-layer="27" to-port="2" />
915
+ <edge from-layer="25" from-port="10" to-layer="27" to-port="3" />
916
+ <edge from-layer="25" from-port="11" to-layer="27" to-port="4" />
917
+ <edge from-layer="25" from-port="12" to-layer="27" to-port="5" />
918
+ <edge from-layer="26" from-port="0" to-layer="27" to-port="6" />
919
+ <edge from-layer="27" from-port="7" to-layer="37" to-port="0" />
920
+ <edge from-layer="27" from-port="11" to-layer="37" to-port="4" />
921
+ <edge from-layer="27" from-port="10" to-layer="37" to-port="3" />
922
+ <edge from-layer="27" from-port="9" to-layer="37" to-port="2" />
923
+ <edge from-layer="27" from-port="8" to-layer="37" to-port="1" />
924
+ <edge from-layer="28" from-port="0" to-layer="29" to-port="0" />
925
+ <edge from-layer="29" from-port="1" to-layer="37" to-port="5" />
926
+ <edge from-layer="29" from-port="2" to-layer="37" to-port="6" />
927
+ <edge from-layer="29" from-port="3" to-layer="37" to-port="7" />
928
+ <edge from-layer="30" from-port="0" to-layer="31" to-port="0" />
929
+ <edge from-layer="31" from-port="1" to-layer="37" to-port="8" />
930
+ <edge from-layer="31" from-port="2" to-layer="37" to-port="9" />
931
+ <edge from-layer="31" from-port="3" to-layer="37" to-port="10" />
932
+ <edge from-layer="32" from-port="0" to-layer="33" to-port="0" />
933
+ <edge from-layer="33" from-port="3" to-layer="37" to-port="13" />
934
+ <edge from-layer="33" from-port="1" to-layer="37" to-port="11" />
935
+ <edge from-layer="33" from-port="2" to-layer="37" to-port="12" />
936
+ <edge from-layer="34" from-port="0" to-layer="35" to-port="0" />
937
+ <edge from-layer="35" from-port="1" to-layer="37" to-port="14" />
938
+ <edge from-layer="35" from-port="3" to-layer="37" to-port="16" />
939
+ <edge from-layer="35" from-port="2" to-layer="37" to-port="15" />
940
+ <edge from-layer="36" from-port="0" to-layer="37" to-port="17" />
941
+ <edge from-layer="37" from-port="19" to-layer="38" to-port="0" />
942
+ <edge from-layer="37" from-port="18" to-layer="38" to-port="1" />
943
+ <edge from-layer="37" from-port="18" to-layer="41" to-port="0" />
944
+ <edge from-layer="37" from-port="18" to-layer="46" to-port="3" />
945
+ <edge from-layer="37" from-port="20" to-layer="46" to-port="5" />
946
+ <edge from-layer="38" from-port="2" to-layer="40" to-port="0" />
947
+ <edge from-layer="39" from-port="0" to-layer="40" to-port="1" />
948
+ <edge from-layer="40" from-port="2" to-layer="41" to-port="1" />
949
+ <edge from-layer="41" from-port="2" to-layer="46" to-port="4" />
950
+ <edge from-layer="42" from-port="0" to-layer="46" to-port="6" />
951
+ <edge from-layer="43" from-port="0" to-layer="46" to-port="7" />
952
+ <edge from-layer="44" from-port="0" to-layer="46" to-port="8" />
953
+ <edge from-layer="45" from-port="0" to-layer="46" to-port="9" />
954
+ <edge from-layer="46" from-port="11" to-layer="51" to-port="1" />
955
+ <edge from-layer="46" from-port="12" to-layer="51" to-port="2" />
956
+ <edge from-layer="46" from-port="10" to-layer="51" to-port="0" />
957
+ <edge from-layer="46" from-port="10" to-layer="47" to-port="1" />
958
+ <edge from-layer="46" from-port="11" to-layer="47" to-port="0" />
959
+ <edge from-layer="47" from-port="2" to-layer="49" to-port="0" />
960
+ <edge from-layer="48" from-port="0" to-layer="49" to-port="1" />
961
+ <edge from-layer="49" from-port="2" to-layer="51" to-port="3" />
962
+ <edge from-layer="50" from-port="0" to-layer="51" to-port="4" />
963
+ <edge from-layer="51" from-port="6" to-layer="52" to-port="0" />
964
+ <edge from-layer="51" from-port="5" to-layer="55" to-port="0" />
965
+ <edge from-layer="52" from-port="1" to-layer="53" to-port="0" />
966
+ <edge from-layer="53" from-port="1" to-layer="54" to-port="0" />
967
+ <edge from-layer="55" from-port="1" to-layer="56" to-port="0" />
968
+ </edges>
969
+ <rt_info>
970
+ <add_attention_mask value="True" />
971
+ <add_prefix_space />
972
+ <add_special_tokens value="True" />
973
+ <bos_token_id value="0" />
974
+ <clean_up_tokenization_spaces />
975
+ <detokenizer_input_type value="i64" />
976
+ <eos_token_id value="1" />
977
+ <handle_special_tokens_with_re />
978
+ <number_of_inputs value="1" />
979
+ <openvino_tokenizers_version value="2025.0.0.0" />
980
+ <openvino_version value="2025.0.0" />
981
+ <original_tokenizer_class value="&lt;class 'transformers.models.clip.tokenization_clip_fast.CLIPTokenizerFast'>" />
982
+ <pad_token_id value="1" />
983
+ <sentencepiece_version value="0.2.0" />
984
+ <skip_special_tokens value="True" />
985
+ <streaming_detokenizer value="False" />
986
+ <tiktoken_version value="0.7.0" />
987
+ <tokenizer_output_type value="i64" />
988
+ <tokenizers_version value="0.21.0" />
989
+ <transformers_version value="4.49.0" />
990
+ <use_max_padding value="False" />
991
+ <use_sentencepiece_backend value="False" />
992
+ <utf8_replace_mode value="replace" />
993
+ <with_detokenizer value="True" />
994
+ </rt_info>
995
+ </net>
tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<|endoftext|>",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<|startoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<|endoftext|>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ }
20
+ },
21
+ "bos_token": "<|startoftext|>",
22
+ "clean_up_tokenization_spaces": false,
23
+ "do_lower_case": true,
24
+ "eos_token": "<|endoftext|>",
25
+ "errors": "replace",
26
+ "extra_special_tokens": {},
27
+ "model_max_length": 77,
28
+ "pad_token": "<|endoftext|>",
29
+ "tokenizer_class": "CLIPTokenizer",
30
+ "unk_token": "<|endoftext|>"
31
+ }
tokenizer/vocab.json ADDED
@@ -0,0 +1,1002 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "!": 2,
3
+ "!</w>": 345,
4
+ "\"": 3,
5
+ "\"</w>": 344,
6
+ "#": 4,
7
+ "#</w>": 325,
8
+ "$": 5,
9
+ "$</w>": 348,
10
+ "%": 6,
11
+ "%</w>": 351,
12
+ "&": 7,
13
+ "&</w>": 352,
14
+ "'": 8,
15
+ "'</w>": 296,
16
+ "(": 9,
17
+ "(</w>": 318,
18
+ ")": 10,
19
+ ")</w>": 330,
20
+ "*": 11,
21
+ "*</w>": 327,
22
+ "+": 12,
23
+ "+</w>": 341,
24
+ ",": 13,
25
+ ",</w>": 279,
26
+ ",@</w>": 754,
27
+ "-": 14,
28
+ "-</w>": 276,
29
+ "-@</w>": 439,
30
+ ".": 15,
31
+ ".</w>": 253,
32
+ ".@</w>": 695,
33
+ "/": 16,
34
+ "/</w>": 350,
35
+ "0": 17,
36
+ "00</w>": 647,
37
+ "0</w>": 216,
38
+ "1": 18,
39
+ "1</w>": 222,
40
+ "2": 19,
41
+ "2</w>": 231,
42
+ "3": 20,
43
+ "3</w>": 243,
44
+ "4": 21,
45
+ "4</w>": 233,
46
+ "5": 22,
47
+ "5</w>": 240,
48
+ "6": 23,
49
+ "6</w>": 226,
50
+ "7": 24,
51
+ "7</w>": 215,
52
+ "8": 25,
53
+ "8</w>": 236,
54
+ "9": 26,
55
+ "9</w>": 242,
56
+ ":": 27,
57
+ ":</w>": 353,
58
+ ";": 28,
59
+ ";</w>": 317,
60
+ "<": 29,
61
+ "<</w>": 340,
62
+ "<|endoftext|>": 1,
63
+ "<|startoftext|>": 0,
64
+ "=": 30,
65
+ "=</w>": 342,
66
+ ">": 31,
67
+ "></w>": 300,
68
+ "?": 32,
69
+ "?</w>": 346,
70
+ "@": 33,
71
+ "@</w>": 320,
72
+ "A": 34,
73
+ "A</w>": 227,
74
+ "B": 35,
75
+ "B</w>": 258,
76
+ "C": 36,
77
+ "C</w>": 239,
78
+ "D": 37,
79
+ "D</w>": 255,
80
+ "E": 38,
81
+ "E</w>": 246,
82
+ "F": 39,
83
+ "F</w>": 213,
84
+ "G": 40,
85
+ "G</w>": 283,
86
+ "H": 41,
87
+ "H</w>": 219,
88
+ "I": 42,
89
+ "I</w>": 237,
90
+ "J": 43,
91
+ "J</w>": 251,
92
+ "K": 44,
93
+ "K</w>": 254,
94
+ "L": 45,
95
+ "L</w>": 218,
96
+ "M": 46,
97
+ "M</w>": 234,
98
+ "N": 47,
99
+ "N</w>": 238,
100
+ "O": 48,
101
+ "O</w>": 265,
102
+ "P": 49,
103
+ "P</w>": 245,
104
+ "Q": 50,
105
+ "Q</w>": 309,
106
+ "R": 51,
107
+ "R</w>": 264,
108
+ "S": 52,
109
+ "S</w>": 230,
110
+ "T": 53,
111
+ "T</w>": 235,
112
+ "U": 54,
113
+ "U</w>": 268,
114
+ "V": 55,
115
+ "V</w>": 248,
116
+ "W": 56,
117
+ "W</w>": 274,
118
+ "X": 57,
119
+ "X</w>": 263,
120
+ "Y": 58,
121
+ "Y</w>": 310,
122
+ "Z": 59,
123
+ "Z</w>": 207,
124
+ "[": 60,
125
+ "[</w>": 270,
126
+ "\\": 61,
127
+ "\\</w>": 338,
128
+ "]": 62,
129
+ "]</w>": 289,
130
+ "^": 63,
131
+ "^</w>": 331,
132
+ "_": 64,
133
+ "_</w>": 334,
134
+ "`": 65,
135
+ "`</w>": 347,
136
+ "a": 66,
137
+ "a</w>": 197,
138
+ "ab": 555,
139
+ "able</w>": 820,
140
+ "ac": 420,
141
+ "ace</w>": 806,
142
+ "ach": 791,
143
+ "ach</w>": 885,
144
+ "ack</w>": 670,
145
+ "act": 929,
146
+ "ad": 508,
147
+ "ad</w>": 860,
148
+ "ade</w>": 771,
149
+ "ag": 511,
150
+ "age</w>": 710,
151
+ "ain": 568,
152
+ "ain</w>": 675,
153
+ "ak": 577,
154
+ "ake</w>": 882,
155
+ "al": 397,
156
+ "al</w>": 405,
157
+ "all": 664,
158
+ "all</w>": 658,
159
+ "ally</w>": 588,
160
+ "als</w>": 796,
161
+ "am": 426,
162
+ "am</w>": 817,
163
+ "ame</w>": 552,
164
+ "ames</w>": 976,
165
+ "amp": 800,
166
+ "an": 384,
167
+ "an</w>": 425,
168
+ "ance</w>": 751,
169
+ "and": 609,
170
+ "and</w>": 780,
171
+ "ang": 816,
172
+ "ans</w>": 844,
173
+ "ant": 837,
174
+ "ant</w>": 753,
175
+ "any</w>": 766,
176
+ "ap": 586,
177
+ "ar": 376,
178
+ "ar</w>": 579,
179
+ "ard</w>": 649,
180
+ "ards</w>": 982,
181
+ "ary</w>": 611,
182
+ "as": 416,
183
+ "as</w>": 404,
184
+ "ase</w>": 849,
185
+ "ased</w>": 814,
186
+ "ason</w>": 865,
187
+ "ass": 792,
188
+ "ast</w>": 661,
189
+ "at": 372,
190
+ "at</w>": 434,
191
+ "ate</w>": 541,
192
+ "ated</w>": 543,
193
+ "ater</w>": 709,
194
+ "ates</w>": 825,
195
+ "ath": 730,
196
+ "ating</w>": 922,
197
+ "ation</w>": 497,
198
+ "ational</w>": 933,
199
+ "ations</w>": 744,
200
+ "att": 903,
201
+ "aus": 858,
202
+ "av": 681,
203
+ "ay": 684,
204
+ "ay</w>": 523,
205
+ "b": 67,
206
+ "b</w>": 212,
207
+ "ber</w>": 593,
208
+ "c": 68,
209
+ "c</w>": 224,
210
+ "cc": 960,
211
+ "ce</w>": 496,
212
+ "ces</w>": 830,
213
+ "ch": 520,
214
+ "ch</w>": 603,
215
+ "ct": 834,
216
+ "d": 69,
217
+ "d</w>": 196,
218
+ "ded</w>": 665,
219
+ "der</w>": 690,
220
+ "ding</w>": 633,
221
+ "ds</w>": 530,
222
+ "duc": 671,
223
+ "e": 70,
224
+ "e</w>": 195,
225
+ "ea": 471,
226
+ "ear": 596,
227
+ "ear</w>": 669,
228
+ "ears</w>": 906,
229
+ "eb": 852,
230
+ "ec": 418,
231
+ "ect": 838,
232
+ "ect</w>": 964,
233
+ "ed": 563,
234
+ "ed</w>": 362,
235
+ "ee": 941,
236
+ "een</w>": 779,
237
+ "ef": 840,
238
+ "eg": 731,
239
+ "el": 407,
240
+ "el</w>": 610,
241
+ "eld</w>": 973,
242
+ "ell": 759,
243
+ "ell</w>": 756,
244
+ "ely</w>": 719,
245
+ "em": 455,
246
+ "ember</w>": 777,
247
+ "ement</w>": 959,
248
+ "emp": 975,
249
+ "en": 375,
250
+ "en</w>": 427,
251
+ "ence</w>": 685,
252
+ "ens": 880,
253
+ "ent": 478,
254
+ "ent</w>": 468,
255
+ "ents</w>": 674,
256
+ "ep": 545,
257
+ "er": 364,
258
+ "er</w>": 374,
259
+ "eral</w>": 793,
260
+ "ere</w>": 481,
261
+ "ered</w>": 748,
262
+ "eric": 921,
263
+ "erm": 861,
264
+ "ern": 887,
265
+ "ern</w>": 977,
266
+ "ers": 598,
267
+ "ers</w>": 486,
268
+ "ert": 986,
269
+ "ery</w>": 805,
270
+ "es": 402,
271
+ "es</w>": 388,
272
+ "ese</w>": 794,
273
+ "ess": 678,
274
+ "ess</w>": 693,
275
+ "est": 606,
276
+ "est</w>": 584,
277
+ "et": 460,
278
+ "et</w>": 594,
279
+ "etw": 824,
280
+ "etween</w>": 886,
281
+ "ev": 493,
282
+ "evel": 980,
283
+ "ever</w>": 855,
284
+ "ew": 687,
285
+ "ew</w>": 612,
286
+ "ex": 938,
287
+ "ey</w>": 713,
288
+ "f": 71,
289
+ "f</w>": 209,
290
+ "fer": 911,
291
+ "ff": 587,
292
+ "for": 728,
293
+ "form": 901,
294
+ "fter</w>": 634,
295
+ "g": 72,
296
+ "g</w>": 214,
297
+ "ge</w>": 592,
298
+ "h": 73,
299
+ "h</w>": 203,
300
+ "i": 74,
301
+ "i</w>": 205,
302
+ "ia</w>": 605,
303
+ "ial</w>": 672,
304
+ "ian</w>": 638,
305
+ "ib": 726,
306
+ "ic": 395,
307
+ "ic</w>": 510,
308
+ "ical</w>": 625,
309
+ "ice</w>": 782,
310
+ "ich</w>": 561,
311
+ "ics</w>": 996,
312
+ "id": 463,
313
+ "id</w>": 613,
314
+ "ide</w>": 739,
315
+ "ie</w>": 974,
316
+ "ied</w>": 812,
317
+ "ies</w>": 516,
318
+ "if": 524,
319
+ "ig": 444,
320
+ "igh": 537,
321
+ "ight</w>": 680,
322
+ "ik": 775,
323
+ "ike</w>": 984,
324
+ "il": 406,
325
+ "il</w>": 714,
326
+ "ile</w>": 721,
327
+ "ill": 608,
328
+ "ill</w>": 789,
329
+ "ily</w>": 950,
330
+ "im": 469,
331
+ "im</w>": 767,
332
+ "ime</w>": 691,
333
+ "in": 358,
334
+ "in</w>": 501,
335
+ "ine</w>": 607,
336
+ "ing": 557,
337
+ "ing</w>": 383,
338
+ "ings</w>": 815,
339
+ "ion": 472,
340
+ "ion</w>": 408,
341
+ "ional</w>": 717,
342
+ "ions</w>": 540,
343
+ "ip": 733,
344
+ "ip</w>": 818,
345
+ "ir": 453,
346
+ "ir</w>": 554,
347
+ "is": 393,
348
+ "is</w>": 441,
349
+ "ish": 694,
350
+ "ish</w>": 654,
351
+ "ished</w>": 942,
352
+ "ision</w>": 944,
353
+ "iss": 876,
354
+ "ist": 550,
355
+ "ist</w>": 811,
356
+ "it": 378,
357
+ "it</w>": 746,
358
+ "ite</w>": 760,
359
+ "ited</w>": 809,
360
+ "ition</w>": 797,
361
+ "ity</w>": 542,
362
+ "iv": 435,
363
+ "ive</w>": 549,
364
+ "ived</w>": 979,
365
+ "iz": 722,
366
+ "j": 75,
367
+ "j</w>": 288,
368
+ "k": 76,
369
+ "k</w>": 210,
370
+ "ked</w>": 810,
371
+ "king</w>": 924,
372
+ "ks</w>": 692,
373
+ "l": 77,
374
+ "l</w>": 201,
375
+ "la": 467,
376
+ "land</w>": 743,
377
+ "ld</w>": 559,
378
+ "le": 536,
379
+ "le</w>": 465,
380
+ "les</w>": 799,
381
+ "lud": 718,
382
+ "ly</w>": 433,
383
+ "m": 78,
384
+ "m</w>": 202,
385
+ "ment</w>": 701,
386
+ "mp": 651,
387
+ "n": 79,
388
+ "n</w>": 199,
389
+ "nd</w>": 369,
390
+ "ned</w>": 758,
391
+ "ning</w>": 843,
392
+ "o": 80,
393
+ "o</w>": 198,
394
+ "ob": 920,
395
+ "oc": 534,
396
+ "od": 575,
397
+ "og": 604,
398
+ "oh": 972,
399
+ "oin": 831,
400
+ "ol": 428,
401
+ "oll": 703,
402
+ "ollow": 928,
403
+ "olog": 932,
404
+ "om": 419,
405
+ "om</w>": 883,
406
+ "ome</w>": 663,
407
+ "on": 382,
408
+ "on</w>": 390,
409
+ "ond</w>": 872,
410
+ "one</w>": 835,
411
+ "ong": 850,
412
+ "ong</w>": 582,
413
+ "oo": 517,
414
+ "ood</w>": 927,
415
+ "ook</w>": 897,
416
+ "op": 531,
417
+ "op</w>": 971,
418
+ "or": 377,
419
+ "or</w>": 424,
420
+ "ore</w>": 571,
421
+ "ors</w>": 917,
422
+ "ort": 768,
423
+ "ort</w>": 752,
424
+ "ory</w>": 737,
425
+ "os": 447,
426
+ "ose</w>": 881,
427
+ "ost</w>": 646,
428
+ "ot": 600,
429
+ "ot</w>": 879,
430
+ "ou": 392,
431
+ "oug": 659,
432
+ "ough</w>": 798,
433
+ "ould</w>": 640,
434
+ "oun": 553,
435
+ "ound</w>": 961,
436
+ "our": 648,
437
+ "our</w>": 772,
438
+ "ous</w>": 712,
439
+ "out</w>": 683,
440
+ "outh</w>": 945,
441
+ "ov": 515,
442
+ "ow": 461,
443
+ "ow</w>": 666,
444
+ "own</w>": 657,
445
+ "oy": 952,
446
+ "p": 81,
447
+ "p</w>": 217,
448
+ "per": 715,
449
+ "ph": 916,
450
+ "pp": 518,
451
+ "q": 82,
452
+ "q</w>": 280,
453
+ "qu": 546,
454
+ "r": 83,
455
+ "r</w>": 204,
456
+ "ra": 457,
457
+ "ran": 624,
458
+ "re": 367,
459
+ "ree</w>": 765,
460
+ "ren": 790,
461
+ "res": 572,
462
+ "res</w>": 747,
463
+ "ri": 487,
464
+ "rib": 804,
465
+ "ric": 745,
466
+ "rit": 589,
467
+ "ro": 385,
468
+ "rom</w>": 498,
469
+ "rop": 826,
470
+ "roug": 803,
471
+ "ru": 951,
472
+ "ruc": 891,
473
+ "ry</w>": 908,
474
+ "s": 84,
475
+ "s</w>": 206,
476
+ "se</w>": 741,
477
+ "sh": 795,
478
+ "so</w>": 630,
479
+ "sp": 992,
480
+ "ss": 673,
481
+ "st": 519,
482
+ "st</w>": 528,
483
+ "t": 85,
484
+ "t</w>": 208,
485
+ "te</w>": 954,
486
+ "ted</w>": 489,
487
+ "ter": 535,
488
+ "ter</w>": 505,
489
+ "th": 449,
490
+ "th</w>": 488,
491
+ "ther</w>": 576,
492
+ "ting</w>": 676,
493
+ "tion</w>": 570,
494
+ "tr": 619,
495
+ "ts</w>": 436,
496
+ "tt": 720,
497
+ "tur": 953,
498
+ "ty</w>": 821,
499
+ "u": 86,
500
+ "u</w>": 229,
501
+ "ub": 591,
502
+ "ubl": 842,
503
+ "uc": 490,
504
+ "ud": 538,
505
+ "ue</w>": 652,
506
+ "ug": 560,
507
+ "uil": 930,
508
+ "ul": 494,
509
+ "um": 532,
510
+ "um</w>": 644,
511
+ "un": 448,
512
+ "und</w>": 828,
513
+ "up": 833,
514
+ "up</w>": 700,
515
+ "ur": 413,
516
+ "ure</w>": 635,
517
+ "uring</w>": 702,
518
+ "ury</w>": 957,
519
+ "us": 438,
520
+ "us</w>": 622,
521
+ "ust</w>": 846,
522
+ "ut": 529,
523
+ "ut</w>": 527,
524
+ "v": 87,
525
+ "v</w>": 232,
526
+ "ve</w>": 567,
527
+ "vi": 866,
528
+ "w": 88,
529
+ "w</w>": 250,
530
+ "way</w>": 970,
531
+ "wn</w>": 999,
532
+ "x": 89,
533
+ "x</w>": 269,
534
+ "y": 90,
535
+ "y</w>": 211,
536
+ "yp": 993,
537
+ "z": 91,
538
+ "z</w>": 228,
539
+ "|": 92,
540
+ "|</w>": 304,
541
+ "}": 93,
542
+ "}</w>": 336,
543
+ "~": 94,
544
+ "~</w>": 343,
545
+ "¡": 95,
546
+ "¡</w>": 220,
547
+ "¢": 96,
548
+ "¢</w>": 306,
549
+ "£": 97,
550
+ "£</w>": 323,
551
+ "¤": 98,
552
+ "¤</w>": 292,
553
+ "¥": 99,
554
+ "¥</w>": 339,
555
+ "¦": 100,
556
+ "¦</w>": 303,
557
+ "§": 101,
558
+ "§</w>": 275,
559
+ "¨": 102,
560
+ "¨</w>": 282,
561
+ "©": 103,
562
+ "©</w>": 259,
563
+ "ª": 104,
564
+ "ª</w>": 286,
565
+ "«": 105,
566
+ "«</w>": 266,
567
+ "¬": 106,
568
+ "¬</w>": 319,
569
+ "®": 107,
570
+ "®</w>": 329,
571
+ "¯": 108,
572
+ "¯</w>": 287,
573
+ "°": 109,
574
+ "°</w>": 298,
575
+ "±": 110,
576
+ "±</w>": 200,
577
+ "²": 111,
578
+ "²</w>": 284,
579
+ "³": 112,
580
+ "³</w>": 272,
581
+ "´": 113,
582
+ "´</w>": 307,
583
+ "µ": 114,
584
+ "µ</w>": 261,
585
+ "¶": 115,
586
+ "¶</w>": 301,
587
+ "·": 116,
588
+ "·</w>": 326,
589
+ "¸": 117,
590
+ "¸</w>": 257,
591
+ "¹": 118,
592
+ "¹</w>": 241,
593
+ "º": 119,
594
+ "º</w>": 260,
595
+ "»": 120,
596
+ "»</w>": 247,
597
+ "¼": 121,
598
+ "¼</w>": 305,
599
+ "½": 122,
600
+ "½</w>": 294,
601
+ "¾": 123,
602
+ "¾</w>": 316,
603
+ "¿": 124,
604
+ "¿</w>": 271,
605
+ "Â": 125,
606
+ "Ã": 126,
607
+ "Ä": 127,
608
+ "Å": 128,
609
+ "Æ": 129,
610
+ "Ç": 130,
611
+ "È": 131,
612
+ "É": 132,
613
+ "Ê": 133,
614
+ "Ë": 134,
615
+ "Ì": 135,
616
+ "Í": 136,
617
+ "Î": 137,
618
+ "Ï": 138,
619
+ "Ð": 139,
620
+ "Ñ": 140,
621
+ "Ö": 141,
622
+ "×": 142,
623
+ "Ø": 143,
624
+ "Ù": 144,
625
+ "Ü": 145,
626
+ "à": 146,
627
+ "á": 147,
628
+ "â": 148,
629
+ "ã": 149,
630
+ "ä": 150,
631
+ "å": 151,
632
+ "æ": 152,
633
+ "ç": 153,
634
+ "è": 154,
635
+ "é": 155,
636
+ "ë": 156,
637
+ "ì": 157,
638
+ "ï": 158,
639
+ "Ċ": 159,
640
+ "Ċ</w>": 349,
641
+ "Ġ": 160,
642
+ "Ġ\"</w>": 401,
643
+ "Ġ'</w>": 431,
644
+ "Ġ(</w>": 475,
645
+ "Ġ)</w>": 474,
646
+ "Ġ,</w>": 360,
647
+ "Ġ.</w>": 365,
648
+ "Ġ0": 847,
649
+ "Ġ1": 411,
650
+ "Ġ18": 769,
651
+ "Ġ19": 492,
652
+ "Ġ199": 893,
653
+ "Ġ1</w>": 778,
654
+ "Ġ2": 462,
655
+ "Ġ20": 522,
656
+ "Ġ200": 620,
657
+ "Ġ201": 734,
658
+ "Ġ2</w>": 813,
659
+ "Ġ3": 735,
660
+ "Ġ3</w>": 888,
661
+ "Ġ4": 870,
662
+ "Ġ5": 907,
663
+ "Ġ5</w>": 990,
664
+ "Ġ:</w>": 637,
665
+ "Ġ;</w>": 615,
666
+ "Ġ</w>": 333,
667
+ "Ġ=</w>": 399,
668
+ "Ġ@": 417,
669
+ "Ġ@,@</w>": 755,
670
+ "Ġ@-@</w>": 440,
671
+ "Ġ@.@</w>": 696,
672
+ "ĠA": 409,
673
+ "ĠA</w>": 807,
674
+ "ĠAl": 716,
675
+ "ĠAm": 829,
676
+ "ĠAmeric": 958,
677
+ "ĠAn": 784,
678
+ "ĠAr": 894,
679
+ "ĠB": 432,
680
+ "ĠC": 410,
681
+ "ĠCh": 581,
682
+ "ĠCom": 904,
683
+ "ĠD": 464,
684
+ "ĠE": 500,
685
+ "ĠEn": 878,
686
+ "ĠF": 470,
687
+ "ĠG": 482,
688
+ "ĠH": 445,
689
+ "ĠHe</w>": 742,
690
+ "ĠI": 442,
691
+ "ĠI</w>": 827,
692
+ "ĠIn": 704,
693
+ "ĠIn</w>": 574,
694
+ "ĠIt</w>": 774,
695
+ "ĠJ": 491,
696
+ "ĠK": 548,
697
+ "ĠL": 484,
698
+ "ĠM": 423,
699
+ "ĠMar": 776,
700
+ "ĠN": 483,
701
+ "ĠO": 504,
702
+ "ĠP": 450,
703
+ "ĠPar": 967,
704
+ "ĠR": 459,
705
+ "ĠS": 403,
706
+ "ĠSh": 750,
707
+ "ĠSt": 590,
708
+ "ĠT": 396,
709
+ "ĠTh": 414,
710
+ "ĠThe</w>": 437,
711
+ "ĠThis</w>": 997,
712
+ "ĠU": 585,
713
+ "ĠUn": 773,
714
+ "ĠV": 617,
715
+ "ĠW": 479,
716
+ "ĠWh": 853,
717
+ "ĠY": 757,
718
+ "Ġa": 356,
719
+ "Ġa</w>": 394,
720
+ "Ġab": 653,
721
+ "Ġabout</w>": 899,
722
+ "Ġac": 583,
723
+ "Ġacc": 874,
724
+ "Ġad": 656,
725
+ "Ġafter</w>": 763,
726
+ "Ġag": 725,
727
+ "Ġal": 476,
728
+ "Ġalb": 991,
729
+ "Ġall</w>": 839,
730
+ "Ġalso</w>": 641,
731
+ "Ġan": 602,
732
+ "Ġan</w>": 562,
733
+ "Ġand</w>": 381,
734
+ "Ġapp": 711,
735
+ "Ġar": 507,
736
+ "Ġare</w>": 601,
737
+ "Ġas</w>": 454,
738
+ "Ġass": 947,
739
+ "Ġat</w>": 514,
740
+ "Ġatt": 788,
741
+ "Ġb": 371,
742
+ "Ġbe": 499,
743
+ "Ġbe</w>": 595,
744
+ "Ġbec": 706,
745
+ "Ġbeen</w>": 686,
746
+ "Ġbeg": 915,
747
+ "Ġbetween</w>": 914,
748
+ "Ġbo": 819,
749
+ "Ġbut</w>": 623,
750
+ "Ġby</w>": 473,
751
+ "Ġc": 368,
752
+ "Ġcent": 823,
753
+ "Ġch": 526,
754
+ "Ġchar": 822,
755
+ "Ġcl": 689,
756
+ "Ġcom": 509,
757
+ "Ġcomm": 707,
758
+ "Ġcomp": 616,
759
+ "Ġcon": 477,
760
+ "Ġcons": 841,
761
+ "Ġcont": 655,
762
+ "Ġcre": 931,
763
+ "Ġd": 387,
764
+ "Ġde": 627,
765
+ "Ġdec": 873,
766
+ "Ġdef": 965,
767
+ "Ġdes": 738,
768
+ "Ġdi": 892,
769
+ "Ġdis": 708,
770
+ "Ġduring</w>": 864,
771
+ "Ġe": 421,
772
+ "Ġear": 854,
773
+ "Ġel": 869,
774
+ "Ġen": 662,
775
+ "Ġev": 682,
776
+ "Ġex": 539,
777
+ "Ġexp": 867,
778
+ "Ġf": 370,
779
+ "Ġfe": 845,
780
+ "Ġfil": 913,
781
+ "Ġfin": 786,
782
+ "Ġfir": 599,
783
+ "Ġfirst</w>": 626,
784
+ "Ġfl": 877,
785
+ "Ġfor": 614,
786
+ "Ġfor</w>": 458,
787
+ "Ġform": 963,
788
+ "Ġfrom</w>": 503,
789
+ "Ġg": 430,
790
+ "Ġgame</w>": 895,
791
+ "Ġgen": 985,
792
+ "Ġgro": 890,
793
+ "Ġh": 380,
794
+ "Ġha": 485,
795
+ "Ġhad</w>": 566,
796
+ "Ġhas</w>": 679,
797
+ "Ġhave</w>": 667,
798
+ "Ġhe</w>": 558,
799
+ "Ġher</w>": 660,
800
+ "Ġhim</w>": 896,
801
+ "Ġhis</w>": 512,
802
+ "Ġi": 366,
803
+ "Ġimp": 909,
804
+ "Ġin": 429,
805
+ "Ġin</w>": 389,
806
+ "Ġinc": 618,
807
+ "Ġinclud": 761,
808
+ "Ġind": 983,
809
+ "Ġint": 628,
810
+ "Ġinter": 832,
811
+ "Ġinto</w>": 785,
812
+ "Ġis</w>": 480,
813
+ "Ġit</w>": 533,
814
+ "Ġits</w>": 642,
815
+ "Ġj": 723,
816
+ "Ġk": 564,
817
+ "Ġkn": 857,
818
+ "Ġl": 398,
819
+ "Ġlar": 962,
820
+ "Ġlater</w>": 936,
821
+ "Ġlea": 868,
822
+ "Ġm": 386,
823
+ "Ġmade</w>": 949,
824
+ "Ġme": 968,
825
+ "Ġmore</w>": 802,
826
+ "Ġmost</w>": 910,
827
+ "Ġmov": 956,
828
+ "Ġmus": 966,
829
+ "Ġn": 415,
830
+ "Ġnew</w>": 989,
831
+ "Ġno": 547,
832
+ "Ġnor": 978,
833
+ "Ġnot</w>": 632,
834
+ "Ġnum": 926,
835
+ "Ġo": 359,
836
+ "Ġof</w>": 373,
837
+ "Ġoff": 875,
838
+ "Ġon": 551,
839
+ "Ġon</w>": 456,
840
+ "Ġone</w>": 677,
841
+ "Ġonly</w>": 871,
842
+ "Ġor": 699,
843
+ "Ġor</w>": 645,
844
+ "Ġother</w>": 787,
845
+ "Ġout</w>": 925,
846
+ "Ġov": 729,
847
+ "Ġover</w>": 856,
848
+ "Ġp": 379,
849
+ "Ġpar": 636,
850
+ "Ġper": 736,
851
+ "Ġpl": 698,
852
+ "Ġpla": 697,
853
+ "Ġplay": 808,
854
+ "Ġpos": 859,
855
+ "Ġpr": 889,
856
+ "Ġpre": 749,
857
+ "Ġpres": 912,
858
+ "Ġpro": 506,
859
+ "Ġproduc": 934,
860
+ "Ġqu": 955,
861
+ "Ġr": 521,
862
+ "Ġra": 863,
863
+ "Ġre": 400,
864
+ "Ġrec": 597,
865
+ "Ġrecor": 919,
866
+ "Ġreg": 943,
867
+ "Ġrel": 900,
868
+ "Ġrele": 946,
869
+ "Ġrem": 848,
870
+ "Ġrep": 762,
871
+ "Ġres": 650,
872
+ "Ġro": 629,
873
+ "Ġs": 361,
874
+ "Ġsa": 905,
875
+ "Ġsc": 732,
876
+ "Ġse": 569,
877
+ "Ġseason</w>": 948,
878
+ "Ġsec": 781,
879
+ "Ġser": 740,
880
+ "Ġsev": 884,
881
+ "Ġsh": 513,
882
+ "Ġshe</w>": 862,
883
+ "Ġsp": 578,
884
+ "Ġspec": 940,
885
+ "Ġst": 446,
886
+ "Ġstar": 939,
887
+ "Ġsu": 770,
888
+ "Ġsub": 969,
889
+ "Ġsuc": 764,
890
+ "Ġsuch</w>": 981,
891
+ "Ġt": 354,
892
+ "Ġth": 355,
893
+ "Ġthan</w>": 918,
894
+ "Ġthat</w>": 452,
895
+ "Ġthe": 502,
896
+ "Ġthe</w>": 357,
897
+ "Ġtheir</w>": 621,
898
+ "Ġthem</w>": 998,
899
+ "Ġthey</w>": 727,
900
+ "Ġthis</w>": 705,
901
+ "Ġthree</w>": 902,
902
+ "Ġthroug": 923,
903
+ "Ġtime</w>": 783,
904
+ "Ġto</w>": 391,
905
+ "Ġtra": 836,
906
+ "Ġtw": 639,
907
+ "Ġtwo</w>": 688,
908
+ "Ġun": 544,
909
+ "Ġup</w>": 898,
910
+ "Ġus": 668,
911
+ "Ġused</w>": 988,
912
+ "Ġv": 495,
913
+ "Ġw": 363,
914
+ "Ġwas</w>": 422,
915
+ "Ġwere</w>": 525,
916
+ "Ġwh": 443,
917
+ "Ġwhen</w>": 851,
918
+ "Ġwhere</w>": 995,
919
+ "Ġwhich</w>": 573,
920
+ "Ġwhile</w>": 935,
921
+ "Ġwho</w>": 724,
922
+ "Ġwit": 451,
923
+ "Ġwith": 994,
924
+ "Ġwith</w>": 466,
925
+ "Ġwor": 643,
926
+ "Ġwould</w>": 801,
927
+ "Ġwrit": 937,
928
+ "Ġy": 580,
929
+ "Ġyear</w>": 987,
930
+ "Ġâ": 556,
931
+ "ĠâĢ": 565,
932
+ "ĠâĢĵ</w>": 631,
933
+ "ĠĊ</w>": 412,
934
+ "Ģ": 161,
935
+ "Ģ</w>": 223,
936
+ "ģ": 162,
937
+ "ģ</w>": 273,
938
+ "Ĥ": 163,
939
+ "Ĥ</w>": 262,
940
+ "ĥ": 164,
941
+ "ĥ</w>": 337,
942
+ "Ħ": 165,
943
+ "Ħ</w>": 278,
944
+ "ħ": 166,
945
+ "ħ</w>": 281,
946
+ "Ĩ": 167,
947
+ "Ĩ</w>": 308,
948
+ "ĩ": 168,
949
+ "ĩ</w>": 225,
950
+ "Ī": 169,
951
+ "Ī</w>": 221,
952
+ "ī": 170,
953
+ "ī</w>": 244,
954
+ "Ĭ": 171,
955
+ "Ĭ</w>": 315,
956
+ "ĭ": 172,
957
+ "ĭ</w>": 321,
958
+ "Į": 173,
959
+ "Į</w>": 324,
960
+ "į": 174,
961
+ "į</w>": 302,
962
+ "İ": 175,
963
+ "İ</w>": 249,
964
+ "ı": 176,
965
+ "ı</w>": 332,
966
+ "IJ": 177,
967
+ "IJ</w>": 295,
968
+ "ij": 178,
969
+ "ij</w>": 313,
970
+ "Ĵ": 179,
971
+ "Ĵ</w>": 328,
972
+ "ĵ": 180,
973
+ "ĵ</w>": 312,
974
+ "Ķ": 181,
975
+ "Ķ</w>": 256,
976
+ "ķ": 182,
977
+ "ķ</w>": 314,
978
+ "ĸ": 183,
979
+ "ĸ</w>": 277,
980
+ "Ĺ": 184,
981
+ "Ĺ</w>": 322,
982
+ "ĺ": 185,
983
+ "ĺ</w>": 285,
984
+ "Ļ": 186,
985
+ "Ļ</w>": 267,
986
+ "ļ": 187,
987
+ "ļ</w>": 290,
988
+ "Ľ": 188,
989
+ "Ľ</w>": 311,
990
+ "ľ": 189,
991
+ "ľ</w>": 299,
992
+ "Ŀ": 190,
993
+ "Ŀ</w>": 291,
994
+ "ŀ": 191,
995
+ "ŀ</w>": 293,
996
+ "Ł": 192,
997
+ "Ł</w>": 335,
998
+ "ł": 193,
999
+ "ł</w>": 252,
1000
+ "Ń": 194,
1001
+ "Ń</w>": 297
1002
+ }
unet/config.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.32.2",
4
+ "_name_or_path": "/home/ea/.cache/huggingface/hub/models--hf-internal-testing--tiny-stable-diffusion-torch/snapshots/a88cdfbd91f96ec7f61eb7484b652ff0f4ee701d/unet",
5
+ "act_fn": "silu",
6
+ "addition_embed_type": null,
7
+ "addition_embed_type_num_heads": 64,
8
+ "addition_time_embed_dim": null,
9
+ "attention_head_dim": 8,
10
+ "attention_type": "default",
11
+ "block_out_channels": [
12
+ 32,
13
+ 64
14
+ ],
15
+ "center_input_sample": false,
16
+ "class_embed_type": null,
17
+ "class_embeddings_concat": false,
18
+ "conv_in_kernel": 3,
19
+ "conv_out_kernel": 3,
20
+ "cross_attention_dim": 32,
21
+ "cross_attention_norm": null,
22
+ "down_block_types": [
23
+ "DownBlock2D",
24
+ "CrossAttnDownBlock2D"
25
+ ],
26
+ "downsample_padding": 1,
27
+ "dropout": 0.0,
28
+ "dual_cross_attention": false,
29
+ "encoder_hid_dim": null,
30
+ "encoder_hid_dim_type": null,
31
+ "flip_sin_to_cos": true,
32
+ "freq_shift": 0,
33
+ "in_channels": 4,
34
+ "layers_per_block": 2,
35
+ "mid_block_only_cross_attention": null,
36
+ "mid_block_scale_factor": 1,
37
+ "mid_block_type": "UNetMidBlock2DCrossAttn",
38
+ "norm_eps": 1e-05,
39
+ "norm_num_groups": 32,
40
+ "num_attention_heads": null,
41
+ "num_class_embeds": null,
42
+ "only_cross_attention": false,
43
+ "out_channels": 4,
44
+ "projection_class_embeddings_input_dim": null,
45
+ "resnet_out_scale_factor": 1.0,
46
+ "resnet_skip_time_act": false,
47
+ "resnet_time_scale_shift": "default",
48
+ "reverse_transformer_layers_per_block": null,
49
+ "sample_size": 64,
50
+ "time_cond_proj_dim": null,
51
+ "time_embedding_act_fn": null,
52
+ "time_embedding_dim": null,
53
+ "time_embedding_type": "positional",
54
+ "timestep_post_act": null,
55
+ "transformer_layers_per_block": 1,
56
+ "up_block_types": [
57
+ "CrossAttnUpBlock2D",
58
+ "UpBlock2D"
59
+ ],
60
+ "upcast_attention": false,
61
+ "use_linear_projection": false
62
+ }
unet/openvino_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9182d2e5c289a9b1ac127a2061c02421ff346f225da141174d671672b55c6e4
3
+ size 5686636
unet/openvino_model.xml ADDED
The diff for this file is too large to render. See raw diff
 
vae_decoder/config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.32.2",
4
+ "_name_or_path": "/home/ea/.cache/huggingface/hub/models--hf-internal-testing--tiny-stable-diffusion-torch/snapshots/a88cdfbd91f96ec7f61eb7484b652ff0f4ee701d/vae",
5
+ "act_fn": "silu",
6
+ "block_out_channels": [
7
+ 32,
8
+ 64
9
+ ],
10
+ "down_block_types": [
11
+ "DownEncoderBlock2D",
12
+ "DownEncoderBlock2D"
13
+ ],
14
+ "force_upcast": true,
15
+ "in_channels": 3,
16
+ "latent_channels": 4,
17
+ "latents_mean": null,
18
+ "latents_std": null,
19
+ "layers_per_block": 1,
20
+ "mid_block_add_attention": true,
21
+ "norm_num_groups": 32,
22
+ "out_channels": 3,
23
+ "sample_size": 128,
24
+ "scaling_factor": 0.18215,
25
+ "shift_factor": null,
26
+ "up_block_types": [
27
+ "UpDecoderBlock2D",
28
+ "UpDecoderBlock2D"
29
+ ],
30
+ "use_post_quant_conv": true,
31
+ "use_quant_conv": true
32
+ }
vae_decoder/openvino_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fdeb182718138bb26eb2477c1529e4cba699049324ae3ce3be03c9e876a710d3
3
+ size 1603080
vae_decoder/openvino_model.xml ADDED
@@ -0,0 +1,3137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <?xml version="1.0"?>
2
+ <net name="Model9" version="11">
3
+ <layers>
4
+ <layer id="0" name="latent_sample" type="Parameter" version="opset1">
5
+ <data shape="?,4,?,?" element_type="f32" />
6
+ <output>
7
+ <port id="0" precision="FP32" names="latent_sample">
8
+ <dim>-1</dim>
9
+ <dim>4</dim>
10
+ <dim>-1</dim>
11
+ <dim>-1</dim>
12
+ </port>
13
+ </output>
14
+ </layer>
15
+ <layer id="1" name="self.post_quant_conv.weight" type="Const" version="opset1">
16
+ <data element_type="f32" shape="4, 4, 1, 1" offset="0" size="64" />
17
+ <output>
18
+ <port id="0" precision="FP32" names="self.post_quant_conv.weight">
19
+ <dim>4</dim>
20
+ <dim>4</dim>
21
+ <dim>1</dim>
22
+ <dim>1</dim>
23
+ </port>
24
+ </output>
25
+ </layer>
26
+ <layer id="2" name="__module.post_quant_conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
27
+ <data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
28
+ <input>
29
+ <port id="0" precision="FP32">
30
+ <dim>-1</dim>
31
+ <dim>4</dim>
32
+ <dim>-1</dim>
33
+ <dim>-1</dim>
34
+ </port>
35
+ <port id="1" precision="FP32">
36
+ <dim>4</dim>
37
+ <dim>4</dim>
38
+ <dim>1</dim>
39
+ <dim>1</dim>
40
+ </port>
41
+ </input>
42
+ <output>
43
+ <port id="2" precision="FP32">
44
+ <dim>-1</dim>
45
+ <dim>4</dim>
46
+ <dim>-1</dim>
47
+ <dim>-1</dim>
48
+ </port>
49
+ </output>
50
+ </layer>
51
+ <layer id="3" name="__module.post_quant_conv/aten::_convolution/Reshape" type="Const" version="opset1">
52
+ <data element_type="f32" shape="1, 4, 1, 1" offset="64" size="16" />
53
+ <output>
54
+ <port id="0" precision="FP32">
55
+ <dim>1</dim>
56
+ <dim>4</dim>
57
+ <dim>1</dim>
58
+ <dim>1</dim>
59
+ </port>
60
+ </output>
61
+ </layer>
62
+ <layer id="4" name="__module.post_quant_conv/aten::_convolution/Add" type="Add" version="opset1">
63
+ <data auto_broadcast="numpy" />
64
+ <input>
65
+ <port id="0" precision="FP32">
66
+ <dim>-1</dim>
67
+ <dim>4</dim>
68
+ <dim>-1</dim>
69
+ <dim>-1</dim>
70
+ </port>
71
+ <port id="1" precision="FP32">
72
+ <dim>1</dim>
73
+ <dim>4</dim>
74
+ <dim>1</dim>
75
+ <dim>1</dim>
76
+ </port>
77
+ </input>
78
+ <output>
79
+ <port id="2" precision="FP32" names="17">
80
+ <dim>-1</dim>
81
+ <dim>4</dim>
82
+ <dim>-1</dim>
83
+ <dim>-1</dim>
84
+ </port>
85
+ </output>
86
+ </layer>
87
+ <layer id="5" name="self.decoder.conv_in.weight" type="Const" version="opset1">
88
+ <data element_type="f32" shape="64, 4, 3, 3" offset="80" size="9216" />
89
+ <output>
90
+ <port id="0" precision="FP32" names="self.decoder.conv_in.weight">
91
+ <dim>64</dim>
92
+ <dim>4</dim>
93
+ <dim>3</dim>
94
+ <dim>3</dim>
95
+ </port>
96
+ </output>
97
+ </layer>
98
+ <layer id="6" name="__module.decoder.conv_in/aten::_convolution/Convolution" type="Convolution" version="opset1">
99
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
100
+ <input>
101
+ <port id="0" precision="FP32">
102
+ <dim>-1</dim>
103
+ <dim>4</dim>
104
+ <dim>-1</dim>
105
+ <dim>-1</dim>
106
+ </port>
107
+ <port id="1" precision="FP32">
108
+ <dim>64</dim>
109
+ <dim>4</dim>
110
+ <dim>3</dim>
111
+ <dim>3</dim>
112
+ </port>
113
+ </input>
114
+ <output>
115
+ <port id="2" precision="FP32">
116
+ <dim>-1</dim>
117
+ <dim>64</dim>
118
+ <dim>-1</dim>
119
+ <dim>-1</dim>
120
+ </port>
121
+ </output>
122
+ </layer>
123
+ <layer id="7" name="__module.decoder.conv_in/aten::_convolution/Reshape" type="Const" version="opset1">
124
+ <data element_type="f32" shape="1, 64, 1, 1" offset="9296" size="256" />
125
+ <output>
126
+ <port id="0" precision="FP32">
127
+ <dim>1</dim>
128
+ <dim>64</dim>
129
+ <dim>1</dim>
130
+ <dim>1</dim>
131
+ </port>
132
+ </output>
133
+ </layer>
134
+ <layer id="8" name="__module.decoder.conv_in/aten::_convolution/Add" type="Add" version="opset1">
135
+ <data auto_broadcast="numpy" />
136
+ <input>
137
+ <port id="0" precision="FP32">
138
+ <dim>-1</dim>
139
+ <dim>64</dim>
140
+ <dim>-1</dim>
141
+ <dim>-1</dim>
142
+ </port>
143
+ <port id="1" precision="FP32">
144
+ <dim>1</dim>
145
+ <dim>64</dim>
146
+ <dim>1</dim>
147
+ <dim>1</dim>
148
+ </port>
149
+ </input>
150
+ <output>
151
+ <port id="2" precision="FP32" names="53,input.1">
152
+ <dim>-1</dim>
153
+ <dim>64</dim>
154
+ <dim>-1</dim>
155
+ <dim>-1</dim>
156
+ </port>
157
+ </output>
158
+ </layer>
159
+ <layer id="9" name="self.decoder.mid_block.resnets.0.norm1.weight" type="Const" version="opset1">
160
+ <data element_type="f32" shape="64" offset="9552" size="256" />
161
+ <output>
162
+ <port id="0" precision="FP32" names="self.decoder.mid_block.resnets.0.norm1.weight">
163
+ <dim>64</dim>
164
+ </port>
165
+ </output>
166
+ </layer>
167
+ <layer id="10" name="self.decoder.mid_block.resnets.0.norm1.bias" type="Const" version="opset1">
168
+ <data element_type="f32" shape="64" offset="9808" size="256" />
169
+ <output>
170
+ <port id="0" precision="FP32" names="self.decoder.mid_block.resnets.0.norm1.bias">
171
+ <dim>64</dim>
172
+ </port>
173
+ </output>
174
+ </layer>
175
+ <layer id="11" name="__module.decoder.mid_block.resnets.0.norm1/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
176
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
177
+ <input>
178
+ <port id="0" precision="FP32">
179
+ <dim>-1</dim>
180
+ <dim>64</dim>
181
+ <dim>-1</dim>
182
+ <dim>-1</dim>
183
+ </port>
184
+ <port id="1" precision="FP32">
185
+ <dim>64</dim>
186
+ </port>
187
+ <port id="2" precision="FP32">
188
+ <dim>64</dim>
189
+ </port>
190
+ </input>
191
+ <output>
192
+ <port id="3" precision="FP32" names="70,input.3">
193
+ <dim>-1</dim>
194
+ <dim>64</dim>
195
+ <dim>-1</dim>
196
+ <dim>-1</dim>
197
+ </port>
198
+ </output>
199
+ </layer>
200
+ <layer id="12" name="__module.decoder.mid_block.resnets.1.nonlinearity/aten::silu/Swish" type="Swish" version="opset4">
201
+ <input>
202
+ <port id="0" precision="FP32">
203
+ <dim>-1</dim>
204
+ <dim>64</dim>
205
+ <dim>-1</dim>
206
+ <dim>-1</dim>
207
+ </port>
208
+ </input>
209
+ <output>
210
+ <port id="1" precision="FP32" names="71">
211
+ <dim>-1</dim>
212
+ <dim>64</dim>
213
+ <dim>-1</dim>
214
+ <dim>-1</dim>
215
+ </port>
216
+ </output>
217
+ </layer>
218
+ <layer id="13" name="self.decoder.mid_block.resnets.0.conv1.weight" type="Const" version="opset1">
219
+ <data element_type="f32" shape="64, 64, 3, 3" offset="10064" size="147456" />
220
+ <output>
221
+ <port id="0" precision="FP32" names="self.decoder.mid_block.resnets.0.conv1.weight">
222
+ <dim>64</dim>
223
+ <dim>64</dim>
224
+ <dim>3</dim>
225
+ <dim>3</dim>
226
+ </port>
227
+ </output>
228
+ </layer>
229
+ <layer id="14" name="__module.decoder.mid_block.resnets.0.conv1/aten::_convolution/Convolution" type="Convolution" version="opset1">
230
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
231
+ <input>
232
+ <port id="0" precision="FP32">
233
+ <dim>-1</dim>
234
+ <dim>64</dim>
235
+ <dim>-1</dim>
236
+ <dim>-1</dim>
237
+ </port>
238
+ <port id="1" precision="FP32">
239
+ <dim>64</dim>
240
+ <dim>64</dim>
241
+ <dim>3</dim>
242
+ <dim>3</dim>
243
+ </port>
244
+ </input>
245
+ <output>
246
+ <port id="2" precision="FP32">
247
+ <dim>-1</dim>
248
+ <dim>64</dim>
249
+ <dim>-1</dim>
250
+ <dim>-1</dim>
251
+ </port>
252
+ </output>
253
+ </layer>
254
+ <layer id="15" name="__module.decoder.mid_block.resnets.0.conv1/aten::_convolution/Reshape" type="Const" version="opset1">
255
+ <data element_type="f32" shape="1, 64, 1, 1" offset="157520" size="256" />
256
+ <output>
257
+ <port id="0" precision="FP32">
258
+ <dim>1</dim>
259
+ <dim>64</dim>
260
+ <dim>1</dim>
261
+ <dim>1</dim>
262
+ </port>
263
+ </output>
264
+ </layer>
265
+ <layer id="16" name="__module.decoder.mid_block.resnets.0.conv1/aten::_convolution/Add" type="Add" version="opset1">
266
+ <data auto_broadcast="numpy" />
267
+ <input>
268
+ <port id="0" precision="FP32">
269
+ <dim>-1</dim>
270
+ <dim>64</dim>
271
+ <dim>-1</dim>
272
+ <dim>-1</dim>
273
+ </port>
274
+ <port id="1" precision="FP32">
275
+ <dim>1</dim>
276
+ <dim>64</dim>
277
+ <dim>1</dim>
278
+ <dim>1</dim>
279
+ </port>
280
+ </input>
281
+ <output>
282
+ <port id="2" precision="FP32" names="78,input.5">
283
+ <dim>-1</dim>
284
+ <dim>64</dim>
285
+ <dim>-1</dim>
286
+ <dim>-1</dim>
287
+ </port>
288
+ </output>
289
+ </layer>
290
+ <layer id="17" name="self.decoder.mid_block.resnets.0.norm2.weight" type="Const" version="opset1">
291
+ <data element_type="f32" shape="64" offset="9552" size="256" />
292
+ <output>
293
+ <port id="0" precision="FP32" names="self.decoder.mid_block.resnets.0.norm2.weight">
294
+ <dim>64</dim>
295
+ </port>
296
+ </output>
297
+ </layer>
298
+ <layer id="18" name="self.decoder.mid_block.resnets.0.norm2.bias" type="Const" version="opset1">
299
+ <data element_type="f32" shape="64" offset="9808" size="256" />
300
+ <output>
301
+ <port id="0" precision="FP32" names="self.decoder.mid_block.resnets.0.norm2.bias">
302
+ <dim>64</dim>
303
+ </port>
304
+ </output>
305
+ </layer>
306
+ <layer id="19" name="__module.decoder.mid_block.resnets.0.norm2/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
307
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
308
+ <input>
309
+ <port id="0" precision="FP32">
310
+ <dim>-1</dim>
311
+ <dim>64</dim>
312
+ <dim>-1</dim>
313
+ <dim>-1</dim>
314
+ </port>
315
+ <port id="1" precision="FP32">
316
+ <dim>64</dim>
317
+ </port>
318
+ <port id="2" precision="FP32">
319
+ <dim>64</dim>
320
+ </port>
321
+ </input>
322
+ <output>
323
+ <port id="3" precision="FP32" names="81,input.7">
324
+ <dim>-1</dim>
325
+ <dim>64</dim>
326
+ <dim>-1</dim>
327
+ <dim>-1</dim>
328
+ </port>
329
+ </output>
330
+ </layer>
331
+ <layer id="20" name="__module.decoder.mid_block.resnets.1.nonlinearity/aten::silu/Swish_1" type="Swish" version="opset4">
332
+ <input>
333
+ <port id="0" precision="FP32">
334
+ <dim>-1</dim>
335
+ <dim>64</dim>
336
+ <dim>-1</dim>
337
+ <dim>-1</dim>
338
+ </port>
339
+ </input>
340
+ <output>
341
+ <port id="1" precision="FP32" names="82,input.9">
342
+ <dim>-1</dim>
343
+ <dim>64</dim>
344
+ <dim>-1</dim>
345
+ <dim>-1</dim>
346
+ </port>
347
+ </output>
348
+ </layer>
349
+ <layer id="21" name="self.decoder.mid_block.resnets.0.conv2.weight" type="Const" version="opset1">
350
+ <data element_type="f32" shape="64, 64, 3, 3" offset="157776" size="147456" />
351
+ <output>
352
+ <port id="0" precision="FP32" names="self.decoder.mid_block.resnets.0.conv2.weight">
353
+ <dim>64</dim>
354
+ <dim>64</dim>
355
+ <dim>3</dim>
356
+ <dim>3</dim>
357
+ </port>
358
+ </output>
359
+ </layer>
360
+ <layer id="22" name="__module.decoder.mid_block.resnets.0.conv2/aten::_convolution/Convolution" type="Convolution" version="opset1">
361
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
362
+ <input>
363
+ <port id="0" precision="FP32">
364
+ <dim>-1</dim>
365
+ <dim>64</dim>
366
+ <dim>-1</dim>
367
+ <dim>-1</dim>
368
+ </port>
369
+ <port id="1" precision="FP32">
370
+ <dim>64</dim>
371
+ <dim>64</dim>
372
+ <dim>3</dim>
373
+ <dim>3</dim>
374
+ </port>
375
+ </input>
376
+ <output>
377
+ <port id="2" precision="FP32">
378
+ <dim>-1</dim>
379
+ <dim>64</dim>
380
+ <dim>-1</dim>
381
+ <dim>-1</dim>
382
+ </port>
383
+ </output>
384
+ </layer>
385
+ <layer id="23" name="__module.decoder.mid_block.resnets.0.conv2/aten::_convolution/Reshape" type="Const" version="opset1">
386
+ <data element_type="f32" shape="1, 64, 1, 1" offset="305232" size="256" />
387
+ <output>
388
+ <port id="0" precision="FP32">
389
+ <dim>1</dim>
390
+ <dim>64</dim>
391
+ <dim>1</dim>
392
+ <dim>1</dim>
393
+ </port>
394
+ </output>
395
+ </layer>
396
+ <layer id="24" name="__module.decoder.mid_block.resnets.0.conv2/aten::_convolution/Add" type="Add" version="opset1">
397
+ <data auto_broadcast="numpy" />
398
+ <input>
399
+ <port id="0" precision="FP32">
400
+ <dim>-1</dim>
401
+ <dim>64</dim>
402
+ <dim>-1</dim>
403
+ <dim>-1</dim>
404
+ </port>
405
+ <port id="1" precision="FP32">
406
+ <dim>1</dim>
407
+ <dim>64</dim>
408
+ <dim>1</dim>
409
+ <dim>1</dim>
410
+ </port>
411
+ </input>
412
+ <output>
413
+ <port id="2" precision="FP32" names="90,hidden_states.1">
414
+ <dim>-1</dim>
415
+ <dim>64</dim>
416
+ <dim>-1</dim>
417
+ <dim>-1</dim>
418
+ </port>
419
+ </output>
420
+ </layer>
421
+ <layer id="25" name="__module.decoder.mid_block.resnets.0/aten::add/Add" type="Add" version="opset1">
422
+ <data auto_broadcast="numpy" />
423
+ <input>
424
+ <port id="0" precision="FP32">
425
+ <dim>-1</dim>
426
+ <dim>64</dim>
427
+ <dim>-1</dim>
428
+ <dim>-1</dim>
429
+ </port>
430
+ <port id="1" precision="FP32">
431
+ <dim>-1</dim>
432
+ <dim>64</dim>
433
+ <dim>-1</dim>
434
+ <dim>-1</dim>
435
+ </port>
436
+ </input>
437
+ <output>
438
+ <port id="2" precision="FP32" names="91,92,hidden_states.3">
439
+ <dim>-1</dim>
440
+ <dim>64</dim>
441
+ <dim>-1</dim>
442
+ <dim>-1</dim>
443
+ </port>
444
+ </output>
445
+ </layer>
446
+ <layer id="26" name="Constant_54139" type="Const" version="opset1">
447
+ <data element_type="i64" shape="3" offset="305488" size="24" />
448
+ <output>
449
+ <port id="0" precision="I64">
450
+ <dim>3</dim>
451
+ </port>
452
+ </output>
453
+ </layer>
454
+ <layer id="27" name="__module.decoder.mid_block.attentions.0/aten::view/Reshape" type="Reshape" version="opset1">
455
+ <data special_zero="true" />
456
+ <input>
457
+ <port id="0" precision="FP32">
458
+ <dim>-1</dim>
459
+ <dim>64</dim>
460
+ <dim>-1</dim>
461
+ <dim>-1</dim>
462
+ </port>
463
+ <port id="1" precision="I64">
464
+ <dim>3</dim>
465
+ </port>
466
+ </input>
467
+ <output>
468
+ <port id="2" precision="FP32" names="108">
469
+ <dim>-1</dim>
470
+ <dim>64</dim>
471
+ <dim>-1</dim>
472
+ </port>
473
+ </output>
474
+ </layer>
475
+ <layer id="28" name="__module.decoder.mid_block.attentions.0/aten::transpose/Constant" type="Const" version="opset1">
476
+ <data element_type="i32" shape="3" offset="305512" size="12" />
477
+ <output>
478
+ <port id="0" precision="I32">
479
+ <dim>3</dim>
480
+ </port>
481
+ </output>
482
+ </layer>
483
+ <layer id="29" name="__module.decoder.mid_block.attentions.0/aten::transpose/Transpose" type="Transpose" version="opset1">
484
+ <input>
485
+ <port id="0" precision="FP32">
486
+ <dim>-1</dim>
487
+ <dim>64</dim>
488
+ <dim>-1</dim>
489
+ </port>
490
+ <port id="1" precision="I32">
491
+ <dim>3</dim>
492
+ </port>
493
+ </input>
494
+ <output>
495
+ <port id="2" precision="FP32" names="109,hidden_states.5">
496
+ <dim>-1</dim>
497
+ <dim>-1</dim>
498
+ <dim>64</dim>
499
+ </port>
500
+ </output>
501
+ </layer>
502
+ <layer id="30" name="__module.decoder.mid_block.attentions.0/aten::transpose/Constant_1" type="Const" version="opset1">
503
+ <data element_type="i32" shape="3" offset="305512" size="12" />
504
+ <output>
505
+ <port id="0" precision="I32">
506
+ <dim>3</dim>
507
+ </port>
508
+ </output>
509
+ </layer>
510
+ <layer id="31" name="__module.decoder.mid_block.attentions.0/aten::transpose/Transpose_1" type="Transpose" version="opset1">
511
+ <input>
512
+ <port id="0" precision="FP32">
513
+ <dim>-1</dim>
514
+ <dim>-1</dim>
515
+ <dim>64</dim>
516
+ </port>
517
+ <port id="1" precision="I32">
518
+ <dim>3</dim>
519
+ </port>
520
+ </input>
521
+ <output>
522
+ <port id="2" precision="FP32" names="111,input.11">
523
+ <dim>-1</dim>
524
+ <dim>64</dim>
525
+ <dim>-1</dim>
526
+ </port>
527
+ </output>
528
+ </layer>
529
+ <layer id="32" name="self.decoder.mid_block.attentions.0.group_norm.weight" type="Const" version="opset1">
530
+ <data element_type="f32" shape="64" offset="9552" size="256" />
531
+ <output>
532
+ <port id="0" precision="FP32" names="self.decoder.mid_block.attentions.0.group_norm.weight">
533
+ <dim>64</dim>
534
+ </port>
535
+ </output>
536
+ </layer>
537
+ <layer id="33" name="self.decoder.mid_block.attentions.0.group_norm.bias" type="Const" version="opset1">
538
+ <data element_type="f32" shape="64" offset="9808" size="256" />
539
+ <output>
540
+ <port id="0" precision="FP32" names="self.decoder.mid_block.attentions.0.group_norm.bias">
541
+ <dim>64</dim>
542
+ </port>
543
+ </output>
544
+ </layer>
545
+ <layer id="34" name="__module.decoder.mid_block.attentions.0.group_norm/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
546
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
547
+ <input>
548
+ <port id="0" precision="FP32">
549
+ <dim>-1</dim>
550
+ <dim>64</dim>
551
+ <dim>-1</dim>
552
+ </port>
553
+ <port id="1" precision="FP32">
554
+ <dim>64</dim>
555
+ </port>
556
+ <port id="2" precision="FP32">
557
+ <dim>64</dim>
558
+ </port>
559
+ </input>
560
+ <output>
561
+ <port id="3" precision="FP32" names="114">
562
+ <dim>-1</dim>
563
+ <dim>64</dim>
564
+ <dim>-1</dim>
565
+ </port>
566
+ </output>
567
+ </layer>
568
+ <layer id="35" name="self.decoder.mid_block.attentions.0.to_q.weight" type="Const" version="opset1">
569
+ <data element_type="f32" shape="64, 64" offset="305524" size="16384" />
570
+ <output>
571
+ <port id="0" precision="FP32" names="self.decoder.mid_block.attentions.0.to_q.weight">
572
+ <dim>64</dim>
573
+ <dim>64</dim>
574
+ </port>
575
+ </output>
576
+ </layer>
577
+ <layer id="36" name="__module.decoder.mid_block.attentions.0.to_q/aten::linear/MatMul" type="MatMul" version="opset1">
578
+ <data transpose_a="true" transpose_b="true" />
579
+ <input>
580
+ <port id="0" precision="FP32">
581
+ <dim>-1</dim>
582
+ <dim>64</dim>
583
+ <dim>-1</dim>
584
+ </port>
585
+ <port id="1" precision="FP32">
586
+ <dim>64</dim>
587
+ <dim>64</dim>
588
+ </port>
589
+ </input>
590
+ <output>
591
+ <port id="2" precision="FP32">
592
+ <dim>-1</dim>
593
+ <dim>-1</dim>
594
+ <dim>64</dim>
595
+ </port>
596
+ </output>
597
+ </layer>
598
+ <layer id="37" name="Constant_54024" type="Const" version="opset1">
599
+ <data element_type="f32" shape="1, 1, 64" offset="321908" size="256" />
600
+ <output>
601
+ <port id="0" precision="FP32">
602
+ <dim>1</dim>
603
+ <dim>1</dim>
604
+ <dim>64</dim>
605
+ </port>
606
+ </output>
607
+ </layer>
608
+ <layer id="38" name="__module.decoder.mid_block.attentions.0.to_q/aten::linear/Add" type="Add" version="opset1">
609
+ <data auto_broadcast="numpy" />
610
+ <input>
611
+ <port id="0" precision="FP32">
612
+ <dim>-1</dim>
613
+ <dim>-1</dim>
614
+ <dim>64</dim>
615
+ </port>
616
+ <port id="1" precision="FP32">
617
+ <dim>1</dim>
618
+ <dim>1</dim>
619
+ <dim>64</dim>
620
+ </port>
621
+ </input>
622
+ <output>
623
+ <port id="2" precision="FP32" names="118,query">
624
+ <dim>-1</dim>
625
+ <dim>-1</dim>
626
+ <dim>64</dim>
627
+ </port>
628
+ </output>
629
+ </layer>
630
+ <layer id="39" name="Constant_54140" type="Const" version="opset1">
631
+ <data element_type="i64" shape="4" offset="322164" size="32" />
632
+ <output>
633
+ <port id="0" precision="I64">
634
+ <dim>4</dim>
635
+ </port>
636
+ </output>
637
+ </layer>
638
+ <layer id="40" name="__module.decoder.mid_block.attentions.0/aten::view/Reshape_1" type="Reshape" version="opset1">
639
+ <data special_zero="true" />
640
+ <input>
641
+ <port id="0" precision="FP32">
642
+ <dim>-1</dim>
643
+ <dim>-1</dim>
644
+ <dim>64</dim>
645
+ </port>
646
+ <port id="1" precision="I64">
647
+ <dim>4</dim>
648
+ </port>
649
+ </input>
650
+ <output>
651
+ <port id="2" precision="FP32" names="132">
652
+ <dim>-1</dim>
653
+ <dim>-1</dim>
654
+ <dim>1</dim>
655
+ <dim>64</dim>
656
+ </port>
657
+ </output>
658
+ </layer>
659
+ <layer id="41" name="Constant_53943" type="Const" version="opset1">
660
+ <data element_type="i64" shape="4" offset="322196" size="32" />
661
+ <output>
662
+ <port id="0" precision="I64">
663
+ <dim>4</dim>
664
+ </port>
665
+ </output>
666
+ </layer>
667
+ <layer id="42" name="__module.decoder.mid_block.attentions.0/aten::transpose/Transpose_3" type="Reshape" version="opset1">
668
+ <data special_zero="true" />
669
+ <input>
670
+ <port id="0" precision="FP32">
671
+ <dim>-1</dim>
672
+ <dim>-1</dim>
673
+ <dim>1</dim>
674
+ <dim>64</dim>
675
+ </port>
676
+ <port id="1" precision="I64">
677
+ <dim>4</dim>
678
+ </port>
679
+ </input>
680
+ <output>
681
+ <port id="2" precision="FP32" names="133">
682
+ <dim>-1</dim>
683
+ <dim>1</dim>
684
+ <dim>-1</dim>
685
+ <dim>64</dim>
686
+ </port>
687
+ </output>
688
+ </layer>
689
+ <layer id="43" name="self.decoder.mid_block.attentions.0.to_k.weight" type="Const" version="opset1">
690
+ <data element_type="f32" shape="64, 64" offset="322228" size="16384" />
691
+ <output>
692
+ <port id="0" precision="FP32" names="self.decoder.mid_block.attentions.0.to_k.weight">
693
+ <dim>64</dim>
694
+ <dim>64</dim>
695
+ </port>
696
+ </output>
697
+ </layer>
698
+ <layer id="44" name="__module.decoder.mid_block.attentions.0.to_k/aten::linear/MatMul" type="MatMul" version="opset1">
699
+ <data transpose_a="true" transpose_b="true" />
700
+ <input>
701
+ <port id="0" precision="FP32">
702
+ <dim>-1</dim>
703
+ <dim>64</dim>
704
+ <dim>-1</dim>
705
+ </port>
706
+ <port id="1" precision="FP32">
707
+ <dim>64</dim>
708
+ <dim>64</dim>
709
+ </port>
710
+ </input>
711
+ <output>
712
+ <port id="2" precision="FP32">
713
+ <dim>-1</dim>
714
+ <dim>-1</dim>
715
+ <dim>64</dim>
716
+ </port>
717
+ </output>
718
+ </layer>
719
+ <layer id="45" name="Constant_54025" type="Const" version="opset1">
720
+ <data element_type="f32" shape="1, 1, 64" offset="338612" size="256" />
721
+ <output>
722
+ <port id="0" precision="FP32">
723
+ <dim>1</dim>
724
+ <dim>1</dim>
725
+ <dim>64</dim>
726
+ </port>
727
+ </output>
728
+ </layer>
729
+ <layer id="46" name="__module.decoder.mid_block.attentions.0.to_k/aten::linear/Add" type="Add" version="opset1">
730
+ <data auto_broadcast="numpy" />
731
+ <input>
732
+ <port id="0" precision="FP32">
733
+ <dim>-1</dim>
734
+ <dim>-1</dim>
735
+ <dim>64</dim>
736
+ </port>
737
+ <port id="1" precision="FP32">
738
+ <dim>1</dim>
739
+ <dim>1</dim>
740
+ <dim>64</dim>
741
+ </port>
742
+ </input>
743
+ <output>
744
+ <port id="2" precision="FP32" names="121,key">
745
+ <dim>-1</dim>
746
+ <dim>-1</dim>
747
+ <dim>64</dim>
748
+ </port>
749
+ </output>
750
+ </layer>
751
+ <layer id="47" name="Constant_54141" type="Const" version="opset1">
752
+ <data element_type="i64" shape="4" offset="322164" size="32" />
753
+ <output>
754
+ <port id="0" precision="I64">
755
+ <dim>4</dim>
756
+ </port>
757
+ </output>
758
+ </layer>
759
+ <layer id="48" name="__module.decoder.mid_block.attentions.0/aten::view/Reshape_2" type="Reshape" version="opset1">
760
+ <data special_zero="true" />
761
+ <input>
762
+ <port id="0" precision="FP32">
763
+ <dim>-1</dim>
764
+ <dim>-1</dim>
765
+ <dim>64</dim>
766
+ </port>
767
+ <port id="1" precision="I64">
768
+ <dim>4</dim>
769
+ </port>
770
+ </input>
771
+ <output>
772
+ <port id="2" precision="FP32" names="135">
773
+ <dim>-1</dim>
774
+ <dim>-1</dim>
775
+ <dim>1</dim>
776
+ <dim>64</dim>
777
+ </port>
778
+ </output>
779
+ </layer>
780
+ <layer id="49" name="Constant_53947" type="Const" version="opset1">
781
+ <data element_type="i64" shape="4" offset="322196" size="32" />
782
+ <output>
783
+ <port id="0" precision="I64">
784
+ <dim>4</dim>
785
+ </port>
786
+ </output>
787
+ </layer>
788
+ <layer id="50" name="__module.decoder.mid_block.attentions.0/aten::transpose/Transpose_4" type="Reshape" version="opset1">
789
+ <data special_zero="true" />
790
+ <input>
791
+ <port id="0" precision="FP32">
792
+ <dim>-1</dim>
793
+ <dim>-1</dim>
794
+ <dim>1</dim>
795
+ <dim>64</dim>
796
+ </port>
797
+ <port id="1" precision="I64">
798
+ <dim>4</dim>
799
+ </port>
800
+ </input>
801
+ <output>
802
+ <port id="2" precision="FP32" names="136">
803
+ <dim>-1</dim>
804
+ <dim>1</dim>
805
+ <dim>-1</dim>
806
+ <dim>64</dim>
807
+ </port>
808
+ </output>
809
+ </layer>
810
+ <layer id="51" name="self.decoder.mid_block.attentions.0.to_v.weight" type="Const" version="opset1">
811
+ <data element_type="f32" shape="64, 64" offset="338868" size="16384" />
812
+ <output>
813
+ <port id="0" precision="FP32" names="self.decoder.mid_block.attentions.0.to_v.weight">
814
+ <dim>64</dim>
815
+ <dim>64</dim>
816
+ </port>
817
+ </output>
818
+ </layer>
819
+ <layer id="52" name="__module.decoder.mid_block.attentions.0.to_v/aten::linear/MatMul" type="MatMul" version="opset1">
820
+ <data transpose_a="true" transpose_b="true" />
821
+ <input>
822
+ <port id="0" precision="FP32">
823
+ <dim>-1</dim>
824
+ <dim>64</dim>
825
+ <dim>-1</dim>
826
+ </port>
827
+ <port id="1" precision="FP32">
828
+ <dim>64</dim>
829
+ <dim>64</dim>
830
+ </port>
831
+ </input>
832
+ <output>
833
+ <port id="2" precision="FP32">
834
+ <dim>-1</dim>
835
+ <dim>-1</dim>
836
+ <dim>64</dim>
837
+ </port>
838
+ </output>
839
+ </layer>
840
+ <layer id="53" name="Constant_54026" type="Const" version="opset1">
841
+ <data element_type="f32" shape="1, 1, 64" offset="355252" size="256" />
842
+ <output>
843
+ <port id="0" precision="FP32">
844
+ <dim>1</dim>
845
+ <dim>1</dim>
846
+ <dim>64</dim>
847
+ </port>
848
+ </output>
849
+ </layer>
850
+ <layer id="54" name="__module.decoder.mid_block.attentions.0.to_v/aten::linear/Add" type="Add" version="opset1">
851
+ <data auto_broadcast="numpy" />
852
+ <input>
853
+ <port id="0" precision="FP32">
854
+ <dim>-1</dim>
855
+ <dim>-1</dim>
856
+ <dim>64</dim>
857
+ </port>
858
+ <port id="1" precision="FP32">
859
+ <dim>1</dim>
860
+ <dim>1</dim>
861
+ <dim>64</dim>
862
+ </port>
863
+ </input>
864
+ <output>
865
+ <port id="2" precision="FP32" names="124,value">
866
+ <dim>-1</dim>
867
+ <dim>-1</dim>
868
+ <dim>64</dim>
869
+ </port>
870
+ </output>
871
+ </layer>
872
+ <layer id="55" name="Constant_54142" type="Const" version="opset1">
873
+ <data element_type="i64" shape="4" offset="322164" size="32" />
874
+ <output>
875
+ <port id="0" precision="I64">
876
+ <dim>4</dim>
877
+ </port>
878
+ </output>
879
+ </layer>
880
+ <layer id="56" name="__module.decoder.mid_block.attentions.0/aten::view/Reshape_3" type="Reshape" version="opset1">
881
+ <data special_zero="true" />
882
+ <input>
883
+ <port id="0" precision="FP32">
884
+ <dim>-1</dim>
885
+ <dim>-1</dim>
886
+ <dim>64</dim>
887
+ </port>
888
+ <port id="1" precision="I64">
889
+ <dim>4</dim>
890
+ </port>
891
+ </input>
892
+ <output>
893
+ <port id="2" precision="FP32" names="138">
894
+ <dim>-1</dim>
895
+ <dim>-1</dim>
896
+ <dim>1</dim>
897
+ <dim>64</dim>
898
+ </port>
899
+ </output>
900
+ </layer>
901
+ <layer id="57" name="Constant_53951" type="Const" version="opset1">
902
+ <data element_type="i64" shape="4" offset="322196" size="32" />
903
+ <output>
904
+ <port id="0" precision="I64">
905
+ <dim>4</dim>
906
+ </port>
907
+ </output>
908
+ </layer>
909
+ <layer id="58" name="__module.decoder.mid_block.attentions.0/aten::transpose/Transpose_5" type="Reshape" version="opset1">
910
+ <data special_zero="true" />
911
+ <input>
912
+ <port id="0" precision="FP32">
913
+ <dim>-1</dim>
914
+ <dim>-1</dim>
915
+ <dim>1</dim>
916
+ <dim>64</dim>
917
+ </port>
918
+ <port id="1" precision="I64">
919
+ <dim>4</dim>
920
+ </port>
921
+ </input>
922
+ <output>
923
+ <port id="2" precision="FP32" names="139">
924
+ <dim>-1</dim>
925
+ <dim>1</dim>
926
+ <dim>-1</dim>
927
+ <dim>64</dim>
928
+ </port>
929
+ </output>
930
+ </layer>
931
+ <layer id="59" name="__module.decoder.mid_block.attentions.0/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
932
+ <data causal="false" />
933
+ <input>
934
+ <port id="0" precision="FP32">
935
+ <dim>-1</dim>
936
+ <dim>1</dim>
937
+ <dim>-1</dim>
938
+ <dim>64</dim>
939
+ </port>
940
+ <port id="1" precision="FP32">
941
+ <dim>-1</dim>
942
+ <dim>1</dim>
943
+ <dim>-1</dim>
944
+ <dim>64</dim>
945
+ </port>
946
+ <port id="2" precision="FP32">
947
+ <dim>-1</dim>
948
+ <dim>1</dim>
949
+ <dim>-1</dim>
950
+ <dim>64</dim>
951
+ </port>
952
+ </input>
953
+ <output>
954
+ <port id="3" precision="FP32" names="140,hidden_states.7">
955
+ <dim>-1</dim>
956
+ <dim>1</dim>
957
+ <dim>-1</dim>
958
+ <dim>64</dim>
959
+ </port>
960
+ </output>
961
+ </layer>
962
+ <layer id="60" name="Constant_53953" type="Const" version="opset1">
963
+ <data element_type="i64" shape="4" offset="355508" size="32" />
964
+ <output>
965
+ <port id="0" precision="I64">
966
+ <dim>4</dim>
967
+ </port>
968
+ </output>
969
+ </layer>
970
+ <layer id="61" name="__module.decoder.mid_block.attentions.0/aten::transpose/Transpose_6" type="Reshape" version="opset1">
971
+ <data special_zero="true" />
972
+ <input>
973
+ <port id="0" precision="FP32">
974
+ <dim>-1</dim>
975
+ <dim>1</dim>
976
+ <dim>-1</dim>
977
+ <dim>64</dim>
978
+ </port>
979
+ <port id="1" precision="I64">
980
+ <dim>4</dim>
981
+ </port>
982
+ </input>
983
+ <output>
984
+ <port id="2" precision="FP32" names="141">
985
+ <dim>-1</dim>
986
+ <dim>-1</dim>
987
+ <dim>1</dim>
988
+ <dim>64</dim>
989
+ </port>
990
+ </output>
991
+ </layer>
992
+ <layer id="62" name="Constant_54143" type="Const" version="opset1">
993
+ <data element_type="i64" shape="3" offset="355540" size="24" />
994
+ <output>
995
+ <port id="0" precision="I64">
996
+ <dim>3</dim>
997
+ </port>
998
+ </output>
999
+ </layer>
1000
+ <layer id="63" name="__module.decoder.mid_block.attentions.0/aten::reshape/Reshape" type="Reshape" version="opset1">
1001
+ <data special_zero="true" />
1002
+ <input>
1003
+ <port id="0" precision="FP32">
1004
+ <dim>-1</dim>
1005
+ <dim>-1</dim>
1006
+ <dim>1</dim>
1007
+ <dim>64</dim>
1008
+ </port>
1009
+ <port id="1" precision="I64">
1010
+ <dim>3</dim>
1011
+ </port>
1012
+ </input>
1013
+ <output>
1014
+ <port id="2" precision="FP32" names="145,146,hidden_states.9">
1015
+ <dim>-1</dim>
1016
+ <dim>-1</dim>
1017
+ <dim>64</dim>
1018
+ </port>
1019
+ </output>
1020
+ </layer>
1021
+ <layer id="64" name="self.decoder.mid_block.attentions.0.to_out.0.weight" type="Const" version="opset1">
1022
+ <data element_type="f32" shape="64, 64" offset="355564" size="16384" />
1023
+ <output>
1024
+ <port id="0" precision="FP32" names="self.decoder.mid_block.attentions.0.to_out.0.weight">
1025
+ <dim>64</dim>
1026
+ <dim>64</dim>
1027
+ </port>
1028
+ </output>
1029
+ </layer>
1030
+ <layer id="65" name="__module.decoder.mid_block.attentions.0.to_out.0/aten::linear/MatMul" type="MatMul" version="opset1">
1031
+ <data transpose_a="false" transpose_b="true" />
1032
+ <input>
1033
+ <port id="0" precision="FP32">
1034
+ <dim>-1</dim>
1035
+ <dim>-1</dim>
1036
+ <dim>64</dim>
1037
+ </port>
1038
+ <port id="1" precision="FP32">
1039
+ <dim>64</dim>
1040
+ <dim>64</dim>
1041
+ </port>
1042
+ </input>
1043
+ <output>
1044
+ <port id="2" precision="FP32">
1045
+ <dim>-1</dim>
1046
+ <dim>-1</dim>
1047
+ <dim>64</dim>
1048
+ </port>
1049
+ </output>
1050
+ </layer>
1051
+ <layer id="66" name="Constant_54027" type="Const" version="opset1">
1052
+ <data element_type="f32" shape="1, 1, 64" offset="371948" size="256" />
1053
+ <output>
1054
+ <port id="0" precision="FP32">
1055
+ <dim>1</dim>
1056
+ <dim>1</dim>
1057
+ <dim>64</dim>
1058
+ </port>
1059
+ </output>
1060
+ </layer>
1061
+ <layer id="67" name="__module.decoder.mid_block.attentions.0.to_out.0/aten::linear/Add" type="Add" version="opset1">
1062
+ <data auto_broadcast="numpy" />
1063
+ <input>
1064
+ <port id="0" precision="FP32">
1065
+ <dim>-1</dim>
1066
+ <dim>-1</dim>
1067
+ <dim>64</dim>
1068
+ </port>
1069
+ <port id="1" precision="FP32">
1070
+ <dim>1</dim>
1071
+ <dim>1</dim>
1072
+ <dim>64</dim>
1073
+ </port>
1074
+ </input>
1075
+ <output>
1076
+ <port id="2" precision="FP32" names="149,input.13">
1077
+ <dim>-1</dim>
1078
+ <dim>-1</dim>
1079
+ <dim>64</dim>
1080
+ </port>
1081
+ </output>
1082
+ </layer>
1083
+ <layer id="68" name="__module.decoder.mid_block.attentions.0/aten::transpose/Constant_7" type="Const" version="opset1">
1084
+ <data element_type="i32" shape="3" offset="305512" size="12" />
1085
+ <output>
1086
+ <port id="0" precision="I32">
1087
+ <dim>3</dim>
1088
+ </port>
1089
+ </output>
1090
+ </layer>
1091
+ <layer id="69" name="__module.decoder.mid_block.attentions.0/aten::transpose/Transpose_7" type="Transpose" version="opset1">
1092
+ <input>
1093
+ <port id="0" precision="FP32">
1094
+ <dim>-1</dim>
1095
+ <dim>-1</dim>
1096
+ <dim>64</dim>
1097
+ </port>
1098
+ <port id="1" precision="I32">
1099
+ <dim>3</dim>
1100
+ </port>
1101
+ </input>
1102
+ <output>
1103
+ <port id="2" precision="FP32" names="151">
1104
+ <dim>-1</dim>
1105
+ <dim>64</dim>
1106
+ <dim>-1</dim>
1107
+ </port>
1108
+ </output>
1109
+ </layer>
1110
+ <layer id="70" name="__module.decoder.mid_block.attentions.0/aten::size/ShapeOf" type="ShapeOf" version="opset3">
1111
+ <data output_type="i64" />
1112
+ <input>
1113
+ <port id="0" precision="FP32">
1114
+ <dim>-1</dim>
1115
+ <dim>64</dim>
1116
+ <dim>-1</dim>
1117
+ <dim>-1</dim>
1118
+ </port>
1119
+ </input>
1120
+ <output>
1121
+ <port id="1" precision="I64">
1122
+ <dim>4</dim>
1123
+ </port>
1124
+ </output>
1125
+ </layer>
1126
+ <layer id="71" name="__module.decoder.mid_block.attentions.0/aten::reshape/Reshape_1" type="Reshape" version="opset1">
1127
+ <data special_zero="false" />
1128
+ <input>
1129
+ <port id="0" precision="FP32">
1130
+ <dim>-1</dim>
1131
+ <dim>64</dim>
1132
+ <dim>-1</dim>
1133
+ </port>
1134
+ <port id="1" precision="I64">
1135
+ <dim>4</dim>
1136
+ </port>
1137
+ </input>
1138
+ <output>
1139
+ <port id="2" precision="FP32" names="153,hidden_states.13">
1140
+ <dim>-1</dim>
1141
+ <dim>64</dim>
1142
+ <dim>-1</dim>
1143
+ <dim>-1</dim>
1144
+ </port>
1145
+ </output>
1146
+ </layer>
1147
+ <layer id="72" name="__module.decoder.mid_block.attentions.0/aten::add/Add" type="Add" version="opset1">
1148
+ <data auto_broadcast="numpy" />
1149
+ <input>
1150
+ <port id="0" precision="FP32">
1151
+ <dim>-1</dim>
1152
+ <dim>64</dim>
1153
+ <dim>-1</dim>
1154
+ <dim>-1</dim>
1155
+ </port>
1156
+ <port id="1" precision="FP32">
1157
+ <dim>-1</dim>
1158
+ <dim>64</dim>
1159
+ <dim>-1</dim>
1160
+ <dim>-1</dim>
1161
+ </port>
1162
+ </input>
1163
+ <output>
1164
+ <port id="2" precision="FP32" names="154,155,hidden_states.15,input.15">
1165
+ <dim>-1</dim>
1166
+ <dim>64</dim>
1167
+ <dim>-1</dim>
1168
+ <dim>-1</dim>
1169
+ </port>
1170
+ </output>
1171
+ </layer>
1172
+ <layer id="73" name="self.decoder.mid_block.resnets.1.norm1.weight" type="Const" version="opset1">
1173
+ <data element_type="f32" shape="64" offset="9552" size="256" />
1174
+ <output>
1175
+ <port id="0" precision="FP32" names="self.decoder.mid_block.resnets.1.norm1.weight">
1176
+ <dim>64</dim>
1177
+ </port>
1178
+ </output>
1179
+ </layer>
1180
+ <layer id="74" name="self.decoder.mid_block.resnets.1.norm1.bias" type="Const" version="opset1">
1181
+ <data element_type="f32" shape="64" offset="9808" size="256" />
1182
+ <output>
1183
+ <port id="0" precision="FP32" names="self.decoder.mid_block.resnets.1.norm1.bias">
1184
+ <dim>64</dim>
1185
+ </port>
1186
+ </output>
1187
+ </layer>
1188
+ <layer id="75" name="__module.decoder.mid_block.resnets.1.norm1/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
1189
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
1190
+ <input>
1191
+ <port id="0" precision="FP32">
1192
+ <dim>-1</dim>
1193
+ <dim>64</dim>
1194
+ <dim>-1</dim>
1195
+ <dim>-1</dim>
1196
+ </port>
1197
+ <port id="1" precision="FP32">
1198
+ <dim>64</dim>
1199
+ </port>
1200
+ <port id="2" precision="FP32">
1201
+ <dim>64</dim>
1202
+ </port>
1203
+ </input>
1204
+ <output>
1205
+ <port id="3" precision="FP32" names="164,input.17">
1206
+ <dim>-1</dim>
1207
+ <dim>64</dim>
1208
+ <dim>-1</dim>
1209
+ <dim>-1</dim>
1210
+ </port>
1211
+ </output>
1212
+ </layer>
1213
+ <layer id="76" name="__module.decoder.mid_block.resnets.1.nonlinearity/aten::silu/Swish_2" type="Swish" version="opset4">
1214
+ <input>
1215
+ <port id="0" precision="FP32">
1216
+ <dim>-1</dim>
1217
+ <dim>64</dim>
1218
+ <dim>-1</dim>
1219
+ <dim>-1</dim>
1220
+ </port>
1221
+ </input>
1222
+ <output>
1223
+ <port id="1" precision="FP32" names="165">
1224
+ <dim>-1</dim>
1225
+ <dim>64</dim>
1226
+ <dim>-1</dim>
1227
+ <dim>-1</dim>
1228
+ </port>
1229
+ </output>
1230
+ </layer>
1231
+ <layer id="77" name="self.decoder.mid_block.resnets.1.conv1.weight" type="Const" version="opset1">
1232
+ <data element_type="f32" shape="64, 64, 3, 3" offset="372204" size="147456" />
1233
+ <output>
1234
+ <port id="0" precision="FP32" names="self.decoder.mid_block.resnets.1.conv1.weight">
1235
+ <dim>64</dim>
1236
+ <dim>64</dim>
1237
+ <dim>3</dim>
1238
+ <dim>3</dim>
1239
+ </port>
1240
+ </output>
1241
+ </layer>
1242
+ <layer id="78" name="__module.decoder.mid_block.resnets.1.conv1/aten::_convolution/Convolution" type="Convolution" version="opset1">
1243
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
1244
+ <input>
1245
+ <port id="0" precision="FP32">
1246
+ <dim>-1</dim>
1247
+ <dim>64</dim>
1248
+ <dim>-1</dim>
1249
+ <dim>-1</dim>
1250
+ </port>
1251
+ <port id="1" precision="FP32">
1252
+ <dim>64</dim>
1253
+ <dim>64</dim>
1254
+ <dim>3</dim>
1255
+ <dim>3</dim>
1256
+ </port>
1257
+ </input>
1258
+ <output>
1259
+ <port id="2" precision="FP32">
1260
+ <dim>-1</dim>
1261
+ <dim>64</dim>
1262
+ <dim>-1</dim>
1263
+ <dim>-1</dim>
1264
+ </port>
1265
+ </output>
1266
+ </layer>
1267
+ <layer id="79" name="__module.decoder.mid_block.resnets.1.conv1/aten::_convolution/Reshape" type="Const" version="opset1">
1268
+ <data element_type="f32" shape="1, 64, 1, 1" offset="519660" size="256" />
1269
+ <output>
1270
+ <port id="0" precision="FP32">
1271
+ <dim>1</dim>
1272
+ <dim>64</dim>
1273
+ <dim>1</dim>
1274
+ <dim>1</dim>
1275
+ </port>
1276
+ </output>
1277
+ </layer>
1278
+ <layer id="80" name="__module.decoder.mid_block.resnets.1.conv1/aten::_convolution/Add" type="Add" version="opset1">
1279
+ <data auto_broadcast="numpy" />
1280
+ <input>
1281
+ <port id="0" precision="FP32">
1282
+ <dim>-1</dim>
1283
+ <dim>64</dim>
1284
+ <dim>-1</dim>
1285
+ <dim>-1</dim>
1286
+ </port>
1287
+ <port id="1" precision="FP32">
1288
+ <dim>1</dim>
1289
+ <dim>64</dim>
1290
+ <dim>1</dim>
1291
+ <dim>1</dim>
1292
+ </port>
1293
+ </input>
1294
+ <output>
1295
+ <port id="2" precision="FP32" names="172,input.19">
1296
+ <dim>-1</dim>
1297
+ <dim>64</dim>
1298
+ <dim>-1</dim>
1299
+ <dim>-1</dim>
1300
+ </port>
1301
+ </output>
1302
+ </layer>
1303
+ <layer id="81" name="self.decoder.mid_block.resnets.1.norm2.weight" type="Const" version="opset1">
1304
+ <data element_type="f32" shape="64" offset="9552" size="256" />
1305
+ <output>
1306
+ <port id="0" precision="FP32" names="self.decoder.mid_block.resnets.1.norm2.weight">
1307
+ <dim>64</dim>
1308
+ </port>
1309
+ </output>
1310
+ </layer>
1311
+ <layer id="82" name="self.decoder.mid_block.resnets.1.norm2.bias" type="Const" version="opset1">
1312
+ <data element_type="f32" shape="64" offset="9808" size="256" />
1313
+ <output>
1314
+ <port id="0" precision="FP32" names="self.decoder.mid_block.resnets.1.norm2.bias">
1315
+ <dim>64</dim>
1316
+ </port>
1317
+ </output>
1318
+ </layer>
1319
+ <layer id="83" name="__module.decoder.mid_block.resnets.1.norm2/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
1320
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
1321
+ <input>
1322
+ <port id="0" precision="FP32">
1323
+ <dim>-1</dim>
1324
+ <dim>64</dim>
1325
+ <dim>-1</dim>
1326
+ <dim>-1</dim>
1327
+ </port>
1328
+ <port id="1" precision="FP32">
1329
+ <dim>64</dim>
1330
+ </port>
1331
+ <port id="2" precision="FP32">
1332
+ <dim>64</dim>
1333
+ </port>
1334
+ </input>
1335
+ <output>
1336
+ <port id="3" precision="FP32" names="175,input.21">
1337
+ <dim>-1</dim>
1338
+ <dim>64</dim>
1339
+ <dim>-1</dim>
1340
+ <dim>-1</dim>
1341
+ </port>
1342
+ </output>
1343
+ </layer>
1344
+ <layer id="84" name="__module.decoder.mid_block.resnets.1.nonlinearity/aten::silu/Swish_3" type="Swish" version="opset4">
1345
+ <input>
1346
+ <port id="0" precision="FP32">
1347
+ <dim>-1</dim>
1348
+ <dim>64</dim>
1349
+ <dim>-1</dim>
1350
+ <dim>-1</dim>
1351
+ </port>
1352
+ </input>
1353
+ <output>
1354
+ <port id="1" precision="FP32" names="176,input.23">
1355
+ <dim>-1</dim>
1356
+ <dim>64</dim>
1357
+ <dim>-1</dim>
1358
+ <dim>-1</dim>
1359
+ </port>
1360
+ </output>
1361
+ </layer>
1362
+ <layer id="85" name="self.decoder.mid_block.resnets.1.conv2.weight" type="Const" version="opset1">
1363
+ <data element_type="f32" shape="64, 64, 3, 3" offset="519916" size="147456" />
1364
+ <output>
1365
+ <port id="0" precision="FP32" names="self.decoder.mid_block.resnets.1.conv2.weight">
1366
+ <dim>64</dim>
1367
+ <dim>64</dim>
1368
+ <dim>3</dim>
1369
+ <dim>3</dim>
1370
+ </port>
1371
+ </output>
1372
+ </layer>
1373
+ <layer id="86" name="__module.decoder.mid_block.resnets.1.conv2/aten::_convolution/Convolution" type="Convolution" version="opset1">
1374
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
1375
+ <input>
1376
+ <port id="0" precision="FP32">
1377
+ <dim>-1</dim>
1378
+ <dim>64</dim>
1379
+ <dim>-1</dim>
1380
+ <dim>-1</dim>
1381
+ </port>
1382
+ <port id="1" precision="FP32">
1383
+ <dim>64</dim>
1384
+ <dim>64</dim>
1385
+ <dim>3</dim>
1386
+ <dim>3</dim>
1387
+ </port>
1388
+ </input>
1389
+ <output>
1390
+ <port id="2" precision="FP32">
1391
+ <dim>-1</dim>
1392
+ <dim>64</dim>
1393
+ <dim>-1</dim>
1394
+ <dim>-1</dim>
1395
+ </port>
1396
+ </output>
1397
+ </layer>
1398
+ <layer id="87" name="__module.decoder.mid_block.resnets.1.conv2/aten::_convolution/Reshape" type="Const" version="opset1">
1399
+ <data element_type="f32" shape="1, 64, 1, 1" offset="667372" size="256" />
1400
+ <output>
1401
+ <port id="0" precision="FP32">
1402
+ <dim>1</dim>
1403
+ <dim>64</dim>
1404
+ <dim>1</dim>
1405
+ <dim>1</dim>
1406
+ </port>
1407
+ </output>
1408
+ </layer>
1409
+ <layer id="88" name="__module.decoder.mid_block.resnets.1.conv2/aten::_convolution/Add" type="Add" version="opset1">
1410
+ <data auto_broadcast="numpy" />
1411
+ <input>
1412
+ <port id="0" precision="FP32">
1413
+ <dim>-1</dim>
1414
+ <dim>64</dim>
1415
+ <dim>-1</dim>
1416
+ <dim>-1</dim>
1417
+ </port>
1418
+ <port id="1" precision="FP32">
1419
+ <dim>1</dim>
1420
+ <dim>64</dim>
1421
+ <dim>1</dim>
1422
+ <dim>1</dim>
1423
+ </port>
1424
+ </input>
1425
+ <output>
1426
+ <port id="2" precision="FP32" names="184,hidden_states.17">
1427
+ <dim>-1</dim>
1428
+ <dim>64</dim>
1429
+ <dim>-1</dim>
1430
+ <dim>-1</dim>
1431
+ </port>
1432
+ </output>
1433
+ </layer>
1434
+ <layer id="89" name="__module.decoder.mid_block.resnets.1/aten::add/Add" type="Add" version="opset1">
1435
+ <data auto_broadcast="numpy" />
1436
+ <input>
1437
+ <port id="0" precision="FP32">
1438
+ <dim>-1</dim>
1439
+ <dim>64</dim>
1440
+ <dim>-1</dim>
1441
+ <dim>-1</dim>
1442
+ </port>
1443
+ <port id="1" precision="FP32">
1444
+ <dim>-1</dim>
1445
+ <dim>64</dim>
1446
+ <dim>-1</dim>
1447
+ <dim>-1</dim>
1448
+ </port>
1449
+ </input>
1450
+ <output>
1451
+ <port id="2" precision="FP32" names="185,186,187,input.25,sample">
1452
+ <dim>-1</dim>
1453
+ <dim>64</dim>
1454
+ <dim>-1</dim>
1455
+ <dim>-1</dim>
1456
+ </port>
1457
+ </output>
1458
+ </layer>
1459
+ <layer id="90" name="self.decoder.up_blocks.0.resnets.0.norm1.weight" type="Const" version="opset1">
1460
+ <data element_type="f32" shape="64" offset="9552" size="256" />
1461
+ <output>
1462
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.0.resnets.0.norm1.weight">
1463
+ <dim>64</dim>
1464
+ </port>
1465
+ </output>
1466
+ </layer>
1467
+ <layer id="91" name="self.decoder.up_blocks.0.resnets.0.norm1.bias" type="Const" version="opset1">
1468
+ <data element_type="f32" shape="64" offset="9808" size="256" />
1469
+ <output>
1470
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.0.resnets.0.norm1.bias">
1471
+ <dim>64</dim>
1472
+ </port>
1473
+ </output>
1474
+ </layer>
1475
+ <layer id="92" name="__module.decoder.up_blocks.0.resnets.0.norm1/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
1476
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
1477
+ <input>
1478
+ <port id="0" precision="FP32">
1479
+ <dim>-1</dim>
1480
+ <dim>64</dim>
1481
+ <dim>-1</dim>
1482
+ <dim>-1</dim>
1483
+ </port>
1484
+ <port id="1" precision="FP32">
1485
+ <dim>64</dim>
1486
+ </port>
1487
+ <port id="2" precision="FP32">
1488
+ <dim>64</dim>
1489
+ </port>
1490
+ </input>
1491
+ <output>
1492
+ <port id="3" precision="FP32" names="201,input.27">
1493
+ <dim>-1</dim>
1494
+ <dim>64</dim>
1495
+ <dim>-1</dim>
1496
+ <dim>-1</dim>
1497
+ </port>
1498
+ </output>
1499
+ </layer>
1500
+ <layer id="93" name="__module.decoder.mid_block.resnets.1.nonlinearity/aten::silu/Swish_4" type="Swish" version="opset4">
1501
+ <input>
1502
+ <port id="0" precision="FP32">
1503
+ <dim>-1</dim>
1504
+ <dim>64</dim>
1505
+ <dim>-1</dim>
1506
+ <dim>-1</dim>
1507
+ </port>
1508
+ </input>
1509
+ <output>
1510
+ <port id="1" precision="FP32" names="202">
1511
+ <dim>-1</dim>
1512
+ <dim>64</dim>
1513
+ <dim>-1</dim>
1514
+ <dim>-1</dim>
1515
+ </port>
1516
+ </output>
1517
+ </layer>
1518
+ <layer id="94" name="self.decoder.up_blocks.0.resnets.0.conv1.weight" type="Const" version="opset1">
1519
+ <data element_type="f32" shape="64, 64, 3, 3" offset="667628" size="147456" />
1520
+ <output>
1521
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.0.resnets.0.conv1.weight">
1522
+ <dim>64</dim>
1523
+ <dim>64</dim>
1524
+ <dim>3</dim>
1525
+ <dim>3</dim>
1526
+ </port>
1527
+ </output>
1528
+ </layer>
1529
+ <layer id="95" name="__module.decoder.up_blocks.0.resnets.0.conv1/aten::_convolution/Convolution" type="Convolution" version="opset1">
1530
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
1531
+ <input>
1532
+ <port id="0" precision="FP32">
1533
+ <dim>-1</dim>
1534
+ <dim>64</dim>
1535
+ <dim>-1</dim>
1536
+ <dim>-1</dim>
1537
+ </port>
1538
+ <port id="1" precision="FP32">
1539
+ <dim>64</dim>
1540
+ <dim>64</dim>
1541
+ <dim>3</dim>
1542
+ <dim>3</dim>
1543
+ </port>
1544
+ </input>
1545
+ <output>
1546
+ <port id="2" precision="FP32">
1547
+ <dim>-1</dim>
1548
+ <dim>64</dim>
1549
+ <dim>-1</dim>
1550
+ <dim>-1</dim>
1551
+ </port>
1552
+ </output>
1553
+ </layer>
1554
+ <layer id="96" name="__module.decoder.up_blocks.0.resnets.0.conv1/aten::_convolution/Reshape" type="Const" version="opset1">
1555
+ <data element_type="f32" shape="1, 64, 1, 1" offset="815084" size="256" />
1556
+ <output>
1557
+ <port id="0" precision="FP32">
1558
+ <dim>1</dim>
1559
+ <dim>64</dim>
1560
+ <dim>1</dim>
1561
+ <dim>1</dim>
1562
+ </port>
1563
+ </output>
1564
+ </layer>
1565
+ <layer id="97" name="__module.decoder.up_blocks.0.resnets.0.conv1/aten::_convolution/Add" type="Add" version="opset1">
1566
+ <data auto_broadcast="numpy" />
1567
+ <input>
1568
+ <port id="0" precision="FP32">
1569
+ <dim>-1</dim>
1570
+ <dim>64</dim>
1571
+ <dim>-1</dim>
1572
+ <dim>-1</dim>
1573
+ </port>
1574
+ <port id="1" precision="FP32">
1575
+ <dim>1</dim>
1576
+ <dim>64</dim>
1577
+ <dim>1</dim>
1578
+ <dim>1</dim>
1579
+ </port>
1580
+ </input>
1581
+ <output>
1582
+ <port id="2" precision="FP32" names="209,input.29">
1583
+ <dim>-1</dim>
1584
+ <dim>64</dim>
1585
+ <dim>-1</dim>
1586
+ <dim>-1</dim>
1587
+ </port>
1588
+ </output>
1589
+ </layer>
1590
+ <layer id="98" name="self.decoder.up_blocks.0.resnets.0.norm2.weight" type="Const" version="opset1">
1591
+ <data element_type="f32" shape="64" offset="9552" size="256" />
1592
+ <output>
1593
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.0.resnets.0.norm2.weight">
1594
+ <dim>64</dim>
1595
+ </port>
1596
+ </output>
1597
+ </layer>
1598
+ <layer id="99" name="self.decoder.up_blocks.0.resnets.0.norm2.bias" type="Const" version="opset1">
1599
+ <data element_type="f32" shape="64" offset="9808" size="256" />
1600
+ <output>
1601
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.0.resnets.0.norm2.bias">
1602
+ <dim>64</dim>
1603
+ </port>
1604
+ </output>
1605
+ </layer>
1606
+ <layer id="100" name="__module.decoder.up_blocks.0.resnets.0.norm2/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
1607
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
1608
+ <input>
1609
+ <port id="0" precision="FP32">
1610
+ <dim>-1</dim>
1611
+ <dim>64</dim>
1612
+ <dim>-1</dim>
1613
+ <dim>-1</dim>
1614
+ </port>
1615
+ <port id="1" precision="FP32">
1616
+ <dim>64</dim>
1617
+ </port>
1618
+ <port id="2" precision="FP32">
1619
+ <dim>64</dim>
1620
+ </port>
1621
+ </input>
1622
+ <output>
1623
+ <port id="3" precision="FP32" names="212,input.31">
1624
+ <dim>-1</dim>
1625
+ <dim>64</dim>
1626
+ <dim>-1</dim>
1627
+ <dim>-1</dim>
1628
+ </port>
1629
+ </output>
1630
+ </layer>
1631
+ <layer id="101" name="__module.decoder.mid_block.resnets.1.nonlinearity/aten::silu/Swish_5" type="Swish" version="opset4">
1632
+ <input>
1633
+ <port id="0" precision="FP32">
1634
+ <dim>-1</dim>
1635
+ <dim>64</dim>
1636
+ <dim>-1</dim>
1637
+ <dim>-1</dim>
1638
+ </port>
1639
+ </input>
1640
+ <output>
1641
+ <port id="1" precision="FP32" names="213,input.33">
1642
+ <dim>-1</dim>
1643
+ <dim>64</dim>
1644
+ <dim>-1</dim>
1645
+ <dim>-1</dim>
1646
+ </port>
1647
+ </output>
1648
+ </layer>
1649
+ <layer id="102" name="self.decoder.up_blocks.0.resnets.0.conv2.weight" type="Const" version="opset1">
1650
+ <data element_type="f32" shape="64, 64, 3, 3" offset="815340" size="147456" />
1651
+ <output>
1652
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.0.resnets.0.conv2.weight">
1653
+ <dim>64</dim>
1654
+ <dim>64</dim>
1655
+ <dim>3</dim>
1656
+ <dim>3</dim>
1657
+ </port>
1658
+ </output>
1659
+ </layer>
1660
+ <layer id="103" name="__module.decoder.up_blocks.0.resnets.0.conv2/aten::_convolution/Convolution" type="Convolution" version="opset1">
1661
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
1662
+ <input>
1663
+ <port id="0" precision="FP32">
1664
+ <dim>-1</dim>
1665
+ <dim>64</dim>
1666
+ <dim>-1</dim>
1667
+ <dim>-1</dim>
1668
+ </port>
1669
+ <port id="1" precision="FP32">
1670
+ <dim>64</dim>
1671
+ <dim>64</dim>
1672
+ <dim>3</dim>
1673
+ <dim>3</dim>
1674
+ </port>
1675
+ </input>
1676
+ <output>
1677
+ <port id="2" precision="FP32">
1678
+ <dim>-1</dim>
1679
+ <dim>64</dim>
1680
+ <dim>-1</dim>
1681
+ <dim>-1</dim>
1682
+ </port>
1683
+ </output>
1684
+ </layer>
1685
+ <layer id="104" name="__module.decoder.up_blocks.0.resnets.0.conv2/aten::_convolution/Reshape" type="Const" version="opset1">
1686
+ <data element_type="f32" shape="1, 64, 1, 1" offset="962796" size="256" />
1687
+ <output>
1688
+ <port id="0" precision="FP32">
1689
+ <dim>1</dim>
1690
+ <dim>64</dim>
1691
+ <dim>1</dim>
1692
+ <dim>1</dim>
1693
+ </port>
1694
+ </output>
1695
+ </layer>
1696
+ <layer id="105" name="__module.decoder.up_blocks.0.resnets.0.conv2/aten::_convolution/Add" type="Add" version="opset1">
1697
+ <data auto_broadcast="numpy" />
1698
+ <input>
1699
+ <port id="0" precision="FP32">
1700
+ <dim>-1</dim>
1701
+ <dim>64</dim>
1702
+ <dim>-1</dim>
1703
+ <dim>-1</dim>
1704
+ </port>
1705
+ <port id="1" precision="FP32">
1706
+ <dim>1</dim>
1707
+ <dim>64</dim>
1708
+ <dim>1</dim>
1709
+ <dim>1</dim>
1710
+ </port>
1711
+ </input>
1712
+ <output>
1713
+ <port id="2" precision="FP32" names="221,hidden_states.19">
1714
+ <dim>-1</dim>
1715
+ <dim>64</dim>
1716
+ <dim>-1</dim>
1717
+ <dim>-1</dim>
1718
+ </port>
1719
+ </output>
1720
+ </layer>
1721
+ <layer id="106" name="__module.decoder.up_blocks.0.resnets.0/aten::add/Add" type="Add" version="opset1">
1722
+ <data auto_broadcast="numpy" />
1723
+ <input>
1724
+ <port id="0" precision="FP32">
1725
+ <dim>-1</dim>
1726
+ <dim>64</dim>
1727
+ <dim>-1</dim>
1728
+ <dim>-1</dim>
1729
+ </port>
1730
+ <port id="1" precision="FP32">
1731
+ <dim>-1</dim>
1732
+ <dim>64</dim>
1733
+ <dim>-1</dim>
1734
+ <dim>-1</dim>
1735
+ </port>
1736
+ </input>
1737
+ <output>
1738
+ <port id="2" precision="FP32" names="222,223,input.35">
1739
+ <dim>-1</dim>
1740
+ <dim>64</dim>
1741
+ <dim>-1</dim>
1742
+ <dim>-1</dim>
1743
+ </port>
1744
+ </output>
1745
+ </layer>
1746
+ <layer id="107" name="self.decoder.up_blocks.0.resnets.1.norm1.weight" type="Const" version="opset1">
1747
+ <data element_type="f32" shape="64" offset="9552" size="256" />
1748
+ <output>
1749
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.0.resnets.1.norm1.weight">
1750
+ <dim>64</dim>
1751
+ </port>
1752
+ </output>
1753
+ </layer>
1754
+ <layer id="108" name="self.decoder.up_blocks.0.resnets.1.norm1.bias" type="Const" version="opset1">
1755
+ <data element_type="f32" shape="64" offset="9808" size="256" />
1756
+ <output>
1757
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.0.resnets.1.norm1.bias">
1758
+ <dim>64</dim>
1759
+ </port>
1760
+ </output>
1761
+ </layer>
1762
+ <layer id="109" name="__module.decoder.up_blocks.0.resnets.1.norm1/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
1763
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
1764
+ <input>
1765
+ <port id="0" precision="FP32">
1766
+ <dim>-1</dim>
1767
+ <dim>64</dim>
1768
+ <dim>-1</dim>
1769
+ <dim>-1</dim>
1770
+ </port>
1771
+ <port id="1" precision="FP32">
1772
+ <dim>64</dim>
1773
+ </port>
1774
+ <port id="2" precision="FP32">
1775
+ <dim>64</dim>
1776
+ </port>
1777
+ </input>
1778
+ <output>
1779
+ <port id="3" precision="FP32" names="231,input.37">
1780
+ <dim>-1</dim>
1781
+ <dim>64</dim>
1782
+ <dim>-1</dim>
1783
+ <dim>-1</dim>
1784
+ </port>
1785
+ </output>
1786
+ </layer>
1787
+ <layer id="110" name="__module.decoder.mid_block.resnets.1.nonlinearity/aten::silu/Swish_6" type="Swish" version="opset4">
1788
+ <input>
1789
+ <port id="0" precision="FP32">
1790
+ <dim>-1</dim>
1791
+ <dim>64</dim>
1792
+ <dim>-1</dim>
1793
+ <dim>-1</dim>
1794
+ </port>
1795
+ </input>
1796
+ <output>
1797
+ <port id="1" precision="FP32" names="232">
1798
+ <dim>-1</dim>
1799
+ <dim>64</dim>
1800
+ <dim>-1</dim>
1801
+ <dim>-1</dim>
1802
+ </port>
1803
+ </output>
1804
+ </layer>
1805
+ <layer id="111" name="self.decoder.up_blocks.0.resnets.1.conv1.weight" type="Const" version="opset1">
1806
+ <data element_type="f32" shape="64, 64, 3, 3" offset="963052" size="147456" />
1807
+ <output>
1808
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.0.resnets.1.conv1.weight">
1809
+ <dim>64</dim>
1810
+ <dim>64</dim>
1811
+ <dim>3</dim>
1812
+ <dim>3</dim>
1813
+ </port>
1814
+ </output>
1815
+ </layer>
1816
+ <layer id="112" name="__module.decoder.up_blocks.0.resnets.1.conv1/aten::_convolution/Convolution" type="Convolution" version="opset1">
1817
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
1818
+ <input>
1819
+ <port id="0" precision="FP32">
1820
+ <dim>-1</dim>
1821
+ <dim>64</dim>
1822
+ <dim>-1</dim>
1823
+ <dim>-1</dim>
1824
+ </port>
1825
+ <port id="1" precision="FP32">
1826
+ <dim>64</dim>
1827
+ <dim>64</dim>
1828
+ <dim>3</dim>
1829
+ <dim>3</dim>
1830
+ </port>
1831
+ </input>
1832
+ <output>
1833
+ <port id="2" precision="FP32">
1834
+ <dim>-1</dim>
1835
+ <dim>64</dim>
1836
+ <dim>-1</dim>
1837
+ <dim>-1</dim>
1838
+ </port>
1839
+ </output>
1840
+ </layer>
1841
+ <layer id="113" name="__module.decoder.up_blocks.0.resnets.1.conv1/aten::_convolution/Reshape" type="Const" version="opset1">
1842
+ <data element_type="f32" shape="1, 64, 1, 1" offset="1110508" size="256" />
1843
+ <output>
1844
+ <port id="0" precision="FP32">
1845
+ <dim>1</dim>
1846
+ <dim>64</dim>
1847
+ <dim>1</dim>
1848
+ <dim>1</dim>
1849
+ </port>
1850
+ </output>
1851
+ </layer>
1852
+ <layer id="114" name="__module.decoder.up_blocks.0.resnets.1.conv1/aten::_convolution/Add" type="Add" version="opset1">
1853
+ <data auto_broadcast="numpy" />
1854
+ <input>
1855
+ <port id="0" precision="FP32">
1856
+ <dim>-1</dim>
1857
+ <dim>64</dim>
1858
+ <dim>-1</dim>
1859
+ <dim>-1</dim>
1860
+ </port>
1861
+ <port id="1" precision="FP32">
1862
+ <dim>1</dim>
1863
+ <dim>64</dim>
1864
+ <dim>1</dim>
1865
+ <dim>1</dim>
1866
+ </port>
1867
+ </input>
1868
+ <output>
1869
+ <port id="2" precision="FP32" names="239,input.39">
1870
+ <dim>-1</dim>
1871
+ <dim>64</dim>
1872
+ <dim>-1</dim>
1873
+ <dim>-1</dim>
1874
+ </port>
1875
+ </output>
1876
+ </layer>
1877
+ <layer id="115" name="self.decoder.up_blocks.0.resnets.1.norm2.weight" type="Const" version="opset1">
1878
+ <data element_type="f32" shape="64" offset="9552" size="256" />
1879
+ <output>
1880
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.0.resnets.1.norm2.weight">
1881
+ <dim>64</dim>
1882
+ </port>
1883
+ </output>
1884
+ </layer>
1885
+ <layer id="116" name="self.decoder.up_blocks.0.resnets.1.norm2.bias" type="Const" version="opset1">
1886
+ <data element_type="f32" shape="64" offset="9808" size="256" />
1887
+ <output>
1888
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.0.resnets.1.norm2.bias">
1889
+ <dim>64</dim>
1890
+ </port>
1891
+ </output>
1892
+ </layer>
1893
+ <layer id="117" name="__module.decoder.up_blocks.0.resnets.1.norm2/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
1894
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
1895
+ <input>
1896
+ <port id="0" precision="FP32">
1897
+ <dim>-1</dim>
1898
+ <dim>64</dim>
1899
+ <dim>-1</dim>
1900
+ <dim>-1</dim>
1901
+ </port>
1902
+ <port id="1" precision="FP32">
1903
+ <dim>64</dim>
1904
+ </port>
1905
+ <port id="2" precision="FP32">
1906
+ <dim>64</dim>
1907
+ </port>
1908
+ </input>
1909
+ <output>
1910
+ <port id="3" precision="FP32" names="242,input.41">
1911
+ <dim>-1</dim>
1912
+ <dim>64</dim>
1913
+ <dim>-1</dim>
1914
+ <dim>-1</dim>
1915
+ </port>
1916
+ </output>
1917
+ </layer>
1918
+ <layer id="118" name="__module.decoder.mid_block.resnets.1.nonlinearity/aten::silu/Swish_7" type="Swish" version="opset4">
1919
+ <input>
1920
+ <port id="0" precision="FP32">
1921
+ <dim>-1</dim>
1922
+ <dim>64</dim>
1923
+ <dim>-1</dim>
1924
+ <dim>-1</dim>
1925
+ </port>
1926
+ </input>
1927
+ <output>
1928
+ <port id="1" precision="FP32" names="243,input.43">
1929
+ <dim>-1</dim>
1930
+ <dim>64</dim>
1931
+ <dim>-1</dim>
1932
+ <dim>-1</dim>
1933
+ </port>
1934
+ </output>
1935
+ </layer>
1936
+ <layer id="119" name="self.decoder.up_blocks.0.resnets.1.conv2.weight" type="Const" version="opset1">
1937
+ <data element_type="f32" shape="64, 64, 3, 3" offset="1110764" size="147456" />
1938
+ <output>
1939
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.0.resnets.1.conv2.weight">
1940
+ <dim>64</dim>
1941
+ <dim>64</dim>
1942
+ <dim>3</dim>
1943
+ <dim>3</dim>
1944
+ </port>
1945
+ </output>
1946
+ </layer>
1947
+ <layer id="120" name="__module.decoder.up_blocks.0.resnets.1.conv2/aten::_convolution/Convolution" type="Convolution" version="opset1">
1948
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
1949
+ <input>
1950
+ <port id="0" precision="FP32">
1951
+ <dim>-1</dim>
1952
+ <dim>64</dim>
1953
+ <dim>-1</dim>
1954
+ <dim>-1</dim>
1955
+ </port>
1956
+ <port id="1" precision="FP32">
1957
+ <dim>64</dim>
1958
+ <dim>64</dim>
1959
+ <dim>3</dim>
1960
+ <dim>3</dim>
1961
+ </port>
1962
+ </input>
1963
+ <output>
1964
+ <port id="2" precision="FP32">
1965
+ <dim>-1</dim>
1966
+ <dim>64</dim>
1967
+ <dim>-1</dim>
1968
+ <dim>-1</dim>
1969
+ </port>
1970
+ </output>
1971
+ </layer>
1972
+ <layer id="121" name="__module.decoder.up_blocks.0.resnets.1.conv2/aten::_convolution/Reshape" type="Const" version="opset1">
1973
+ <data element_type="f32" shape="1, 64, 1, 1" offset="1258220" size="256" />
1974
+ <output>
1975
+ <port id="0" precision="FP32">
1976
+ <dim>1</dim>
1977
+ <dim>64</dim>
1978
+ <dim>1</dim>
1979
+ <dim>1</dim>
1980
+ </port>
1981
+ </output>
1982
+ </layer>
1983
+ <layer id="122" name="__module.decoder.up_blocks.0.resnets.1.conv2/aten::_convolution/Add" type="Add" version="opset1">
1984
+ <data auto_broadcast="numpy" />
1985
+ <input>
1986
+ <port id="0" precision="FP32">
1987
+ <dim>-1</dim>
1988
+ <dim>64</dim>
1989
+ <dim>-1</dim>
1990
+ <dim>-1</dim>
1991
+ </port>
1992
+ <port id="1" precision="FP32">
1993
+ <dim>1</dim>
1994
+ <dim>64</dim>
1995
+ <dim>1</dim>
1996
+ <dim>1</dim>
1997
+ </port>
1998
+ </input>
1999
+ <output>
2000
+ <port id="2" precision="FP32" names="251,hidden_states.21">
2001
+ <dim>-1</dim>
2002
+ <dim>64</dim>
2003
+ <dim>-1</dim>
2004
+ <dim>-1</dim>
2005
+ </port>
2006
+ </output>
2007
+ </layer>
2008
+ <layer id="123" name="__module.decoder.up_blocks.0.resnets.1/aten::add/Add" type="Add" version="opset1">
2009
+ <data auto_broadcast="numpy" />
2010
+ <input>
2011
+ <port id="0" precision="FP32">
2012
+ <dim>-1</dim>
2013
+ <dim>64</dim>
2014
+ <dim>-1</dim>
2015
+ <dim>-1</dim>
2016
+ </port>
2017
+ <port id="1" precision="FP32">
2018
+ <dim>-1</dim>
2019
+ <dim>64</dim>
2020
+ <dim>-1</dim>
2021
+ <dim>-1</dim>
2022
+ </port>
2023
+ </input>
2024
+ <output>
2025
+ <port id="2" precision="FP32" names="252,253,hidden_states.23">
2026
+ <dim>-1</dim>
2027
+ <dim>64</dim>
2028
+ <dim>-1</dim>
2029
+ <dim>-1</dim>
2030
+ </port>
2031
+ </output>
2032
+ </layer>
2033
+ <layer id="124" name="__module.decoder.up_blocks.0.upsamplers.0/aten::upsample_nearest2d/Multiply" type="Const" version="opset1">
2034
+ <data element_type="f32" shape="2" offset="1258476" size="8" />
2035
+ <output>
2036
+ <port id="0" precision="FP32">
2037
+ <dim>2</dim>
2038
+ </port>
2039
+ </output>
2040
+ </layer>
2041
+ <layer id="125" name="Constant_50364" type="Const" version="opset1">
2042
+ <data element_type="i32" shape="2" offset="1258484" size="8" />
2043
+ <output>
2044
+ <port id="0" precision="I32">
2045
+ <dim>2</dim>
2046
+ </port>
2047
+ </output>
2048
+ </layer>
2049
+ <layer id="126" name="__module.decoder.up_blocks.0.upsamplers.0/aten::upsample_nearest2d/Interpolate" type="Interpolate" version="opset11">
2050
+ <data mode="nearest" shape_calculation_mode="scales" coordinate_transformation_mode="asymmetric" nearest_mode="floor" antialias="false" pads_begin="0, 0, 0, 0" pads_end="0, 0, 0, 0" cube_coeff="-0.75" />
2051
+ <input>
2052
+ <port id="0" precision="FP32">
2053
+ <dim>-1</dim>
2054
+ <dim>64</dim>
2055
+ <dim>-1</dim>
2056
+ <dim>-1</dim>
2057
+ </port>
2058
+ <port id="1" precision="FP32">
2059
+ <dim>2</dim>
2060
+ </port>
2061
+ <port id="2" precision="I32">
2062
+ <dim>2</dim>
2063
+ </port>
2064
+ </input>
2065
+ <output>
2066
+ <port id="3" precision="FP32" names="256">
2067
+ <dim>-1</dim>
2068
+ <dim>64</dim>
2069
+ <dim>-1</dim>
2070
+ <dim>-1</dim>
2071
+ </port>
2072
+ </output>
2073
+ </layer>
2074
+ <layer id="127" name="self.decoder.up_blocks.0.upsamplers.0.conv.weight" type="Const" version="opset1">
2075
+ <data element_type="f32" shape="64, 64, 3, 3" offset="1258492" size="147456" />
2076
+ <output>
2077
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.0.upsamplers.0.conv.weight">
2078
+ <dim>64</dim>
2079
+ <dim>64</dim>
2080
+ <dim>3</dim>
2081
+ <dim>3</dim>
2082
+ </port>
2083
+ </output>
2084
+ </layer>
2085
+ <layer id="128" name="__module.decoder.up_blocks.0.upsamplers.0.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
2086
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
2087
+ <input>
2088
+ <port id="0" precision="FP32">
2089
+ <dim>-1</dim>
2090
+ <dim>64</dim>
2091
+ <dim>-1</dim>
2092
+ <dim>-1</dim>
2093
+ </port>
2094
+ <port id="1" precision="FP32">
2095
+ <dim>64</dim>
2096
+ <dim>64</dim>
2097
+ <dim>3</dim>
2098
+ <dim>3</dim>
2099
+ </port>
2100
+ </input>
2101
+ <output>
2102
+ <port id="2" precision="FP32">
2103
+ <dim>-1</dim>
2104
+ <dim>64</dim>
2105
+ <dim>-1</dim>
2106
+ <dim>-1</dim>
2107
+ </port>
2108
+ </output>
2109
+ </layer>
2110
+ <layer id="129" name="__module.decoder.up_blocks.0.upsamplers.0.conv/aten::_convolution/Reshape" type="Const" version="opset1">
2111
+ <data element_type="f32" shape="1, 64, 1, 1" offset="1405948" size="256" />
2112
+ <output>
2113
+ <port id="0" precision="FP32">
2114
+ <dim>1</dim>
2115
+ <dim>64</dim>
2116
+ <dim>1</dim>
2117
+ <dim>1</dim>
2118
+ </port>
2119
+ </output>
2120
+ </layer>
2121
+ <layer id="130" name="__module.decoder.up_blocks.0.upsamplers.0.conv/aten::_convolution/Add" type="Add" version="opset1">
2122
+ <data auto_broadcast="numpy" />
2123
+ <input>
2124
+ <port id="0" precision="FP32">
2125
+ <dim>-1</dim>
2126
+ <dim>64</dim>
2127
+ <dim>-1</dim>
2128
+ <dim>-1</dim>
2129
+ </port>
2130
+ <port id="1" precision="FP32">
2131
+ <dim>1</dim>
2132
+ <dim>64</dim>
2133
+ <dim>1</dim>
2134
+ <dim>1</dim>
2135
+ </port>
2136
+ </input>
2137
+ <output>
2138
+ <port id="2" precision="FP32" names="263,input.45">
2139
+ <dim>-1</dim>
2140
+ <dim>64</dim>
2141
+ <dim>-1</dim>
2142
+ <dim>-1</dim>
2143
+ </port>
2144
+ </output>
2145
+ </layer>
2146
+ <layer id="131" name="self.decoder.up_blocks.1.resnets.0.conv_shortcut.weight" type="Const" version="opset1">
2147
+ <data element_type="f32" shape="32, 64, 1, 1" offset="1406204" size="8192" />
2148
+ <output>
2149
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.1.resnets.0.conv_shortcut.weight">
2150
+ <dim>32</dim>
2151
+ <dim>64</dim>
2152
+ <dim>1</dim>
2153
+ <dim>1</dim>
2154
+ </port>
2155
+ </output>
2156
+ </layer>
2157
+ <layer id="132" name="__module.decoder.up_blocks.1.resnets.0.conv_shortcut/aten::_convolution/Convolution" type="Convolution" version="opset1">
2158
+ <data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
2159
+ <input>
2160
+ <port id="0" precision="FP32">
2161
+ <dim>-1</dim>
2162
+ <dim>64</dim>
2163
+ <dim>-1</dim>
2164
+ <dim>-1</dim>
2165
+ </port>
2166
+ <port id="1" precision="FP32">
2167
+ <dim>32</dim>
2168
+ <dim>64</dim>
2169
+ <dim>1</dim>
2170
+ <dim>1</dim>
2171
+ </port>
2172
+ </input>
2173
+ <output>
2174
+ <port id="2" precision="FP32">
2175
+ <dim>-1</dim>
2176
+ <dim>32</dim>
2177
+ <dim>-1</dim>
2178
+ <dim>-1</dim>
2179
+ </port>
2180
+ </output>
2181
+ </layer>
2182
+ <layer id="133" name="__module.decoder.up_blocks.1.resnets.0.conv_shortcut/aten::_convolution/Reshape" type="Const" version="opset1">
2183
+ <data element_type="f32" shape="1, 32, 1, 1" offset="1414396" size="128" />
2184
+ <output>
2185
+ <port id="0" precision="FP32">
2186
+ <dim>1</dim>
2187
+ <dim>32</dim>
2188
+ <dim>1</dim>
2189
+ <dim>1</dim>
2190
+ </port>
2191
+ </output>
2192
+ </layer>
2193
+ <layer id="134" name="__module.decoder.up_blocks.1.resnets.0.conv_shortcut/aten::_convolution/Add" type="Add" version="opset1">
2194
+ <data auto_broadcast="numpy" />
2195
+ <input>
2196
+ <port id="0" precision="FP32">
2197
+ <dim>-1</dim>
2198
+ <dim>32</dim>
2199
+ <dim>-1</dim>
2200
+ <dim>-1</dim>
2201
+ </port>
2202
+ <port id="1" precision="FP32">
2203
+ <dim>1</dim>
2204
+ <dim>32</dim>
2205
+ <dim>1</dim>
2206
+ <dim>1</dim>
2207
+ </port>
2208
+ </input>
2209
+ <output>
2210
+ <port id="2" precision="FP32" names="303,input_tensor">
2211
+ <dim>-1</dim>
2212
+ <dim>32</dim>
2213
+ <dim>-1</dim>
2214
+ <dim>-1</dim>
2215
+ </port>
2216
+ </output>
2217
+ </layer>
2218
+ <layer id="135" name="self.decoder.up_blocks.1.resnets.0.norm1.weight" type="Const" version="opset1">
2219
+ <data element_type="f32" shape="64" offset="9552" size="256" />
2220
+ <output>
2221
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.1.resnets.0.norm1.weight">
2222
+ <dim>64</dim>
2223
+ </port>
2224
+ </output>
2225
+ </layer>
2226
+ <layer id="136" name="self.decoder.up_blocks.1.resnets.0.norm1.bias" type="Const" version="opset1">
2227
+ <data element_type="f32" shape="64" offset="9808" size="256" />
2228
+ <output>
2229
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.1.resnets.0.norm1.bias">
2230
+ <dim>64</dim>
2231
+ </port>
2232
+ </output>
2233
+ </layer>
2234
+ <layer id="137" name="__module.decoder.up_blocks.1.resnets.0.norm1/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
2235
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
2236
+ <input>
2237
+ <port id="0" precision="FP32">
2238
+ <dim>-1</dim>
2239
+ <dim>64</dim>
2240
+ <dim>-1</dim>
2241
+ <dim>-1</dim>
2242
+ </port>
2243
+ <port id="1" precision="FP32">
2244
+ <dim>64</dim>
2245
+ </port>
2246
+ <port id="2" precision="FP32">
2247
+ <dim>64</dim>
2248
+ </port>
2249
+ </input>
2250
+ <output>
2251
+ <port id="3" precision="FP32" names="276,input.47">
2252
+ <dim>-1</dim>
2253
+ <dim>64</dim>
2254
+ <dim>-1</dim>
2255
+ <dim>-1</dim>
2256
+ </port>
2257
+ </output>
2258
+ </layer>
2259
+ <layer id="138" name="__module.decoder.mid_block.resnets.1.nonlinearity/aten::silu/Swish_8" type="Swish" version="opset4">
2260
+ <input>
2261
+ <port id="0" precision="FP32">
2262
+ <dim>-1</dim>
2263
+ <dim>64</dim>
2264
+ <dim>-1</dim>
2265
+ <dim>-1</dim>
2266
+ </port>
2267
+ </input>
2268
+ <output>
2269
+ <port id="1" precision="FP32" names="277">
2270
+ <dim>-1</dim>
2271
+ <dim>64</dim>
2272
+ <dim>-1</dim>
2273
+ <dim>-1</dim>
2274
+ </port>
2275
+ </output>
2276
+ </layer>
2277
+ <layer id="139" name="self.decoder.up_blocks.1.resnets.0.conv1.weight" type="Const" version="opset1">
2278
+ <data element_type="f32" shape="32, 64, 3, 3" offset="1414524" size="73728" />
2279
+ <output>
2280
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.1.resnets.0.conv1.weight">
2281
+ <dim>32</dim>
2282
+ <dim>64</dim>
2283
+ <dim>3</dim>
2284
+ <dim>3</dim>
2285
+ </port>
2286
+ </output>
2287
+ </layer>
2288
+ <layer id="140" name="__module.decoder.up_blocks.1.resnets.0.conv1/aten::_convolution/Convolution" type="Convolution" version="opset1">
2289
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
2290
+ <input>
2291
+ <port id="0" precision="FP32">
2292
+ <dim>-1</dim>
2293
+ <dim>64</dim>
2294
+ <dim>-1</dim>
2295
+ <dim>-1</dim>
2296
+ </port>
2297
+ <port id="1" precision="FP32">
2298
+ <dim>32</dim>
2299
+ <dim>64</dim>
2300
+ <dim>3</dim>
2301
+ <dim>3</dim>
2302
+ </port>
2303
+ </input>
2304
+ <output>
2305
+ <port id="2" precision="FP32">
2306
+ <dim>-1</dim>
2307
+ <dim>32</dim>
2308
+ <dim>-1</dim>
2309
+ <dim>-1</dim>
2310
+ </port>
2311
+ </output>
2312
+ </layer>
2313
+ <layer id="141" name="__module.decoder.up_blocks.1.resnets.0.conv1/aten::_convolution/Reshape" type="Const" version="opset1">
2314
+ <data element_type="f32" shape="1, 32, 1, 1" offset="1488252" size="128" />
2315
+ <output>
2316
+ <port id="0" precision="FP32">
2317
+ <dim>1</dim>
2318
+ <dim>32</dim>
2319
+ <dim>1</dim>
2320
+ <dim>1</dim>
2321
+ </port>
2322
+ </output>
2323
+ </layer>
2324
+ <layer id="142" name="__module.decoder.up_blocks.1.resnets.0.conv1/aten::_convolution/Add" type="Add" version="opset1">
2325
+ <data auto_broadcast="numpy" />
2326
+ <input>
2327
+ <port id="0" precision="FP32">
2328
+ <dim>-1</dim>
2329
+ <dim>32</dim>
2330
+ <dim>-1</dim>
2331
+ <dim>-1</dim>
2332
+ </port>
2333
+ <port id="1" precision="FP32">
2334
+ <dim>1</dim>
2335
+ <dim>32</dim>
2336
+ <dim>1</dim>
2337
+ <dim>1</dim>
2338
+ </port>
2339
+ </input>
2340
+ <output>
2341
+ <port id="2" precision="FP32" names="284,input.49">
2342
+ <dim>-1</dim>
2343
+ <dim>32</dim>
2344
+ <dim>-1</dim>
2345
+ <dim>-1</dim>
2346
+ </port>
2347
+ </output>
2348
+ </layer>
2349
+ <layer id="143" name="self.decoder.up_blocks.1.resnets.0.norm2.weight" type="Const" version="opset1">
2350
+ <data element_type="f32" shape="32" offset="1488380" size="128" />
2351
+ <output>
2352
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.1.resnets.0.norm2.weight">
2353
+ <dim>32</dim>
2354
+ </port>
2355
+ </output>
2356
+ </layer>
2357
+ <layer id="144" name="self.decoder.up_blocks.1.resnets.0.norm2.bias" type="Const" version="opset1">
2358
+ <data element_type="f32" shape="32" offset="1488508" size="128" />
2359
+ <output>
2360
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.1.resnets.0.norm2.bias">
2361
+ <dim>32</dim>
2362
+ </port>
2363
+ </output>
2364
+ </layer>
2365
+ <layer id="145" name="__module.decoder.up_blocks.1.resnets.0.norm2/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
2366
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
2367
+ <input>
2368
+ <port id="0" precision="FP32">
2369
+ <dim>-1</dim>
2370
+ <dim>32</dim>
2371
+ <dim>-1</dim>
2372
+ <dim>-1</dim>
2373
+ </port>
2374
+ <port id="1" precision="FP32">
2375
+ <dim>32</dim>
2376
+ </port>
2377
+ <port id="2" precision="FP32">
2378
+ <dim>32</dim>
2379
+ </port>
2380
+ </input>
2381
+ <output>
2382
+ <port id="3" precision="FP32" names="287,input.51">
2383
+ <dim>-1</dim>
2384
+ <dim>32</dim>
2385
+ <dim>-1</dim>
2386
+ <dim>-1</dim>
2387
+ </port>
2388
+ </output>
2389
+ </layer>
2390
+ <layer id="146" name="__module.decoder.mid_block.resnets.1.nonlinearity/aten::silu/Swish_9" type="Swish" version="opset4">
2391
+ <input>
2392
+ <port id="0" precision="FP32">
2393
+ <dim>-1</dim>
2394
+ <dim>32</dim>
2395
+ <dim>-1</dim>
2396
+ <dim>-1</dim>
2397
+ </port>
2398
+ </input>
2399
+ <output>
2400
+ <port id="1" precision="FP32" names="288,input.53">
2401
+ <dim>-1</dim>
2402
+ <dim>32</dim>
2403
+ <dim>-1</dim>
2404
+ <dim>-1</dim>
2405
+ </port>
2406
+ </output>
2407
+ </layer>
2408
+ <layer id="147" name="self.decoder.up_blocks.1.resnets.0.conv2.weight" type="Const" version="opset1">
2409
+ <data element_type="f32" shape="32, 32, 3, 3" offset="1488636" size="36864" />
2410
+ <output>
2411
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.1.resnets.0.conv2.weight">
2412
+ <dim>32</dim>
2413
+ <dim>32</dim>
2414
+ <dim>3</dim>
2415
+ <dim>3</dim>
2416
+ </port>
2417
+ </output>
2418
+ </layer>
2419
+ <layer id="148" name="__module.decoder.up_blocks.1.resnets.0.conv2/aten::_convolution/Convolution" type="Convolution" version="opset1">
2420
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
2421
+ <input>
2422
+ <port id="0" precision="FP32">
2423
+ <dim>-1</dim>
2424
+ <dim>32</dim>
2425
+ <dim>-1</dim>
2426
+ <dim>-1</dim>
2427
+ </port>
2428
+ <port id="1" precision="FP32">
2429
+ <dim>32</dim>
2430
+ <dim>32</dim>
2431
+ <dim>3</dim>
2432
+ <dim>3</dim>
2433
+ </port>
2434
+ </input>
2435
+ <output>
2436
+ <port id="2" precision="FP32">
2437
+ <dim>-1</dim>
2438
+ <dim>32</dim>
2439
+ <dim>-1</dim>
2440
+ <dim>-1</dim>
2441
+ </port>
2442
+ </output>
2443
+ </layer>
2444
+ <layer id="149" name="__module.decoder.up_blocks.1.resnets.0.conv2/aten::_convolution/Reshape" type="Const" version="opset1">
2445
+ <data element_type="f32" shape="1, 32, 1, 1" offset="1525500" size="128" />
2446
+ <output>
2447
+ <port id="0" precision="FP32">
2448
+ <dim>1</dim>
2449
+ <dim>32</dim>
2450
+ <dim>1</dim>
2451
+ <dim>1</dim>
2452
+ </port>
2453
+ </output>
2454
+ </layer>
2455
+ <layer id="150" name="__module.decoder.up_blocks.1.resnets.0.conv2/aten::_convolution/Add" type="Add" version="opset1">
2456
+ <data auto_broadcast="numpy" />
2457
+ <input>
2458
+ <port id="0" precision="FP32">
2459
+ <dim>-1</dim>
2460
+ <dim>32</dim>
2461
+ <dim>-1</dim>
2462
+ <dim>-1</dim>
2463
+ </port>
2464
+ <port id="1" precision="FP32">
2465
+ <dim>1</dim>
2466
+ <dim>32</dim>
2467
+ <dim>1</dim>
2468
+ <dim>1</dim>
2469
+ </port>
2470
+ </input>
2471
+ <output>
2472
+ <port id="2" precision="FP32" names="296,hidden_states.25">
2473
+ <dim>-1</dim>
2474
+ <dim>32</dim>
2475
+ <dim>-1</dim>
2476
+ <dim>-1</dim>
2477
+ </port>
2478
+ </output>
2479
+ </layer>
2480
+ <layer id="151" name="__module.decoder.up_blocks.1.resnets.0/aten::add/Add" type="Add" version="opset1">
2481
+ <data auto_broadcast="numpy" />
2482
+ <input>
2483
+ <port id="0" precision="FP32">
2484
+ <dim>-1</dim>
2485
+ <dim>32</dim>
2486
+ <dim>-1</dim>
2487
+ <dim>-1</dim>
2488
+ </port>
2489
+ <port id="1" precision="FP32">
2490
+ <dim>-1</dim>
2491
+ <dim>32</dim>
2492
+ <dim>-1</dim>
2493
+ <dim>-1</dim>
2494
+ </port>
2495
+ </input>
2496
+ <output>
2497
+ <port id="2" precision="FP32" names="304,305,input.55">
2498
+ <dim>-1</dim>
2499
+ <dim>32</dim>
2500
+ <dim>-1</dim>
2501
+ <dim>-1</dim>
2502
+ </port>
2503
+ </output>
2504
+ </layer>
2505
+ <layer id="152" name="self.decoder.up_blocks.1.resnets.1.norm1.weight" type="Const" version="opset1">
2506
+ <data element_type="f32" shape="32" offset="1488380" size="128" />
2507
+ <output>
2508
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.1.resnets.1.norm1.weight">
2509
+ <dim>32</dim>
2510
+ </port>
2511
+ </output>
2512
+ </layer>
2513
+ <layer id="153" name="self.decoder.up_blocks.1.resnets.1.norm1.bias" type="Const" version="opset1">
2514
+ <data element_type="f32" shape="32" offset="1488508" size="128" />
2515
+ <output>
2516
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.1.resnets.1.norm1.bias">
2517
+ <dim>32</dim>
2518
+ </port>
2519
+ </output>
2520
+ </layer>
2521
+ <layer id="154" name="__module.decoder.up_blocks.1.resnets.1.norm1/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
2522
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
2523
+ <input>
2524
+ <port id="0" precision="FP32">
2525
+ <dim>-1</dim>
2526
+ <dim>32</dim>
2527
+ <dim>-1</dim>
2528
+ <dim>-1</dim>
2529
+ </port>
2530
+ <port id="1" precision="FP32">
2531
+ <dim>32</dim>
2532
+ </port>
2533
+ <port id="2" precision="FP32">
2534
+ <dim>32</dim>
2535
+ </port>
2536
+ </input>
2537
+ <output>
2538
+ <port id="3" precision="FP32" names="313,input.57">
2539
+ <dim>-1</dim>
2540
+ <dim>32</dim>
2541
+ <dim>-1</dim>
2542
+ <dim>-1</dim>
2543
+ </port>
2544
+ </output>
2545
+ </layer>
2546
+ <layer id="155" name="__module.decoder.mid_block.resnets.1.nonlinearity/aten::silu/Swish_10" type="Swish" version="opset4">
2547
+ <input>
2548
+ <port id="0" precision="FP32">
2549
+ <dim>-1</dim>
2550
+ <dim>32</dim>
2551
+ <dim>-1</dim>
2552
+ <dim>-1</dim>
2553
+ </port>
2554
+ </input>
2555
+ <output>
2556
+ <port id="1" precision="FP32" names="314">
2557
+ <dim>-1</dim>
2558
+ <dim>32</dim>
2559
+ <dim>-1</dim>
2560
+ <dim>-1</dim>
2561
+ </port>
2562
+ </output>
2563
+ </layer>
2564
+ <layer id="156" name="self.decoder.up_blocks.1.resnets.1.conv1.weight" type="Const" version="opset1">
2565
+ <data element_type="f32" shape="32, 32, 3, 3" offset="1525628" size="36864" />
2566
+ <output>
2567
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.1.resnets.1.conv1.weight">
2568
+ <dim>32</dim>
2569
+ <dim>32</dim>
2570
+ <dim>3</dim>
2571
+ <dim>3</dim>
2572
+ </port>
2573
+ </output>
2574
+ </layer>
2575
+ <layer id="157" name="__module.decoder.up_blocks.1.resnets.1.conv1/aten::_convolution/Convolution" type="Convolution" version="opset1">
2576
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
2577
+ <input>
2578
+ <port id="0" precision="FP32">
2579
+ <dim>-1</dim>
2580
+ <dim>32</dim>
2581
+ <dim>-1</dim>
2582
+ <dim>-1</dim>
2583
+ </port>
2584
+ <port id="1" precision="FP32">
2585
+ <dim>32</dim>
2586
+ <dim>32</dim>
2587
+ <dim>3</dim>
2588
+ <dim>3</dim>
2589
+ </port>
2590
+ </input>
2591
+ <output>
2592
+ <port id="2" precision="FP32">
2593
+ <dim>-1</dim>
2594
+ <dim>32</dim>
2595
+ <dim>-1</dim>
2596
+ <dim>-1</dim>
2597
+ </port>
2598
+ </output>
2599
+ </layer>
2600
+ <layer id="158" name="__module.decoder.up_blocks.1.resnets.1.conv1/aten::_convolution/Reshape" type="Const" version="opset1">
2601
+ <data element_type="f32" shape="1, 32, 1, 1" offset="1562492" size="128" />
2602
+ <output>
2603
+ <port id="0" precision="FP32">
2604
+ <dim>1</dim>
2605
+ <dim>32</dim>
2606
+ <dim>1</dim>
2607
+ <dim>1</dim>
2608
+ </port>
2609
+ </output>
2610
+ </layer>
2611
+ <layer id="159" name="__module.decoder.up_blocks.1.resnets.1.conv1/aten::_convolution/Add" type="Add" version="opset1">
2612
+ <data auto_broadcast="numpy" />
2613
+ <input>
2614
+ <port id="0" precision="FP32">
2615
+ <dim>-1</dim>
2616
+ <dim>32</dim>
2617
+ <dim>-1</dim>
2618
+ <dim>-1</dim>
2619
+ </port>
2620
+ <port id="1" precision="FP32">
2621
+ <dim>1</dim>
2622
+ <dim>32</dim>
2623
+ <dim>1</dim>
2624
+ <dim>1</dim>
2625
+ </port>
2626
+ </input>
2627
+ <output>
2628
+ <port id="2" precision="FP32" names="321,input.59">
2629
+ <dim>-1</dim>
2630
+ <dim>32</dim>
2631
+ <dim>-1</dim>
2632
+ <dim>-1</dim>
2633
+ </port>
2634
+ </output>
2635
+ </layer>
2636
+ <layer id="160" name="self.decoder.up_blocks.1.resnets.1.norm2.weight" type="Const" version="opset1">
2637
+ <data element_type="f32" shape="32" offset="1488380" size="128" />
2638
+ <output>
2639
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.1.resnets.1.norm2.weight">
2640
+ <dim>32</dim>
2641
+ </port>
2642
+ </output>
2643
+ </layer>
2644
+ <layer id="161" name="self.decoder.up_blocks.1.resnets.1.norm2.bias" type="Const" version="opset1">
2645
+ <data element_type="f32" shape="32" offset="1488508" size="128" />
2646
+ <output>
2647
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.1.resnets.1.norm2.bias">
2648
+ <dim>32</dim>
2649
+ </port>
2650
+ </output>
2651
+ </layer>
2652
+ <layer id="162" name="__module.decoder.up_blocks.1.resnets.1.norm2/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
2653
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
2654
+ <input>
2655
+ <port id="0" precision="FP32">
2656
+ <dim>-1</dim>
2657
+ <dim>32</dim>
2658
+ <dim>-1</dim>
2659
+ <dim>-1</dim>
2660
+ </port>
2661
+ <port id="1" precision="FP32">
2662
+ <dim>32</dim>
2663
+ </port>
2664
+ <port id="2" precision="FP32">
2665
+ <dim>32</dim>
2666
+ </port>
2667
+ </input>
2668
+ <output>
2669
+ <port id="3" precision="FP32" names="324,input.61">
2670
+ <dim>-1</dim>
2671
+ <dim>32</dim>
2672
+ <dim>-1</dim>
2673
+ <dim>-1</dim>
2674
+ </port>
2675
+ </output>
2676
+ </layer>
2677
+ <layer id="163" name="__module.decoder.mid_block.resnets.1.nonlinearity/aten::silu/Swish_11" type="Swish" version="opset4">
2678
+ <input>
2679
+ <port id="0" precision="FP32">
2680
+ <dim>-1</dim>
2681
+ <dim>32</dim>
2682
+ <dim>-1</dim>
2683
+ <dim>-1</dim>
2684
+ </port>
2685
+ </input>
2686
+ <output>
2687
+ <port id="1" precision="FP32" names="325,input.63">
2688
+ <dim>-1</dim>
2689
+ <dim>32</dim>
2690
+ <dim>-1</dim>
2691
+ <dim>-1</dim>
2692
+ </port>
2693
+ </output>
2694
+ </layer>
2695
+ <layer id="164" name="self.decoder.up_blocks.1.resnets.1.conv2.weight" type="Const" version="opset1">
2696
+ <data element_type="f32" shape="32, 32, 3, 3" offset="1562620" size="36864" />
2697
+ <output>
2698
+ <port id="0" precision="FP32" names="self.decoder.up_blocks.1.resnets.1.conv2.weight">
2699
+ <dim>32</dim>
2700
+ <dim>32</dim>
2701
+ <dim>3</dim>
2702
+ <dim>3</dim>
2703
+ </port>
2704
+ </output>
2705
+ </layer>
2706
+ <layer id="165" name="__module.decoder.up_blocks.1.resnets.1.conv2/aten::_convolution/Convolution" type="Convolution" version="opset1">
2707
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
2708
+ <input>
2709
+ <port id="0" precision="FP32">
2710
+ <dim>-1</dim>
2711
+ <dim>32</dim>
2712
+ <dim>-1</dim>
2713
+ <dim>-1</dim>
2714
+ </port>
2715
+ <port id="1" precision="FP32">
2716
+ <dim>32</dim>
2717
+ <dim>32</dim>
2718
+ <dim>3</dim>
2719
+ <dim>3</dim>
2720
+ </port>
2721
+ </input>
2722
+ <output>
2723
+ <port id="2" precision="FP32">
2724
+ <dim>-1</dim>
2725
+ <dim>32</dim>
2726
+ <dim>-1</dim>
2727
+ <dim>-1</dim>
2728
+ </port>
2729
+ </output>
2730
+ </layer>
2731
+ <layer id="166" name="__module.decoder.up_blocks.1.resnets.1.conv2/aten::_convolution/Reshape" type="Const" version="opset1">
2732
+ <data element_type="f32" shape="1, 32, 1, 1" offset="1599484" size="128" />
2733
+ <output>
2734
+ <port id="0" precision="FP32">
2735
+ <dim>1</dim>
2736
+ <dim>32</dim>
2737
+ <dim>1</dim>
2738
+ <dim>1</dim>
2739
+ </port>
2740
+ </output>
2741
+ </layer>
2742
+ <layer id="167" name="__module.decoder.up_blocks.1.resnets.1.conv2/aten::_convolution/Add" type="Add" version="opset1">
2743
+ <data auto_broadcast="numpy" />
2744
+ <input>
2745
+ <port id="0" precision="FP32">
2746
+ <dim>-1</dim>
2747
+ <dim>32</dim>
2748
+ <dim>-1</dim>
2749
+ <dim>-1</dim>
2750
+ </port>
2751
+ <port id="1" precision="FP32">
2752
+ <dim>1</dim>
2753
+ <dim>32</dim>
2754
+ <dim>1</dim>
2755
+ <dim>1</dim>
2756
+ </port>
2757
+ </input>
2758
+ <output>
2759
+ <port id="2" precision="FP32" names="333,hidden_states">
2760
+ <dim>-1</dim>
2761
+ <dim>32</dim>
2762
+ <dim>-1</dim>
2763
+ <dim>-1</dim>
2764
+ </port>
2765
+ </output>
2766
+ </layer>
2767
+ <layer id="168" name="__module.decoder.up_blocks.1.resnets.1/aten::add/Add" type="Add" version="opset1">
2768
+ <data auto_broadcast="numpy" />
2769
+ <input>
2770
+ <port id="0" precision="FP32">
2771
+ <dim>-1</dim>
2772
+ <dim>32</dim>
2773
+ <dim>-1</dim>
2774
+ <dim>-1</dim>
2775
+ </port>
2776
+ <port id="1" precision="FP32">
2777
+ <dim>-1</dim>
2778
+ <dim>32</dim>
2779
+ <dim>-1</dim>
2780
+ <dim>-1</dim>
2781
+ </port>
2782
+ </input>
2783
+ <output>
2784
+ <port id="2" precision="FP32" names="334,335,input.65">
2785
+ <dim>-1</dim>
2786
+ <dim>32</dim>
2787
+ <dim>-1</dim>
2788
+ <dim>-1</dim>
2789
+ </port>
2790
+ </output>
2791
+ </layer>
2792
+ <layer id="169" name="self.decoder.conv_norm_out.weight" type="Const" version="opset1">
2793
+ <data element_type="f32" shape="32" offset="1488380" size="128" />
2794
+ <output>
2795
+ <port id="0" precision="FP32" names="self.decoder.conv_norm_out.weight">
2796
+ <dim>32</dim>
2797
+ </port>
2798
+ </output>
2799
+ </layer>
2800
+ <layer id="170" name="self.decoder.conv_norm_out.bias" type="Const" version="opset1">
2801
+ <data element_type="f32" shape="32" offset="1488508" size="128" />
2802
+ <output>
2803
+ <port id="0" precision="FP32" names="self.decoder.conv_norm_out.bias">
2804
+ <dim>32</dim>
2805
+ </port>
2806
+ </output>
2807
+ </layer>
2808
+ <layer id="171" name="__module.decoder.conv_norm_out/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
2809
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
2810
+ <input>
2811
+ <port id="0" precision="FP32">
2812
+ <dim>-1</dim>
2813
+ <dim>32</dim>
2814
+ <dim>-1</dim>
2815
+ <dim>-1</dim>
2816
+ </port>
2817
+ <port id="1" precision="FP32">
2818
+ <dim>32</dim>
2819
+ </port>
2820
+ <port id="2" precision="FP32">
2821
+ <dim>32</dim>
2822
+ </port>
2823
+ </input>
2824
+ <output>
2825
+ <port id="3" precision="FP32" names="338,input">
2826
+ <dim>-1</dim>
2827
+ <dim>32</dim>
2828
+ <dim>-1</dim>
2829
+ <dim>-1</dim>
2830
+ </port>
2831
+ </output>
2832
+ </layer>
2833
+ <layer id="172" name="__module.decoder.conv_act/aten::silu/Swish" type="Swish" version="opset4">
2834
+ <input>
2835
+ <port id="0" precision="FP32">
2836
+ <dim>-1</dim>
2837
+ <dim>32</dim>
2838
+ <dim>-1</dim>
2839
+ <dim>-1</dim>
2840
+ </port>
2841
+ </input>
2842
+ <output>
2843
+ <port id="1" precision="FP32" names="339">
2844
+ <dim>-1</dim>
2845
+ <dim>32</dim>
2846
+ <dim>-1</dim>
2847
+ <dim>-1</dim>
2848
+ </port>
2849
+ </output>
2850
+ </layer>
2851
+ <layer id="173" name="self.decoder.conv_out.weight" type="Const" version="opset1">
2852
+ <data element_type="f32" shape="3, 32, 3, 3" offset="1599612" size="3456" />
2853
+ <output>
2854
+ <port id="0" precision="FP32" names="self.decoder.conv_out.weight">
2855
+ <dim>3</dim>
2856
+ <dim>32</dim>
2857
+ <dim>3</dim>
2858
+ <dim>3</dim>
2859
+ </port>
2860
+ </output>
2861
+ </layer>
2862
+ <layer id="174" name="__module.decoder.conv_out/aten::_convolution/Convolution" type="Convolution" version="opset1">
2863
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
2864
+ <input>
2865
+ <port id="0" precision="FP32">
2866
+ <dim>-1</dim>
2867
+ <dim>32</dim>
2868
+ <dim>-1</dim>
2869
+ <dim>-1</dim>
2870
+ </port>
2871
+ <port id="1" precision="FP32">
2872
+ <dim>3</dim>
2873
+ <dim>32</dim>
2874
+ <dim>3</dim>
2875
+ <dim>3</dim>
2876
+ </port>
2877
+ </input>
2878
+ <output>
2879
+ <port id="2" precision="FP32">
2880
+ <dim>-1</dim>
2881
+ <dim>3</dim>
2882
+ <dim>-1</dim>
2883
+ <dim>-1</dim>
2884
+ </port>
2885
+ </output>
2886
+ </layer>
2887
+ <layer id="175" name="__module.decoder.conv_out/aten::_convolution/Reshape" type="Const" version="opset1">
2888
+ <data element_type="f32" shape="1, 3, 1, 1" offset="1603068" size="12" />
2889
+ <output>
2890
+ <port id="0" precision="FP32">
2891
+ <dim>1</dim>
2892
+ <dim>3</dim>
2893
+ <dim>1</dim>
2894
+ <dim>1</dim>
2895
+ </port>
2896
+ </output>
2897
+ </layer>
2898
+ <layer id="176" name="__module.decoder.conv_out/aten::_convolution/Add" type="Add" version="opset1">
2899
+ <data auto_broadcast="numpy" />
2900
+ <input>
2901
+ <port id="0" precision="FP32">
2902
+ <dim>-1</dim>
2903
+ <dim>3</dim>
2904
+ <dim>-1</dim>
2905
+ <dim>-1</dim>
2906
+ </port>
2907
+ <port id="1" precision="FP32">
2908
+ <dim>1</dim>
2909
+ <dim>3</dim>
2910
+ <dim>1</dim>
2911
+ <dim>1</dim>
2912
+ </port>
2913
+ </input>
2914
+ <output>
2915
+ <port id="2" precision="FP32" names="sample">
2916
+ <dim>-1</dim>
2917
+ <dim>3</dim>
2918
+ <dim>-1</dim>
2919
+ <dim>-1</dim>
2920
+ </port>
2921
+ </output>
2922
+ </layer>
2923
+ <layer id="177" name="Result_50803" type="Result" version="opset1">
2924
+ <input>
2925
+ <port id="0" precision="FP32">
2926
+ <dim>-1</dim>
2927
+ <dim>3</dim>
2928
+ <dim>-1</dim>
2929
+ <dim>-1</dim>
2930
+ </port>
2931
+ </input>
2932
+ </layer>
2933
+ </layers>
2934
+ <edges>
2935
+ <edge from-layer="0" from-port="0" to-layer="2" to-port="0" />
2936
+ <edge from-layer="1" from-port="0" to-layer="2" to-port="1" />
2937
+ <edge from-layer="2" from-port="2" to-layer="4" to-port="0" />
2938
+ <edge from-layer="3" from-port="0" to-layer="4" to-port="1" />
2939
+ <edge from-layer="4" from-port="2" to-layer="6" to-port="0" />
2940
+ <edge from-layer="5" from-port="0" to-layer="6" to-port="1" />
2941
+ <edge from-layer="6" from-port="2" to-layer="8" to-port="0" />
2942
+ <edge from-layer="7" from-port="0" to-layer="8" to-port="1" />
2943
+ <edge from-layer="8" from-port="2" to-layer="25" to-port="0" />
2944
+ <edge from-layer="8" from-port="2" to-layer="11" to-port="0" />
2945
+ <edge from-layer="9" from-port="0" to-layer="11" to-port="1" />
2946
+ <edge from-layer="10" from-port="0" to-layer="11" to-port="2" />
2947
+ <edge from-layer="11" from-port="3" to-layer="12" to-port="0" />
2948
+ <edge from-layer="12" from-port="1" to-layer="14" to-port="0" />
2949
+ <edge from-layer="13" from-port="0" to-layer="14" to-port="1" />
2950
+ <edge from-layer="14" from-port="2" to-layer="16" to-port="0" />
2951
+ <edge from-layer="15" from-port="0" to-layer="16" to-port="1" />
2952
+ <edge from-layer="16" from-port="2" to-layer="19" to-port="0" />
2953
+ <edge from-layer="17" from-port="0" to-layer="19" to-port="1" />
2954
+ <edge from-layer="18" from-port="0" to-layer="19" to-port="2" />
2955
+ <edge from-layer="19" from-port="3" to-layer="20" to-port="0" />
2956
+ <edge from-layer="20" from-port="1" to-layer="22" to-port="0" />
2957
+ <edge from-layer="21" from-port="0" to-layer="22" to-port="1" />
2958
+ <edge from-layer="22" from-port="2" to-layer="24" to-port="0" />
2959
+ <edge from-layer="23" from-port="0" to-layer="24" to-port="1" />
2960
+ <edge from-layer="24" from-port="2" to-layer="25" to-port="1" />
2961
+ <edge from-layer="25" from-port="2" to-layer="27" to-port="0" />
2962
+ <edge from-layer="25" from-port="2" to-layer="70" to-port="0" />
2963
+ <edge from-layer="25" from-port="2" to-layer="72" to-port="1" />
2964
+ <edge from-layer="26" from-port="0" to-layer="27" to-port="1" />
2965
+ <edge from-layer="27" from-port="2" to-layer="29" to-port="0" />
2966
+ <edge from-layer="28" from-port="0" to-layer="29" to-port="1" />
2967
+ <edge from-layer="29" from-port="2" to-layer="31" to-port="0" />
2968
+ <edge from-layer="30" from-port="0" to-layer="31" to-port="1" />
2969
+ <edge from-layer="31" from-port="2" to-layer="34" to-port="0" />
2970
+ <edge from-layer="32" from-port="0" to-layer="34" to-port="1" />
2971
+ <edge from-layer="33" from-port="0" to-layer="34" to-port="2" />
2972
+ <edge from-layer="34" from-port="3" to-layer="52" to-port="0" />
2973
+ <edge from-layer="34" from-port="3" to-layer="44" to-port="0" />
2974
+ <edge from-layer="34" from-port="3" to-layer="36" to-port="0" />
2975
+ <edge from-layer="35" from-port="0" to-layer="36" to-port="1" />
2976
+ <edge from-layer="36" from-port="2" to-layer="38" to-port="0" />
2977
+ <edge from-layer="37" from-port="0" to-layer="38" to-port="1" />
2978
+ <edge from-layer="38" from-port="2" to-layer="40" to-port="0" />
2979
+ <edge from-layer="39" from-port="0" to-layer="40" to-port="1" />
2980
+ <edge from-layer="40" from-port="2" to-layer="42" to-port="0" />
2981
+ <edge from-layer="41" from-port="0" to-layer="42" to-port="1" />
2982
+ <edge from-layer="42" from-port="2" to-layer="59" to-port="0" />
2983
+ <edge from-layer="43" from-port="0" to-layer="44" to-port="1" />
2984
+ <edge from-layer="44" from-port="2" to-layer="46" to-port="0" />
2985
+ <edge from-layer="45" from-port="0" to-layer="46" to-port="1" />
2986
+ <edge from-layer="46" from-port="2" to-layer="48" to-port="0" />
2987
+ <edge from-layer="47" from-port="0" to-layer="48" to-port="1" />
2988
+ <edge from-layer="48" from-port="2" to-layer="50" to-port="0" />
2989
+ <edge from-layer="49" from-port="0" to-layer="50" to-port="1" />
2990
+ <edge from-layer="50" from-port="2" to-layer="59" to-port="1" />
2991
+ <edge from-layer="51" from-port="0" to-layer="52" to-port="1" />
2992
+ <edge from-layer="52" from-port="2" to-layer="54" to-port="0" />
2993
+ <edge from-layer="53" from-port="0" to-layer="54" to-port="1" />
2994
+ <edge from-layer="54" from-port="2" to-layer="56" to-port="0" />
2995
+ <edge from-layer="55" from-port="0" to-layer="56" to-port="1" />
2996
+ <edge from-layer="56" from-port="2" to-layer="58" to-port="0" />
2997
+ <edge from-layer="57" from-port="0" to-layer="58" to-port="1" />
2998
+ <edge from-layer="58" from-port="2" to-layer="59" to-port="2" />
2999
+ <edge from-layer="59" from-port="3" to-layer="61" to-port="0" />
3000
+ <edge from-layer="60" from-port="0" to-layer="61" to-port="1" />
3001
+ <edge from-layer="61" from-port="2" to-layer="63" to-port="0" />
3002
+ <edge from-layer="62" from-port="0" to-layer="63" to-port="1" />
3003
+ <edge from-layer="63" from-port="2" to-layer="65" to-port="0" />
3004
+ <edge from-layer="64" from-port="0" to-layer="65" to-port="1" />
3005
+ <edge from-layer="65" from-port="2" to-layer="67" to-port="0" />
3006
+ <edge from-layer="66" from-port="0" to-layer="67" to-port="1" />
3007
+ <edge from-layer="67" from-port="2" to-layer="69" to-port="0" />
3008
+ <edge from-layer="68" from-port="0" to-layer="69" to-port="1" />
3009
+ <edge from-layer="69" from-port="2" to-layer="71" to-port="0" />
3010
+ <edge from-layer="70" from-port="1" to-layer="71" to-port="1" />
3011
+ <edge from-layer="71" from-port="2" to-layer="72" to-port="0" />
3012
+ <edge from-layer="72" from-port="2" to-layer="89" to-port="0" />
3013
+ <edge from-layer="72" from-port="2" to-layer="75" to-port="0" />
3014
+ <edge from-layer="73" from-port="0" to-layer="75" to-port="1" />
3015
+ <edge from-layer="74" from-port="0" to-layer="75" to-port="2" />
3016
+ <edge from-layer="75" from-port="3" to-layer="76" to-port="0" />
3017
+ <edge from-layer="76" from-port="1" to-layer="78" to-port="0" />
3018
+ <edge from-layer="77" from-port="0" to-layer="78" to-port="1" />
3019
+ <edge from-layer="78" from-port="2" to-layer="80" to-port="0" />
3020
+ <edge from-layer="79" from-port="0" to-layer="80" to-port="1" />
3021
+ <edge from-layer="80" from-port="2" to-layer="83" to-port="0" />
3022
+ <edge from-layer="81" from-port="0" to-layer="83" to-port="1" />
3023
+ <edge from-layer="82" from-port="0" to-layer="83" to-port="2" />
3024
+ <edge from-layer="83" from-port="3" to-layer="84" to-port="0" />
3025
+ <edge from-layer="84" from-port="1" to-layer="86" to-port="0" />
3026
+ <edge from-layer="85" from-port="0" to-layer="86" to-port="1" />
3027
+ <edge from-layer="86" from-port="2" to-layer="88" to-port="0" />
3028
+ <edge from-layer="87" from-port="0" to-layer="88" to-port="1" />
3029
+ <edge from-layer="88" from-port="2" to-layer="89" to-port="1" />
3030
+ <edge from-layer="89" from-port="2" to-layer="106" to-port="0" />
3031
+ <edge from-layer="89" from-port="2" to-layer="92" to-port="0" />
3032
+ <edge from-layer="90" from-port="0" to-layer="92" to-port="1" />
3033
+ <edge from-layer="91" from-port="0" to-layer="92" to-port="2" />
3034
+ <edge from-layer="92" from-port="3" to-layer="93" to-port="0" />
3035
+ <edge from-layer="93" from-port="1" to-layer="95" to-port="0" />
3036
+ <edge from-layer="94" from-port="0" to-layer="95" to-port="1" />
3037
+ <edge from-layer="95" from-port="2" to-layer="97" to-port="0" />
3038
+ <edge from-layer="96" from-port="0" to-layer="97" to-port="1" />
3039
+ <edge from-layer="97" from-port="2" to-layer="100" to-port="0" />
3040
+ <edge from-layer="98" from-port="0" to-layer="100" to-port="1" />
3041
+ <edge from-layer="99" from-port="0" to-layer="100" to-port="2" />
3042
+ <edge from-layer="100" from-port="3" to-layer="101" to-port="0" />
3043
+ <edge from-layer="101" from-port="1" to-layer="103" to-port="0" />
3044
+ <edge from-layer="102" from-port="0" to-layer="103" to-port="1" />
3045
+ <edge from-layer="103" from-port="2" to-layer="105" to-port="0" />
3046
+ <edge from-layer="104" from-port="0" to-layer="105" to-port="1" />
3047
+ <edge from-layer="105" from-port="2" to-layer="106" to-port="1" />
3048
+ <edge from-layer="106" from-port="2" to-layer="109" to-port="0" />
3049
+ <edge from-layer="106" from-port="2" to-layer="123" to-port="0" />
3050
+ <edge from-layer="107" from-port="0" to-layer="109" to-port="1" />
3051
+ <edge from-layer="108" from-port="0" to-layer="109" to-port="2" />
3052
+ <edge from-layer="109" from-port="3" to-layer="110" to-port="0" />
3053
+ <edge from-layer="110" from-port="1" to-layer="112" to-port="0" />
3054
+ <edge from-layer="111" from-port="0" to-layer="112" to-port="1" />
3055
+ <edge from-layer="112" from-port="2" to-layer="114" to-port="0" />
3056
+ <edge from-layer="113" from-port="0" to-layer="114" to-port="1" />
3057
+ <edge from-layer="114" from-port="2" to-layer="117" to-port="0" />
3058
+ <edge from-layer="115" from-port="0" to-layer="117" to-port="1" />
3059
+ <edge from-layer="116" from-port="0" to-layer="117" to-port="2" />
3060
+ <edge from-layer="117" from-port="3" to-layer="118" to-port="0" />
3061
+ <edge from-layer="118" from-port="1" to-layer="120" to-port="0" />
3062
+ <edge from-layer="119" from-port="0" to-layer="120" to-port="1" />
3063
+ <edge from-layer="120" from-port="2" to-layer="122" to-port="0" />
3064
+ <edge from-layer="121" from-port="0" to-layer="122" to-port="1" />
3065
+ <edge from-layer="122" from-port="2" to-layer="123" to-port="1" />
3066
+ <edge from-layer="123" from-port="2" to-layer="126" to-port="0" />
3067
+ <edge from-layer="124" from-port="0" to-layer="126" to-port="1" />
3068
+ <edge from-layer="125" from-port="0" to-layer="126" to-port="2" />
3069
+ <edge from-layer="126" from-port="3" to-layer="128" to-port="0" />
3070
+ <edge from-layer="127" from-port="0" to-layer="128" to-port="1" />
3071
+ <edge from-layer="128" from-port="2" to-layer="130" to-port="0" />
3072
+ <edge from-layer="129" from-port="0" to-layer="130" to-port="1" />
3073
+ <edge from-layer="130" from-port="2" to-layer="132" to-port="0" />
3074
+ <edge from-layer="130" from-port="2" to-layer="137" to-port="0" />
3075
+ <edge from-layer="131" from-port="0" to-layer="132" to-port="1" />
3076
+ <edge from-layer="132" from-port="2" to-layer="134" to-port="0" />
3077
+ <edge from-layer="133" from-port="0" to-layer="134" to-port="1" />
3078
+ <edge from-layer="134" from-port="2" to-layer="151" to-port="0" />
3079
+ <edge from-layer="135" from-port="0" to-layer="137" to-port="1" />
3080
+ <edge from-layer="136" from-port="0" to-layer="137" to-port="2" />
3081
+ <edge from-layer="137" from-port="3" to-layer="138" to-port="0" />
3082
+ <edge from-layer="138" from-port="1" to-layer="140" to-port="0" />
3083
+ <edge from-layer="139" from-port="0" to-layer="140" to-port="1" />
3084
+ <edge from-layer="140" from-port="2" to-layer="142" to-port="0" />
3085
+ <edge from-layer="141" from-port="0" to-layer="142" to-port="1" />
3086
+ <edge from-layer="142" from-port="2" to-layer="145" to-port="0" />
3087
+ <edge from-layer="143" from-port="0" to-layer="145" to-port="1" />
3088
+ <edge from-layer="144" from-port="0" to-layer="145" to-port="2" />
3089
+ <edge from-layer="145" from-port="3" to-layer="146" to-port="0" />
3090
+ <edge from-layer="146" from-port="1" to-layer="148" to-port="0" />
3091
+ <edge from-layer="147" from-port="0" to-layer="148" to-port="1" />
3092
+ <edge from-layer="148" from-port="2" to-layer="150" to-port="0" />
3093
+ <edge from-layer="149" from-port="0" to-layer="150" to-port="1" />
3094
+ <edge from-layer="150" from-port="2" to-layer="151" to-port="1" />
3095
+ <edge from-layer="151" from-port="2" to-layer="154" to-port="0" />
3096
+ <edge from-layer="151" from-port="2" to-layer="168" to-port="0" />
3097
+ <edge from-layer="152" from-port="0" to-layer="154" to-port="1" />
3098
+ <edge from-layer="153" from-port="0" to-layer="154" to-port="2" />
3099
+ <edge from-layer="154" from-port="3" to-layer="155" to-port="0" />
3100
+ <edge from-layer="155" from-port="1" to-layer="157" to-port="0" />
3101
+ <edge from-layer="156" from-port="0" to-layer="157" to-port="1" />
3102
+ <edge from-layer="157" from-port="2" to-layer="159" to-port="0" />
3103
+ <edge from-layer="158" from-port="0" to-layer="159" to-port="1" />
3104
+ <edge from-layer="159" from-port="2" to-layer="162" to-port="0" />
3105
+ <edge from-layer="160" from-port="0" to-layer="162" to-port="1" />
3106
+ <edge from-layer="161" from-port="0" to-layer="162" to-port="2" />
3107
+ <edge from-layer="162" from-port="3" to-layer="163" to-port="0" />
3108
+ <edge from-layer="163" from-port="1" to-layer="165" to-port="0" />
3109
+ <edge from-layer="164" from-port="0" to-layer="165" to-port="1" />
3110
+ <edge from-layer="165" from-port="2" to-layer="167" to-port="0" />
3111
+ <edge from-layer="166" from-port="0" to-layer="167" to-port="1" />
3112
+ <edge from-layer="167" from-port="2" to-layer="168" to-port="1" />
3113
+ <edge from-layer="168" from-port="2" to-layer="171" to-port="0" />
3114
+ <edge from-layer="169" from-port="0" to-layer="171" to-port="1" />
3115
+ <edge from-layer="170" from-port="0" to-layer="171" to-port="2" />
3116
+ <edge from-layer="171" from-port="3" to-layer="172" to-port="0" />
3117
+ <edge from-layer="172" from-port="1" to-layer="174" to-port="0" />
3118
+ <edge from-layer="173" from-port="0" to-layer="174" to-port="1" />
3119
+ <edge from-layer="174" from-port="2" to-layer="176" to-port="0" />
3120
+ <edge from-layer="175" from-port="0" to-layer="176" to-port="1" />
3121
+ <edge from-layer="176" from-port="2" to-layer="177" to-port="0" />
3122
+ </edges>
3123
+ <rt_info>
3124
+ <Runtime_version value="2025.0.0-17942-1f68be9f594-releases/2025/0" />
3125
+ <conversion_parameters>
3126
+ <framework value="pytorch" />
3127
+ <is_python_object value="True" />
3128
+ </conversion_parameters>
3129
+ <optimum>
3130
+ <diffusers_version value="0.32.2" />
3131
+ <optimum_intel_version value="1.23.0.dev0+5ac35448" />
3132
+ <optimum_version value="1.24.0" />
3133
+ <pytorch_version value="2.5.0+cpu" />
3134
+ <transformers_version value="4.49.0" />
3135
+ </optimum>
3136
+ </rt_info>
3137
+ </net>
vae_encoder/config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.32.2",
4
+ "_name_or_path": "/home/ea/.cache/huggingface/hub/models--hf-internal-testing--tiny-stable-diffusion-torch/snapshots/a88cdfbd91f96ec7f61eb7484b652ff0f4ee701d/vae",
5
+ "act_fn": "silu",
6
+ "block_out_channels": [
7
+ 32,
8
+ 64
9
+ ],
10
+ "down_block_types": [
11
+ "DownEncoderBlock2D",
12
+ "DownEncoderBlock2D"
13
+ ],
14
+ "force_upcast": true,
15
+ "in_channels": 3,
16
+ "latent_channels": 4,
17
+ "latents_mean": null,
18
+ "latents_std": null,
19
+ "layers_per_block": 1,
20
+ "mid_block_add_attention": true,
21
+ "norm_num_groups": 32,
22
+ "out_channels": 3,
23
+ "sample_size": 128,
24
+ "scaling_factor": 0.18215,
25
+ "shift_factor": null,
26
+ "up_block_types": [
27
+ "UpDecoderBlock2D",
28
+ "UpDecoderBlock2D"
29
+ ],
30
+ "use_post_quant_conv": true,
31
+ "use_quant_conv": true
32
+ }
vae_encoder/openvino_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:113fda04b98bbfc59a7461f558a908ebc5e0364fafa893fe49363f0d6920587d
3
+ size 1021788
vae_encoder/openvino_model.xml ADDED
@@ -0,0 +1,2483 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <?xml version="1.0"?>
2
+ <net name="Model6" version="11">
3
+ <layers>
4
+ <layer id="0" name="sample" type="Parameter" version="opset1">
5
+ <data shape="?,3,?,?" element_type="f32" />
6
+ <output>
7
+ <port id="0" precision="FP32" names="sample">
8
+ <dim>-1</dim>
9
+ <dim>3</dim>
10
+ <dim>-1</dim>
11
+ <dim>-1</dim>
12
+ </port>
13
+ </output>
14
+ </layer>
15
+ <layer id="1" name="self.encoder.conv_in.weight" type="Const" version="opset1">
16
+ <data element_type="f32" shape="32, 3, 3, 3" offset="0" size="3456" />
17
+ <output>
18
+ <port id="0" precision="FP32" names="self.encoder.conv_in.weight">
19
+ <dim>32</dim>
20
+ <dim>3</dim>
21
+ <dim>3</dim>
22
+ <dim>3</dim>
23
+ </port>
24
+ </output>
25
+ </layer>
26
+ <layer id="2" name="__module.encoder.conv_in/aten::_convolution/Convolution" type="Convolution" version="opset1">
27
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
28
+ <input>
29
+ <port id="0" precision="FP32">
30
+ <dim>-1</dim>
31
+ <dim>3</dim>
32
+ <dim>-1</dim>
33
+ <dim>-1</dim>
34
+ </port>
35
+ <port id="1" precision="FP32">
36
+ <dim>32</dim>
37
+ <dim>3</dim>
38
+ <dim>3</dim>
39
+ <dim>3</dim>
40
+ </port>
41
+ </input>
42
+ <output>
43
+ <port id="2" precision="FP32">
44
+ <dim>-1</dim>
45
+ <dim>32</dim>
46
+ <dim>-1</dim>
47
+ <dim>-1</dim>
48
+ </port>
49
+ </output>
50
+ </layer>
51
+ <layer id="3" name="__module.encoder.conv_in/aten::_convolution/Reshape" type="Const" version="opset1">
52
+ <data element_type="f32" shape="1, 32, 1, 1" offset="3456" size="128" />
53
+ <output>
54
+ <port id="0" precision="FP32">
55
+ <dim>1</dim>
56
+ <dim>32</dim>
57
+ <dim>1</dim>
58
+ <dim>1</dim>
59
+ </port>
60
+ </output>
61
+ </layer>
62
+ <layer id="4" name="__module.encoder.conv_in/aten::_convolution/Add" type="Add" version="opset1">
63
+ <data auto_broadcast="numpy" />
64
+ <input>
65
+ <port id="0" precision="FP32">
66
+ <dim>-1</dim>
67
+ <dim>32</dim>
68
+ <dim>-1</dim>
69
+ <dim>-1</dim>
70
+ </port>
71
+ <port id="1" precision="FP32">
72
+ <dim>1</dim>
73
+ <dim>32</dim>
74
+ <dim>1</dim>
75
+ <dim>1</dim>
76
+ </port>
77
+ </input>
78
+ <output>
79
+ <port id="2" precision="FP32" names="43,input.1">
80
+ <dim>-1</dim>
81
+ <dim>32</dim>
82
+ <dim>-1</dim>
83
+ <dim>-1</dim>
84
+ </port>
85
+ </output>
86
+ </layer>
87
+ <layer id="5" name="self.encoder.down_blocks.0.resnets.0.norm1.weight" type="Const" version="opset1">
88
+ <data element_type="f32" shape="32" offset="3584" size="128" />
89
+ <output>
90
+ <port id="0" precision="FP32" names="self.encoder.down_blocks.0.resnets.0.norm1.weight">
91
+ <dim>32</dim>
92
+ </port>
93
+ </output>
94
+ </layer>
95
+ <layer id="6" name="self.encoder.down_blocks.0.resnets.0.norm1.bias" type="Const" version="opset1">
96
+ <data element_type="f32" shape="32" offset="3712" size="128" />
97
+ <output>
98
+ <port id="0" precision="FP32" names="self.encoder.down_blocks.0.resnets.0.norm1.bias">
99
+ <dim>32</dim>
100
+ </port>
101
+ </output>
102
+ </layer>
103
+ <layer id="7" name="__module.encoder.down_blocks.0.resnets.0.norm1/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
104
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
105
+ <input>
106
+ <port id="0" precision="FP32">
107
+ <dim>-1</dim>
108
+ <dim>32</dim>
109
+ <dim>-1</dim>
110
+ <dim>-1</dim>
111
+ </port>
112
+ <port id="1" precision="FP32">
113
+ <dim>32</dim>
114
+ </port>
115
+ <port id="2" precision="FP32">
116
+ <dim>32</dim>
117
+ </port>
118
+ </input>
119
+ <output>
120
+ <port id="3" precision="FP32" names="55,input.3">
121
+ <dim>-1</dim>
122
+ <dim>32</dim>
123
+ <dim>-1</dim>
124
+ <dim>-1</dim>
125
+ </port>
126
+ </output>
127
+ </layer>
128
+ <layer id="8" name="__module.decoder.mid_block.resnets.1.nonlinearity/aten::silu/Swish" type="Swish" version="opset4">
129
+ <input>
130
+ <port id="0" precision="FP32">
131
+ <dim>-1</dim>
132
+ <dim>32</dim>
133
+ <dim>-1</dim>
134
+ <dim>-1</dim>
135
+ </port>
136
+ </input>
137
+ <output>
138
+ <port id="1" precision="FP32" names="56">
139
+ <dim>-1</dim>
140
+ <dim>32</dim>
141
+ <dim>-1</dim>
142
+ <dim>-1</dim>
143
+ </port>
144
+ </output>
145
+ </layer>
146
+ <layer id="9" name="self.encoder.down_blocks.0.resnets.0.conv1.weight" type="Const" version="opset1">
147
+ <data element_type="f32" shape="32, 32, 3, 3" offset="3840" size="36864" />
148
+ <output>
149
+ <port id="0" precision="FP32" names="self.encoder.down_blocks.0.resnets.0.conv1.weight">
150
+ <dim>32</dim>
151
+ <dim>32</dim>
152
+ <dim>3</dim>
153
+ <dim>3</dim>
154
+ </port>
155
+ </output>
156
+ </layer>
157
+ <layer id="10" name="__module.encoder.down_blocks.0.resnets.0.conv1/aten::_convolution/Convolution" type="Convolution" version="opset1">
158
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
159
+ <input>
160
+ <port id="0" precision="FP32">
161
+ <dim>-1</dim>
162
+ <dim>32</dim>
163
+ <dim>-1</dim>
164
+ <dim>-1</dim>
165
+ </port>
166
+ <port id="1" precision="FP32">
167
+ <dim>32</dim>
168
+ <dim>32</dim>
169
+ <dim>3</dim>
170
+ <dim>3</dim>
171
+ </port>
172
+ </input>
173
+ <output>
174
+ <port id="2" precision="FP32">
175
+ <dim>-1</dim>
176
+ <dim>32</dim>
177
+ <dim>-1</dim>
178
+ <dim>-1</dim>
179
+ </port>
180
+ </output>
181
+ </layer>
182
+ <layer id="11" name="__module.encoder.down_blocks.0.resnets.0.conv1/aten::_convolution/Reshape" type="Const" version="opset1">
183
+ <data element_type="f32" shape="1, 32, 1, 1" offset="40704" size="128" />
184
+ <output>
185
+ <port id="0" precision="FP32">
186
+ <dim>1</dim>
187
+ <dim>32</dim>
188
+ <dim>1</dim>
189
+ <dim>1</dim>
190
+ </port>
191
+ </output>
192
+ </layer>
193
+ <layer id="12" name="__module.encoder.down_blocks.0.resnets.0.conv1/aten::_convolution/Add" type="Add" version="opset1">
194
+ <data auto_broadcast="numpy" />
195
+ <input>
196
+ <port id="0" precision="FP32">
197
+ <dim>-1</dim>
198
+ <dim>32</dim>
199
+ <dim>-1</dim>
200
+ <dim>-1</dim>
201
+ </port>
202
+ <port id="1" precision="FP32">
203
+ <dim>1</dim>
204
+ <dim>32</dim>
205
+ <dim>1</dim>
206
+ <dim>1</dim>
207
+ </port>
208
+ </input>
209
+ <output>
210
+ <port id="2" precision="FP32" names="63,input.5">
211
+ <dim>-1</dim>
212
+ <dim>32</dim>
213
+ <dim>-1</dim>
214
+ <dim>-1</dim>
215
+ </port>
216
+ </output>
217
+ </layer>
218
+ <layer id="13" name="self.encoder.down_blocks.0.resnets.0.norm2.weight" type="Const" version="opset1">
219
+ <data element_type="f32" shape="32" offset="3584" size="128" />
220
+ <output>
221
+ <port id="0" precision="FP32" names="self.encoder.down_blocks.0.resnets.0.norm2.weight">
222
+ <dim>32</dim>
223
+ </port>
224
+ </output>
225
+ </layer>
226
+ <layer id="14" name="self.encoder.down_blocks.0.resnets.0.norm2.bias" type="Const" version="opset1">
227
+ <data element_type="f32" shape="32" offset="3712" size="128" />
228
+ <output>
229
+ <port id="0" precision="FP32" names="self.encoder.down_blocks.0.resnets.0.norm2.bias">
230
+ <dim>32</dim>
231
+ </port>
232
+ </output>
233
+ </layer>
234
+ <layer id="15" name="__module.encoder.down_blocks.0.resnets.0.norm2/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
235
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
236
+ <input>
237
+ <port id="0" precision="FP32">
238
+ <dim>-1</dim>
239
+ <dim>32</dim>
240
+ <dim>-1</dim>
241
+ <dim>-1</dim>
242
+ </port>
243
+ <port id="1" precision="FP32">
244
+ <dim>32</dim>
245
+ </port>
246
+ <port id="2" precision="FP32">
247
+ <dim>32</dim>
248
+ </port>
249
+ </input>
250
+ <output>
251
+ <port id="3" precision="FP32" names="66,input.7">
252
+ <dim>-1</dim>
253
+ <dim>32</dim>
254
+ <dim>-1</dim>
255
+ <dim>-1</dim>
256
+ </port>
257
+ </output>
258
+ </layer>
259
+ <layer id="16" name="__module.decoder.mid_block.resnets.1.nonlinearity/aten::silu/Swish_1" type="Swish" version="opset4">
260
+ <input>
261
+ <port id="0" precision="FP32">
262
+ <dim>-1</dim>
263
+ <dim>32</dim>
264
+ <dim>-1</dim>
265
+ <dim>-1</dim>
266
+ </port>
267
+ </input>
268
+ <output>
269
+ <port id="1" precision="FP32" names="67,input.9">
270
+ <dim>-1</dim>
271
+ <dim>32</dim>
272
+ <dim>-1</dim>
273
+ <dim>-1</dim>
274
+ </port>
275
+ </output>
276
+ </layer>
277
+ <layer id="17" name="self.encoder.down_blocks.0.resnets.0.conv2.weight" type="Const" version="opset1">
278
+ <data element_type="f32" shape="32, 32, 3, 3" offset="40832" size="36864" />
279
+ <output>
280
+ <port id="0" precision="FP32" names="self.encoder.down_blocks.0.resnets.0.conv2.weight">
281
+ <dim>32</dim>
282
+ <dim>32</dim>
283
+ <dim>3</dim>
284
+ <dim>3</dim>
285
+ </port>
286
+ </output>
287
+ </layer>
288
+ <layer id="18" name="__module.encoder.down_blocks.0.resnets.0.conv2/aten::_convolution/Convolution" type="Convolution" version="opset1">
289
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
290
+ <input>
291
+ <port id="0" precision="FP32">
292
+ <dim>-1</dim>
293
+ <dim>32</dim>
294
+ <dim>-1</dim>
295
+ <dim>-1</dim>
296
+ </port>
297
+ <port id="1" precision="FP32">
298
+ <dim>32</dim>
299
+ <dim>32</dim>
300
+ <dim>3</dim>
301
+ <dim>3</dim>
302
+ </port>
303
+ </input>
304
+ <output>
305
+ <port id="2" precision="FP32">
306
+ <dim>-1</dim>
307
+ <dim>32</dim>
308
+ <dim>-1</dim>
309
+ <dim>-1</dim>
310
+ </port>
311
+ </output>
312
+ </layer>
313
+ <layer id="19" name="__module.encoder.down_blocks.0.resnets.0.conv2/aten::_convolution/Reshape" type="Const" version="opset1">
314
+ <data element_type="f32" shape="1, 32, 1, 1" offset="77696" size="128" />
315
+ <output>
316
+ <port id="0" precision="FP32">
317
+ <dim>1</dim>
318
+ <dim>32</dim>
319
+ <dim>1</dim>
320
+ <dim>1</dim>
321
+ </port>
322
+ </output>
323
+ </layer>
324
+ <layer id="20" name="__module.encoder.down_blocks.0.resnets.0.conv2/aten::_convolution/Add" type="Add" version="opset1">
325
+ <data auto_broadcast="numpy" />
326
+ <input>
327
+ <port id="0" precision="FP32">
328
+ <dim>-1</dim>
329
+ <dim>32</dim>
330
+ <dim>-1</dim>
331
+ <dim>-1</dim>
332
+ </port>
333
+ <port id="1" precision="FP32">
334
+ <dim>1</dim>
335
+ <dim>32</dim>
336
+ <dim>1</dim>
337
+ <dim>1</dim>
338
+ </port>
339
+ </input>
340
+ <output>
341
+ <port id="2" precision="FP32" names="75,hidden_states.1">
342
+ <dim>-1</dim>
343
+ <dim>32</dim>
344
+ <dim>-1</dim>
345
+ <dim>-1</dim>
346
+ </port>
347
+ </output>
348
+ </layer>
349
+ <layer id="21" name="__module.encoder.down_blocks.0.resnets.0/aten::add/Add" type="Add" version="opset1">
350
+ <data auto_broadcast="numpy" />
351
+ <input>
352
+ <port id="0" precision="FP32">
353
+ <dim>-1</dim>
354
+ <dim>32</dim>
355
+ <dim>-1</dim>
356
+ <dim>-1</dim>
357
+ </port>
358
+ <port id="1" precision="FP32">
359
+ <dim>-1</dim>
360
+ <dim>32</dim>
361
+ <dim>-1</dim>
362
+ <dim>-1</dim>
363
+ </port>
364
+ </input>
365
+ <output>
366
+ <port id="2" precision="FP32" names="76,77,hidden_states.3">
367
+ <dim>-1</dim>
368
+ <dim>32</dim>
369
+ <dim>-1</dim>
370
+ <dim>-1</dim>
371
+ </port>
372
+ </output>
373
+ </layer>
374
+ <layer id="22" name="self.encoder.down_blocks.0.downsamplers.0.conv.weight" type="Const" version="opset1">
375
+ <data element_type="f32" shape="32, 32, 3, 3" offset="77824" size="36864" />
376
+ <output>
377
+ <port id="0" precision="FP32" names="self.encoder.down_blocks.0.downsamplers.0.conv.weight">
378
+ <dim>32</dim>
379
+ <dim>32</dim>
380
+ <dim>3</dim>
381
+ <dim>3</dim>
382
+ </port>
383
+ </output>
384
+ </layer>
385
+ <layer id="23" name="__module.encoder.down_blocks.0.downsamplers.0.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
386
+ <data strides="2, 2" dilations="1, 1" pads_begin="0, 0" pads_end="1, 1" auto_pad="explicit" />
387
+ <input>
388
+ <port id="0" precision="FP32">
389
+ <dim>-1</dim>
390
+ <dim>32</dim>
391
+ <dim>-1</dim>
392
+ <dim>-1</dim>
393
+ </port>
394
+ <port id="1" precision="FP32">
395
+ <dim>32</dim>
396
+ <dim>32</dim>
397
+ <dim>3</dim>
398
+ <dim>3</dim>
399
+ </port>
400
+ </input>
401
+ <output>
402
+ <port id="2" precision="FP32">
403
+ <dim>-1</dim>
404
+ <dim>32</dim>
405
+ <dim>-1</dim>
406
+ <dim>-1</dim>
407
+ </port>
408
+ </output>
409
+ </layer>
410
+ <layer id="24" name="__module.encoder.down_blocks.0.downsamplers.0.conv/aten::_convolution/Reshape" type="Const" version="opset1">
411
+ <data element_type="f32" shape="1, 32, 1, 1" offset="114688" size="128" />
412
+ <output>
413
+ <port id="0" precision="FP32">
414
+ <dim>1</dim>
415
+ <dim>32</dim>
416
+ <dim>1</dim>
417
+ <dim>1</dim>
418
+ </port>
419
+ </output>
420
+ </layer>
421
+ <layer id="25" name="__module.encoder.down_blocks.0.downsamplers.0.conv/aten::_convolution/Add" type="Add" version="opset1">
422
+ <data auto_broadcast="numpy" />
423
+ <input>
424
+ <port id="0" precision="FP32">
425
+ <dim>-1</dim>
426
+ <dim>32</dim>
427
+ <dim>-1</dim>
428
+ <dim>-1</dim>
429
+ </port>
430
+ <port id="1" precision="FP32">
431
+ <dim>1</dim>
432
+ <dim>32</dim>
433
+ <dim>1</dim>
434
+ <dim>1</dim>
435
+ </port>
436
+ </input>
437
+ <output>
438
+ <port id="2" precision="FP32" names="87,input.11">
439
+ <dim>-1</dim>
440
+ <dim>32</dim>
441
+ <dim>-1</dim>
442
+ <dim>-1</dim>
443
+ </port>
444
+ </output>
445
+ </layer>
446
+ <layer id="26" name="self.encoder.down_blocks.1.resnets.0.conv_shortcut.weight" type="Const" version="opset1">
447
+ <data element_type="f32" shape="64, 32, 1, 1" offset="114816" size="8192" />
448
+ <output>
449
+ <port id="0" precision="FP32" names="self.encoder.down_blocks.1.resnets.0.conv_shortcut.weight">
450
+ <dim>64</dim>
451
+ <dim>32</dim>
452
+ <dim>1</dim>
453
+ <dim>1</dim>
454
+ </port>
455
+ </output>
456
+ </layer>
457
+ <layer id="27" name="__module.encoder.down_blocks.1.resnets.0.conv_shortcut/aten::_convolution/Convolution" type="Convolution" version="opset1">
458
+ <data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
459
+ <input>
460
+ <port id="0" precision="FP32">
461
+ <dim>-1</dim>
462
+ <dim>32</dim>
463
+ <dim>-1</dim>
464
+ <dim>-1</dim>
465
+ </port>
466
+ <port id="1" precision="FP32">
467
+ <dim>64</dim>
468
+ <dim>32</dim>
469
+ <dim>1</dim>
470
+ <dim>1</dim>
471
+ </port>
472
+ </input>
473
+ <output>
474
+ <port id="2" precision="FP32">
475
+ <dim>-1</dim>
476
+ <dim>64</dim>
477
+ <dim>-1</dim>
478
+ <dim>-1</dim>
479
+ </port>
480
+ </output>
481
+ </layer>
482
+ <layer id="28" name="__module.encoder.down_blocks.1.resnets.0.conv_shortcut/aten::_convolution/Reshape" type="Const" version="opset1">
483
+ <data element_type="f32" shape="1, 64, 1, 1" offset="123008" size="256" />
484
+ <output>
485
+ <port id="0" precision="FP32">
486
+ <dim>1</dim>
487
+ <dim>64</dim>
488
+ <dim>1</dim>
489
+ <dim>1</dim>
490
+ </port>
491
+ </output>
492
+ </layer>
493
+ <layer id="29" name="__module.encoder.down_blocks.1.resnets.0.conv_shortcut/aten::_convolution/Add" type="Add" version="opset1">
494
+ <data auto_broadcast="numpy" />
495
+ <input>
496
+ <port id="0" precision="FP32">
497
+ <dim>-1</dim>
498
+ <dim>64</dim>
499
+ <dim>-1</dim>
500
+ <dim>-1</dim>
501
+ </port>
502
+ <port id="1" precision="FP32">
503
+ <dim>1</dim>
504
+ <dim>64</dim>
505
+ <dim>1</dim>
506
+ <dim>1</dim>
507
+ </port>
508
+ </input>
509
+ <output>
510
+ <port id="2" precision="FP32" names="125,input_tensor">
511
+ <dim>-1</dim>
512
+ <dim>64</dim>
513
+ <dim>-1</dim>
514
+ <dim>-1</dim>
515
+ </port>
516
+ </output>
517
+ </layer>
518
+ <layer id="30" name="self.encoder.down_blocks.1.resnets.0.norm1.weight" type="Const" version="opset1">
519
+ <data element_type="f32" shape="32" offset="3584" size="128" />
520
+ <output>
521
+ <port id="0" precision="FP32" names="self.encoder.down_blocks.1.resnets.0.norm1.weight">
522
+ <dim>32</dim>
523
+ </port>
524
+ </output>
525
+ </layer>
526
+ <layer id="31" name="self.encoder.down_blocks.1.resnets.0.norm1.bias" type="Const" version="opset1">
527
+ <data element_type="f32" shape="32" offset="3712" size="128" />
528
+ <output>
529
+ <port id="0" precision="FP32" names="self.encoder.down_blocks.1.resnets.0.norm1.bias">
530
+ <dim>32</dim>
531
+ </port>
532
+ </output>
533
+ </layer>
534
+ <layer id="32" name="__module.encoder.down_blocks.1.resnets.0.norm1/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
535
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
536
+ <input>
537
+ <port id="0" precision="FP32">
538
+ <dim>-1</dim>
539
+ <dim>32</dim>
540
+ <dim>-1</dim>
541
+ <dim>-1</dim>
542
+ </port>
543
+ <port id="1" precision="FP32">
544
+ <dim>32</dim>
545
+ </port>
546
+ <port id="2" precision="FP32">
547
+ <dim>32</dim>
548
+ </port>
549
+ </input>
550
+ <output>
551
+ <port id="3" precision="FP32" names="98,input.13">
552
+ <dim>-1</dim>
553
+ <dim>32</dim>
554
+ <dim>-1</dim>
555
+ <dim>-1</dim>
556
+ </port>
557
+ </output>
558
+ </layer>
559
+ <layer id="33" name="__module.decoder.mid_block.resnets.1.nonlinearity/aten::silu/Swish_2" type="Swish" version="opset4">
560
+ <input>
561
+ <port id="0" precision="FP32">
562
+ <dim>-1</dim>
563
+ <dim>32</dim>
564
+ <dim>-1</dim>
565
+ <dim>-1</dim>
566
+ </port>
567
+ </input>
568
+ <output>
569
+ <port id="1" precision="FP32" names="99">
570
+ <dim>-1</dim>
571
+ <dim>32</dim>
572
+ <dim>-1</dim>
573
+ <dim>-1</dim>
574
+ </port>
575
+ </output>
576
+ </layer>
577
+ <layer id="34" name="self.encoder.down_blocks.1.resnets.0.conv1.weight" type="Const" version="opset1">
578
+ <data element_type="f32" shape="64, 32, 3, 3" offset="123264" size="73728" />
579
+ <output>
580
+ <port id="0" precision="FP32" names="self.encoder.down_blocks.1.resnets.0.conv1.weight">
581
+ <dim>64</dim>
582
+ <dim>32</dim>
583
+ <dim>3</dim>
584
+ <dim>3</dim>
585
+ </port>
586
+ </output>
587
+ </layer>
588
+ <layer id="35" name="__module.encoder.down_blocks.1.resnets.0.conv1/aten::_convolution/Convolution" type="Convolution" version="opset1">
589
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
590
+ <input>
591
+ <port id="0" precision="FP32">
592
+ <dim>-1</dim>
593
+ <dim>32</dim>
594
+ <dim>-1</dim>
595
+ <dim>-1</dim>
596
+ </port>
597
+ <port id="1" precision="FP32">
598
+ <dim>64</dim>
599
+ <dim>32</dim>
600
+ <dim>3</dim>
601
+ <dim>3</dim>
602
+ </port>
603
+ </input>
604
+ <output>
605
+ <port id="2" precision="FP32">
606
+ <dim>-1</dim>
607
+ <dim>64</dim>
608
+ <dim>-1</dim>
609
+ <dim>-1</dim>
610
+ </port>
611
+ </output>
612
+ </layer>
613
+ <layer id="36" name="__module.encoder.down_blocks.1.resnets.0.conv1/aten::_convolution/Reshape" type="Const" version="opset1">
614
+ <data element_type="f32" shape="1, 64, 1, 1" offset="196992" size="256" />
615
+ <output>
616
+ <port id="0" precision="FP32">
617
+ <dim>1</dim>
618
+ <dim>64</dim>
619
+ <dim>1</dim>
620
+ <dim>1</dim>
621
+ </port>
622
+ </output>
623
+ </layer>
624
+ <layer id="37" name="__module.encoder.down_blocks.1.resnets.0.conv1/aten::_convolution/Add" type="Add" version="opset1">
625
+ <data auto_broadcast="numpy" />
626
+ <input>
627
+ <port id="0" precision="FP32">
628
+ <dim>-1</dim>
629
+ <dim>64</dim>
630
+ <dim>-1</dim>
631
+ <dim>-1</dim>
632
+ </port>
633
+ <port id="1" precision="FP32">
634
+ <dim>1</dim>
635
+ <dim>64</dim>
636
+ <dim>1</dim>
637
+ <dim>1</dim>
638
+ </port>
639
+ </input>
640
+ <output>
641
+ <port id="2" precision="FP32" names="106,input.15">
642
+ <dim>-1</dim>
643
+ <dim>64</dim>
644
+ <dim>-1</dim>
645
+ <dim>-1</dim>
646
+ </port>
647
+ </output>
648
+ </layer>
649
+ <layer id="38" name="self.encoder.down_blocks.1.resnets.0.norm2.weight" type="Const" version="opset1">
650
+ <data element_type="f32" shape="64" offset="197248" size="256" />
651
+ <output>
652
+ <port id="0" precision="FP32" names="self.encoder.down_blocks.1.resnets.0.norm2.weight">
653
+ <dim>64</dim>
654
+ </port>
655
+ </output>
656
+ </layer>
657
+ <layer id="39" name="self.encoder.down_blocks.1.resnets.0.norm2.bias" type="Const" version="opset1">
658
+ <data element_type="f32" shape="64" offset="197504" size="256" />
659
+ <output>
660
+ <port id="0" precision="FP32" names="self.encoder.down_blocks.1.resnets.0.norm2.bias">
661
+ <dim>64</dim>
662
+ </port>
663
+ </output>
664
+ </layer>
665
+ <layer id="40" name="__module.encoder.down_blocks.1.resnets.0.norm2/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
666
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
667
+ <input>
668
+ <port id="0" precision="FP32">
669
+ <dim>-1</dim>
670
+ <dim>64</dim>
671
+ <dim>-1</dim>
672
+ <dim>-1</dim>
673
+ </port>
674
+ <port id="1" precision="FP32">
675
+ <dim>64</dim>
676
+ </port>
677
+ <port id="2" precision="FP32">
678
+ <dim>64</dim>
679
+ </port>
680
+ </input>
681
+ <output>
682
+ <port id="3" precision="FP32" names="109,input.17">
683
+ <dim>-1</dim>
684
+ <dim>64</dim>
685
+ <dim>-1</dim>
686
+ <dim>-1</dim>
687
+ </port>
688
+ </output>
689
+ </layer>
690
+ <layer id="41" name="__module.decoder.mid_block.resnets.1.nonlinearity/aten::silu/Swish_3" type="Swish" version="opset4">
691
+ <input>
692
+ <port id="0" precision="FP32">
693
+ <dim>-1</dim>
694
+ <dim>64</dim>
695
+ <dim>-1</dim>
696
+ <dim>-1</dim>
697
+ </port>
698
+ </input>
699
+ <output>
700
+ <port id="1" precision="FP32" names="110,input.19">
701
+ <dim>-1</dim>
702
+ <dim>64</dim>
703
+ <dim>-1</dim>
704
+ <dim>-1</dim>
705
+ </port>
706
+ </output>
707
+ </layer>
708
+ <layer id="42" name="self.encoder.down_blocks.1.resnets.0.conv2.weight" type="Const" version="opset1">
709
+ <data element_type="f32" shape="64, 64, 3, 3" offset="197760" size="147456" />
710
+ <output>
711
+ <port id="0" precision="FP32" names="self.encoder.down_blocks.1.resnets.0.conv2.weight">
712
+ <dim>64</dim>
713
+ <dim>64</dim>
714
+ <dim>3</dim>
715
+ <dim>3</dim>
716
+ </port>
717
+ </output>
718
+ </layer>
719
+ <layer id="43" name="__module.encoder.down_blocks.1.resnets.0.conv2/aten::_convolution/Convolution" type="Convolution" version="opset1">
720
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
721
+ <input>
722
+ <port id="0" precision="FP32">
723
+ <dim>-1</dim>
724
+ <dim>64</dim>
725
+ <dim>-1</dim>
726
+ <dim>-1</dim>
727
+ </port>
728
+ <port id="1" precision="FP32">
729
+ <dim>64</dim>
730
+ <dim>64</dim>
731
+ <dim>3</dim>
732
+ <dim>3</dim>
733
+ </port>
734
+ </input>
735
+ <output>
736
+ <port id="2" precision="FP32">
737
+ <dim>-1</dim>
738
+ <dim>64</dim>
739
+ <dim>-1</dim>
740
+ <dim>-1</dim>
741
+ </port>
742
+ </output>
743
+ </layer>
744
+ <layer id="44" name="__module.encoder.down_blocks.1.resnets.0.conv2/aten::_convolution/Reshape" type="Const" version="opset1">
745
+ <data element_type="f32" shape="1, 64, 1, 1" offset="345216" size="256" />
746
+ <output>
747
+ <port id="0" precision="FP32">
748
+ <dim>1</dim>
749
+ <dim>64</dim>
750
+ <dim>1</dim>
751
+ <dim>1</dim>
752
+ </port>
753
+ </output>
754
+ </layer>
755
+ <layer id="45" name="__module.encoder.down_blocks.1.resnets.0.conv2/aten::_convolution/Add" type="Add" version="opset1">
756
+ <data auto_broadcast="numpy" />
757
+ <input>
758
+ <port id="0" precision="FP32">
759
+ <dim>-1</dim>
760
+ <dim>64</dim>
761
+ <dim>-1</dim>
762
+ <dim>-1</dim>
763
+ </port>
764
+ <port id="1" precision="FP32">
765
+ <dim>1</dim>
766
+ <dim>64</dim>
767
+ <dim>1</dim>
768
+ <dim>1</dim>
769
+ </port>
770
+ </input>
771
+ <output>
772
+ <port id="2" precision="FP32" names="118,hidden_states.7">
773
+ <dim>-1</dim>
774
+ <dim>64</dim>
775
+ <dim>-1</dim>
776
+ <dim>-1</dim>
777
+ </port>
778
+ </output>
779
+ </layer>
780
+ <layer id="46" name="__module.encoder.down_blocks.1.resnets.0/aten::add/Add" type="Add" version="opset1">
781
+ <data auto_broadcast="numpy" />
782
+ <input>
783
+ <port id="0" precision="FP32">
784
+ <dim>-1</dim>
785
+ <dim>64</dim>
786
+ <dim>-1</dim>
787
+ <dim>-1</dim>
788
+ </port>
789
+ <port id="1" precision="FP32">
790
+ <dim>-1</dim>
791
+ <dim>64</dim>
792
+ <dim>-1</dim>
793
+ <dim>-1</dim>
794
+ </port>
795
+ </input>
796
+ <output>
797
+ <port id="2" precision="FP32" names="126,127,input.21">
798
+ <dim>-1</dim>
799
+ <dim>64</dim>
800
+ <dim>-1</dim>
801
+ <dim>-1</dim>
802
+ </port>
803
+ </output>
804
+ </layer>
805
+ <layer id="47" name="self.encoder.mid_block.resnets.0.norm1.weight" type="Const" version="opset1">
806
+ <data element_type="f32" shape="64" offset="197248" size="256" />
807
+ <output>
808
+ <port id="0" precision="FP32" names="self.encoder.mid_block.resnets.0.norm1.weight">
809
+ <dim>64</dim>
810
+ </port>
811
+ </output>
812
+ </layer>
813
+ <layer id="48" name="self.encoder.mid_block.resnets.0.norm1.bias" type="Const" version="opset1">
814
+ <data element_type="f32" shape="64" offset="197504" size="256" />
815
+ <output>
816
+ <port id="0" precision="FP32" names="self.encoder.mid_block.resnets.0.norm1.bias">
817
+ <dim>64</dim>
818
+ </port>
819
+ </output>
820
+ </layer>
821
+ <layer id="49" name="__module.encoder.mid_block.resnets.0.norm1/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
822
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
823
+ <input>
824
+ <port id="0" precision="FP32">
825
+ <dim>-1</dim>
826
+ <dim>64</dim>
827
+ <dim>-1</dim>
828
+ <dim>-1</dim>
829
+ </port>
830
+ <port id="1" precision="FP32">
831
+ <dim>64</dim>
832
+ </port>
833
+ <port id="2" precision="FP32">
834
+ <dim>64</dim>
835
+ </port>
836
+ </input>
837
+ <output>
838
+ <port id="3" precision="FP32" names="141,input.23">
839
+ <dim>-1</dim>
840
+ <dim>64</dim>
841
+ <dim>-1</dim>
842
+ <dim>-1</dim>
843
+ </port>
844
+ </output>
845
+ </layer>
846
+ <layer id="50" name="__module.decoder.mid_block.resnets.1.nonlinearity/aten::silu/Swish_4" type="Swish" version="opset4">
847
+ <input>
848
+ <port id="0" precision="FP32">
849
+ <dim>-1</dim>
850
+ <dim>64</dim>
851
+ <dim>-1</dim>
852
+ <dim>-1</dim>
853
+ </port>
854
+ </input>
855
+ <output>
856
+ <port id="1" precision="FP32" names="142">
857
+ <dim>-1</dim>
858
+ <dim>64</dim>
859
+ <dim>-1</dim>
860
+ <dim>-1</dim>
861
+ </port>
862
+ </output>
863
+ </layer>
864
+ <layer id="51" name="self.encoder.mid_block.resnets.0.conv1.weight" type="Const" version="opset1">
865
+ <data element_type="f32" shape="64, 64, 3, 3" offset="345472" size="147456" />
866
+ <output>
867
+ <port id="0" precision="FP32" names="self.encoder.mid_block.resnets.0.conv1.weight">
868
+ <dim>64</dim>
869
+ <dim>64</dim>
870
+ <dim>3</dim>
871
+ <dim>3</dim>
872
+ </port>
873
+ </output>
874
+ </layer>
875
+ <layer id="52" name="__module.encoder.mid_block.resnets.0.conv1/aten::_convolution/Convolution" type="Convolution" version="opset1">
876
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
877
+ <input>
878
+ <port id="0" precision="FP32">
879
+ <dim>-1</dim>
880
+ <dim>64</dim>
881
+ <dim>-1</dim>
882
+ <dim>-1</dim>
883
+ </port>
884
+ <port id="1" precision="FP32">
885
+ <dim>64</dim>
886
+ <dim>64</dim>
887
+ <dim>3</dim>
888
+ <dim>3</dim>
889
+ </port>
890
+ </input>
891
+ <output>
892
+ <port id="2" precision="FP32">
893
+ <dim>-1</dim>
894
+ <dim>64</dim>
895
+ <dim>-1</dim>
896
+ <dim>-1</dim>
897
+ </port>
898
+ </output>
899
+ </layer>
900
+ <layer id="53" name="__module.encoder.mid_block.resnets.0.conv1/aten::_convolution/Reshape" type="Const" version="opset1">
901
+ <data element_type="f32" shape="1, 64, 1, 1" offset="492928" size="256" />
902
+ <output>
903
+ <port id="0" precision="FP32">
904
+ <dim>1</dim>
905
+ <dim>64</dim>
906
+ <dim>1</dim>
907
+ <dim>1</dim>
908
+ </port>
909
+ </output>
910
+ </layer>
911
+ <layer id="54" name="__module.encoder.mid_block.resnets.0.conv1/aten::_convolution/Add" type="Add" version="opset1">
912
+ <data auto_broadcast="numpy" />
913
+ <input>
914
+ <port id="0" precision="FP32">
915
+ <dim>-1</dim>
916
+ <dim>64</dim>
917
+ <dim>-1</dim>
918
+ <dim>-1</dim>
919
+ </port>
920
+ <port id="1" precision="FP32">
921
+ <dim>1</dim>
922
+ <dim>64</dim>
923
+ <dim>1</dim>
924
+ <dim>1</dim>
925
+ </port>
926
+ </input>
927
+ <output>
928
+ <port id="2" precision="FP32" names="149,input.25">
929
+ <dim>-1</dim>
930
+ <dim>64</dim>
931
+ <dim>-1</dim>
932
+ <dim>-1</dim>
933
+ </port>
934
+ </output>
935
+ </layer>
936
+ <layer id="55" name="self.encoder.mid_block.resnets.0.norm2.weight" type="Const" version="opset1">
937
+ <data element_type="f32" shape="64" offset="197248" size="256" />
938
+ <output>
939
+ <port id="0" precision="FP32" names="self.encoder.mid_block.resnets.0.norm2.weight">
940
+ <dim>64</dim>
941
+ </port>
942
+ </output>
943
+ </layer>
944
+ <layer id="56" name="self.encoder.mid_block.resnets.0.norm2.bias" type="Const" version="opset1">
945
+ <data element_type="f32" shape="64" offset="197504" size="256" />
946
+ <output>
947
+ <port id="0" precision="FP32" names="self.encoder.mid_block.resnets.0.norm2.bias">
948
+ <dim>64</dim>
949
+ </port>
950
+ </output>
951
+ </layer>
952
+ <layer id="57" name="__module.encoder.mid_block.resnets.0.norm2/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
953
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
954
+ <input>
955
+ <port id="0" precision="FP32">
956
+ <dim>-1</dim>
957
+ <dim>64</dim>
958
+ <dim>-1</dim>
959
+ <dim>-1</dim>
960
+ </port>
961
+ <port id="1" precision="FP32">
962
+ <dim>64</dim>
963
+ </port>
964
+ <port id="2" precision="FP32">
965
+ <dim>64</dim>
966
+ </port>
967
+ </input>
968
+ <output>
969
+ <port id="3" precision="FP32" names="152,input.27">
970
+ <dim>-1</dim>
971
+ <dim>64</dim>
972
+ <dim>-1</dim>
973
+ <dim>-1</dim>
974
+ </port>
975
+ </output>
976
+ </layer>
977
+ <layer id="58" name="__module.decoder.mid_block.resnets.1.nonlinearity/aten::silu/Swish_5" type="Swish" version="opset4">
978
+ <input>
979
+ <port id="0" precision="FP32">
980
+ <dim>-1</dim>
981
+ <dim>64</dim>
982
+ <dim>-1</dim>
983
+ <dim>-1</dim>
984
+ </port>
985
+ </input>
986
+ <output>
987
+ <port id="1" precision="FP32" names="153,input.29">
988
+ <dim>-1</dim>
989
+ <dim>64</dim>
990
+ <dim>-1</dim>
991
+ <dim>-1</dim>
992
+ </port>
993
+ </output>
994
+ </layer>
995
+ <layer id="59" name="self.encoder.mid_block.resnets.0.conv2.weight" type="Const" version="opset1">
996
+ <data element_type="f32" shape="64, 64, 3, 3" offset="493184" size="147456" />
997
+ <output>
998
+ <port id="0" precision="FP32" names="self.encoder.mid_block.resnets.0.conv2.weight">
999
+ <dim>64</dim>
1000
+ <dim>64</dim>
1001
+ <dim>3</dim>
1002
+ <dim>3</dim>
1003
+ </port>
1004
+ </output>
1005
+ </layer>
1006
+ <layer id="60" name="__module.encoder.mid_block.resnets.0.conv2/aten::_convolution/Convolution" type="Convolution" version="opset1">
1007
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
1008
+ <input>
1009
+ <port id="0" precision="FP32">
1010
+ <dim>-1</dim>
1011
+ <dim>64</dim>
1012
+ <dim>-1</dim>
1013
+ <dim>-1</dim>
1014
+ </port>
1015
+ <port id="1" precision="FP32">
1016
+ <dim>64</dim>
1017
+ <dim>64</dim>
1018
+ <dim>3</dim>
1019
+ <dim>3</dim>
1020
+ </port>
1021
+ </input>
1022
+ <output>
1023
+ <port id="2" precision="FP32">
1024
+ <dim>-1</dim>
1025
+ <dim>64</dim>
1026
+ <dim>-1</dim>
1027
+ <dim>-1</dim>
1028
+ </port>
1029
+ </output>
1030
+ </layer>
1031
+ <layer id="61" name="__module.encoder.mid_block.resnets.0.conv2/aten::_convolution/Reshape" type="Const" version="opset1">
1032
+ <data element_type="f32" shape="1, 64, 1, 1" offset="640640" size="256" />
1033
+ <output>
1034
+ <port id="0" precision="FP32">
1035
+ <dim>1</dim>
1036
+ <dim>64</dim>
1037
+ <dim>1</dim>
1038
+ <dim>1</dim>
1039
+ </port>
1040
+ </output>
1041
+ </layer>
1042
+ <layer id="62" name="__module.encoder.mid_block.resnets.0.conv2/aten::_convolution/Add" type="Add" version="opset1">
1043
+ <data auto_broadcast="numpy" />
1044
+ <input>
1045
+ <port id="0" precision="FP32">
1046
+ <dim>-1</dim>
1047
+ <dim>64</dim>
1048
+ <dim>-1</dim>
1049
+ <dim>-1</dim>
1050
+ </port>
1051
+ <port id="1" precision="FP32">
1052
+ <dim>1</dim>
1053
+ <dim>64</dim>
1054
+ <dim>1</dim>
1055
+ <dim>1</dim>
1056
+ </port>
1057
+ </input>
1058
+ <output>
1059
+ <port id="2" precision="FP32" names="161,hidden_states.9">
1060
+ <dim>-1</dim>
1061
+ <dim>64</dim>
1062
+ <dim>-1</dim>
1063
+ <dim>-1</dim>
1064
+ </port>
1065
+ </output>
1066
+ </layer>
1067
+ <layer id="63" name="__module.encoder.mid_block.resnets.0/aten::add/Add" type="Add" version="opset1">
1068
+ <data auto_broadcast="numpy" />
1069
+ <input>
1070
+ <port id="0" precision="FP32">
1071
+ <dim>-1</dim>
1072
+ <dim>64</dim>
1073
+ <dim>-1</dim>
1074
+ <dim>-1</dim>
1075
+ </port>
1076
+ <port id="1" precision="FP32">
1077
+ <dim>-1</dim>
1078
+ <dim>64</dim>
1079
+ <dim>-1</dim>
1080
+ <dim>-1</dim>
1081
+ </port>
1082
+ </input>
1083
+ <output>
1084
+ <port id="2" precision="FP32" names="162,163,hidden_states.11">
1085
+ <dim>-1</dim>
1086
+ <dim>64</dim>
1087
+ <dim>-1</dim>
1088
+ <dim>-1</dim>
1089
+ </port>
1090
+ </output>
1091
+ </layer>
1092
+ <layer id="64" name="Constant_48490" type="Const" version="opset1">
1093
+ <data element_type="i64" shape="3" offset="640896" size="24" />
1094
+ <output>
1095
+ <port id="0" precision="I64">
1096
+ <dim>3</dim>
1097
+ </port>
1098
+ </output>
1099
+ </layer>
1100
+ <layer id="65" name="__module.encoder.mid_block.attentions.0/aten::view/Reshape" type="Reshape" version="opset1">
1101
+ <data special_zero="true" />
1102
+ <input>
1103
+ <port id="0" precision="FP32">
1104
+ <dim>-1</dim>
1105
+ <dim>64</dim>
1106
+ <dim>-1</dim>
1107
+ <dim>-1</dim>
1108
+ </port>
1109
+ <port id="1" precision="I64">
1110
+ <dim>3</dim>
1111
+ </port>
1112
+ </input>
1113
+ <output>
1114
+ <port id="2" precision="FP32" names="179">
1115
+ <dim>-1</dim>
1116
+ <dim>64</dim>
1117
+ <dim>-1</dim>
1118
+ </port>
1119
+ </output>
1120
+ </layer>
1121
+ <layer id="66" name="__module.encoder.mid_block.attentions.0/aten::transpose/Constant" type="Const" version="opset1">
1122
+ <data element_type="i32" shape="3" offset="640920" size="12" />
1123
+ <output>
1124
+ <port id="0" precision="I32">
1125
+ <dim>3</dim>
1126
+ </port>
1127
+ </output>
1128
+ </layer>
1129
+ <layer id="67" name="__module.encoder.mid_block.attentions.0/aten::transpose/Transpose" type="Transpose" version="opset1">
1130
+ <input>
1131
+ <port id="0" precision="FP32">
1132
+ <dim>-1</dim>
1133
+ <dim>64</dim>
1134
+ <dim>-1</dim>
1135
+ </port>
1136
+ <port id="1" precision="I32">
1137
+ <dim>3</dim>
1138
+ </port>
1139
+ </input>
1140
+ <output>
1141
+ <port id="2" precision="FP32" names="180,hidden_states.13">
1142
+ <dim>-1</dim>
1143
+ <dim>-1</dim>
1144
+ <dim>64</dim>
1145
+ </port>
1146
+ </output>
1147
+ </layer>
1148
+ <layer id="68" name="__module.encoder.mid_block.attentions.0/aten::transpose/Constant_1" type="Const" version="opset1">
1149
+ <data element_type="i32" shape="3" offset="640920" size="12" />
1150
+ <output>
1151
+ <port id="0" precision="I32">
1152
+ <dim>3</dim>
1153
+ </port>
1154
+ </output>
1155
+ </layer>
1156
+ <layer id="69" name="__module.encoder.mid_block.attentions.0/aten::transpose/Transpose_1" type="Transpose" version="opset1">
1157
+ <input>
1158
+ <port id="0" precision="FP32">
1159
+ <dim>-1</dim>
1160
+ <dim>-1</dim>
1161
+ <dim>64</dim>
1162
+ </port>
1163
+ <port id="1" precision="I32">
1164
+ <dim>3</dim>
1165
+ </port>
1166
+ </input>
1167
+ <output>
1168
+ <port id="2" precision="FP32" names="182,input.31">
1169
+ <dim>-1</dim>
1170
+ <dim>64</dim>
1171
+ <dim>-1</dim>
1172
+ </port>
1173
+ </output>
1174
+ </layer>
1175
+ <layer id="70" name="self.encoder.mid_block.attentions.0.group_norm.weight" type="Const" version="opset1">
1176
+ <data element_type="f32" shape="64" offset="197248" size="256" />
1177
+ <output>
1178
+ <port id="0" precision="FP32" names="self.encoder.mid_block.attentions.0.group_norm.weight">
1179
+ <dim>64</dim>
1180
+ </port>
1181
+ </output>
1182
+ </layer>
1183
+ <layer id="71" name="self.encoder.mid_block.attentions.0.group_norm.bias" type="Const" version="opset1">
1184
+ <data element_type="f32" shape="64" offset="197504" size="256" />
1185
+ <output>
1186
+ <port id="0" precision="FP32" names="self.encoder.mid_block.attentions.0.group_norm.bias">
1187
+ <dim>64</dim>
1188
+ </port>
1189
+ </output>
1190
+ </layer>
1191
+ <layer id="72" name="__module.encoder.mid_block.attentions.0.group_norm/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
1192
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
1193
+ <input>
1194
+ <port id="0" precision="FP32">
1195
+ <dim>-1</dim>
1196
+ <dim>64</dim>
1197
+ <dim>-1</dim>
1198
+ </port>
1199
+ <port id="1" precision="FP32">
1200
+ <dim>64</dim>
1201
+ </port>
1202
+ <port id="2" precision="FP32">
1203
+ <dim>64</dim>
1204
+ </port>
1205
+ </input>
1206
+ <output>
1207
+ <port id="3" precision="FP32" names="185">
1208
+ <dim>-1</dim>
1209
+ <dim>64</dim>
1210
+ <dim>-1</dim>
1211
+ </port>
1212
+ </output>
1213
+ </layer>
1214
+ <layer id="73" name="self.encoder.mid_block.attentions.0.to_q.weight" type="Const" version="opset1">
1215
+ <data element_type="f32" shape="64, 64" offset="640932" size="16384" />
1216
+ <output>
1217
+ <port id="0" precision="FP32" names="self.encoder.mid_block.attentions.0.to_q.weight">
1218
+ <dim>64</dim>
1219
+ <dim>64</dim>
1220
+ </port>
1221
+ </output>
1222
+ </layer>
1223
+ <layer id="74" name="__module.encoder.mid_block.attentions.0.to_q/aten::linear/MatMul" type="MatMul" version="opset1">
1224
+ <data transpose_a="true" transpose_b="true" />
1225
+ <input>
1226
+ <port id="0" precision="FP32">
1227
+ <dim>-1</dim>
1228
+ <dim>64</dim>
1229
+ <dim>-1</dim>
1230
+ </port>
1231
+ <port id="1" precision="FP32">
1232
+ <dim>64</dim>
1233
+ <dim>64</dim>
1234
+ </port>
1235
+ </input>
1236
+ <output>
1237
+ <port id="2" precision="FP32">
1238
+ <dim>-1</dim>
1239
+ <dim>-1</dim>
1240
+ <dim>64</dim>
1241
+ </port>
1242
+ </output>
1243
+ </layer>
1244
+ <layer id="75" name="Constant_48375" type="Const" version="opset1">
1245
+ <data element_type="f32" shape="1, 1, 64" offset="657316" size="256" />
1246
+ <output>
1247
+ <port id="0" precision="FP32">
1248
+ <dim>1</dim>
1249
+ <dim>1</dim>
1250
+ <dim>64</dim>
1251
+ </port>
1252
+ </output>
1253
+ </layer>
1254
+ <layer id="76" name="__module.encoder.mid_block.attentions.0.to_q/aten::linear/Add" type="Add" version="opset1">
1255
+ <data auto_broadcast="numpy" />
1256
+ <input>
1257
+ <port id="0" precision="FP32">
1258
+ <dim>-1</dim>
1259
+ <dim>-1</dim>
1260
+ <dim>64</dim>
1261
+ </port>
1262
+ <port id="1" precision="FP32">
1263
+ <dim>1</dim>
1264
+ <dim>1</dim>
1265
+ <dim>64</dim>
1266
+ </port>
1267
+ </input>
1268
+ <output>
1269
+ <port id="2" precision="FP32" names="189,query">
1270
+ <dim>-1</dim>
1271
+ <dim>-1</dim>
1272
+ <dim>64</dim>
1273
+ </port>
1274
+ </output>
1275
+ </layer>
1276
+ <layer id="77" name="Constant_48491" type="Const" version="opset1">
1277
+ <data element_type="i64" shape="4" offset="657572" size="32" />
1278
+ <output>
1279
+ <port id="0" precision="I64">
1280
+ <dim>4</dim>
1281
+ </port>
1282
+ </output>
1283
+ </layer>
1284
+ <layer id="78" name="__module.encoder.mid_block.attentions.0/aten::view/Reshape_1" type="Reshape" version="opset1">
1285
+ <data special_zero="true" />
1286
+ <input>
1287
+ <port id="0" precision="FP32">
1288
+ <dim>-1</dim>
1289
+ <dim>-1</dim>
1290
+ <dim>64</dim>
1291
+ </port>
1292
+ <port id="1" precision="I64">
1293
+ <dim>4</dim>
1294
+ </port>
1295
+ </input>
1296
+ <output>
1297
+ <port id="2" precision="FP32" names="203">
1298
+ <dim>-1</dim>
1299
+ <dim>-1</dim>
1300
+ <dim>1</dim>
1301
+ <dim>64</dim>
1302
+ </port>
1303
+ </output>
1304
+ </layer>
1305
+ <layer id="79" name="Constant_48320" type="Const" version="opset1">
1306
+ <data element_type="i64" shape="4" offset="657604" size="32" />
1307
+ <output>
1308
+ <port id="0" precision="I64">
1309
+ <dim>4</dim>
1310
+ </port>
1311
+ </output>
1312
+ </layer>
1313
+ <layer id="80" name="__module.encoder.mid_block.attentions.0/aten::transpose/Transpose_3" type="Reshape" version="opset1">
1314
+ <data special_zero="true" />
1315
+ <input>
1316
+ <port id="0" precision="FP32">
1317
+ <dim>-1</dim>
1318
+ <dim>-1</dim>
1319
+ <dim>1</dim>
1320
+ <dim>64</dim>
1321
+ </port>
1322
+ <port id="1" precision="I64">
1323
+ <dim>4</dim>
1324
+ </port>
1325
+ </input>
1326
+ <output>
1327
+ <port id="2" precision="FP32" names="204">
1328
+ <dim>-1</dim>
1329
+ <dim>1</dim>
1330
+ <dim>-1</dim>
1331
+ <dim>64</dim>
1332
+ </port>
1333
+ </output>
1334
+ </layer>
1335
+ <layer id="81" name="self.encoder.mid_block.attentions.0.to_k.weight" type="Const" version="opset1">
1336
+ <data element_type="f32" shape="64, 64" offset="657636" size="16384" />
1337
+ <output>
1338
+ <port id="0" precision="FP32" names="self.encoder.mid_block.attentions.0.to_k.weight">
1339
+ <dim>64</dim>
1340
+ <dim>64</dim>
1341
+ </port>
1342
+ </output>
1343
+ </layer>
1344
+ <layer id="82" name="__module.encoder.mid_block.attentions.0.to_k/aten::linear/MatMul" type="MatMul" version="opset1">
1345
+ <data transpose_a="true" transpose_b="true" />
1346
+ <input>
1347
+ <port id="0" precision="FP32">
1348
+ <dim>-1</dim>
1349
+ <dim>64</dim>
1350
+ <dim>-1</dim>
1351
+ </port>
1352
+ <port id="1" precision="FP32">
1353
+ <dim>64</dim>
1354
+ <dim>64</dim>
1355
+ </port>
1356
+ </input>
1357
+ <output>
1358
+ <port id="2" precision="FP32">
1359
+ <dim>-1</dim>
1360
+ <dim>-1</dim>
1361
+ <dim>64</dim>
1362
+ </port>
1363
+ </output>
1364
+ </layer>
1365
+ <layer id="83" name="Constant_48376" type="Const" version="opset1">
1366
+ <data element_type="f32" shape="1, 1, 64" offset="674020" size="256" />
1367
+ <output>
1368
+ <port id="0" precision="FP32">
1369
+ <dim>1</dim>
1370
+ <dim>1</dim>
1371
+ <dim>64</dim>
1372
+ </port>
1373
+ </output>
1374
+ </layer>
1375
+ <layer id="84" name="__module.encoder.mid_block.attentions.0.to_k/aten::linear/Add" type="Add" version="opset1">
1376
+ <data auto_broadcast="numpy" />
1377
+ <input>
1378
+ <port id="0" precision="FP32">
1379
+ <dim>-1</dim>
1380
+ <dim>-1</dim>
1381
+ <dim>64</dim>
1382
+ </port>
1383
+ <port id="1" precision="FP32">
1384
+ <dim>1</dim>
1385
+ <dim>1</dim>
1386
+ <dim>64</dim>
1387
+ </port>
1388
+ </input>
1389
+ <output>
1390
+ <port id="2" precision="FP32" names="192,key">
1391
+ <dim>-1</dim>
1392
+ <dim>-1</dim>
1393
+ <dim>64</dim>
1394
+ </port>
1395
+ </output>
1396
+ </layer>
1397
+ <layer id="85" name="Constant_48492" type="Const" version="opset1">
1398
+ <data element_type="i64" shape="4" offset="657572" size="32" />
1399
+ <output>
1400
+ <port id="0" precision="I64">
1401
+ <dim>4</dim>
1402
+ </port>
1403
+ </output>
1404
+ </layer>
1405
+ <layer id="86" name="__module.encoder.mid_block.attentions.0/aten::view/Reshape_2" type="Reshape" version="opset1">
1406
+ <data special_zero="true" />
1407
+ <input>
1408
+ <port id="0" precision="FP32">
1409
+ <dim>-1</dim>
1410
+ <dim>-1</dim>
1411
+ <dim>64</dim>
1412
+ </port>
1413
+ <port id="1" precision="I64">
1414
+ <dim>4</dim>
1415
+ </port>
1416
+ </input>
1417
+ <output>
1418
+ <port id="2" precision="FP32" names="206">
1419
+ <dim>-1</dim>
1420
+ <dim>-1</dim>
1421
+ <dim>1</dim>
1422
+ <dim>64</dim>
1423
+ </port>
1424
+ </output>
1425
+ </layer>
1426
+ <layer id="87" name="Constant_48324" type="Const" version="opset1">
1427
+ <data element_type="i64" shape="4" offset="657604" size="32" />
1428
+ <output>
1429
+ <port id="0" precision="I64">
1430
+ <dim>4</dim>
1431
+ </port>
1432
+ </output>
1433
+ </layer>
1434
+ <layer id="88" name="__module.encoder.mid_block.attentions.0/aten::transpose/Transpose_4" type="Reshape" version="opset1">
1435
+ <data special_zero="true" />
1436
+ <input>
1437
+ <port id="0" precision="FP32">
1438
+ <dim>-1</dim>
1439
+ <dim>-1</dim>
1440
+ <dim>1</dim>
1441
+ <dim>64</dim>
1442
+ </port>
1443
+ <port id="1" precision="I64">
1444
+ <dim>4</dim>
1445
+ </port>
1446
+ </input>
1447
+ <output>
1448
+ <port id="2" precision="FP32" names="207">
1449
+ <dim>-1</dim>
1450
+ <dim>1</dim>
1451
+ <dim>-1</dim>
1452
+ <dim>64</dim>
1453
+ </port>
1454
+ </output>
1455
+ </layer>
1456
+ <layer id="89" name="self.encoder.mid_block.attentions.0.to_v.weight" type="Const" version="opset1">
1457
+ <data element_type="f32" shape="64, 64" offset="674276" size="16384" />
1458
+ <output>
1459
+ <port id="0" precision="FP32" names="self.encoder.mid_block.attentions.0.to_v.weight">
1460
+ <dim>64</dim>
1461
+ <dim>64</dim>
1462
+ </port>
1463
+ </output>
1464
+ </layer>
1465
+ <layer id="90" name="__module.encoder.mid_block.attentions.0.to_v/aten::linear/MatMul" type="MatMul" version="opset1">
1466
+ <data transpose_a="true" transpose_b="true" />
1467
+ <input>
1468
+ <port id="0" precision="FP32">
1469
+ <dim>-1</dim>
1470
+ <dim>64</dim>
1471
+ <dim>-1</dim>
1472
+ </port>
1473
+ <port id="1" precision="FP32">
1474
+ <dim>64</dim>
1475
+ <dim>64</dim>
1476
+ </port>
1477
+ </input>
1478
+ <output>
1479
+ <port id="2" precision="FP32">
1480
+ <dim>-1</dim>
1481
+ <dim>-1</dim>
1482
+ <dim>64</dim>
1483
+ </port>
1484
+ </output>
1485
+ </layer>
1486
+ <layer id="91" name="Constant_48377" type="Const" version="opset1">
1487
+ <data element_type="f32" shape="1, 1, 64" offset="690660" size="256" />
1488
+ <output>
1489
+ <port id="0" precision="FP32">
1490
+ <dim>1</dim>
1491
+ <dim>1</dim>
1492
+ <dim>64</dim>
1493
+ </port>
1494
+ </output>
1495
+ </layer>
1496
+ <layer id="92" name="__module.encoder.mid_block.attentions.0.to_v/aten::linear/Add" type="Add" version="opset1">
1497
+ <data auto_broadcast="numpy" />
1498
+ <input>
1499
+ <port id="0" precision="FP32">
1500
+ <dim>-1</dim>
1501
+ <dim>-1</dim>
1502
+ <dim>64</dim>
1503
+ </port>
1504
+ <port id="1" precision="FP32">
1505
+ <dim>1</dim>
1506
+ <dim>1</dim>
1507
+ <dim>64</dim>
1508
+ </port>
1509
+ </input>
1510
+ <output>
1511
+ <port id="2" precision="FP32" names="195,value">
1512
+ <dim>-1</dim>
1513
+ <dim>-1</dim>
1514
+ <dim>64</dim>
1515
+ </port>
1516
+ </output>
1517
+ </layer>
1518
+ <layer id="93" name="Constant_48493" type="Const" version="opset1">
1519
+ <data element_type="i64" shape="4" offset="657572" size="32" />
1520
+ <output>
1521
+ <port id="0" precision="I64">
1522
+ <dim>4</dim>
1523
+ </port>
1524
+ </output>
1525
+ </layer>
1526
+ <layer id="94" name="__module.encoder.mid_block.attentions.0/aten::view/Reshape_3" type="Reshape" version="opset1">
1527
+ <data special_zero="true" />
1528
+ <input>
1529
+ <port id="0" precision="FP32">
1530
+ <dim>-1</dim>
1531
+ <dim>-1</dim>
1532
+ <dim>64</dim>
1533
+ </port>
1534
+ <port id="1" precision="I64">
1535
+ <dim>4</dim>
1536
+ </port>
1537
+ </input>
1538
+ <output>
1539
+ <port id="2" precision="FP32" names="209">
1540
+ <dim>-1</dim>
1541
+ <dim>-1</dim>
1542
+ <dim>1</dim>
1543
+ <dim>64</dim>
1544
+ </port>
1545
+ </output>
1546
+ </layer>
1547
+ <layer id="95" name="Constant_48328" type="Const" version="opset1">
1548
+ <data element_type="i64" shape="4" offset="657604" size="32" />
1549
+ <output>
1550
+ <port id="0" precision="I64">
1551
+ <dim>4</dim>
1552
+ </port>
1553
+ </output>
1554
+ </layer>
1555
+ <layer id="96" name="__module.encoder.mid_block.attentions.0/aten::transpose/Transpose_5" type="Reshape" version="opset1">
1556
+ <data special_zero="true" />
1557
+ <input>
1558
+ <port id="0" precision="FP32">
1559
+ <dim>-1</dim>
1560
+ <dim>-1</dim>
1561
+ <dim>1</dim>
1562
+ <dim>64</dim>
1563
+ </port>
1564
+ <port id="1" precision="I64">
1565
+ <dim>4</dim>
1566
+ </port>
1567
+ </input>
1568
+ <output>
1569
+ <port id="2" precision="FP32" names="210">
1570
+ <dim>-1</dim>
1571
+ <dim>1</dim>
1572
+ <dim>-1</dim>
1573
+ <dim>64</dim>
1574
+ </port>
1575
+ </output>
1576
+ </layer>
1577
+ <layer id="97" name="__module.encoder.mid_block.attentions.0/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
1578
+ <data causal="false" />
1579
+ <input>
1580
+ <port id="0" precision="FP32">
1581
+ <dim>-1</dim>
1582
+ <dim>1</dim>
1583
+ <dim>-1</dim>
1584
+ <dim>64</dim>
1585
+ </port>
1586
+ <port id="1" precision="FP32">
1587
+ <dim>-1</dim>
1588
+ <dim>1</dim>
1589
+ <dim>-1</dim>
1590
+ <dim>64</dim>
1591
+ </port>
1592
+ <port id="2" precision="FP32">
1593
+ <dim>-1</dim>
1594
+ <dim>1</dim>
1595
+ <dim>-1</dim>
1596
+ <dim>64</dim>
1597
+ </port>
1598
+ </input>
1599
+ <output>
1600
+ <port id="3" precision="FP32" names="211,hidden_states.15">
1601
+ <dim>-1</dim>
1602
+ <dim>1</dim>
1603
+ <dim>-1</dim>
1604
+ <dim>64</dim>
1605
+ </port>
1606
+ </output>
1607
+ </layer>
1608
+ <layer id="98" name="Constant_48330" type="Const" version="opset1">
1609
+ <data element_type="i64" shape="4" offset="690916" size="32" />
1610
+ <output>
1611
+ <port id="0" precision="I64">
1612
+ <dim>4</dim>
1613
+ </port>
1614
+ </output>
1615
+ </layer>
1616
+ <layer id="99" name="__module.encoder.mid_block.attentions.0/aten::transpose/Transpose_6" type="Reshape" version="opset1">
1617
+ <data special_zero="true" />
1618
+ <input>
1619
+ <port id="0" precision="FP32">
1620
+ <dim>-1</dim>
1621
+ <dim>1</dim>
1622
+ <dim>-1</dim>
1623
+ <dim>64</dim>
1624
+ </port>
1625
+ <port id="1" precision="I64">
1626
+ <dim>4</dim>
1627
+ </port>
1628
+ </input>
1629
+ <output>
1630
+ <port id="2" precision="FP32" names="212">
1631
+ <dim>-1</dim>
1632
+ <dim>-1</dim>
1633
+ <dim>1</dim>
1634
+ <dim>64</dim>
1635
+ </port>
1636
+ </output>
1637
+ </layer>
1638
+ <layer id="100" name="Constant_48494" type="Const" version="opset1">
1639
+ <data element_type="i64" shape="3" offset="690948" size="24" />
1640
+ <output>
1641
+ <port id="0" precision="I64">
1642
+ <dim>3</dim>
1643
+ </port>
1644
+ </output>
1645
+ </layer>
1646
+ <layer id="101" name="__module.encoder.mid_block.attentions.0/aten::reshape/Reshape" type="Reshape" version="opset1">
1647
+ <data special_zero="true" />
1648
+ <input>
1649
+ <port id="0" precision="FP32">
1650
+ <dim>-1</dim>
1651
+ <dim>-1</dim>
1652
+ <dim>1</dim>
1653
+ <dim>64</dim>
1654
+ </port>
1655
+ <port id="1" precision="I64">
1656
+ <dim>3</dim>
1657
+ </port>
1658
+ </input>
1659
+ <output>
1660
+ <port id="2" precision="FP32" names="216,217,hidden_states.17">
1661
+ <dim>-1</dim>
1662
+ <dim>-1</dim>
1663
+ <dim>64</dim>
1664
+ </port>
1665
+ </output>
1666
+ </layer>
1667
+ <layer id="102" name="self.encoder.mid_block.attentions.0.to_out.0.weight" type="Const" version="opset1">
1668
+ <data element_type="f32" shape="64, 64" offset="690972" size="16384" />
1669
+ <output>
1670
+ <port id="0" precision="FP32" names="self.encoder.mid_block.attentions.0.to_out.0.weight">
1671
+ <dim>64</dim>
1672
+ <dim>64</dim>
1673
+ </port>
1674
+ </output>
1675
+ </layer>
1676
+ <layer id="103" name="__module.encoder.mid_block.attentions.0.to_out.0/aten::linear/MatMul" type="MatMul" version="opset1">
1677
+ <data transpose_a="false" transpose_b="true" />
1678
+ <input>
1679
+ <port id="0" precision="FP32">
1680
+ <dim>-1</dim>
1681
+ <dim>-1</dim>
1682
+ <dim>64</dim>
1683
+ </port>
1684
+ <port id="1" precision="FP32">
1685
+ <dim>64</dim>
1686
+ <dim>64</dim>
1687
+ </port>
1688
+ </input>
1689
+ <output>
1690
+ <port id="2" precision="FP32">
1691
+ <dim>-1</dim>
1692
+ <dim>-1</dim>
1693
+ <dim>64</dim>
1694
+ </port>
1695
+ </output>
1696
+ </layer>
1697
+ <layer id="104" name="Constant_48378" type="Const" version="opset1">
1698
+ <data element_type="f32" shape="1, 1, 64" offset="707356" size="256" />
1699
+ <output>
1700
+ <port id="0" precision="FP32">
1701
+ <dim>1</dim>
1702
+ <dim>1</dim>
1703
+ <dim>64</dim>
1704
+ </port>
1705
+ </output>
1706
+ </layer>
1707
+ <layer id="105" name="__module.encoder.mid_block.attentions.0.to_out.0/aten::linear/Add" type="Add" version="opset1">
1708
+ <data auto_broadcast="numpy" />
1709
+ <input>
1710
+ <port id="0" precision="FP32">
1711
+ <dim>-1</dim>
1712
+ <dim>-1</dim>
1713
+ <dim>64</dim>
1714
+ </port>
1715
+ <port id="1" precision="FP32">
1716
+ <dim>1</dim>
1717
+ <dim>1</dim>
1718
+ <dim>64</dim>
1719
+ </port>
1720
+ </input>
1721
+ <output>
1722
+ <port id="2" precision="FP32" names="220,input.33">
1723
+ <dim>-1</dim>
1724
+ <dim>-1</dim>
1725
+ <dim>64</dim>
1726
+ </port>
1727
+ </output>
1728
+ </layer>
1729
+ <layer id="106" name="__module.encoder.mid_block.attentions.0/aten::transpose/Constant_7" type="Const" version="opset1">
1730
+ <data element_type="i32" shape="3" offset="640920" size="12" />
1731
+ <output>
1732
+ <port id="0" precision="I32">
1733
+ <dim>3</dim>
1734
+ </port>
1735
+ </output>
1736
+ </layer>
1737
+ <layer id="107" name="__module.encoder.mid_block.attentions.0/aten::transpose/Transpose_7" type="Transpose" version="opset1">
1738
+ <input>
1739
+ <port id="0" precision="FP32">
1740
+ <dim>-1</dim>
1741
+ <dim>-1</dim>
1742
+ <dim>64</dim>
1743
+ </port>
1744
+ <port id="1" precision="I32">
1745
+ <dim>3</dim>
1746
+ </port>
1747
+ </input>
1748
+ <output>
1749
+ <port id="2" precision="FP32" names="222">
1750
+ <dim>-1</dim>
1751
+ <dim>64</dim>
1752
+ <dim>-1</dim>
1753
+ </port>
1754
+ </output>
1755
+ </layer>
1756
+ <layer id="108" name="__module.encoder.mid_block.attentions.0/aten::size/ShapeOf" type="ShapeOf" version="opset3">
1757
+ <data output_type="i64" />
1758
+ <input>
1759
+ <port id="0" precision="FP32">
1760
+ <dim>-1</dim>
1761
+ <dim>64</dim>
1762
+ <dim>-1</dim>
1763
+ <dim>-1</dim>
1764
+ </port>
1765
+ </input>
1766
+ <output>
1767
+ <port id="1" precision="I64">
1768
+ <dim>4</dim>
1769
+ </port>
1770
+ </output>
1771
+ </layer>
1772
+ <layer id="109" name="__module.encoder.mid_block.attentions.0/aten::reshape/Reshape_1" type="Reshape" version="opset1">
1773
+ <data special_zero="false" />
1774
+ <input>
1775
+ <port id="0" precision="FP32">
1776
+ <dim>-1</dim>
1777
+ <dim>64</dim>
1778
+ <dim>-1</dim>
1779
+ </port>
1780
+ <port id="1" precision="I64">
1781
+ <dim>4</dim>
1782
+ </port>
1783
+ </input>
1784
+ <output>
1785
+ <port id="2" precision="FP32" names="224,hidden_states.21">
1786
+ <dim>-1</dim>
1787
+ <dim>64</dim>
1788
+ <dim>-1</dim>
1789
+ <dim>-1</dim>
1790
+ </port>
1791
+ </output>
1792
+ </layer>
1793
+ <layer id="110" name="__module.encoder.mid_block.attentions.0/aten::add/Add" type="Add" version="opset1">
1794
+ <data auto_broadcast="numpy" />
1795
+ <input>
1796
+ <port id="0" precision="FP32">
1797
+ <dim>-1</dim>
1798
+ <dim>64</dim>
1799
+ <dim>-1</dim>
1800
+ <dim>-1</dim>
1801
+ </port>
1802
+ <port id="1" precision="FP32">
1803
+ <dim>-1</dim>
1804
+ <dim>64</dim>
1805
+ <dim>-1</dim>
1806
+ <dim>-1</dim>
1807
+ </port>
1808
+ </input>
1809
+ <output>
1810
+ <port id="2" precision="FP32" names="225,226,hidden_states.23,input.35">
1811
+ <dim>-1</dim>
1812
+ <dim>64</dim>
1813
+ <dim>-1</dim>
1814
+ <dim>-1</dim>
1815
+ </port>
1816
+ </output>
1817
+ </layer>
1818
+ <layer id="111" name="self.encoder.mid_block.resnets.1.norm1.weight" type="Const" version="opset1">
1819
+ <data element_type="f32" shape="64" offset="197248" size="256" />
1820
+ <output>
1821
+ <port id="0" precision="FP32" names="self.encoder.mid_block.resnets.1.norm1.weight">
1822
+ <dim>64</dim>
1823
+ </port>
1824
+ </output>
1825
+ </layer>
1826
+ <layer id="112" name="self.encoder.mid_block.resnets.1.norm1.bias" type="Const" version="opset1">
1827
+ <data element_type="f32" shape="64" offset="197504" size="256" />
1828
+ <output>
1829
+ <port id="0" precision="FP32" names="self.encoder.mid_block.resnets.1.norm1.bias">
1830
+ <dim>64</dim>
1831
+ </port>
1832
+ </output>
1833
+ </layer>
1834
+ <layer id="113" name="__module.encoder.mid_block.resnets.1.norm1/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
1835
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
1836
+ <input>
1837
+ <port id="0" precision="FP32">
1838
+ <dim>-1</dim>
1839
+ <dim>64</dim>
1840
+ <dim>-1</dim>
1841
+ <dim>-1</dim>
1842
+ </port>
1843
+ <port id="1" precision="FP32">
1844
+ <dim>64</dim>
1845
+ </port>
1846
+ <port id="2" precision="FP32">
1847
+ <dim>64</dim>
1848
+ </port>
1849
+ </input>
1850
+ <output>
1851
+ <port id="3" precision="FP32" names="234,input.37">
1852
+ <dim>-1</dim>
1853
+ <dim>64</dim>
1854
+ <dim>-1</dim>
1855
+ <dim>-1</dim>
1856
+ </port>
1857
+ </output>
1858
+ </layer>
1859
+ <layer id="114" name="__module.decoder.mid_block.resnets.1.nonlinearity/aten::silu/Swish_6" type="Swish" version="opset4">
1860
+ <input>
1861
+ <port id="0" precision="FP32">
1862
+ <dim>-1</dim>
1863
+ <dim>64</dim>
1864
+ <dim>-1</dim>
1865
+ <dim>-1</dim>
1866
+ </port>
1867
+ </input>
1868
+ <output>
1869
+ <port id="1" precision="FP32" names="235">
1870
+ <dim>-1</dim>
1871
+ <dim>64</dim>
1872
+ <dim>-1</dim>
1873
+ <dim>-1</dim>
1874
+ </port>
1875
+ </output>
1876
+ </layer>
1877
+ <layer id="115" name="self.encoder.mid_block.resnets.1.conv1.weight" type="Const" version="opset1">
1878
+ <data element_type="f32" shape="64, 64, 3, 3" offset="707612" size="147456" />
1879
+ <output>
1880
+ <port id="0" precision="FP32" names="self.encoder.mid_block.resnets.1.conv1.weight">
1881
+ <dim>64</dim>
1882
+ <dim>64</dim>
1883
+ <dim>3</dim>
1884
+ <dim>3</dim>
1885
+ </port>
1886
+ </output>
1887
+ </layer>
1888
+ <layer id="116" name="__module.encoder.mid_block.resnets.1.conv1/aten::_convolution/Convolution" type="Convolution" version="opset1">
1889
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
1890
+ <input>
1891
+ <port id="0" precision="FP32">
1892
+ <dim>-1</dim>
1893
+ <dim>64</dim>
1894
+ <dim>-1</dim>
1895
+ <dim>-1</dim>
1896
+ </port>
1897
+ <port id="1" precision="FP32">
1898
+ <dim>64</dim>
1899
+ <dim>64</dim>
1900
+ <dim>3</dim>
1901
+ <dim>3</dim>
1902
+ </port>
1903
+ </input>
1904
+ <output>
1905
+ <port id="2" precision="FP32">
1906
+ <dim>-1</dim>
1907
+ <dim>64</dim>
1908
+ <dim>-1</dim>
1909
+ <dim>-1</dim>
1910
+ </port>
1911
+ </output>
1912
+ </layer>
1913
+ <layer id="117" name="__module.encoder.mid_block.resnets.1.conv1/aten::_convolution/Reshape" type="Const" version="opset1">
1914
+ <data element_type="f32" shape="1, 64, 1, 1" offset="855068" size="256" />
1915
+ <output>
1916
+ <port id="0" precision="FP32">
1917
+ <dim>1</dim>
1918
+ <dim>64</dim>
1919
+ <dim>1</dim>
1920
+ <dim>1</dim>
1921
+ </port>
1922
+ </output>
1923
+ </layer>
1924
+ <layer id="118" name="__module.encoder.mid_block.resnets.1.conv1/aten::_convolution/Add" type="Add" version="opset1">
1925
+ <data auto_broadcast="numpy" />
1926
+ <input>
1927
+ <port id="0" precision="FP32">
1928
+ <dim>-1</dim>
1929
+ <dim>64</dim>
1930
+ <dim>-1</dim>
1931
+ <dim>-1</dim>
1932
+ </port>
1933
+ <port id="1" precision="FP32">
1934
+ <dim>1</dim>
1935
+ <dim>64</dim>
1936
+ <dim>1</dim>
1937
+ <dim>1</dim>
1938
+ </port>
1939
+ </input>
1940
+ <output>
1941
+ <port id="2" precision="FP32" names="242,input.39">
1942
+ <dim>-1</dim>
1943
+ <dim>64</dim>
1944
+ <dim>-1</dim>
1945
+ <dim>-1</dim>
1946
+ </port>
1947
+ </output>
1948
+ </layer>
1949
+ <layer id="119" name="self.encoder.mid_block.resnets.1.norm2.weight" type="Const" version="opset1">
1950
+ <data element_type="f32" shape="64" offset="197248" size="256" />
1951
+ <output>
1952
+ <port id="0" precision="FP32" names="self.encoder.mid_block.resnets.1.norm2.weight">
1953
+ <dim>64</dim>
1954
+ </port>
1955
+ </output>
1956
+ </layer>
1957
+ <layer id="120" name="self.encoder.mid_block.resnets.1.norm2.bias" type="Const" version="opset1">
1958
+ <data element_type="f32" shape="64" offset="197504" size="256" />
1959
+ <output>
1960
+ <port id="0" precision="FP32" names="self.encoder.mid_block.resnets.1.norm2.bias">
1961
+ <dim>64</dim>
1962
+ </port>
1963
+ </output>
1964
+ </layer>
1965
+ <layer id="121" name="__module.encoder.mid_block.resnets.1.norm2/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
1966
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
1967
+ <input>
1968
+ <port id="0" precision="FP32">
1969
+ <dim>-1</dim>
1970
+ <dim>64</dim>
1971
+ <dim>-1</dim>
1972
+ <dim>-1</dim>
1973
+ </port>
1974
+ <port id="1" precision="FP32">
1975
+ <dim>64</dim>
1976
+ </port>
1977
+ <port id="2" precision="FP32">
1978
+ <dim>64</dim>
1979
+ </port>
1980
+ </input>
1981
+ <output>
1982
+ <port id="3" precision="FP32" names="245,input.41">
1983
+ <dim>-1</dim>
1984
+ <dim>64</dim>
1985
+ <dim>-1</dim>
1986
+ <dim>-1</dim>
1987
+ </port>
1988
+ </output>
1989
+ </layer>
1990
+ <layer id="122" name="__module.decoder.mid_block.resnets.1.nonlinearity/aten::silu/Swish_7" type="Swish" version="opset4">
1991
+ <input>
1992
+ <port id="0" precision="FP32">
1993
+ <dim>-1</dim>
1994
+ <dim>64</dim>
1995
+ <dim>-1</dim>
1996
+ <dim>-1</dim>
1997
+ </port>
1998
+ </input>
1999
+ <output>
2000
+ <port id="1" precision="FP32" names="246,input.43">
2001
+ <dim>-1</dim>
2002
+ <dim>64</dim>
2003
+ <dim>-1</dim>
2004
+ <dim>-1</dim>
2005
+ </port>
2006
+ </output>
2007
+ </layer>
2008
+ <layer id="123" name="self.encoder.mid_block.resnets.1.conv2.weight" type="Const" version="opset1">
2009
+ <data element_type="f32" shape="64, 64, 3, 3" offset="855324" size="147456" />
2010
+ <output>
2011
+ <port id="0" precision="FP32" names="self.encoder.mid_block.resnets.1.conv2.weight">
2012
+ <dim>64</dim>
2013
+ <dim>64</dim>
2014
+ <dim>3</dim>
2015
+ <dim>3</dim>
2016
+ </port>
2017
+ </output>
2018
+ </layer>
2019
+ <layer id="124" name="__module.encoder.mid_block.resnets.1.conv2/aten::_convolution/Convolution" type="Convolution" version="opset1">
2020
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
2021
+ <input>
2022
+ <port id="0" precision="FP32">
2023
+ <dim>-1</dim>
2024
+ <dim>64</dim>
2025
+ <dim>-1</dim>
2026
+ <dim>-1</dim>
2027
+ </port>
2028
+ <port id="1" precision="FP32">
2029
+ <dim>64</dim>
2030
+ <dim>64</dim>
2031
+ <dim>3</dim>
2032
+ <dim>3</dim>
2033
+ </port>
2034
+ </input>
2035
+ <output>
2036
+ <port id="2" precision="FP32">
2037
+ <dim>-1</dim>
2038
+ <dim>64</dim>
2039
+ <dim>-1</dim>
2040
+ <dim>-1</dim>
2041
+ </port>
2042
+ </output>
2043
+ </layer>
2044
+ <layer id="125" name="__module.encoder.mid_block.resnets.1.conv2/aten::_convolution/Reshape" type="Const" version="opset1">
2045
+ <data element_type="f32" shape="1, 64, 1, 1" offset="1002780" size="256" />
2046
+ <output>
2047
+ <port id="0" precision="FP32">
2048
+ <dim>1</dim>
2049
+ <dim>64</dim>
2050
+ <dim>1</dim>
2051
+ <dim>1</dim>
2052
+ </port>
2053
+ </output>
2054
+ </layer>
2055
+ <layer id="126" name="__module.encoder.mid_block.resnets.1.conv2/aten::_convolution/Add" type="Add" version="opset1">
2056
+ <data auto_broadcast="numpy" />
2057
+ <input>
2058
+ <port id="0" precision="FP32">
2059
+ <dim>-1</dim>
2060
+ <dim>64</dim>
2061
+ <dim>-1</dim>
2062
+ <dim>-1</dim>
2063
+ </port>
2064
+ <port id="1" precision="FP32">
2065
+ <dim>1</dim>
2066
+ <dim>64</dim>
2067
+ <dim>1</dim>
2068
+ <dim>1</dim>
2069
+ </port>
2070
+ </input>
2071
+ <output>
2072
+ <port id="2" precision="FP32" names="254,hidden_states">
2073
+ <dim>-1</dim>
2074
+ <dim>64</dim>
2075
+ <dim>-1</dim>
2076
+ <dim>-1</dim>
2077
+ </port>
2078
+ </output>
2079
+ </layer>
2080
+ <layer id="127" name="__module.encoder.mid_block.resnets.1/aten::add/Add" type="Add" version="opset1">
2081
+ <data auto_broadcast="numpy" />
2082
+ <input>
2083
+ <port id="0" precision="FP32">
2084
+ <dim>-1</dim>
2085
+ <dim>64</dim>
2086
+ <dim>-1</dim>
2087
+ <dim>-1</dim>
2088
+ </port>
2089
+ <port id="1" precision="FP32">
2090
+ <dim>-1</dim>
2091
+ <dim>64</dim>
2092
+ <dim>-1</dim>
2093
+ <dim>-1</dim>
2094
+ </port>
2095
+ </input>
2096
+ <output>
2097
+ <port id="2" precision="FP32" names="255,256,input.45">
2098
+ <dim>-1</dim>
2099
+ <dim>64</dim>
2100
+ <dim>-1</dim>
2101
+ <dim>-1</dim>
2102
+ </port>
2103
+ </output>
2104
+ </layer>
2105
+ <layer id="128" name="self.encoder.conv_norm_out.weight" type="Const" version="opset1">
2106
+ <data element_type="f32" shape="64" offset="197248" size="256" />
2107
+ <output>
2108
+ <port id="0" precision="FP32" names="self.encoder.conv_norm_out.weight">
2109
+ <dim>64</dim>
2110
+ </port>
2111
+ </output>
2112
+ </layer>
2113
+ <layer id="129" name="self.encoder.conv_norm_out.bias" type="Const" version="opset1">
2114
+ <data element_type="f32" shape="64" offset="197504" size="256" />
2115
+ <output>
2116
+ <port id="0" precision="FP32" names="self.encoder.conv_norm_out.bias">
2117
+ <dim>64</dim>
2118
+ </port>
2119
+ </output>
2120
+ </layer>
2121
+ <layer id="130" name="__module.encoder.conv_norm_out/aten::group_norm/GroupNormalization" type="GroupNormalization" version="opset12">
2122
+ <data num_groups="32" epsilon="9.9999999747524271e-07" />
2123
+ <input>
2124
+ <port id="0" precision="FP32">
2125
+ <dim>-1</dim>
2126
+ <dim>64</dim>
2127
+ <dim>-1</dim>
2128
+ <dim>-1</dim>
2129
+ </port>
2130
+ <port id="1" precision="FP32">
2131
+ <dim>64</dim>
2132
+ </port>
2133
+ <port id="2" precision="FP32">
2134
+ <dim>64</dim>
2135
+ </port>
2136
+ </input>
2137
+ <output>
2138
+ <port id="3" precision="FP32" names="259,input">
2139
+ <dim>-1</dim>
2140
+ <dim>64</dim>
2141
+ <dim>-1</dim>
2142
+ <dim>-1</dim>
2143
+ </port>
2144
+ </output>
2145
+ </layer>
2146
+ <layer id="131" name="__module.encoder.conv_act/aten::silu/Swish" type="Swish" version="opset4">
2147
+ <input>
2148
+ <port id="0" precision="FP32">
2149
+ <dim>-1</dim>
2150
+ <dim>64</dim>
2151
+ <dim>-1</dim>
2152
+ <dim>-1</dim>
2153
+ </port>
2154
+ </input>
2155
+ <output>
2156
+ <port id="1" precision="FP32" names="260">
2157
+ <dim>-1</dim>
2158
+ <dim>64</dim>
2159
+ <dim>-1</dim>
2160
+ <dim>-1</dim>
2161
+ </port>
2162
+ </output>
2163
+ </layer>
2164
+ <layer id="132" name="self.encoder.conv_out.weight" type="Const" version="opset1">
2165
+ <data element_type="f32" shape="8, 64, 3, 3" offset="1003036" size="18432" />
2166
+ <output>
2167
+ <port id="0" precision="FP32" names="self.encoder.conv_out.weight">
2168
+ <dim>8</dim>
2169
+ <dim>64</dim>
2170
+ <dim>3</dim>
2171
+ <dim>3</dim>
2172
+ </port>
2173
+ </output>
2174
+ </layer>
2175
+ <layer id="133" name="__module.encoder.conv_out/aten::_convolution/Convolution" type="Convolution" version="opset1">
2176
+ <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
2177
+ <input>
2178
+ <port id="0" precision="FP32">
2179
+ <dim>-1</dim>
2180
+ <dim>64</dim>
2181
+ <dim>-1</dim>
2182
+ <dim>-1</dim>
2183
+ </port>
2184
+ <port id="1" precision="FP32">
2185
+ <dim>8</dim>
2186
+ <dim>64</dim>
2187
+ <dim>3</dim>
2188
+ <dim>3</dim>
2189
+ </port>
2190
+ </input>
2191
+ <output>
2192
+ <port id="2" precision="FP32">
2193
+ <dim>-1</dim>
2194
+ <dim>8</dim>
2195
+ <dim>-1</dim>
2196
+ <dim>-1</dim>
2197
+ </port>
2198
+ </output>
2199
+ </layer>
2200
+ <layer id="134" name="__module.encoder.conv_out/aten::_convolution/Reshape" type="Const" version="opset1">
2201
+ <data element_type="f32" shape="1, 8, 1, 1" offset="1021468" size="32" />
2202
+ <output>
2203
+ <port id="0" precision="FP32">
2204
+ <dim>1</dim>
2205
+ <dim>8</dim>
2206
+ <dim>1</dim>
2207
+ <dim>1</dim>
2208
+ </port>
2209
+ </output>
2210
+ </layer>
2211
+ <layer id="135" name="__module.encoder.conv_out/aten::_convolution/Add" type="Add" version="opset1">
2212
+ <data auto_broadcast="numpy" />
2213
+ <input>
2214
+ <port id="0" precision="FP32">
2215
+ <dim>-1</dim>
2216
+ <dim>8</dim>
2217
+ <dim>-1</dim>
2218
+ <dim>-1</dim>
2219
+ </port>
2220
+ <port id="1" precision="FP32">
2221
+ <dim>1</dim>
2222
+ <dim>8</dim>
2223
+ <dim>1</dim>
2224
+ <dim>1</dim>
2225
+ </port>
2226
+ </input>
2227
+ <output>
2228
+ <port id="2" precision="FP32" names="267">
2229
+ <dim>-1</dim>
2230
+ <dim>8</dim>
2231
+ <dim>-1</dim>
2232
+ <dim>-1</dim>
2233
+ </port>
2234
+ </output>
2235
+ </layer>
2236
+ <layer id="136" name="self.quant_conv.weight" type="Const" version="opset1">
2237
+ <data element_type="f32" shape="8, 8, 1, 1" offset="1021500" size="256" />
2238
+ <output>
2239
+ <port id="0" precision="FP32" names="self.quant_conv.weight">
2240
+ <dim>8</dim>
2241
+ <dim>8</dim>
2242
+ <dim>1</dim>
2243
+ <dim>1</dim>
2244
+ </port>
2245
+ </output>
2246
+ </layer>
2247
+ <layer id="137" name="__module.quant_conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
2248
+ <data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
2249
+ <input>
2250
+ <port id="0" precision="FP32">
2251
+ <dim>-1</dim>
2252
+ <dim>8</dim>
2253
+ <dim>-1</dim>
2254
+ <dim>-1</dim>
2255
+ </port>
2256
+ <port id="1" precision="FP32">
2257
+ <dim>8</dim>
2258
+ <dim>8</dim>
2259
+ <dim>1</dim>
2260
+ <dim>1</dim>
2261
+ </port>
2262
+ </input>
2263
+ <output>
2264
+ <port id="2" precision="FP32">
2265
+ <dim>-1</dim>
2266
+ <dim>8</dim>
2267
+ <dim>-1</dim>
2268
+ <dim>-1</dim>
2269
+ </port>
2270
+ </output>
2271
+ </layer>
2272
+ <layer id="138" name="__module.quant_conv/aten::_convolution/Reshape" type="Const" version="opset1">
2273
+ <data element_type="f32" shape="1, 8, 1, 1" offset="1021756" size="32" />
2274
+ <output>
2275
+ <port id="0" precision="FP32">
2276
+ <dim>1</dim>
2277
+ <dim>8</dim>
2278
+ <dim>1</dim>
2279
+ <dim>1</dim>
2280
+ </port>
2281
+ </output>
2282
+ </layer>
2283
+ <layer id="139" name="__module.quant_conv/aten::_convolution/Add" type="Add" version="opset1">
2284
+ <data auto_broadcast="numpy" />
2285
+ <input>
2286
+ <port id="0" precision="FP32">
2287
+ <dim>-1</dim>
2288
+ <dim>8</dim>
2289
+ <dim>-1</dim>
2290
+ <dim>-1</dim>
2291
+ </port>
2292
+ <port id="1" precision="FP32">
2293
+ <dim>1</dim>
2294
+ <dim>8</dim>
2295
+ <dim>1</dim>
2296
+ <dim>1</dim>
2297
+ </port>
2298
+ </input>
2299
+ <output>
2300
+ <port id="2" precision="FP32" names="latent_parameters">
2301
+ <dim>-1</dim>
2302
+ <dim>8</dim>
2303
+ <dim>-1</dim>
2304
+ <dim>-1</dim>
2305
+ </port>
2306
+ </output>
2307
+ </layer>
2308
+ <layer id="140" name="Result_45354" type="Result" version="opset1">
2309
+ <input>
2310
+ <port id="0" precision="FP32">
2311
+ <dim>-1</dim>
2312
+ <dim>8</dim>
2313
+ <dim>-1</dim>
2314
+ <dim>-1</dim>
2315
+ </port>
2316
+ </input>
2317
+ </layer>
2318
+ </layers>
2319
+ <edges>
2320
+ <edge from-layer="0" from-port="0" to-layer="2" to-port="0" />
2321
+ <edge from-layer="1" from-port="0" to-layer="2" to-port="1" />
2322
+ <edge from-layer="2" from-port="2" to-layer="4" to-port="0" />
2323
+ <edge from-layer="3" from-port="0" to-layer="4" to-port="1" />
2324
+ <edge from-layer="4" from-port="2" to-layer="21" to-port="0" />
2325
+ <edge from-layer="4" from-port="2" to-layer="7" to-port="0" />
2326
+ <edge from-layer="5" from-port="0" to-layer="7" to-port="1" />
2327
+ <edge from-layer="6" from-port="0" to-layer="7" to-port="2" />
2328
+ <edge from-layer="7" from-port="3" to-layer="8" to-port="0" />
2329
+ <edge from-layer="8" from-port="1" to-layer="10" to-port="0" />
2330
+ <edge from-layer="9" from-port="0" to-layer="10" to-port="1" />
2331
+ <edge from-layer="10" from-port="2" to-layer="12" to-port="0" />
2332
+ <edge from-layer="11" from-port="0" to-layer="12" to-port="1" />
2333
+ <edge from-layer="12" from-port="2" to-layer="15" to-port="0" />
2334
+ <edge from-layer="13" from-port="0" to-layer="15" to-port="1" />
2335
+ <edge from-layer="14" from-port="0" to-layer="15" to-port="2" />
2336
+ <edge from-layer="15" from-port="3" to-layer="16" to-port="0" />
2337
+ <edge from-layer="16" from-port="1" to-layer="18" to-port="0" />
2338
+ <edge from-layer="17" from-port="0" to-layer="18" to-port="1" />
2339
+ <edge from-layer="18" from-port="2" to-layer="20" to-port="0" />
2340
+ <edge from-layer="19" from-port="0" to-layer="20" to-port="1" />
2341
+ <edge from-layer="20" from-port="2" to-layer="21" to-port="1" />
2342
+ <edge from-layer="21" from-port="2" to-layer="23" to-port="0" />
2343
+ <edge from-layer="22" from-port="0" to-layer="23" to-port="1" />
2344
+ <edge from-layer="23" from-port="2" to-layer="25" to-port="0" />
2345
+ <edge from-layer="24" from-port="0" to-layer="25" to-port="1" />
2346
+ <edge from-layer="25" from-port="2" to-layer="27" to-port="0" />
2347
+ <edge from-layer="25" from-port="2" to-layer="32" to-port="0" />
2348
+ <edge from-layer="26" from-port="0" to-layer="27" to-port="1" />
2349
+ <edge from-layer="27" from-port="2" to-layer="29" to-port="0" />
2350
+ <edge from-layer="28" from-port="0" to-layer="29" to-port="1" />
2351
+ <edge from-layer="29" from-port="2" to-layer="46" to-port="0" />
2352
+ <edge from-layer="30" from-port="0" to-layer="32" to-port="1" />
2353
+ <edge from-layer="31" from-port="0" to-layer="32" to-port="2" />
2354
+ <edge from-layer="32" from-port="3" to-layer="33" to-port="0" />
2355
+ <edge from-layer="33" from-port="1" to-layer="35" to-port="0" />
2356
+ <edge from-layer="34" from-port="0" to-layer="35" to-port="1" />
2357
+ <edge from-layer="35" from-port="2" to-layer="37" to-port="0" />
2358
+ <edge from-layer="36" from-port="0" to-layer="37" to-port="1" />
2359
+ <edge from-layer="37" from-port="2" to-layer="40" to-port="0" />
2360
+ <edge from-layer="38" from-port="0" to-layer="40" to-port="1" />
2361
+ <edge from-layer="39" from-port="0" to-layer="40" to-port="2" />
2362
+ <edge from-layer="40" from-port="3" to-layer="41" to-port="0" />
2363
+ <edge from-layer="41" from-port="1" to-layer="43" to-port="0" />
2364
+ <edge from-layer="42" from-port="0" to-layer="43" to-port="1" />
2365
+ <edge from-layer="43" from-port="2" to-layer="45" to-port="0" />
2366
+ <edge from-layer="44" from-port="0" to-layer="45" to-port="1" />
2367
+ <edge from-layer="45" from-port="2" to-layer="46" to-port="1" />
2368
+ <edge from-layer="46" from-port="2" to-layer="63" to-port="0" />
2369
+ <edge from-layer="46" from-port="2" to-layer="49" to-port="0" />
2370
+ <edge from-layer="47" from-port="0" to-layer="49" to-port="1" />
2371
+ <edge from-layer="48" from-port="0" to-layer="49" to-port="2" />
2372
+ <edge from-layer="49" from-port="3" to-layer="50" to-port="0" />
2373
+ <edge from-layer="50" from-port="1" to-layer="52" to-port="0" />
2374
+ <edge from-layer="51" from-port="0" to-layer="52" to-port="1" />
2375
+ <edge from-layer="52" from-port="2" to-layer="54" to-port="0" />
2376
+ <edge from-layer="53" from-port="0" to-layer="54" to-port="1" />
2377
+ <edge from-layer="54" from-port="2" to-layer="57" to-port="0" />
2378
+ <edge from-layer="55" from-port="0" to-layer="57" to-port="1" />
2379
+ <edge from-layer="56" from-port="0" to-layer="57" to-port="2" />
2380
+ <edge from-layer="57" from-port="3" to-layer="58" to-port="0" />
2381
+ <edge from-layer="58" from-port="1" to-layer="60" to-port="0" />
2382
+ <edge from-layer="59" from-port="0" to-layer="60" to-port="1" />
2383
+ <edge from-layer="60" from-port="2" to-layer="62" to-port="0" />
2384
+ <edge from-layer="61" from-port="0" to-layer="62" to-port="1" />
2385
+ <edge from-layer="62" from-port="2" to-layer="63" to-port="1" />
2386
+ <edge from-layer="63" from-port="2" to-layer="108" to-port="0" />
2387
+ <edge from-layer="63" from-port="2" to-layer="110" to-port="1" />
2388
+ <edge from-layer="63" from-port="2" to-layer="65" to-port="0" />
2389
+ <edge from-layer="64" from-port="0" to-layer="65" to-port="1" />
2390
+ <edge from-layer="65" from-port="2" to-layer="67" to-port="0" />
2391
+ <edge from-layer="66" from-port="0" to-layer="67" to-port="1" />
2392
+ <edge from-layer="67" from-port="2" to-layer="69" to-port="0" />
2393
+ <edge from-layer="68" from-port="0" to-layer="69" to-port="1" />
2394
+ <edge from-layer="69" from-port="2" to-layer="72" to-port="0" />
2395
+ <edge from-layer="70" from-port="0" to-layer="72" to-port="1" />
2396
+ <edge from-layer="71" from-port="0" to-layer="72" to-port="2" />
2397
+ <edge from-layer="72" from-port="3" to-layer="74" to-port="0" />
2398
+ <edge from-layer="72" from-port="3" to-layer="90" to-port="0" />
2399
+ <edge from-layer="72" from-port="3" to-layer="82" to-port="0" />
2400
+ <edge from-layer="73" from-port="0" to-layer="74" to-port="1" />
2401
+ <edge from-layer="74" from-port="2" to-layer="76" to-port="0" />
2402
+ <edge from-layer="75" from-port="0" to-layer="76" to-port="1" />
2403
+ <edge from-layer="76" from-port="2" to-layer="78" to-port="0" />
2404
+ <edge from-layer="77" from-port="0" to-layer="78" to-port="1" />
2405
+ <edge from-layer="78" from-port="2" to-layer="80" to-port="0" />
2406
+ <edge from-layer="79" from-port="0" to-layer="80" to-port="1" />
2407
+ <edge from-layer="80" from-port="2" to-layer="97" to-port="0" />
2408
+ <edge from-layer="81" from-port="0" to-layer="82" to-port="1" />
2409
+ <edge from-layer="82" from-port="2" to-layer="84" to-port="0" />
2410
+ <edge from-layer="83" from-port="0" to-layer="84" to-port="1" />
2411
+ <edge from-layer="84" from-port="2" to-layer="86" to-port="0" />
2412
+ <edge from-layer="85" from-port="0" to-layer="86" to-port="1" />
2413
+ <edge from-layer="86" from-port="2" to-layer="88" to-port="0" />
2414
+ <edge from-layer="87" from-port="0" to-layer="88" to-port="1" />
2415
+ <edge from-layer="88" from-port="2" to-layer="97" to-port="1" />
2416
+ <edge from-layer="89" from-port="0" to-layer="90" to-port="1" />
2417
+ <edge from-layer="90" from-port="2" to-layer="92" to-port="0" />
2418
+ <edge from-layer="91" from-port="0" to-layer="92" to-port="1" />
2419
+ <edge from-layer="92" from-port="2" to-layer="94" to-port="0" />
2420
+ <edge from-layer="93" from-port="0" to-layer="94" to-port="1" />
2421
+ <edge from-layer="94" from-port="2" to-layer="96" to-port="0" />
2422
+ <edge from-layer="95" from-port="0" to-layer="96" to-port="1" />
2423
+ <edge from-layer="96" from-port="2" to-layer="97" to-port="2" />
2424
+ <edge from-layer="97" from-port="3" to-layer="99" to-port="0" />
2425
+ <edge from-layer="98" from-port="0" to-layer="99" to-port="1" />
2426
+ <edge from-layer="99" from-port="2" to-layer="101" to-port="0" />
2427
+ <edge from-layer="100" from-port="0" to-layer="101" to-port="1" />
2428
+ <edge from-layer="101" from-port="2" to-layer="103" to-port="0" />
2429
+ <edge from-layer="102" from-port="0" to-layer="103" to-port="1" />
2430
+ <edge from-layer="103" from-port="2" to-layer="105" to-port="0" />
2431
+ <edge from-layer="104" from-port="0" to-layer="105" to-port="1" />
2432
+ <edge from-layer="105" from-port="2" to-layer="107" to-port="0" />
2433
+ <edge from-layer="106" from-port="0" to-layer="107" to-port="1" />
2434
+ <edge from-layer="107" from-port="2" to-layer="109" to-port="0" />
2435
+ <edge from-layer="108" from-port="1" to-layer="109" to-port="1" />
2436
+ <edge from-layer="109" from-port="2" to-layer="110" to-port="0" />
2437
+ <edge from-layer="110" from-port="2" to-layer="127" to-port="0" />
2438
+ <edge from-layer="110" from-port="2" to-layer="113" to-port="0" />
2439
+ <edge from-layer="111" from-port="0" to-layer="113" to-port="1" />
2440
+ <edge from-layer="112" from-port="0" to-layer="113" to-port="2" />
2441
+ <edge from-layer="113" from-port="3" to-layer="114" to-port="0" />
2442
+ <edge from-layer="114" from-port="1" to-layer="116" to-port="0" />
2443
+ <edge from-layer="115" from-port="0" to-layer="116" to-port="1" />
2444
+ <edge from-layer="116" from-port="2" to-layer="118" to-port="0" />
2445
+ <edge from-layer="117" from-port="0" to-layer="118" to-port="1" />
2446
+ <edge from-layer="118" from-port="2" to-layer="121" to-port="0" />
2447
+ <edge from-layer="119" from-port="0" to-layer="121" to-port="1" />
2448
+ <edge from-layer="120" from-port="0" to-layer="121" to-port="2" />
2449
+ <edge from-layer="121" from-port="3" to-layer="122" to-port="0" />
2450
+ <edge from-layer="122" from-port="1" to-layer="124" to-port="0" />
2451
+ <edge from-layer="123" from-port="0" to-layer="124" to-port="1" />
2452
+ <edge from-layer="124" from-port="2" to-layer="126" to-port="0" />
2453
+ <edge from-layer="125" from-port="0" to-layer="126" to-port="1" />
2454
+ <edge from-layer="126" from-port="2" to-layer="127" to-port="1" />
2455
+ <edge from-layer="127" from-port="2" to-layer="130" to-port="0" />
2456
+ <edge from-layer="128" from-port="0" to-layer="130" to-port="1" />
2457
+ <edge from-layer="129" from-port="0" to-layer="130" to-port="2" />
2458
+ <edge from-layer="130" from-port="3" to-layer="131" to-port="0" />
2459
+ <edge from-layer="131" from-port="1" to-layer="133" to-port="0" />
2460
+ <edge from-layer="132" from-port="0" to-layer="133" to-port="1" />
2461
+ <edge from-layer="133" from-port="2" to-layer="135" to-port="0" />
2462
+ <edge from-layer="134" from-port="0" to-layer="135" to-port="1" />
2463
+ <edge from-layer="135" from-port="2" to-layer="137" to-port="0" />
2464
+ <edge from-layer="136" from-port="0" to-layer="137" to-port="1" />
2465
+ <edge from-layer="137" from-port="2" to-layer="139" to-port="0" />
2466
+ <edge from-layer="138" from-port="0" to-layer="139" to-port="1" />
2467
+ <edge from-layer="139" from-port="2" to-layer="140" to-port="0" />
2468
+ </edges>
2469
+ <rt_info>
2470
+ <Runtime_version value="2025.0.0-17942-1f68be9f594-releases/2025/0" />
2471
+ <conversion_parameters>
2472
+ <framework value="pytorch" />
2473
+ <is_python_object value="True" />
2474
+ </conversion_parameters>
2475
+ <optimum>
2476
+ <diffusers_version value="0.32.2" />
2477
+ <optimum_intel_version value="1.23.0.dev0+5ac35448" />
2478
+ <optimum_version value="1.24.0" />
2479
+ <pytorch_version value="2.5.0+cpu" />
2480
+ <transformers_version value="4.49.0" />
2481
+ </optimum>
2482
+ </rt_info>
2483
+ </net>