arisha07 commited on
Commit
771f3ec
·
1 Parent(s): 33ffef9

Upload 5 files

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ INT8/unet_int8_NPU.blob filter=lfs diff=lfs merge=lfs -text
INT8/unet_int8.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ab6b2f6fbdf5590e7653b02d661e17ffd33ffde28ec613219df1f6fd73d2d99
3
+ size 862858284
INT8/unet_int8.xml ADDED
The diff for this file is too large to render. See raw diff
 
INT8/unet_int8_NPU.blob ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c85295896802904d197bd040ddbacaa08ec9110744f2ffb92dc3ece33c322307
3
+ size 924927432
INT8/unet_time_proj.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53f57a66e6b1892ccf586610d5b23d5a9b2477914b051066304de335860653a3
3
+ size 425680
INT8/unet_time_proj.xml ADDED
@@ -0,0 +1,563 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <?xml version="1.0"?>
2
+ <net name="Model2" version="11">
3
+ <layers>
4
+ <layer id="0" name="t" type="Parameter" version="opset1">
5
+ <data shape="1" element_type="f32" />
6
+ <output>
7
+ <port id="0" precision="FP32" names="t">
8
+ <dim>1</dim>
9
+ </port>
10
+ </output>
11
+ </layer>
12
+ <layer id="1" name="/Where3704" type="Const" version="opset1">
13
+ <data element_type="i64" shape="1" offset="0" size="8" />
14
+ <output>
15
+ <port id="0" precision="I64" names="/Where_output_0">
16
+ <dim>1</dim>
17
+ </port>
18
+ </output>
19
+ </layer>
20
+ <layer id="2" name="/Expand" type="Broadcast" version="opset3">
21
+ <data mode="bidirectional" />
22
+ <input>
23
+ <port id="0" precision="FP32">
24
+ <dim>1</dim>
25
+ </port>
26
+ <port id="1" precision="I64">
27
+ <dim>1</dim>
28
+ </port>
29
+ </input>
30
+ <output>
31
+ <port id="2" precision="FP32" names="/Expand_output_0">
32
+ <dim>1</dim>
33
+ </port>
34
+ </output>
35
+ </layer>
36
+ <layer id="3" name="/time_proj/Constant370674469" type="Const" version="opset1">
37
+ <data element_type="i64" shape="1" offset="0" size="8" />
38
+ <output>
39
+ <port id="0" precision="I64" names="/time_proj/Constant_output_0">
40
+ <dim>1</dim>
41
+ </port>
42
+ </output>
43
+ </layer>
44
+ <layer id="4" name="/time_proj/Unsqueeze" type="Unsqueeze" version="opset1">
45
+ <input>
46
+ <port id="0" precision="FP32">
47
+ <dim>1</dim>
48
+ </port>
49
+ <port id="1" precision="I64">
50
+ <dim>1</dim>
51
+ </port>
52
+ </input>
53
+ <output>
54
+ <port id="2" precision="FP32" names="/time_proj/Cast_output_0,/time_proj/Unsqueeze_output_0">
55
+ <dim>1</dim>
56
+ <dim>1</dim>
57
+ </port>
58
+ </output>
59
+ </layer>
60
+ <layer id="5" name="/time_proj/Constant_1370876734" type="Const" version="opset1">
61
+ <data element_type="f32" shape="1, 160" offset="8" size="640" />
62
+ <output>
63
+ <port id="0" precision="FP32" names="/time_proj/Constant_1_output_0">
64
+ <dim>1</dim>
65
+ <dim>160</dim>
66
+ </port>
67
+ </output>
68
+ </layer>
69
+ <layer id="6" name="/time_proj/Mul" type="Multiply" version="opset1">
70
+ <data auto_broadcast="numpy" />
71
+ <input>
72
+ <port id="0" precision="FP32">
73
+ <dim>1</dim>
74
+ <dim>1</dim>
75
+ </port>
76
+ <port id="1" precision="FP32">
77
+ <dim>1</dim>
78
+ <dim>160</dim>
79
+ </port>
80
+ </input>
81
+ <output>
82
+ <port id="2" precision="FP32" names="/time_proj/Mul_output_0">
83
+ <dim>1</dim>
84
+ <dim>160</dim>
85
+ </port>
86
+ </output>
87
+ </layer>
88
+ <layer id="7" name="/time_proj/Sin" type="Sin" version="opset1">
89
+ <input>
90
+ <port id="0" precision="FP32">
91
+ <dim>1</dim>
92
+ <dim>160</dim>
93
+ </port>
94
+ </input>
95
+ <output>
96
+ <port id="1" precision="FP32" names="/time_proj/Sin_output_0">
97
+ <dim>1</dim>
98
+ <dim>160</dim>
99
+ </port>
100
+ </output>
101
+ </layer>
102
+ <layer id="8" name="254012540576818" type="Const" version="opset1">
103
+ <data element_type="f32" shape="1, 1" offset="648" size="4" />
104
+ <output>
105
+ <port id="0" precision="FP32">
106
+ <dim>1</dim>
107
+ <dim>1</dim>
108
+ </port>
109
+ </output>
110
+ </layer>
111
+ <layer id="9" name="254022540673599" type="Const" version="opset1">
112
+ <data element_type="f32" shape="1, 1" offset="652" size="4" />
113
+ <output>
114
+ <port id="0" precision="FP32">
115
+ <dim>1</dim>
116
+ <dim>1</dim>
117
+ </port>
118
+ </output>
119
+ </layer>
120
+ <layer id="10" name="254032540780985" type="Const" version="opset1">
121
+ <data element_type="f32" shape="1, 1" offset="648" size="4" />
122
+ <output>
123
+ <port id="0" precision="FP32">
124
+ <dim>1</dim>
125
+ <dim>1</dim>
126
+ </port>
127
+ </output>
128
+ </layer>
129
+ <layer id="11" name="254042540872462" type="Const" version="opset1">
130
+ <data element_type="f32" shape="1, 1" offset="652" size="4" />
131
+ <output>
132
+ <port id="0" precision="FP32">
133
+ <dim>1</dim>
134
+ <dim>1</dim>
135
+ </port>
136
+ </output>
137
+ </layer>
138
+ <layer id="12" name="/time_proj/Concat/fq_input_0" type="FakeQuantize" version="opset1">
139
+ <data levels="256" auto_broadcast="numpy" />
140
+ <input>
141
+ <port id="0" precision="FP32">
142
+ <dim>1</dim>
143
+ <dim>160</dim>
144
+ </port>
145
+ <port id="1" precision="FP32">
146
+ <dim>1</dim>
147
+ <dim>1</dim>
148
+ </port>
149
+ <port id="2" precision="FP32">
150
+ <dim>1</dim>
151
+ <dim>1</dim>
152
+ </port>
153
+ <port id="3" precision="FP32">
154
+ <dim>1</dim>
155
+ <dim>1</dim>
156
+ </port>
157
+ <port id="4" precision="FP32">
158
+ <dim>1</dim>
159
+ <dim>1</dim>
160
+ </port>
161
+ </input>
162
+ <output>
163
+ <port id="5" precision="FP32">
164
+ <dim>1</dim>
165
+ <dim>160</dim>
166
+ </port>
167
+ </output>
168
+ </layer>
169
+ <layer id="13" name="/time_proj/Cos" type="Cos" version="opset1">
170
+ <input>
171
+ <port id="0" precision="FP32">
172
+ <dim>1</dim>
173
+ <dim>160</dim>
174
+ </port>
175
+ </input>
176
+ <output>
177
+ <port id="1" precision="FP32" names="/time_proj/Cos_output_0">
178
+ <dim>1</dim>
179
+ <dim>160</dim>
180
+ </port>
181
+ </output>
182
+ </layer>
183
+ <layer id="14" name="254112541573602" type="Const" version="opset1">
184
+ <data element_type="f32" shape="1, 1" offset="648" size="4" />
185
+ <output>
186
+ <port id="0" precision="FP32">
187
+ <dim>1</dim>
188
+ <dim>1</dim>
189
+ </port>
190
+ </output>
191
+ </layer>
192
+ <layer id="15" name="254122541672060" type="Const" version="opset1">
193
+ <data element_type="f32" shape="1, 1" offset="652" size="4" />
194
+ <output>
195
+ <port id="0" precision="FP32">
196
+ <dim>1</dim>
197
+ <dim>1</dim>
198
+ </port>
199
+ </output>
200
+ </layer>
201
+ <layer id="16" name="254132541778846" type="Const" version="opset1">
202
+ <data element_type="f32" shape="1, 1" offset="648" size="4" />
203
+ <output>
204
+ <port id="0" precision="FP32">
205
+ <dim>1</dim>
206
+ <dim>1</dim>
207
+ </port>
208
+ </output>
209
+ </layer>
210
+ <layer id="17" name="254142541880802" type="Const" version="opset1">
211
+ <data element_type="f32" shape="1, 1" offset="652" size="4" />
212
+ <output>
213
+ <port id="0" precision="FP32">
214
+ <dim>1</dim>
215
+ <dim>1</dim>
216
+ </port>
217
+ </output>
218
+ </layer>
219
+ <layer id="18" name="/time_proj/Concat/fq_input_1" type="FakeQuantize" version="opset1">
220
+ <data levels="256" auto_broadcast="numpy" />
221
+ <input>
222
+ <port id="0" precision="FP32">
223
+ <dim>1</dim>
224
+ <dim>160</dim>
225
+ </port>
226
+ <port id="1" precision="FP32">
227
+ <dim>1</dim>
228
+ <dim>1</dim>
229
+ </port>
230
+ <port id="2" precision="FP32">
231
+ <dim>1</dim>
232
+ <dim>1</dim>
233
+ </port>
234
+ <port id="3" precision="FP32">
235
+ <dim>1</dim>
236
+ <dim>1</dim>
237
+ </port>
238
+ <port id="4" precision="FP32">
239
+ <dim>1</dim>
240
+ <dim>1</dim>
241
+ </port>
242
+ </input>
243
+ <output>
244
+ <port id="5" precision="FP32">
245
+ <dim>1</dim>
246
+ <dim>160</dim>
247
+ </port>
248
+ </output>
249
+ </layer>
250
+ <layer id="19" name="/time_proj/Concat" type="Concat" version="opset1">
251
+ <data axis="1" />
252
+ <input>
253
+ <port id="0" precision="FP32">
254
+ <dim>1</dim>
255
+ <dim>160</dim>
256
+ </port>
257
+ <port id="1" precision="FP32">
258
+ <dim>1</dim>
259
+ <dim>160</dim>
260
+ </port>
261
+ </input>
262
+ <output>
263
+ <port id="2" precision="FP32" names="/time_proj/Concat_output_0">
264
+ <dim>1</dim>
265
+ <dim>320</dim>
266
+ </port>
267
+ </output>
268
+ </layer>
269
+ <layer id="20" name="Constant_640423713" type="Const" version="opset1">
270
+ <data element_type="i64" shape="2" offset="656" size="16" />
271
+ <output>
272
+ <port id="0" precision="I64">
273
+ <dim>2</dim>
274
+ </port>
275
+ </output>
276
+ </layer>
277
+ <layer id="21" name="Constant_640453714" type="Const" version="opset1">
278
+ <data element_type="i64" shape="2" offset="672" size="16" />
279
+ <output>
280
+ <port id="0" precision="I64">
281
+ <dim>2</dim>
282
+ </port>
283
+ </output>
284
+ </layer>
285
+ <layer id="22" name="Constant_640483715" type="Const" version="opset1">
286
+ <data element_type="i64" shape="2" offset="688" size="16" />
287
+ <output>
288
+ <port id="0" precision="I64">
289
+ <dim>2</dim>
290
+ </port>
291
+ </output>
292
+ </layer>
293
+ <layer id="23" name="/time_proj/Slice" type="StridedSlice" version="opset1">
294
+ <data begin_mask="1, 0" end_mask="1, 0" new_axis_mask="" shrink_axis_mask="" ellipsis_mask="" />
295
+ <input>
296
+ <port id="0" precision="FP32">
297
+ <dim>1</dim>
298
+ <dim>320</dim>
299
+ </port>
300
+ <port id="1" precision="I64">
301
+ <dim>2</dim>
302
+ </port>
303
+ <port id="2" precision="I64">
304
+ <dim>2</dim>
305
+ </port>
306
+ <port id="3" precision="I64">
307
+ <dim>2</dim>
308
+ </port>
309
+ </input>
310
+ <output>
311
+ <port id="4" precision="FP32" names="/time_proj/Slice_output_0">
312
+ <dim>1</dim>
313
+ <dim>160</dim>
314
+ </port>
315
+ </output>
316
+ </layer>
317
+ <layer id="24" name="Constant_640543717" type="Const" version="opset1">
318
+ <data element_type="i64" shape="2" offset="704" size="16" />
319
+ <output>
320
+ <port id="0" precision="I64">
321
+ <dim>2</dim>
322
+ </port>
323
+ </output>
324
+ </layer>
325
+ <layer id="25" name="Constant_640573718" type="Const" version="opset1">
326
+ <data element_type="i64" shape="2" offset="656" size="16" />
327
+ <output>
328
+ <port id="0" precision="I64">
329
+ <dim>2</dim>
330
+ </port>
331
+ </output>
332
+ </layer>
333
+ <layer id="26" name="Constant_640603719" type="Const" version="opset1">
334
+ <data element_type="i64" shape="2" offset="688" size="16" />
335
+ <output>
336
+ <port id="0" precision="I64">
337
+ <dim>2</dim>
338
+ </port>
339
+ </output>
340
+ </layer>
341
+ <layer id="27" name="/time_proj/Slice_1" type="StridedSlice" version="opset1">
342
+ <data begin_mask="1, 0" end_mask="1, 0" new_axis_mask="" shrink_axis_mask="" ellipsis_mask="" />
343
+ <input>
344
+ <port id="0" precision="FP32">
345
+ <dim>1</dim>
346
+ <dim>320</dim>
347
+ </port>
348
+ <port id="1" precision="I64">
349
+ <dim>2</dim>
350
+ </port>
351
+ <port id="2" precision="I64">
352
+ <dim>2</dim>
353
+ </port>
354
+ <port id="3" precision="I64">
355
+ <dim>2</dim>
356
+ </port>
357
+ </input>
358
+ <output>
359
+ <port id="4" precision="FP32" names="/time_proj/Slice_1_output_0">
360
+ <dim>1</dim>
361
+ <dim>160</dim>
362
+ </port>
363
+ </output>
364
+ </layer>
365
+ <layer id="28" name="/time_proj/Concat_1" type="Concat" version="opset1">
366
+ <data axis="1" />
367
+ <input>
368
+ <port id="0" precision="FP32">
369
+ <dim>1</dim>
370
+ <dim>160</dim>
371
+ </port>
372
+ <port id="1" precision="FP32">
373
+ <dim>1</dim>
374
+ <dim>160</dim>
375
+ </port>
376
+ </input>
377
+ <output>
378
+ <port id="2" precision="FP32" names="/Cast_output_0,/time_proj/Concat_1_output_0">
379
+ <dim>1</dim>
380
+ <dim>320</dim>
381
+ </port>
382
+ </output>
383
+ </layer>
384
+ <layer id="29" name="time_embedding.linear_1.weight372242271/quantized4893876872" type="Const" version="opset1">
385
+ <data element_type="i8" shape="1280, 320" offset="720" size="409600" />
386
+ <output>
387
+ <port id="0" precision="I8">
388
+ <dim>1280</dim>
389
+ <dim>320</dim>
390
+ </port>
391
+ </output>
392
+ </layer>
393
+ <layer id="30" name="time_embedding.linear_1.weight372242271/quantized/to_f32" type="Convert" version="opset1">
394
+ <data destination_type="f32" />
395
+ <input>
396
+ <port id="0" precision="I8">
397
+ <dim>1280</dim>
398
+ <dim>320</dim>
399
+ </port>
400
+ </input>
401
+ <output>
402
+ <port id="1" precision="FP32">
403
+ <dim>1280</dim>
404
+ <dim>320</dim>
405
+ </port>
406
+ </output>
407
+ </layer>
408
+ <layer id="31" name="/time_embedding/linear_1/Gemm/WithoutBiases/fq_weights_1/zero_point4895772690" type="Const" version="opset1">
409
+ <data element_type="f32" shape="1280, 1" offset="410320" size="5120" />
410
+ <output>
411
+ <port id="0" precision="FP32">
412
+ <dim>1280</dim>
413
+ <dim>1</dim>
414
+ </port>
415
+ </output>
416
+ </layer>
417
+ <layer id="32" name="/time_embedding/linear_1/Gemm/WithoutBiases/fq_weights_1/minus_zp" type="Subtract" version="opset1">
418
+ <data auto_broadcast="numpy" />
419
+ <input>
420
+ <port id="0" precision="FP32">
421
+ <dim>1280</dim>
422
+ <dim>320</dim>
423
+ </port>
424
+ <port id="1" precision="FP32">
425
+ <dim>1280</dim>
426
+ <dim>1</dim>
427
+ </port>
428
+ </input>
429
+ <output>
430
+ <port id="2" precision="FP32">
431
+ <dim>1280</dim>
432
+ <dim>320</dim>
433
+ </port>
434
+ </output>
435
+ </layer>
436
+ <layer id="33" name="/time_embedding/linear_1/Gemm/WithoutBiases/fq_weights_1/scale4894678903" type="Const" version="opset1">
437
+ <data element_type="f32" shape="1280, 1" offset="415440" size="5120" />
438
+ <output>
439
+ <port id="0" precision="FP32">
440
+ <dim>1280</dim>
441
+ <dim>1</dim>
442
+ </port>
443
+ </output>
444
+ </layer>
445
+ <layer id="34" name="/time_embedding/linear_1/Gemm/WithoutBiases/fq_weights_1/mulpiply_by_scale" type="Multiply" version="opset1">
446
+ <data auto_broadcast="numpy" />
447
+ <input>
448
+ <port id="0" precision="FP32">
449
+ <dim>1280</dim>
450
+ <dim>320</dim>
451
+ </port>
452
+ <port id="1" precision="FP32">
453
+ <dim>1280</dim>
454
+ <dim>1</dim>
455
+ </port>
456
+ </input>
457
+ <output>
458
+ <port id="2" precision="FP32">
459
+ <dim>1280</dim>
460
+ <dim>320</dim>
461
+ </port>
462
+ </output>
463
+ </layer>
464
+ <layer id="35" name="/time_embedding/linear_1/Gemm/WithoutBiases" type="MatMul" version="opset1">
465
+ <data transpose_a="false" transpose_b="true" />
466
+ <input>
467
+ <port id="0" precision="FP32">
468
+ <dim>1</dim>
469
+ <dim>320</dim>
470
+ </port>
471
+ <port id="1" precision="FP32">
472
+ <dim>1280</dim>
473
+ <dim>320</dim>
474
+ </port>
475
+ </input>
476
+ <output>
477
+ <port id="2" precision="FP32">
478
+ <dim>1</dim>
479
+ <dim>1280</dim>
480
+ </port>
481
+ </output>
482
+ </layer>
483
+ <layer id="36" name="Constant_75023372475417" type="Const" version="opset1">
484
+ <data element_type="f32" shape="1, 1280" offset="420560" size="5120" />
485
+ <output>
486
+ <port id="0" precision="FP32">
487
+ <dim>1</dim>
488
+ <dim>1280</dim>
489
+ </port>
490
+ </output>
491
+ </layer>
492
+ <layer id="37" name="/time_embedding/linear_1/Gemm" type="Add" version="opset1">
493
+ <data auto_broadcast="numpy" />
494
+ <input>
495
+ <port id="0" precision="FP32">
496
+ <dim>1</dim>
497
+ <dim>1280</dim>
498
+ </port>
499
+ <port id="1" precision="FP32">
500
+ <dim>1</dim>
501
+ <dim>1280</dim>
502
+ </port>
503
+ </input>
504
+ <output>
505
+ <port id="2" precision="FP32" names="/time_embedding/linear_1/Gemm_output_0">
506
+ <dim>1</dim>
507
+ <dim>1280</dim>
508
+ </port>
509
+ </output>
510
+ </layer>
511
+ <layer id="38" name="/time_embedding/linear_1/Gemm0" type="Result" version="opset1">
512
+ <input>
513
+ <port id="0" precision="FP32">
514
+ <dim>1</dim>
515
+ <dim>1280</dim>
516
+ </port>
517
+ </input>
518
+ </layer>
519
+ </layers>
520
+ <edges>
521
+ <edge from-layer="0" from-port="0" to-layer="2" to-port="0" />
522
+ <edge from-layer="1" from-port="0" to-layer="2" to-port="1" />
523
+ <edge from-layer="2" from-port="2" to-layer="4" to-port="0" />
524
+ <edge from-layer="3" from-port="0" to-layer="4" to-port="1" />
525
+ <edge from-layer="4" from-port="2" to-layer="6" to-port="0" />
526
+ <edge from-layer="5" from-port="0" to-layer="6" to-port="1" />
527
+ <edge from-layer="6" from-port="2" to-layer="7" to-port="0" />
528
+ <edge from-layer="6" from-port="2" to-layer="13" to-port="0" />
529
+ <edge from-layer="7" from-port="1" to-layer="12" to-port="0" />
530
+ <edge from-layer="8" from-port="0" to-layer="12" to-port="1" />
531
+ <edge from-layer="9" from-port="0" to-layer="12" to-port="2" />
532
+ <edge from-layer="10" from-port="0" to-layer="12" to-port="3" />
533
+ <edge from-layer="11" from-port="0" to-layer="12" to-port="4" />
534
+ <edge from-layer="12" from-port="5" to-layer="19" to-port="0" />
535
+ <edge from-layer="13" from-port="1" to-layer="18" to-port="0" />
536
+ <edge from-layer="14" from-port="0" to-layer="18" to-port="1" />
537
+ <edge from-layer="15" from-port="0" to-layer="18" to-port="2" />
538
+ <edge from-layer="16" from-port="0" to-layer="18" to-port="3" />
539
+ <edge from-layer="17" from-port="0" to-layer="18" to-port="4" />
540
+ <edge from-layer="18" from-port="5" to-layer="19" to-port="1" />
541
+ <edge from-layer="19" from-port="2" to-layer="23" to-port="0" />
542
+ <edge from-layer="19" from-port="2" to-layer="27" to-port="0" />
543
+ <edge from-layer="20" from-port="0" to-layer="23" to-port="1" />
544
+ <edge from-layer="21" from-port="0" to-layer="23" to-port="2" />
545
+ <edge from-layer="22" from-port="0" to-layer="23" to-port="3" />
546
+ <edge from-layer="23" from-port="4" to-layer="28" to-port="0" />
547
+ <edge from-layer="24" from-port="0" to-layer="27" to-port="1" />
548
+ <edge from-layer="25" from-port="0" to-layer="27" to-port="2" />
549
+ <edge from-layer="26" from-port="0" to-layer="27" to-port="3" />
550
+ <edge from-layer="27" from-port="4" to-layer="28" to-port="1" />
551
+ <edge from-layer="28" from-port="2" to-layer="35" to-port="0" />
552
+ <edge from-layer="29" from-port="0" to-layer="30" to-port="0" />
553
+ <edge from-layer="30" from-port="1" to-layer="32" to-port="0" />
554
+ <edge from-layer="31" from-port="0" to-layer="32" to-port="1" />
555
+ <edge from-layer="32" from-port="2" to-layer="34" to-port="0" />
556
+ <edge from-layer="33" from-port="0" to-layer="34" to-port="1" />
557
+ <edge from-layer="34" from-port="2" to-layer="35" to-port="1" />
558
+ <edge from-layer="35" from-port="2" to-layer="37" to-port="0" />
559
+ <edge from-layer="36" from-port="0" to-layer="37" to-port="1" />
560
+ <edge from-layer="37" from-port="2" to-layer="38" to-port="0" />
561
+ </edges>
562
+ <rt_info />
563
+ </net>