-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathconvert_model.py
More file actions
752 lines (603 loc) · 27.5 KB
/
convert_model.py
File metadata and controls
752 lines (603 loc) · 27.5 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
import argparse
import base64
import datetime
import hashlib
import json
from typing import Tuple, Union, Mapping, List, Callable, Set, Any, TypeVar
import numpy as np
import numpy.typing
from keras import backend as K, Layer
from keras.layers import Input, Embedding, CategoryEncoding
from keras.models import Model, load_model
from keras.src import Functional
__author__ = "Tobias Hermann"
__copyright__ = "Copyright 2017, Tobias Hermann"
__license__ = "MIT"
__maintainer__ = "Tobias Hermann, https://github.com/Dobiasd/frugally-deep"
__email__ = "editgym@gmail.com"
NDFloat32Array = np.typing.NDArray[np.float32]
NDUInt32Array = np.typing.NDArray[np.int32]
Shape = Tuple[int, ...]
LayerConfig = Union[None, Mapping[str, Union[float, list[str], list[list[str]]]]]
TensorRepr = Mapping[str, Union[Shape, List[str]]]
TypeT = TypeVar('TypeT')
def as_list(value_or_values: Union[TypeT, List[TypeT]]) -> List[TypeT]:
"""Leave lists untouched, convert non-list types to a singleton list"""
if isinstance(value_or_values, list):
return value_or_values
return [value_or_values]
def keras_shape_to_fdeep_tensor_shape(raw_shape: Shape) -> Shape:
"""Convert a keras shape to an fdeep shape"""
return raw_shape[1:]
def get_layer_input_shape(layer: Layer) -> Shape:
"""It is stored in a different property depending on the situation."""
if hasattr(layer, 'batch_shape'):
return tuple(layer.batch_shape)
return tuple(layer.input.shape)
def get_layer_input_shape_tensor_shape(layer: Layer) -> Shape:
"""Convert layer input shape to an fdeep shape"""
return keras_shape_to_fdeep_tensor_shape(get_layer_input_shape(layer))
def show_tensor(tens: NDFloat32Array) -> TensorRepr:
"""Serialize 3-tensor to a dict"""
return {
'shape': tens.shape[1:],
'values': encode_floats(tens.flatten())
}
def get_model_input_layers(model: Model) -> List[Layer]:
"""Gets the input layers from model.layers in the correct input order."""
if len(model.inputs) == 1:
from keras.src.layers.core.input_layer import InputLayer
input_layers = []
for layer in model.layers:
if isinstance(layer, InputLayer):
input_layers.append(layer)
return input_layers
input_layer_names = [model_input.name for model_input in model.inputs]
model_layers = {layer.name: layer for layer in model.layers}
return [model_layers[layer_names] for layer_names in input_layer_names]
def measure_predict(model: Model, data_in: List[NDFloat32Array]) -> Tuple[List[NDFloat32Array], float]:
"""Returns output and duration in seconds"""
start_time = datetime.datetime.now()
data_out = model.predict(data_in)
end_time = datetime.datetime.now()
duration = end_time - start_time
print('Forward pass took {} s.'.format(duration.total_seconds()))
return data_out, duration.total_seconds()
def replace_none_with(value: int, shape: Shape) -> Shape:
"""Replace every None with a fixed value."""
return tuple(list(map(lambda x: x if x is not None else value, shape)))
def get_first_outbound_op(layer: Layer) -> Functional:
"""Determine primary outbound operation"""
return layer._outbound_nodes[0].operation
def are_embedding_and_category_encoding_layer_positions_ok_for_testing(model: Model) -> bool:
"""
Test data can only be generated if all Embedding layers
and CategoryEncoding layers are positioned directly behind the input nodes.
"""
def embedding_layer_names(model: Model) -> Set[str]:
layers = model.layers
result = set()
for layer in layers:
if isinstance(layer, Embedding):
result.add(layer.name)
layer_type = type(layer).__name__
if layer_type in ['Model', 'Sequential', 'Functional']:
result.union(embedding_layer_names(layer))
return result
def embedding_layer_names_at_input_nodes(model: Model) -> Set[str]:
result = set()
for input_layer in get_model_input_layers(model):
if input_layer._outbound_nodes and (
isinstance(get_first_outbound_op(input_layer), Embedding) or
isinstance(get_first_outbound_op(input_layer), CategoryEncoding)):
result.add(get_first_outbound_op(input_layer).name)
return result
return embedding_layer_names(model) == embedding_layer_names_at_input_nodes(model)
def gen_test_data(model: Model) -> Mapping[str, List[TensorRepr]]:
"""Generate data for model verification test."""
def set_shape_idx_0_to_1_if_none(shape: Shape) -> Shape:
"""Change first element in tuple to 1."""
if shape[0] is not None:
return shape
shape_lst = list(shape)
shape_lst[0] = 1
shape = tuple(shape_lst)
return shape
def generate_input_data(input_layer: Layer) -> NDFloat32Array:
"""Random data fitting the input shape of a layer."""
random_fn: Callable[[Shape], Union[NDFloat32Array, NDUInt32Array]]
if input_layer._outbound_nodes and isinstance(
get_first_outbound_op(input_layer), Embedding):
random_fn = lambda size: np.random.randint(
0, get_first_outbound_op(input_layer).input_dim, size)
elif input_layer._outbound_nodes and isinstance(
get_first_outbound_op(input_layer), CategoryEncoding):
random_fn = lambda size: np.random.randint(
0, get_first_outbound_op(input_layer).num_tokens, size)
else:
random_fn = lambda size: np.random.normal(size=size).astype(np.float32)
shape = get_layer_input_shape(input_layer)
data = random_fn(replace_none_with(32, set_shape_idx_0_to_1_if_none(shape))).astype(np.float32)
return data
assert are_embedding_and_category_encoding_layer_positions_ok_for_testing(
model), 'Test data can only be generated if embedding layers are positioned directly after input nodes.'
data_in: List[NDFloat32Array] = list(map(generate_input_data, get_model_input_layers(model)))
data_out_test: List[NDFloat32Array]
warm_up_runs = 3
test_runs = 5
for i in range(warm_up_runs):
if i == 0:
data_out_test_raw, _ = measure_predict(model, data_in)
data_out_test = as_list(data_out_test_raw)
else:
measure_predict(model, data_in)
duration_sum = 0.0
print('Starting performance measurements.')
for _ in range(test_runs):
data_out, duration = measure_predict(model, data_in)
duration_sum = duration_sum + duration
duration_avg = duration_sum / test_runs
print('Forward pass took {} s on average.'.format(duration_avg))
return {
'inputs': list(map(show_tensor, data_in)),
'outputs': list(map(show_tensor, data_out_test))
}
def split_every(size: int, seq: str) -> List[str]:
"""Split a sequence every seq elements."""
return [seq[pos:pos + size] for pos in range(0, len(seq), size)]
def encode_floats(arr: NDFloat32Array) -> List[str]:
"""Serialize a sequence of floats."""
return list(split_every(1024, base64.b64encode(arr).decode('ascii')))
def prepare_filter_weights_conv_2d(weights: NDFloat32Array) -> NDFloat32Array:
"""Change dimension order of 2d filter weights to the one used in fdeep"""
assert len(weights.shape) == 4
return np.moveaxis(weights, [0, 1, 2, 3], [1, 2, 3, 0]).flatten()
def prepare_filter_weights_slice_conv_2d(weights: NDFloat32Array) -> NDFloat32Array:
"""Change dimension order of 2d filter weights to the one used in fdeep"""
assert len(weights.shape) == 4
return np.moveaxis(weights, [0, 1, 2, 3], [1, 2, 0, 3]).flatten()
def prepare_filter_weights_conv_1d(weights: NDFloat32Array) -> NDFloat32Array:
"""Change dimension order of 1d filter weights to the one used in fdeep"""
assert len(weights.shape) == 3
return np.moveaxis(weights, [0, 1, 2], [1, 2, 0]).flatten()
def prepare_filter_weights_conv_1d_transpose(weights: NDFloat32Array) -> NDFloat32Array:
"""Change dimension order of 1d filter weights to the one used in fdeep"""
assert len(weights.shape) == 3
return np.moveaxis(weights, [0, 1, 2], [1, 0, 2]).flatten()
def prepare_filter_weights_conv_2d_transpose(weights: NDFloat32Array) -> NDFloat32Array:
"""Change dimension order of 2d filter weights to the one used in fdeep"""
assert len(weights.shape) == 4
return np.moveaxis(weights, [0, 1, 2, 3], [1, 2, 0, 3]).flatten()
def show_conv_1d_layer(layer: Layer) -> Mapping[str, list[str]]:
"""Serialize Conv1D layer to dict"""
weights = layer.get_weights()
assert len(weights) == 1 or len(weights) == 2
assert len(weights[0].shape) == 3
weights_flat = prepare_filter_weights_conv_1d(weights[0])
assert layer.padding in ['valid', 'same', 'causal']
assert layer.groups == 1
assert len(get_layer_input_shape(layer)) == 3
assert get_layer_input_shape(layer)[0] in {None, 1}
result = {
'weights': encode_floats(weights_flat)
}
if len(weights) == 2:
bias = weights[1]
result['bias'] = encode_floats(bias)
return result
def show_conv_2d_layer(layer: Layer) -> Mapping[str, list[str]]:
"""Serialize Conv2D layer to dict"""
weights = layer.get_weights()
assert len(weights) == 1 or len(weights) == 2
assert len(weights[0].shape) == 4
weights_flat = prepare_filter_weights_conv_2d(weights[0])
assert layer.padding in ['valid', 'same']
assert layer.groups == 1
assert len(get_layer_input_shape(layer)) == 4
assert get_layer_input_shape(layer)[0] in {None, 1}
result = {
'weights': encode_floats(weights_flat)
}
if len(weights) == 2:
bias = weights[1]
result['bias'] = encode_floats(bias)
return result
def show_separable_conv_2d_layer(layer: Layer) -> Mapping[str, list[str]]:
"""Serialize SeparableConv2D layer to dict"""
weights = layer.get_weights()
assert layer.depth_multiplier == 1
assert len(weights) == 2 or len(weights) == 3
assert len(weights[0].shape) == 4
assert len(weights[1].shape) == 4
# probably incorrect for depth_multiplier > 1?
slice_weights = prepare_filter_weights_slice_conv_2d(weights[0])
stack_weights = prepare_filter_weights_conv_2d(weights[1])
assert layer.padding in ['valid', 'same']
assert len(get_layer_input_shape(layer)) == 4
assert get_layer_input_shape(layer)[0] in {None, 1}
result = {
'slice_weights': encode_floats(slice_weights),
'stack_weights': encode_floats(stack_weights),
}
if len(weights) == 3:
bias = weights[2]
result['bias'] = encode_floats(bias)
return result
def show_depthwise_conv_2d_layer(layer: Layer) -> Mapping[str, list[str]]:
"""Serialize DepthwiseConv2D layer to dict"""
weights = layer.get_weights()
assert layer.depth_multiplier == 1
assert len(weights) in [1, 2]
assert len(weights[0].shape) == 4
# probably incorrect for depth_multiplier > 1?
slice_weights = prepare_filter_weights_slice_conv_2d(weights[0])
assert layer.padding in ['valid', 'same']
assert len(get_layer_input_shape(layer)) == 4
assert get_layer_input_shape(layer)[0] in {None, 1}
result = {
'slice_weights': encode_floats(slice_weights),
}
if len(weights) == 2:
bias = weights[1]
result['bias'] = encode_floats(bias)
return result
def show_conv_1d_transpose_layer(layer: Layer) -> Mapping[str, list[str]]:
"""Serialize Conv1D transpose layer to dict"""
weights = layer.get_weights()
assert len(weights) == 1 or len(weights) == 2
assert len(weights[0].shape) == 3
weights_flat = prepare_filter_weights_conv_1d_transpose(weights[0])
assert layer.padding in ['valid', 'same', 'causal']
assert layer.strides[0] <= layer.kernel_size[0]
assert len(get_layer_input_shape(layer)) == 3
assert get_layer_input_shape(layer)[0] in {None, 1}
result = {
'weights': encode_floats(weights_flat)
}
if len(weights) == 2:
bias = weights[1]
result['bias'] = encode_floats(bias)
return result
def show_conv_2d_transpose_layer(layer: Layer) -> Mapping[str, list[str]]:
"""Serialize Conv2D transpose layer to dict"""
weights = layer.get_weights()
assert len(weights) == 1 or len(weights) == 2
assert len(weights[0].shape) == 4
weights_flat = prepare_filter_weights_conv_2d_transpose(weights[0])
assert layer.padding in ['valid', 'same']
assert layer.strides[0] <= layer.kernel_size[0]
assert layer.strides[1] <= layer.kernel_size[1]
assert sum([
layer.dilation_rate[0] == layer.dilation_rate[1],
layer.strides[0] == layer.strides[1],
layer.kernel_size[0] == layer.kernel_size[1]]) >= 2
assert len(get_layer_input_shape(layer)) == 4
assert get_layer_input_shape(layer)[0] in {None, 1}
result = {
'weights': encode_floats(weights_flat)
}
if len(weights) == 2:
bias = weights[1]
result['bias'] = encode_floats(bias)
return result
def show_batch_normalization_layer(layer: Layer) -> Mapping[str, list[str]]:
"""Serialize batch normalization layer to dict"""
moving_mean = layer.moving_mean.numpy()
moving_variance = layer.moving_variance.numpy()
result = {}
result['moving_mean'] = encode_floats(moving_mean)
result['moving_variance'] = encode_floats(moving_variance)
if layer.center:
beta = layer.beta.numpy()
result['beta'] = encode_floats(beta)
if layer.scale:
gamma = layer.gamma.numpy()
result['gamma'] = encode_floats(gamma)
return result
def show_layer_normalization_layer(layer: Layer) -> Mapping[str, list[str]]:
"""Serialize layer normalization layer to dict"""
result = {}
if layer.center:
beta = layer.beta.numpy()
result['beta'] = encode_floats(beta)
if layer.scale:
gamma = layer.gamma.numpy()
result['gamma'] = encode_floats(gamma)
return result
def show_dense_layer(layer: Layer) -> Mapping[str, list[str]]:
"""Serialize dense layer to dict"""
weights = layer.get_weights()
assert len(weights) == 1 or len(weights) == 2
assert len(weights[0].shape) == 2
weights_flat = weights[0].flatten()
result = {
'weights': encode_floats(weights_flat)
}
if len(weights) == 2:
bias = weights[1]
result['bias'] = encode_floats(bias)
return result
def show_dot_layer(layer: Layer) -> None:
"""Check valid configuration of Dot layer"""
assert len(get_layer_input_shape(layer)) == 2
assert isinstance(layer.axes, int) or (isinstance(layer.axes, list) and len(layer.axes) == 2)
assert layer.input.shape[0][0] is None
assert layer.input.shape[1][0] is None
assert len(layer.output_shape) <= 5
def show_prelu_layer(layer: Layer) -> Mapping[str, list[str]]:
"""Serialize prelu layer to dict"""
weights = layer.get_weights()
assert len(weights) == 1
weights_flat = weights[0].flatten()
result = {
'alpha': encode_floats(weights_flat)
}
return result
def show_embedding_layer(layer: Layer) -> Mapping[str, list[str]]:
"""Serialize Embedding layer to dict"""
weights = layer.get_weights()
assert len(weights) == 1
result = {
'weights': encode_floats(weights[0])
}
return result
def show_input_layer(layer: Layer) -> None:
"""Serialize input layer to dict"""
assert not layer.sparse
def show_softmax_layer(layer: Layer) -> None:
"""Serialize softmax layer to dict"""
assert layer.axis == -1
def show_normalization_layer(layer: Layer) -> Mapping[str, list[str]]:
"""Serialize normalization layer to dict"""
assert len(layer.axis) <= 1, 'Multiple normalization axes are not supported'
if len(layer.axis) == 1:
assert layer.axis[0] in (-1, 1, 2, 3, 4, 5), 'Invalid axis for Normalization layer.'
return {
'mean': encode_floats(layer.mean),
'variance': encode_floats(layer.variance)
}
def show_upsampling2d_layer(layer: Layer) -> None:
"""Serialize UpSampling2D layer to dict"""
assert layer.interpolation in ['nearest', 'bilinear']
def show_resizing_layer(layer: Layer) -> None:
"""Serialize Resizing layer to dict"""
assert layer.interpolation in ['nearest', 'bilinear', 'area']
def show_rescaling_layer(layer: Layer) -> None:
"""Serialize Rescaling layer to dict"""
assert isinstance(layer.scale, float)
def show_category_encoding_layer(layer: Layer) -> None:
"""Serialize CategoryEncoding layer to dict"""
assert layer.output_mode in ['multi_hot', 'count', 'one_hot']
def show_attention_layer(layer: Layer) -> Mapping[str, float]:
"""Serialize Attention layer to dict"""
assert layer.score_mode in ['dot', 'concat']
data = {}
if layer.scale is not None:
data['scale'] = float(layer.scale.numpy())
if layer.score_mode == 'concat':
data['concat_score_weight'] = float(layer.concat_score_weight.numpy())
return data
def show_additive_attention_layer(layer: Layer) -> Mapping[str, List[str]]:
"""Serialize AdditiveAttention layer to dict"""
data = {}
if layer.scale is not None:
data['scale'] = encode_floats(layer.scale.numpy())
return data
def show_multi_head_attention_layer(layer: Layer) -> Mapping[str, List[list[str]]]:
"""Serialize MultiHeadAttention layer to dict"""
assert layer._output_shape is None
assert layer._attention_axes == (1,), 'MultiHeadAttention supported only with attention_axes=None'
return {
'weight_shapes': list(map(lambda w: list(w.shape), layer.weights)),
'weights': list(map(lambda w: encode_floats(w.numpy()), layer.weights)),
}
def get_layer_functions_dict() -> Mapping[str, Callable[[Layer], LayerConfig]]:
return {
'Conv1D': show_conv_1d_layer,
'Conv2D': show_conv_2d_layer,
'Conv1DTranspose': show_conv_1d_transpose_layer,
'Conv2DTranspose': show_conv_2d_transpose_layer,
'SeparableConv2D': show_separable_conv_2d_layer,
'DepthwiseConv2D': show_depthwise_conv_2d_layer,
'BatchNormalization': show_batch_normalization_layer,
'Dense': show_dense_layer,
'Dot': show_dot_layer,
'PReLU': show_prelu_layer,
'Embedding': show_embedding_layer,
'LayerNormalization': show_layer_normalization_layer,
'TimeDistributed': show_time_distributed_layer,
'Input': show_input_layer,
'Softmax': show_softmax_layer,
'Normalization': show_normalization_layer,
'UpSampling2D': show_upsampling2d_layer,
'Resizing': show_resizing_layer,
'Rescaling': show_rescaling_layer,
'CategoryEncoding': show_category_encoding_layer,
'Attention': show_attention_layer,
'AdditiveAttention': show_additive_attention_layer,
'MultiHeadAttention': show_multi_head_attention_layer,
}
def show_time_distributed_layer(layer: Layer) -> Union[None, LayerConfig]:
show_layer_functions = get_layer_functions_dict()
config = layer.get_config()
class_name = str(config['layer']['class_name'])
if show_layer_functions and class_name in show_layer_functions:
input_shape_new: Shape
if len(get_layer_input_shape(layer)) == 3:
input_shape_new = (get_layer_input_shape(layer)[0], get_layer_input_shape(layer)[2])
elif len(get_layer_input_shape(layer)) == 4:
input_shape_new = (
get_layer_input_shape(layer)[0], get_layer_input_shape(layer)[2], get_layer_input_shape(layer)[3])
elif len(get_layer_input_shape(layer)) == 5:
input_shape_new = (
get_layer_input_shape(layer)[0], get_layer_input_shape(layer)[2], get_layer_input_shape(layer)[3],
get_layer_input_shape(layer)[4])
elif len(get_layer_input_shape(layer)) == 6:
input_shape_new = (
get_layer_input_shape(layer)[0], get_layer_input_shape(layer)[2], get_layer_input_shape(layer)[3],
get_layer_input_shape(layer)[4],
get_layer_input_shape(layer)[5])
else:
raise Exception('Wrong input shape')
layer_function = show_layer_functions[class_name]
attributes = dir(layer.layer)
class CopiedLayer:
pass
copied_layer = CopiedLayer()
for attr in attributes:
try:
if attr not in ['batch_shape', '__class__']:
setattr(copied_layer, attr, getattr(layer.layer, attr))
except Exception:
continue
setattr(copied_layer, 'batch_shape', input_shape_new)
setattr(copied_layer, 'output_shape', layer.output.shape)
return layer_function(copied_layer)
else:
return None
def get_dict_keys(d: Mapping[str, LayerConfig]) -> list[str]:
"""Return keys of a dictionary"""
return [key for key in d]
def merge_two_disjunct_dicts(x: Mapping[str, LayerConfig], y: Mapping[str, LayerConfig]) -> Mapping[str, LayerConfig]:
"""Given two dicts, merge them into a new dict as a shallow copy.
No Key is allowed to be present in both dictionaries.
"""
assert set(get_dict_keys(x)).isdisjoint(get_dict_keys(y))
assert isinstance(x, dict) and isinstance(y, dict)
z = x.copy()
z.update(y)
return z
def is_ascii(some_string: str) -> bool:
"""Check if a string only contains ascii characters"""
try:
some_string.encode('ascii')
except UnicodeEncodeError:
return False
else:
return True
def get_layer_weights(layer: Layer, name: str) -> Mapping[str, LayerConfig]:
"""Serialize all weights of a single normal layer"""
result: dict[str, LayerConfig] = {}
layer_type = type(layer).__name__
if hasattr(layer, 'data_format'):
assert layer.data_format == 'channels_last'
show_func = get_layer_functions_dict().get(layer_type, None)
shown_layer = None
if show_func:
shown_layer = show_func(layer)
if shown_layer:
result[name] = shown_layer
if show_func and layer_type == 'TimeDistributed':
result[name] = {'td_input_len': encode_floats(
np.array([len(get_layer_input_shape(layer)) - 1], dtype=np.float32)),
'td_output_len': encode_floats(np.array([len(layer.output.shape) - 1], dtype=np.float32))}
return result
def get_all_weights(model: Model, prefix: str) -> Mapping[str, LayerConfig]:
"""Serialize all weights of the models layers"""
result: dict[str, LayerConfig] = {}
layers = model.layers
assert K.image_data_format() == 'channels_last'
for layer in layers:
layer_type = type(layer).__name__
for node in layer._inbound_nodes:
if 'training' in node.arguments.kwargs:
is_layer_with_accidental_training_flag = layer_type in ('CenterCrop', 'Resizing')
has_training = node.arguments.kwargs['training'] is True
assert not has_training or is_layer_with_accidental_training_flag, \
'training=true is not supported, see https://github.com/Dobiasd/frugally-deep/issues/284'
name = prefix + layer.name
assert is_ascii(name)
if name in result:
raise ValueError('duplicate layer name ' + name)
if layer_type in ['Model', 'Sequential', 'Functional']:
result = dict(merge_two_disjunct_dicts(result, get_all_weights(layer, name + '_')))
elif layer_type in ['TimeDistributed'] and type(layer.layer).__name__ in ['Model', 'Sequential', 'Functional']:
inner_layer = layer.layer
result = dict(merge_two_disjunct_dicts(result, get_layer_weights(layer, name)))
result = dict(merge_two_disjunct_dicts(result, get_all_weights(inner_layer, name + '_')))
else:
result = dict(merge_two_disjunct_dicts(result, get_layer_weights(layer, name)))
return result
def get_model_name(model: Model) -> str:
"""Return .name or ._name"""
if hasattr(model, 'name'):
return str(model.name)
return str(model._name)
def convert_sequential_to_model(model: Model) -> Model:
"""Convert a sequential model to the underlying functional format"""
if type(model).__name__ in ['Sequential']:
name = get_model_name(model)
inbound_nodes = model._inbound_nodes
input_layer = Input(batch_shape=get_layer_input_shape(model.layers[0]))
prev_layer = input_layer
for layer in model.layers:
layer._inbound_nodes = []
prev_layer = layer(prev_layer)
funcmodel = Model([input_layer], [prev_layer], name=name)
model = funcmodel
model._inbound_nodes = inbound_nodes
if type(model).__name__ == 'TimeDistributed':
model.layer = convert_sequential_to_model(model.layer)
if type(model).__name__ in ['Model', 'Functional']:
for i in range(len(model.layers)):
new_layer = convert_sequential_to_model(model.layers[i])
if new_layer == model.layers[i]:
continue
model._operations[i] = new_layer
assert model.layers[i] == new_layer
return model
def get_shapes(tensors: List[Mapping[str, Shape]]) -> List[Shape]:
"""Return shapes of a list of tensors"""
return [t['shape'] for t in tensors]
def calculate_hash(model: Model) -> str:
layers = model.layers
hash_m = hashlib.sha256()
for layer in layers:
for weights in layer.get_weights():
if isinstance(weights, np.ndarray):
hash_m.update(weights.tobytes())
hash_m.update(layer.name.encode('ascii'))
return hash_m.hexdigest()
def model_to_fdeep_json(model: Model, no_tests: bool = False) -> Mapping[str, Any]:
"""Convert any Keras model to the frugally-deep model format."""
model.compile(loss='mse', optimizer='sgd')
model = convert_sequential_to_model(model)
test_data = None if no_tests else gen_test_data(model)
json_output = {}
print('Converting model architecture.')
json_output['architecture'] = json.loads(model.to_json())
json_output['image_data_format'] = K.image_data_format()
json_output['input_shapes'] = list(map(get_layer_input_shape_tensor_shape, get_model_input_layers(model)))
json_output['output_shapes'] = list(map(keras_shape_to_fdeep_tensor_shape, as_list(model.output_shape)))
if test_data:
json_output['tests'] = [test_data]
print('Converting model weights.')
json_output['trainable_params'] = get_all_weights(model, '')
print('Done converting model weights.')
print('Calculating model hash.')
json_output['hash'] = calculate_hash(model)
print('Model conversion finished.')
return json_output
def assert_model_type(model: Model) -> None:
import keras
assert type(model) in [keras.src.models.sequential.Sequential, keras.src.models.functional.Functional]
def convert(in_path: str, out_path: str, no_tests: bool = False) -> None:
"""Convert any (h5-)stored Keras model to the frugally-deep model format."""
print('loading {}'.format(in_path))
model = load_model(in_path, compile=False)
json_output = model_to_fdeep_json(model, no_tests)
print('writing {}'.format(out_path))
with open(out_path, 'w') as f:
json.dump(json_output, f, allow_nan=False, separators=(',', ':'))
def main() -> None:
"""Parse command line and convert model."""
parser = argparse.ArgumentParser(
prog='frugally-deep model converter',
description='Converts models from Keras\' .keras format to frugally-deep\'s .json format.')
parser.add_argument('input_path', type=str)
parser.add_argument('output_path', type=str)
parser.add_argument('--no-tests', action='store_true')
args = parser.parse_args()
convert(args.input_path, args.output_path, args.no_tests)
if __name__ == '__main__':
main()