-
Notifications
You must be signed in to change notification settings - Fork 82
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Fix the bug of compile model to tensorrt
Closes #491 Signed-off-by: zhangkaili <zhang.kaili@zte.com.cn>
- Loading branch information
1 parent
01b8e7f
commit 7e9aa0c
Showing
84 changed files
with
6,070 additions
and
5,923 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
131 changes: 66 additions & 65 deletions
131
model_compiler/src/model_compiler/compilers/caffe_model_file_to_onnx_model.py
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,65 +1,66 @@ | ||
# Copyright 2019 ZTE corporation. All Rights Reserved. | ||
# SPDX-License-Identifier: Apache-2.0 | ||
|
||
import os | ||
from typing import Any, Mapping, NamedTuple, Optional, Sequence, List | ||
import numpy as np | ||
import caffe2.python.onnx.frontend | ||
from caffe2.proto import caffe2_pb2 | ||
from . import repository | ||
from .. import utilities | ||
from ..models.data_format import DataFormat | ||
from ..models.data_type import DataType | ||
from ..models.irs.onnx_model import OnnxModel | ||
from ..models.sources.caffe_model_file import CaffeModelFile | ||
|
||
|
||
class Config(NamedTuple): | ||
input_names: Sequence[str] | ||
input_formats: Sequence[Optional[DataFormat]] | ||
input_shapes: List[List] | ||
input_type: np.dtype | ||
max_batch_size: int | ||
|
||
@staticmethod | ||
def from_json(value: Mapping[str, Any]) -> 'Config': | ||
return Config(input_names=value['input_names'], | ||
input_formats=utilities.get_data_formats(value.get('input_formats')), | ||
input_shapes=utilities.get_input_shapes(value['input_shapes']), | ||
input_type=DataType.from_caffe_data_type(value['data_type']).to_onnx_data_type(), | ||
max_batch_size=value['max_batch_size']) | ||
|
||
@staticmethod | ||
def from_env(env: Mapping[str, str]) -> 'Config': | ||
return Config(input_names=env['INPUT_NAMES'].split(','), | ||
input_formats=utilities.get_data_formats(utilities.split_by(env.get('INPUT_FORMATS'), ',')), | ||
input_shapes=utilities.get_input_shapes( | ||
utilities.get_input_shapes_from_env(env.get('INPUT_SHAPES')) | ||
), | ||
input_type=DataType.from_caffe_data_type(env['DATA_TYPE']).to_onnx_data_type(), | ||
max_batch_size=int(env['MAX_BATCH_SIZE'])) | ||
|
||
|
||
def parse_caffe_net(net, pb_path): | ||
with open(pb_path, 'rb') as file: | ||
net.ParseFromString(file.read()) | ||
return net | ||
|
||
|
||
@repository.REPOSITORY.register(source_type=CaffeModelFile, target_type=OnnxModel, config_type=Config) | ||
def compile_source(source: CaffeModelFile, config: Config) -> OnnxModel: | ||
predict_net = parse_caffe_net(caffe2_pb2.NetDef(), os.path.join(source.model_path, 'predict_net.pb')) | ||
predict_net.name = "model" if predict_net.name == "" else predict_net.name # pylint: disable=no-member | ||
init_net = parse_caffe_net(caffe2_pb2.NetDef(), os.path.join(source.model_path, 'init_net.pb')) | ||
|
||
value_info = {} | ||
for i, input_shape in enumerate(config.input_shapes): | ||
input_shape.insert(0, config.max_batch_size) | ||
value_info[config.input_names[i]] = (config.input_type, input_shape) | ||
|
||
onnx_model = caffe2.python.onnx.frontend.caffe2_net_to_onnx_model(predict_net, init_net, value_info) | ||
|
||
graph = onnx_model.graph # pylint: disable=no-member | ||
return OnnxModel(model_proto=onnx_model, | ||
input_data_formats=utilities.get_onnx_model_input_data_formats(graph, | ||
config.input_formats)) | ||
# # Copyright 2019 ZTE corporation. All Rights Reserved. | ||
# # SPDX-License-Identifier: Apache-2.0 | ||
# | ||
# import os | ||
# from typing import Any, Mapping, NamedTuple, Optional, Sequence, List | ||
# import numpy as np | ||
# import caffe2.python.onnx.frontend | ||
# from caffe2.proto import caffe2_pb2 | ||
# from . import repository | ||
# from .. import utilities | ||
# from ..models.data_format import DataFormat | ||
# from ..models.data_type import DataType | ||
# from ..models.irs.onnx_model import OnnxModel | ||
# from ..models.sources.caffe_model_file import CaffeModelFile | ||
# | ||
# | ||
# class Config(NamedTuple): | ||
# input_names: Sequence[str] | ||
# input_formats: Sequence[Optional[DataFormat]] | ||
# input_shapes: List[List] | ||
# input_type: np.dtype | ||
# max_batch_size: int | ||
# | ||
# @staticmethod | ||
# def from_json(value: Mapping[str, Any]) -> 'Config': | ||
# return Config(input_names=value['input_names'], | ||
# input_formats=utilities.get_data_formats(value.get('input_formats')), | ||
# input_shapes=utilities.get_input_shapes(value['input_shapes']), | ||
# input_type=DataType.from_caffe_data_type(value['data_type']).to_onnx_data_type(), | ||
# max_batch_size=value['max_batch_size']) | ||
# | ||
# @staticmethod | ||
# def from_env(env: Mapping[str, str]) -> 'Config': | ||
# return Config(input_names=env['INPUT_NAMES'].split(','), | ||
# input_formats=utilities.get_data_formats(utilities.split_by(env.get('INPUT_FORMATS'), ',')), | ||
# input_shapes=utilities.get_input_shapes( | ||
# utilities.get_input_shapes_from_env(env.get('INPUT_SHAPES')) | ||
# ), | ||
# input_type=DataType.from_caffe_data_type(env['DATA_TYPE']).to_onnx_data_type(), | ||
# max_batch_size=int(env['MAX_BATCH_SIZE'])) | ||
# | ||
# | ||
# def parse_caffe_net(net, pb_path): | ||
# with open(pb_path, 'rb') as file: | ||
# net.ParseFromString(file.read()) | ||
# return net | ||
# | ||
# | ||
# @repository.REPOSITORY.register(source_type=CaffeModelFile, target_type=OnnxModel, config_type=Config) | ||
# def compile_source(source: CaffeModelFile, config: Config) -> OnnxModel: | ||
# predict_net = parse_caffe_net(caffe2_pb2.NetDef(), os.path.join(source.model_path, 'predict_net.pb')) | ||
# predict_net.name = "model" if predict_net.name == "" else predict_net.name # pylint: disable=no-member | ||
# init_net = parse_caffe_net(caffe2_pb2.NetDef(), os.path.join(source.model_path, 'init_net.pb')) | ||
# | ||
# value_info = {} | ||
# for i, input_shape in enumerate(config.input_shapes): | ||
# input_shape.insert(0, config.max_batch_size) | ||
# value_info[config.input_names[i]] = (config.input_type, input_shape) | ||
# | ||
# from caffe2.python.onnx.frontend import caffe2_net_to_onnx_model | ||
# onnx_model = caffe2_net_to_onnx_model(predict_net, init_net, value_info) | ||
# | ||
# graph = onnx_model.graph # pylint: disable=no-member | ||
# return OnnxModel(model_proto=onnx_model, | ||
# input_data_formats=utilities.get_onnx_model_input_data_formats(graph, | ||
# config.input_formats)) |
54 changes: 27 additions & 27 deletions
54
model_compiler/src/model_compiler/compilers/keras_model_file_to_keras_model.py
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,27 +1,27 @@ | ||
# Copyright 2019 ZTE corporation. All Rights Reserved. | ||
# SPDX-License-Identifier: Apache-2.0 | ||
|
||
import tensorflow as tf | ||
from tensorflow import keras | ||
|
||
from . import repository | ||
from .. import utilities | ||
from .. import keras_util | ||
from ..models.irs.keras_model import KerasModel | ||
from ..models.sources.keras_model_file import KerasModelFile | ||
|
||
|
||
@repository.REPOSITORY.register(source_type=KerasModelFile, target_type=KerasModel) | ||
def compile_source(source: KerasModelFile) -> KerasModel: | ||
with tf.Graph().as_default(): | ||
if source.script_path: | ||
with tf.compat.v1.Session(graph=tf.Graph(), config=utilities.get_tf_cpu_only_config()): | ||
custom_objects = keras_util.get_custom_objects(source.script_path) | ||
else: | ||
custom_objects = None | ||
|
||
with tf.compat.v1.Session(config=utilities.get_tf_cpu_only_config()).as_default() as session: | ||
keras.backend.set_learning_phase(0) | ||
model = keras.models.load_model(source.model_path, custom_objects=custom_objects, compile=False) | ||
|
||
return KerasModel(model=model, session=session) | ||
# # Copyright 2019 ZTE corporation. All Rights Reserved. | ||
# # SPDX-License-Identifier: Apache-2.0 | ||
# | ||
# import tensorflow as tf | ||
# from tensorflow import keras | ||
# | ||
# from . import repository | ||
# from .. import utilities | ||
# from .. import keras_util | ||
# from ..models.irs.keras_model import KerasModel | ||
# from ..models.sources.keras_model_file import KerasModelFile | ||
# | ||
# | ||
# @repository.REPOSITORY.register(source_type=KerasModelFile, target_type=KerasModel) | ||
# def compile_source(source: KerasModelFile) -> KerasModel: | ||
# with tf.Graph().as_default(): | ||
# if source.script_path: | ||
# with tf.compat.v1.Session(graph=tf.Graph(), config=utilities.get_tf_cpu_only_config()): | ||
# custom_objects = keras_util.get_custom_objects(source.script_path) | ||
# else: | ||
# custom_objects = None | ||
# | ||
# with tf.compat.v1.Session(config=utilities.get_tf_cpu_only_config()).as_default() as session: | ||
# keras.backend.set_learning_phase(0) | ||
# model = keras.models.load_model(source.model_path, custom_objects=custom_objects, compile=False) | ||
# | ||
# return KerasModel(model=model, session=session) |
48 changes: 24 additions & 24 deletions
48
model_compiler/src/model_compiler/compilers/keras_model_file_to_tflite_model.py
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,24 +1,24 @@ | ||
# Copyright 2019 ZTE corporation. All Rights Reserved. | ||
# SPDX-License-Identifier: Apache-2.0 | ||
|
||
import tensorflow as tf | ||
|
||
from . import repository | ||
from ..models.sources.keras_model_file import KerasModelFile | ||
|
||
from ..models.targets.tflite_model import TfLiteModel | ||
from .. import tflite_util | ||
from .. import keras_util | ||
|
||
|
||
@repository.REPOSITORY.register(source_type=KerasModelFile, target_type=TfLiteModel, config_type=tflite_util.Config) | ||
def compile_source(source: KerasModelFile, config: tflite_util.Config) -> TfLiteModel: | ||
if source.script_path: | ||
custom_objects = keras_util.get_custom_objects(source.script_path) | ||
else: | ||
custom_objects = None | ||
|
||
model = tf.keras.models.load_model(filepath=source.model_path, custom_objects=custom_objects, compile=False) | ||
converter = tf.lite.TFLiteConverter.from_keras_model(model) | ||
tflite_model = tflite_util.get_tflite_model(converter, config) | ||
return TfLiteModel(tflite_model, config.input_formats) | ||
# # Copyright 2019 ZTE corporation. All Rights Reserved. | ||
# # SPDX-License-Identifier: Apache-2.0 | ||
# | ||
# import tensorflow as tf | ||
# | ||
# from . import repository | ||
# from ..models.sources.keras_model_file import KerasModelFile | ||
# | ||
# from ..models.targets.tflite_model import TfLiteModel | ||
# from .. import tflite_util | ||
# from .. import keras_util | ||
# | ||
# | ||
# @repository.REPOSITORY.register(source_type=KerasModelFile, target_type=TfLiteModel, config_type=tflite_util.Config) | ||
# def compile_source(source: KerasModelFile, config: tflite_util.Config) -> TfLiteModel: | ||
# if source.script_path: | ||
# custom_objects = keras_util.get_custom_objects(source.script_path) | ||
# else: | ||
# custom_objects = None | ||
# | ||
# model = tf.keras.models.load_model(filepath=source.model_path, custom_objects=custom_objects, compile=False) | ||
# converter = tf.lite.TFLiteConverter.from_keras_model(model) | ||
# tflite_model = tflite_util.get_tflite_model(converter, config) | ||
# return TfLiteModel(tflite_model, config.input_formats) |
86 changes: 43 additions & 43 deletions
86
model_compiler/src/model_compiler/compilers/keras_model_file_to_tvm_model.py
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,43 +1,43 @@ | ||
# Copyright 2019 ZTE corporation. All Rights Reserved. | ||
# SPDX-License-Identifier: Apache-2.0 | ||
|
||
import tensorflow as tf | ||
import tvm | ||
import tvm.relay as relay | ||
|
||
from . import repository | ||
from ..models.sources.keras_model_file import KerasModelFile | ||
from ..models.targets.tvm_model import TvmModel, Input, Output | ||
from ..keras_util import Config, get_inputs, get_outputs, DataFormat | ||
|
||
|
||
def _get_shape_dict(model_inputs, max_batch_size): | ||
shape_dict = {} | ||
for input_tensor, data_format in model_inputs: | ||
tensor_shape = list(input_tensor.shape) | ||
tensor_shape.pop(0) | ||
tensor_shape.insert(0, max_batch_size) | ||
if data_format == DataFormat.CHANNELS_LAST: | ||
tensor_shape[1], tensor_shape[3] = tensor_shape[3], tensor_shape[1] | ||
shape_dict[input_tensor.name] = tensor_shape | ||
return shape_dict | ||
|
||
|
||
@repository.REPOSITORY.register(source_type=KerasModelFile, target_type=TvmModel, config_type=Config) | ||
def compile_source(source: KerasModelFile, config: Config) -> TvmModel: | ||
tf.keras.backend.set_learning_phase(0) | ||
source_model = tf.keras.models.load_model(source.model_path, compile=False) | ||
model_inputs = get_inputs(source_model, config.input_nodes) | ||
|
||
shape_dict = _get_shape_dict(model_inputs, config.max_batch_size) | ||
model, params = relay.frontend.from_keras(source_model, shape_dict) | ||
compiled_lib = relay.build(model, tvm.target.create("llvm"), params=params) | ||
return TvmModel(tvm_model=compiled_lib, | ||
model_inputs=[Input(name=tensor.name, | ||
shape=shape_dict[tensor.name], | ||
data_type=tensor.dtype.as_datatype_enum, | ||
data_format=DataFormat.CHANNELS_FIRST) for tensor, _ in model_inputs], | ||
model_outputs=[Output(name=tensor.name, | ||
shape=list(tensor.shape), | ||
data_type=tensor.dtype.as_datatype_enum) | ||
for tensor in get_outputs(source_model, config.output_nodes)]) | ||
# # Copyright 2019 ZTE corporation. All Rights Reserved. | ||
# # SPDX-License-Identifier: Apache-2.0 | ||
# | ||
# import tensorflow as tf | ||
# import tvm | ||
# import tvm.relay as relay | ||
# | ||
# from . import repository | ||
# from ..models.sources.keras_model_file import KerasModelFile | ||
# from ..models.targets.tvm_model import TvmModel, Input, Output | ||
# from ..keras_util import Config, get_inputs, get_outputs, DataFormat | ||
# | ||
# | ||
# def _get_shape_dict(model_inputs, max_batch_size): | ||
# shape_dict = {} | ||
# for input_tensor, data_format in model_inputs: | ||
# tensor_shape = list(input_tensor.shape) | ||
# tensor_shape.pop(0) | ||
# tensor_shape.insert(0, max_batch_size) | ||
# if data_format == DataFormat.CHANNELS_LAST: | ||
# tensor_shape[1], tensor_shape[3] = tensor_shape[3], tensor_shape[1] | ||
# shape_dict[input_tensor.name] = tensor_shape | ||
# return shape_dict | ||
# | ||
# | ||
# @repository.REPOSITORY.register(source_type=KerasModelFile, target_type=TvmModel, config_type=Config) | ||
# def compile_source(source: KerasModelFile, config: Config) -> TvmModel: | ||
# tf.keras.backend.set_learning_phase(0) | ||
# source_model = tf.keras.models.load_model(source.model_path, compile=False) | ||
# model_inputs = get_inputs(source_model, config.input_nodes) | ||
# | ||
# shape_dict = _get_shape_dict(model_inputs, config.max_batch_size) | ||
# model, params = relay.frontend.from_keras(source_model, shape_dict) | ||
# compiled_lib = relay.build(model, tvm.target.create("llvm"), params=params) | ||
# return TvmModel(tvm_model=compiled_lib, | ||
# model_inputs=[Input(name=tensor.name, | ||
# shape=shape_dict[tensor.name], | ||
# data_type=tensor.dtype.as_datatype_enum, | ||
# data_format=DataFormat.CHANNELS_FIRST) for tensor, _ in model_inputs], | ||
# model_outputs=[Output(name=tensor.name, | ||
# shape=list(tensor.shape), | ||
# data_type=tensor.dtype.as_datatype_enum) | ||
# for tensor in get_outputs(source_model, config.output_nodes)]) |
Oops, something went wrong.