Skip to content

Commit

Permalink
Edge TPU tf.lite.experimental.load_delegate fix
Browse files Browse the repository at this point in the history
Fix attempt for #6535
  • Loading branch information
glenn-jocher committed Feb 4, 2022
1 parent 9bc72a3 commit 8676333
Showing 1 changed file with 6 additions and 5 deletions.
11 changes: 6 additions & 5 deletions models/common.py
Original file line number Diff line number Diff line change
Expand Up @@ -374,19 +374,20 @@ def wrap_frozen_graph(gd, inputs, outputs):
graph_def.ParseFromString(open(w, 'rb').read())
frozen_func = wrap_frozen_graph(gd=graph_def, inputs="x:0", outputs="Identity:0")
elif tflite: # https://www.tensorflow.org/lite/guide/python#install_tensorflow_lite_for_python
try:
import tflite_runtime.interpreter as tfl # prefer tflite_runtime if installed
try: # prefer tflite_runtime if installed
from tflite_runtime.interpreter import Interpreter, load_delegate
except ImportError:
import tensorflow.lite as tfl
import tensorflow.lite.Interpreter as Interpreter
import tensorflow.lite.experimental.load_delegate as load_delegate
if 'edgetpu' in w.lower(): # Edge TPU https://coral.ai/software/#edgetpu-runtime
LOGGER.info(f'Loading {w} for TensorFlow Lite Edge TPU inference...')
delegate = {'Linux': 'libedgetpu.so.1',
'Darwin': 'libedgetpu.1.dylib',
'Windows': 'edgetpu.dll'}[platform.system()]
interpreter = tfl.Interpreter(model_path=w, experimental_delegates=[tfl.load_delegate(delegate)])
interpreter = Interpreter(model_path=w, experimental_delegates=[load_delegate(delegate)])
else: # Lite
LOGGER.info(f'Loading {w} for TensorFlow Lite inference...')
interpreter = tfl.Interpreter(model_path=w) # load TFLite model
interpreter = Interpreter(model_path=w) # load TFLite model
interpreter.allocate_tensors() # allocate
input_details = interpreter.get_input_details() # inputs
output_details = interpreter.get_output_details() # outputs
Expand Down

0 comments on commit 8676333

Please sign in to comment.