Edge TPU compiler comment (#6196)
* Edge TPU compiler comment * 7 to 8 fix
This commit is contained in:
parent
9e9219fe17
commit
5402753a53
|
|
@ -17,7 +17,7 @@ TensorFlow Edge TPU | `edgetpu` | yolov5s_edgetpu.tf
|
||||||
TensorFlow.js | `tfjs` | yolov5s_web_model/
|
TensorFlow.js | `tfjs` | yolov5s_web_model/
|
||||||
|
|
||||||
Usage:
|
Usage:
|
||||||
$ python path/to/export.py --weights yolov5s.pt --include torchscript onnx coreml openvino saved_model tflite tfjs
|
$ python path/to/export.py --weights yolov5s.pt --include torchscript onnx openvino engine coreml tflite ...
|
||||||
|
|
||||||
Inference:
|
Inference:
|
||||||
$ python path/to/detect.py --weights yolov5s.pt # PyTorch
|
$ python path/to/detect.py --weights yolov5s.pt # PyTorch
|
||||||
|
|
@ -179,7 +179,7 @@ def export_engine(model, im, file, train, half, simplify, workspace=4, verbose=F
|
||||||
export_onnx(model, im, file, 12, train, False, simplify) # opset 12
|
export_onnx(model, im, file, 12, train, False, simplify) # opset 12
|
||||||
model.model[-1].anchor_grid = grid
|
model.model[-1].anchor_grid = grid
|
||||||
else: # TensorRT >= 8
|
else: # TensorRT >= 8
|
||||||
check_version(trt.__version__, '7.0.0', hard=True) # require tensorrt>=8.0.0
|
check_version(trt.__version__, '8.0.0', hard=True) # require tensorrt>=8.0.0
|
||||||
export_onnx(model, im, file, 13, train, False, simplify) # opset 13
|
export_onnx(model, im, file, 13, train, False, simplify) # opset 13
|
||||||
onnx = file.with_suffix('.onnx')
|
onnx = file.with_suffix('.onnx')
|
||||||
assert onnx.exists(), f'failed to export ONNX file: {onnx}'
|
assert onnx.exists(), f'failed to export ONNX file: {onnx}'
|
||||||
|
|
@ -308,7 +308,7 @@ def export_tflite(keras_model, im, file, int8, data, ncalib, prefix=colorstr('Te
|
||||||
def export_edgetpu(keras_model, im, file, prefix=colorstr('Edge TPU:')):
|
def export_edgetpu(keras_model, im, file, prefix=colorstr('Edge TPU:')):
|
||||||
# YOLOv5 Edge TPU export https://coral.ai/docs/edgetpu/models-intro/
|
# YOLOv5 Edge TPU export https://coral.ai/docs/edgetpu/models-intro/
|
||||||
try:
|
try:
|
||||||
cmd = 'edgetpu_compiler --version'
|
cmd = 'edgetpu_compiler --version' # install https://coral.ai/docs/edgetpu/compiler/
|
||||||
out = subprocess.run(cmd, shell=True, capture_output=True, check=True)
|
out = subprocess.run(cmd, shell=True, capture_output=True, check=True)
|
||||||
ver = out.stdout.decode().split()[-1]
|
ver = out.stdout.decode().split()[-1]
|
||||||
LOGGER.info(f'\n{prefix} starting export with Edge TPU compiler {ver}...')
|
LOGGER.info(f'\n{prefix} starting export with Edge TPU compiler {ver}...')
|
||||||
|
|
|
||||||
|
|
@ -376,8 +376,8 @@ class DetectMultiBackend(nn.Module):
|
||||||
elif tflite: # https://www.tensorflow.org/lite/guide/python#install_tensorflow_lite_for_python
|
elif tflite: # https://www.tensorflow.org/lite/guide/python#install_tensorflow_lite_for_python
|
||||||
if 'edgetpu' in w.lower(): # Edge TPU
|
if 'edgetpu' in w.lower(): # Edge TPU
|
||||||
LOGGER.info(f'Loading {w} for TensorFlow Lite Edge TPU inference...')
|
LOGGER.info(f'Loading {w} for TensorFlow Lite Edge TPU inference...')
|
||||||
import tflite_runtime.interpreter as tfli
|
import tflite_runtime.interpreter as tfli # install https://coral.ai/software/#edgetpu-runtime
|
||||||
delegate = {'Linux': 'libedgetpu.so.1', # install https://coral.ai/software/#edgetpu-runtime
|
delegate = {'Linux': 'libedgetpu.so.1',
|
||||||
'Darwin': 'libedgetpu.1.dylib',
|
'Darwin': 'libedgetpu.1.dylib',
|
||||||
'Windows': 'edgetpu.dll'}[platform.system()]
|
'Windows': 'edgetpu.dll'}[platform.system()]
|
||||||
interpreter = tfli.Interpreter(model_path=w, experimental_delegates=[tfli.load_delegate(delegate)])
|
interpreter = tfli.Interpreter(model_path=w, experimental_delegates=[tfli.load_delegate(delegate)])
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue