Nevar pievienot vairāk kā 25 tēmas Tēmai ir jāsākas ar burtu vai ciparu, tā var saturēt domu zīmes ('-') un var būt līdz 35 simboliem gara.

precommit: yapf (#5494) * precommit: yapf * align isort * fix # Conflicts: # utils/plots.py * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * Update setup.cfg * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * Update setup.cfg * Update setup.cfg * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * Update wandb_utils.py * Update augmentations.py * Update setup.cfg * Update yolo.py * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * Update val.py * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * simplify colorstr * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * val run fix * export.py last comma * Update export.py * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * Update hubconf.py * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * PyTorch Hub tuple fix * PyTorch Hub tuple fix2 * PyTorch Hub tuple fix3 * Update setup Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com> Co-authored-by: Glenn Jocher <glenn.jocher@ultralytics.com>
pirms 2 gadiem
precommit: yapf (#5494) * precommit: yapf * align isort * fix # Conflicts: # utils/plots.py * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * Update setup.cfg * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * Update setup.cfg * Update setup.cfg * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * Update wandb_utils.py * Update augmentations.py * Update setup.cfg * Update yolo.py * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * Update val.py * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * simplify colorstr * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * val run fix * export.py last comma * Update export.py * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * Update hubconf.py * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * PyTorch Hub tuple fix * PyTorch Hub tuple fix2 * PyTorch Hub tuple fix3 * Update setup Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com> Co-authored-by: Glenn Jocher <glenn.jocher@ultralytics.com>
pirms 2 gadiem
123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148
  1. # YOLOv5 🚀 by Ultralytics, GPL-3.0 license
  2. """
  3. Run YOLOv5 benchmarks on all supported export formats
  4. Format | `export.py --include` | Model
  5. --- | --- | ---
  6. PyTorch | - | yolov5s.pt
  7. TorchScript | `torchscript` | yolov5s.torchscript
  8. ONNX | `onnx` | yolov5s.onnx
  9. OpenVINO | `openvino` | yolov5s_openvino_model/
  10. TensorRT | `engine` | yolov5s.engine
  11. CoreML | `coreml` | yolov5s.mlmodel
  12. TensorFlow SavedModel | `saved_model` | yolov5s_saved_model/
  13. TensorFlow GraphDef | `pb` | yolov5s.pb
  14. TensorFlow Lite | `tflite` | yolov5s.tflite
  15. TensorFlow Edge TPU | `edgetpu` | yolov5s_edgetpu.tflite
  16. TensorFlow.js | `tfjs` | yolov5s_web_model/
  17. Requirements:
  18. $ pip install -r requirements.txt coremltools onnx onnx-simplifier onnxruntime openvino-dev tensorflow-cpu # CPU
  19. $ pip install -r requirements.txt coremltools onnx onnx-simplifier onnxruntime-gpu openvino-dev tensorflow # GPU
  20. $ pip install -U nvidia-tensorrt --index-url https://pypi.ngc.nvidia.com # TensorRT
  21. Usage:
  22. $ python utils/benchmarks.py --weights yolov5s.pt --img 640
  23. """
  24. import argparse
  25. import sys
  26. import time
  27. from pathlib import Path
  28. import pandas as pd
  29. FILE = Path(__file__).resolve()
  30. ROOT = FILE.parents[1] # YOLOv5 root directory
  31. if str(ROOT) not in sys.path:
  32. sys.path.append(str(ROOT)) # add ROOT to PATH
  33. # ROOT = ROOT.relative_to(Path.cwd()) # relative
  34. import export
  35. import val
  36. from utils import notebook_init
  37. from utils.general import LOGGER, check_yaml, print_args
  38. from utils.torch_utils import select_device
  39. def run(
  40. weights=ROOT / 'yolov5s.pt', # weights path
  41. imgsz=640, # inference size (pixels)
  42. batch_size=1, # batch size
  43. data=ROOT / 'data/coco128.yaml', # dataset.yaml path
  44. device='', # cuda device, i.e. 0 or 0,1,2,3 or cpu
  45. half=False, # use FP16 half-precision inference
  46. test=False, # test exports only
  47. pt_only=False, # test PyTorch only
  48. ):
  49. y, t = [], time.time()
  50. device = select_device(device)
  51. for i, (name, f, suffix, gpu) in export.export_formats().iterrows(): # index, (name, file, suffix, gpu-capable)
  52. try:
  53. assert i != 9, 'Edge TPU not supported'
  54. assert i != 10, 'TF.js not supported'
  55. if device.type != 'cpu':
  56. assert gpu, f'{name} inference not supported on GPU'
  57. # Export
  58. if f == '-':
  59. w = weights # PyTorch format
  60. else:
  61. w = export.run(weights=weights, imgsz=[imgsz], include=[f], device=device, half=half)[-1] # all others
  62. assert suffix in str(w), 'export failed'
  63. # Validate
  64. result = val.run(data, w, batch_size, imgsz, plots=False, device=device, task='benchmark', half=half)
  65. metrics = result[0] # metrics (mp, mr, map50, map, *losses(box, obj, cls))
  66. speeds = result[2] # times (preprocess, inference, postprocess)
  67. y.append([name, round(metrics[3], 4), round(speeds[1], 2)]) # mAP, t_inference
  68. except Exception as e:
  69. LOGGER.warning(f'WARNING: Benchmark failure for {name}: {e}')
  70. y.append([name, None, None]) # mAP, t_inference
  71. if pt_only and i == 0:
  72. break # break after PyTorch
  73. # Print results
  74. LOGGER.info('\n')
  75. parse_opt()
  76. notebook_init() # print system info
  77. py = pd.DataFrame(y, columns=['Format', 'mAP@0.5:0.95', 'Inference time (ms)'] if map else ['Format', 'Export', ''])
  78. LOGGER.info(f'\nBenchmarks complete ({time.time() - t:.2f}s)')
  79. LOGGER.info(str(py if map else py.iloc[:, :2]))
  80. return py
  81. def test(
  82. weights=ROOT / 'yolov5s.pt', # weights path
  83. imgsz=640, # inference size (pixels)
  84. batch_size=1, # batch size
  85. data=ROOT / 'data/coco128.yaml', # dataset.yaml path
  86. device='', # cuda device, i.e. 0 or 0,1,2,3 or cpu
  87. half=False, # use FP16 half-precision inference
  88. test=False, # test exports only
  89. pt_only=False, # test PyTorch only
  90. ):
  91. y, t = [], time.time()
  92. device = select_device(device)
  93. for i, (name, f, suffix, gpu) in export.export_formats().iterrows(): # index, (name, file, suffix, gpu-capable)
  94. try:
  95. w = weights if f == '-' else \
  96. export.run(weights=weights, imgsz=[imgsz], include=[f], device=device, half=half)[-1] # weights
  97. assert suffix in str(w), 'export failed'
  98. y.append([name, True])
  99. except Exception:
  100. y.append([name, False]) # mAP, t_inference
  101. # Print results
  102. LOGGER.info('\n')
  103. parse_opt()
  104. notebook_init() # print system info
  105. py = pd.DataFrame(y, columns=['Format', 'Export'])
  106. LOGGER.info(f'\nExports complete ({time.time() - t:.2f}s)')
  107. LOGGER.info(str(py))
  108. return py
  109. def parse_opt():
  110. parser = argparse.ArgumentParser()
  111. parser.add_argument('--weights', type=str, default=ROOT / 'yolov5s.pt', help='weights path')
  112. parser.add_argument('--imgsz', '--img', '--img-size', type=int, default=640, help='inference size (pixels)')
  113. parser.add_argument('--batch-size', type=int, default=1, help='batch size')
  114. parser.add_argument('--data', type=str, default=ROOT / 'data/coco128.yaml', help='dataset.yaml path')
  115. parser.add_argument('--device', default='', help='cuda device, i.e. 0 or 0,1,2,3 or cpu')
  116. parser.add_argument('--half', action='store_true', help='use FP16 half-precision inference')
  117. parser.add_argument('--test', action='store_true', help='test exports only')
  118. parser.add_argument('--pt-only', action='store_true', help='test PyTorch only')
  119. opt = parser.parse_args()
  120. opt.data = check_yaml(opt.data) # check YAML
  121. print_args(vars(opt))
  122. return opt
  123. def main(opt):
  124. test(**vars(opt)) if opt.test else run(**vars(opt))
  125. if __name__ == "__main__":
  126. opt = parse_opt()
  127. main(opt)