from models.experimental import attempt_load import tensorrt as trt import torch import sys from segutils.trtUtils import segPreProcess_image,segTrtForward,segPreProcess_image_torch from segutils.model_stages import BiSeNet_STDC import time,cv2 import numpy as np class stdcModel(object): def __init__(self, weights=None, par={'modelSize':(640,360),'dynamic':False,'nclass':2,'predResize':True,'mean':(0.485, 0.456, 0.406),'std' :(0.229, 0.224, 0.225),'numpy':False, 'RGB_convert_first':True} ): self.par = par self.device = 'cuda:0' self.half =True if 'dynamic' not in par.keys(): self.dynamic=False else: self.dynamic=par['dynamic'] if weights.endswith('.engine'): self. infer_type ='trt' elif weights.endswith('.pth') or weights.endswith('.pt') : self. infer_type ='pth' else: print('#########ERROR:',weights,': no registered inference type, exit') sys.exit(0) if self.infer_type=='trt': if self.dynamic : print('####################ERROR##########,STDC动态模型不能采用trt格式########') logger = trt.Logger(trt.Logger.ERROR) with open(weights, "rb") as f, trt.Runtime(logger) as runtime: self.model=runtime.deserialize_cuda_engine(f.read())# 输入trt本地文件,返回ICudaEngine对象 elif self.infer_type=='pth': if self.dynamic: modelSize=None else: modelSize=( self.par['modelSize'][1], self.par['modelSize'][0] ) self.model = BiSeNet_STDC(backbone='STDCNet813', n_classes=par['seg_nclass'], use_boundary_2=False, use_boundary_4=False, use_boundary_8=True, use_boundary_16=False, use_conv_last=False, modelSize = modelSize ) self.model.load_state_dict(torch.load(weights, map_location=torch.device(self.device) )) self.model= self.model.to(self.device) print('#########加载模型:',weights,' 类型:',self.infer_type) def preprocess_image(self,image): image = self.RB_convert(image) if self.dynamic: H,W=image.shape[0:2]; yscale = self.par['modelSize'][1]/H xscale = self.par['modelSize'][0]/W dscale = min(yscale,xscale) re_size = ( int((dscale*W)//4*4), int( (dscale*H)//4*4 ) ) else: re_size = self.par['modelSize'] #print('####line 58:,', re_size,image.shape) image = cv2.resize(image,re_size, interpolation=cv2.INTER_LINEAR) image = image.astype(np.float32) image /= 255.0 image[:, :, 0] -= self.par['mean'][0] image[:, :, 1] -= self.par['mean'][1] image[:, :, 2] -= self.par['mean'][2] image[:, :, 0] /= self.par['std'][0] image[:, :, 1] /= self.par['std'][1] image[:, :, 2] /= self.par['std'][2] image = np.transpose(image, (2, 0, 1)) image = torch.from_numpy(image).float() image = image.unsqueeze(0) if self.device != 'cpu': image = image.to(self.device) return image def RB_convert(self,image): image_c = image.copy() image_c[:,:,0] = image[:,:,2] image_c[:,:,2] = image[:,:,0] return image_c def get_ms(self,t1,t0): return (t1-t0)*1000.0 def eval(self,image): time0 = time.time() imageH, imageW, _ = image.shape img = self.preprocess_image(image) time1 = time.time() if self.infer_type=='trt': pred=segTrtForward(self.model,[img]) elif self.infer_type=='pth': self.model.eval() with torch.no_grad(): pred = self.model(img) time2 = time.time() pred=torch.argmax(pred,dim=1).cpu().numpy()[0] time3 = time.time() pred = cv2.resize(pred.astype(np.uint8),(imageW,imageH)) time4 = time.time() outstr= 'pre-precess:%.1f ,infer:%.1f ,post-cpu-argmax:%.1f ,post-resize:%.1f, total:%.1f \n '%( self.get_ms(time1,time0),self.get_ms(time2,time1),self.get_ms(time3,time2),self.get_ms(time4,time3),self.get_ms(time4,time0) ) return pred,outstr