赞
踩
https://github.com/ultralytics/yolov5/tree/v5.0https://github.com/ultralytics/yolov5/tree/v5.0
master分支,tag是v5.0
按照文档跑了一个demo,用的是最简模型,效果不错:
- import cv2
- import torch
- from PIL import Image
-
- # Model
- model = torch.hub.load('ultralytics/yolov5', 'yolov5s')
-
- # Images
- # for f in 'zidane.jpg', 'bus.jpg':
- # torch.hub.download_url_to_file('https://ultralytics.com/images/' + f, f) # download 2 images
- im1 = Image.open('e:\\images\\w.jpg') # PIL image
- im2 = cv2.imread('e:\\images\\1.jpg')[..., ::-1] # OpenCV image (BGR to RGB)
-
- # Inference
- results = model([im1, im2], size=640) # batch of images
-
- # Results
- results.print()
- results.save() # or .show()
-
- results.xyxy[0] # im1 predictions (tensor)
- results.pandas().xyxy[0] # im1 predictions (pandas)
其中的一张原图w.jpg:
识别后:
制作自己的训练集
1.LabelImg的安装
在Windows10系统下使用Anaconda来安装LabelImg,步骤如下:
创建一个新环境来安装LabelImg
conda create -n labelimg
pip install labelimg
安装好后,输入LabelImg
出现这个界面
我们可以标注,并制定到处yolo格式的txt文件,save后
然后给yolo继续操作吧
进行训练train.py
目录结构
训练后会出现best.pt最好的结果模型和last.pt最后一次结果模型文件
编写自己的yaml文件指向images和labels文件夹如:
- train: e:/project/yolov5/panda/train
- val: e:/project/yolov5/panda/valid
-
- nc: 1
- names: ['panda']
改写train.py的参数
- def parse_opt(known=False):
- parser = argparse.ArgumentParser()
- parser.add_argument('--weights', type=str, default=ROOT / 'weights/yolov5s.pt', help='initial weights path')
- parser.add_argument('--cfg', type=str, default='models/yolov5s.yaml', help='model.yaml path')
- parser.add_argument('--data', type=str, default=ROOT / 'panda/data.yaml', help='dataset.yaml path')
- parser.add_argument('--hyp', type=str, default=ROOT / 'data/hyps/hyp.scratch-low.yaml', help='hyperparameters path')
- parser.add_argument('--epochs', type=int, default=300, help='total training epochs')
- parser.add_argument('--batch-size', type=int, default=2, help='total batch size for all GPUs, -1 for autobatch')
- parser.add_argument('--imgsz', '--img', '--img-size', type=int, default=640, help='train, val image size (pixels)')
- parser.add_argument('--rect', action='store_true', help='rectangular training')
- parser.add_argument('--resume', nargs='?', const=True, default=False, help='resume most recent training')
- parser.add_argument('--nosave', action='store_true', help='only save final checkpoint')
- parser.add_argument('--noval', action='store_true', help='only validate final epoch')
- parser.add_argument('--noautoanchor', action='store_true', help='disable AutoAnchor')
- parser.add_argument('--noplots', action='store_true', help='save no plot files')
- parser.add_argument('--evolve', type=int, nargs='?', const=300, help='evolve hyperparameters for x generations')
- parser.add_argument('--bucket', type=str, default='', help='gsutil bucket')
- parser.add_argument('--cache', type=str, nargs='?', const='ram', help='--cache images in "ram" (default) or "disk"')
- parser.add_argument('--image-weights', action='store_true', help='use weighted image selection for training')
- parser.add_argument('--device', default='', help='cuda device, i.e. 0 or 0,1,2,3 or cpu')
- parser.add_argument('--multi-scale', action='store_true', help='vary img-size +/- 50%%')
- parser.add_argument('--single-cls', action='store_true', help='train multi-class data as single-class')
- parser.add_argument('--optimizer', type=str, choices=['SGD', 'Adam', 'AdamW'], default='SGD', help='optimizer')
- parser.add_argument('--sync-bn', action='store_true', help='use SyncBatchNorm, only available in DDP mode')
- parser.add_argument('--workers', type=int, default=8, help='max dataloader workers (per RANK in DDP mode)')
- parser.add_argument('--project', default=ROOT / 'runs/train', help='save to project/name')
- parser.add_argument('--name', default='exp', help='save to project/name')
- parser.add_argument('--exist-ok', action='store_true', help='existing project/name ok, do not increment')
- parser.add_argument('--quad', action='store_true', help='quad dataloader')
- parser.add_argument('--cos-lr', action='store_true', help='cosine LR scheduler')
- parser.add_argument('--label-smoothing', type=float, default=0.0, help='Label smoothing epsilon')
- parser.add_argument('--patience', type=int, default=100, help='EarlyStopping patience (epochs without improvement)')
- parser.add_argument('--freeze', nargs='+', type=int, default=[0], help='Freeze layers: backbone=10, first3=0 1 2')
- parser.add_argument('--save-period', type=int, default=-1, help='Save checkpoint every x epochs (disabled if < 1)')
- parser.add_argument('--seed', type=int, default=0, help='Global training seed')
- parser.add_argument('--local_rank', type=int, default=-1, help='Automatic DDP Multi-GPU argument, do not modify')
-
- # Logger arguments
- parser.add_argument('--entity', default=None, help='Entity')
- parser.add_argument('--upload_dataset', nargs='?', const=True, default=False, help='Upload data, "val" option')
- parser.add_argument('--bbox_interval', type=int, default=-1, help='Set bounding-box image logging interval')
- parser.add_argument('--artifact_alias', type=str, default='latest', help='Version of dataset artifact to use')
-
- return parser.parse_known_args()[0] if known else parser.parse_args()
-
运行train.py
后台输出
训练完成时,保存在runs\train\exp3中
可视化页面查看
运行python命令开启可视化服务,路径写到自己的runs\train路径
tensorboard --logdir E:\project\yolov5\runs\train
访问localhost:6006
实时查看训练过程,训练好后可以进行自己的模型检测了,
detect.py修改文件参数,使用自己的训练好的模型,对test的images文件夹进行识别
- def parse_opt():
- parser = argparse.ArgumentParser()
- parser.add_argument('--weights', nargs='+', type=str, default=' E:/project/yolov5/runs/train/exp3/weights/best.pt', help='model path or triton URL')
- parser.add_argument('--source', type=str, default='E:\\project\\yolov5\\panda\\test\\images', help='file/dir/URL/glob/screen/0(webcam)')
- parser.add_argument('--data', type=str, default=ROOT / 'panda/data.yaml', help='(optional) dataset.yaml path')
- parser.add_argument('--imgsz', '--img', '--img-size', nargs='+', type=int, default=[640], help='inference size h,w')
- parser.add_argument('--conf-thres', type=float, default=0.25, help='confidence threshold')
- parser.add_argument('--iou-thres', type=float, default=0.45, help='NMS IoU threshold')
- parser.add_argument('--max-det', type=int, default=1000, help='maximum detections per image')
- parser.add_argument('--device', default='', help='cuda device, i.e. 0 or 0,1,2,3 or cpu')
- parser.add_argument('--view-img', action='store_true', help='show results')
- parser.add_argument('--save-txt', action='store_true', help='save results to *.txt')
- parser.add_argument('--save-conf', action='store_true', help='save confidences in --save-txt labels')
- parser.add_argument('--save-crop', action='store_true', help='save cropped prediction boxes')
- parser.add_argument('--nosave', action='store_true', help='do not save images/videos')
- parser.add_argument('--classes', nargs='+', type=int, help='filter by class: --classes 0, or --classes 0 2 3')
- parser.add_argument('--agnostic-nms', action='store_true', help='class-agnostic NMS')
- parser.add_argument('--augment', action='store_true', help='augmented inference')
- parser.add_argument('--visualize', action='store_true', help='visualize features')
- parser.add_argument('--update', action='store_true', help='update all models')
- parser.add_argument('--project', default=ROOT / 'runs/detect', help='save results to project/name')
- parser.add_argument('--name', default='exp', help='save results to project/name')
- parser.add_argument('--exist-ok', action='store_true', help='existing project/name ok, do not increment')
- parser.add_argument('--line-thickness', default=3, type=int, help='bounding box thickness (pixels)')
- parser.add_argument('--hide-labels', default=False, action='store_true', help='hide labels')
- parser.add_argument('--hide-conf', default=False, action='store_true', help='hide confidences')
- parser.add_argument('--half', action='store_true', help='use FP16 half-precision inference')
- parser.add_argument('--dnn', action='store_true', help='use OpenCV DNN for ONNX inference')
- parser.add_argument('--vid-stride', type=int, default=1, help='video frame-rate stride')
- opt = parser.parse_args()
- opt.imgsz *= 2 if len(opt.imgsz) == 1 else 1 # expand
- print_args(vars(opt))
- return opt
run detect.py文件会在runs\detect\exp3中保存检测后的图片
成功使用自己的模型【panda】识别出来图片,打上了标签【panda】,更多的图片和训练,识别会更精准,还需要不断优化。
Copyright © 2003-2013 www.wpsshop.cn 版权所有,并保留所有权利。