当前位置:   article > 正文

yolov5机器学习,训练自己的数据集_video frame-rate stride

video frame-rate stride

https://github.com/ultralytics/yolov5/tree/v5.0https://github.com/ultralytics/yolov5/tree/v5.0

 master分支,tag是v5.0

按照文档跑了一个demo,用的是最简模型,效果不错:

  1. import cv2
  2. import torch
  3. from PIL import Image
  4. # Model
  5. model = torch.hub.load('ultralytics/yolov5', 'yolov5s')
  6. # Images
  7. # for f in 'zidane.jpg', 'bus.jpg':
  8. # torch.hub.download_url_to_file('https://ultralytics.com/images/' + f, f) # download 2 images
  9. im1 = Image.open('e:\\images\\w.jpg') # PIL image
  10. im2 = cv2.imread('e:\\images\\1.jpg')[..., ::-1] # OpenCV image (BGR to RGB)
  11. # Inference
  12. results = model([im1, im2], size=640) # batch of images
  13. # Results
  14. results.print()
  15. results.save() # or .show()
  16. results.xyxy[0] # im1 predictions (tensor)
  17. results.pandas().xyxy[0] # im1 predictions (pandas)

其中的一张原图w.jpg:

 识别后:

制作自己的训练集

1.LabelImg的安装

在Windows10系统下使用Anaconda来安装LabelImg,步骤如下:

创建一个新环境来安装LabelImg

conda create -n labelimg 
pip install labelimg

 安装好后,输入LabelImg

出现这个界面

 

 

我们可以标注,并制定到处yolo格式的txt文件,save后

 

 然后给yolo继续操作吧

进行训练train.py

目录结构

 训练后会出现best.pt最好的结果模型和last.pt最后一次结果模型文件

编写自己的yaml文件指向images和labels文件夹如:

  1. train: e:/project/yolov5/panda/train
  2. val: e:/project/yolov5/panda/valid
  3. nc: 1
  4. names: ['panda']

改写train.py的参数

  1. def parse_opt(known=False):
  2. parser = argparse.ArgumentParser()
  3. parser.add_argument('--weights', type=str, default=ROOT / 'weights/yolov5s.pt', help='initial weights path')
  4. parser.add_argument('--cfg', type=str, default='models/yolov5s.yaml', help='model.yaml path')
  5. parser.add_argument('--data', type=str, default=ROOT / 'panda/data.yaml', help='dataset.yaml path')
  6. parser.add_argument('--hyp', type=str, default=ROOT / 'data/hyps/hyp.scratch-low.yaml', help='hyperparameters path')
  7. parser.add_argument('--epochs', type=int, default=300, help='total training epochs')
  8. parser.add_argument('--batch-size', type=int, default=2, help='total batch size for all GPUs, -1 for autobatch')
  9. parser.add_argument('--imgsz', '--img', '--img-size', type=int, default=640, help='train, val image size (pixels)')
  10. parser.add_argument('--rect', action='store_true', help='rectangular training')
  11. parser.add_argument('--resume', nargs='?', const=True, default=False, help='resume most recent training')
  12. parser.add_argument('--nosave', action='store_true', help='only save final checkpoint')
  13. parser.add_argument('--noval', action='store_true', help='only validate final epoch')
  14. parser.add_argument('--noautoanchor', action='store_true', help='disable AutoAnchor')
  15. parser.add_argument('--noplots', action='store_true', help='save no plot files')
  16. parser.add_argument('--evolve', type=int, nargs='?', const=300, help='evolve hyperparameters for x generations')
  17. parser.add_argument('--bucket', type=str, default='', help='gsutil bucket')
  18. parser.add_argument('--cache', type=str, nargs='?', const='ram', help='--cache images in "ram" (default) or "disk"')
  19. parser.add_argument('--image-weights', action='store_true', help='use weighted image selection for training')
  20. parser.add_argument('--device', default='', help='cuda device, i.e. 0 or 0,1,2,3 or cpu')
  21. parser.add_argument('--multi-scale', action='store_true', help='vary img-size +/- 50%%')
  22. parser.add_argument('--single-cls', action='store_true', help='train multi-class data as single-class')
  23. parser.add_argument('--optimizer', type=str, choices=['SGD', 'Adam', 'AdamW'], default='SGD', help='optimizer')
  24. parser.add_argument('--sync-bn', action='store_true', help='use SyncBatchNorm, only available in DDP mode')
  25. parser.add_argument('--workers', type=int, default=8, help='max dataloader workers (per RANK in DDP mode)')
  26. parser.add_argument('--project', default=ROOT / 'runs/train', help='save to project/name')
  27. parser.add_argument('--name', default='exp', help='save to project/name')
  28. parser.add_argument('--exist-ok', action='store_true', help='existing project/name ok, do not increment')
  29. parser.add_argument('--quad', action='store_true', help='quad dataloader')
  30. parser.add_argument('--cos-lr', action='store_true', help='cosine LR scheduler')
  31. parser.add_argument('--label-smoothing', type=float, default=0.0, help='Label smoothing epsilon')
  32. parser.add_argument('--patience', type=int, default=100, help='EarlyStopping patience (epochs without improvement)')
  33. parser.add_argument('--freeze', nargs='+', type=int, default=[0], help='Freeze layers: backbone=10, first3=0 1 2')
  34. parser.add_argument('--save-period', type=int, default=-1, help='Save checkpoint every x epochs (disabled if < 1)')
  35. parser.add_argument('--seed', type=int, default=0, help='Global training seed')
  36. parser.add_argument('--local_rank', type=int, default=-1, help='Automatic DDP Multi-GPU argument, do not modify')
  37. # Logger arguments
  38. parser.add_argument('--entity', default=None, help='Entity')
  39. parser.add_argument('--upload_dataset', nargs='?', const=True, default=False, help='Upload data, "val" option')
  40. parser.add_argument('--bbox_interval', type=int, default=-1, help='Set bounding-box image logging interval')
  41. parser.add_argument('--artifact_alias', type=str, default='latest', help='Version of dataset artifact to use')
  42. return parser.parse_known_args()[0] if known else parser.parse_args()

运行train.py

后台输出

 

训练完成时,保存在runs\train\exp3中

可视化页面查看

运行python命令开启可视化服务,路径写到自己的runs\train路径

tensorboard --logdir E:\project\yolov5\runs\train

访问localhost:6006

 实时查看训练过程,训练好后可以进行自己的模型检测了,

detect.py修改文件参数,使用自己的训练好的模型,对test的images文件夹进行识别
  1. def parse_opt():
  2. parser = argparse.ArgumentParser()
  3. parser.add_argument('--weights', nargs='+', type=str, default=' E:/project/yolov5/runs/train/exp3/weights/best.pt', help='model path or triton URL')
  4. parser.add_argument('--source', type=str, default='E:\\project\\yolov5\\panda\\test\\images', help='file/dir/URL/glob/screen/0(webcam)')
  5. parser.add_argument('--data', type=str, default=ROOT / 'panda/data.yaml', help='(optional) dataset.yaml path')
  6. parser.add_argument('--imgsz', '--img', '--img-size', nargs='+', type=int, default=[640], help='inference size h,w')
  7. parser.add_argument('--conf-thres', type=float, default=0.25, help='confidence threshold')
  8. parser.add_argument('--iou-thres', type=float, default=0.45, help='NMS IoU threshold')
  9. parser.add_argument('--max-det', type=int, default=1000, help='maximum detections per image')
  10. parser.add_argument('--device', default='', help='cuda device, i.e. 0 or 0,1,2,3 or cpu')
  11. parser.add_argument('--view-img', action='store_true', help='show results')
  12. parser.add_argument('--save-txt', action='store_true', help='save results to *.txt')
  13. parser.add_argument('--save-conf', action='store_true', help='save confidences in --save-txt labels')
  14. parser.add_argument('--save-crop', action='store_true', help='save cropped prediction boxes')
  15. parser.add_argument('--nosave', action='store_true', help='do not save images/videos')
  16. parser.add_argument('--classes', nargs='+', type=int, help='filter by class: --classes 0, or --classes 0 2 3')
  17. parser.add_argument('--agnostic-nms', action='store_true', help='class-agnostic NMS')
  18. parser.add_argument('--augment', action='store_true', help='augmented inference')
  19. parser.add_argument('--visualize', action='store_true', help='visualize features')
  20. parser.add_argument('--update', action='store_true', help='update all models')
  21. parser.add_argument('--project', default=ROOT / 'runs/detect', help='save results to project/name')
  22. parser.add_argument('--name', default='exp', help='save results to project/name')
  23. parser.add_argument('--exist-ok', action='store_true', help='existing project/name ok, do not increment')
  24. parser.add_argument('--line-thickness', default=3, type=int, help='bounding box thickness (pixels)')
  25. parser.add_argument('--hide-labels', default=False, action='store_true', help='hide labels')
  26. parser.add_argument('--hide-conf', default=False, action='store_true', help='hide confidences')
  27. parser.add_argument('--half', action='store_true', help='use FP16 half-precision inference')
  28. parser.add_argument('--dnn', action='store_true', help='use OpenCV DNN for ONNX inference')
  29. parser.add_argument('--vid-stride', type=int, default=1, help='video frame-rate stride')
  30. opt = parser.parse_args()
  31. opt.imgsz *= 2 if len(opt.imgsz) == 1 else 1 # expand
  32. print_args(vars(opt))
  33. return opt

run detect.py文件会在runs\detect\exp3中保存检测后的图片

 成功使用自己的模型【panda】识别出来图片,打上了标签【panda】,更多的图片和训练,识别会更精准,还需要不断优化。

声明:本文内容由网友自发贡献,不代表【wpsshop博客】立场,版权归原作者所有,本站不承担相应法律责任。如您发现有侵权的内容,请联系我们。转载请注明出处:https://www.wpsshop.cn/w/木道寻08/article/detail/806882
推荐阅读
相关标签
  

闽ICP备14008679号