赞
踩
单目标跟踪:
直接调用opencv中封装的tracker即可。
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Sun Jan 5 17:50:47 2020
第四章 kcf跟踪
@author: youxinlin
"""
import cv2
from items import MessageItem
import time
import numpy as np
'''
监视者模块,负责入侵检测,目标跟踪
'''
class WatchDog(object):
#入侵检测者模块,用于入侵检测
def __init__(self,frame=None):
#运动检测器构造函数
self._background = None
if frame is not None:
self._background = cv2.GaussianBlur(cv2.cvtColor(frame,cv2.COLOR_BGR2GRAY),(21,21),0)
self.es = cv2.getStructuringElement(cv2.MORPH_ELLIPSE, (10, 10))
def isWorking(self):
#运动检测器是否工作
return self._background is not None
def startWorking(self,frame):
#运动检测器开始工作
if frame is not None:
self._background = cv2.GaussianBlur(cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY), (21, 21), 0)
def stopWorking(self):
#运动检测器结束工作
self._background = None
def analyze(self,frame):
#运动检测
if frame is None or self._background is None:
return
sample_frame = cv2.GaussianBlur(cv2.cvtColor(frame,cv2.COLOR_BGR2GRAY),(21,21),0)
diff = cv2.absdiff(self._background,sample_frame)
diff = cv2.threshold(diff, 25, 255, cv2.THRESH_BINARY)[1]
diff = cv2.dilate(diff, self.es, iterations=2)
image, cnts, hierarchy = cv2.findContours(diff.copy(),cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE)
coordinate = []
bigC = None
bigMulti = 0
for c in cnts:
if cv2.contourArea(c) < 1500:
continue
(x,y,w,h) = cv2.boundingRect(c)
if w * h > bigMulti:
bigMulti = w * h
bigC = ((x,y),(x+w,y+h))
if bigC:
cv2.rectangle(frame, bigC[0],bigC[1], (255,0,0), 2, 1)
coordinate.append(bigC)
message = {"coord":coordinate}
message['msg'] = None
return MessageItem(frame,message)
class Tracker(object):
'''
追踪者模块,用于追踪指定目标
'''
def __init__(self,tracker_type = "BOOSTING",draw_coord = True):
'''
初始化追踪器种类
'''
#获得opencv版本
(major_ver, minor_ver, subminor_ver) = (cv2.__version__).split('.')
self.tracker_types = ['BOOSTING', 'MIL','KCF', 'TLD', 'MEDIANFLOW', 'GOTURN']
self.tracker_type = tracker_type
self.isWorking = False
self.draw_coord = draw_coord
#构造追踪器
if int(minor_ver) < 3:
self.tracker = cv2.Tracker_create(tracker_type)
else:
if tracker_type == 'BOOSTING':
self.tracker = cv2.TrackerBoosting_create()
if tracker_type == 'MIL':
self.tracker = cv2.TrackerMIL_create()
if tracker_type == 'KCF':
self.tracker = cv2.TrackerKCF_create()
if tracker_type == 'TLD':
self.tracker = cv2.TrackerTLD_create()
if tracker_type == 'MEDIANFLOW':
self.tracker = cv2.TrackerMedianFlow_create()
if tracker_type == 'GOTURN':
self.tracker = cv2.TrackerGOTURN_create()
def initWorking(self,frame,box):
'''
追踪器工作初始化
frame:初始化追踪画面
box:追踪的区域
'''
if not self.tracker:
raise Exception("追踪器未初始化")
status = self.tracker.init(frame,box)
if not status:
raise Exception("追踪器工作初始化失败")
self.coord = box
self.isWorking = True
def track(self,frame):
'''
开启追踪
'''
message = None
if self.isWorking:
status,self.coord = self.tracker.update(frame)
if status:
message = {"coord":[((int(self.coord[0]), int(self.coord[1])),(int(self.coord[0] + self.coord[2]), int(self.coord[1] + self.coord[3])))]}
if self.draw_coord:
p1 = (int(self.coord[0]), int(self.coord[1]))
p2 = (int(self.coord[0] + self.coord[2]), int(self.coord[1] + self.coord[3]))
cv2.rectangle(frame, p1, p2, (255,0,0), 2, 1)
message['msg'] = "is tracking"
return MessageItem(frame,message)
class ObjectTracker(object):
def __init__(self,dataSet):
self.cascade = cv2.CascadeClassifier(dataSet)
def track(self,frame):
gray = cv2.cvtColor(frame,cv2.COLOR_BGR2GRAY)
faces = self.cascade.detectMultiScale(gray,1.03,5)
for (x,y,w,h) in faces:
cv2.rectangle(frame,(x,y),(x+w,y+h),(0,0,255),2)
return frame
if __name__ == '__main__' :
# tracker_types = ['BOOSTING', 'MIL','KCF', 'TLD', 'MEDIANFLOW', 'GOTURN']
tracker = Tracker(tracker_type="KCF")
# video = cv2.VideoCapture(0)
# video = cv2.VideoCapture("complex1.mov")
video = cv2.VideoCapture(r"/Users/youxinlin/Desktop/video_data/complex1.MOV")
ok, frame = video.read()
bbox = cv2.selectROI(frame, False)
tracker.initWorking(frame,bbox)
while True:
_,frame = video.read();
if(_):
item = tracker.track(frame);
cv2.imshow("track",item.getFrame())
k = cv2.waitKey(1) & 0xff
if k == 27:
break
附带items.py,放在同个文件夹下:
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Sun Jan 5 17:51:04 2020
@author: youxinlin
"""
import json
from utils import IOUtil
'''
信息封装类
'''
class MessageItem(object):
#用于封装信息的类,包含图片和其他信息
def __init__(self,frame,message):
self._frame = frame
self._message = message
def getFrame(self):
#图片信息
return self._frame
def getMessage(self):
#文字信息,json格式
return self._message
def getBase64Frame(self):
#返回base64格式的图片,将BGR图像转化为RGB图像
jepg = IOUtil.array_to_bytes(self._frame[...,::-1])
return IOUtil.bytes_to_base64(jepg)
def getBase64FrameByte(self):
#返回base64格式图片的bytes
return bytes(self.getBase64Frame())
def getJson(self):
#获得json数据格式
dicdata = {"frame":self.getBase64Frame().decode(),"message":self.getMessage()}
return json.dumps(dicdata)
def getBinaryFrame(self):
return IOUtil.array_to_bytes(self._frame[...,::-1])
utils.py:也放在同一个文件夹下。
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Sun Jan 5 17:51:40 2020
@author: youxinlin
"""
import time
import numpy
import base64
import os
import logging
import sys
from PIL import Image
from io import BytesIO
#工具类
class IOUtil(object):
#流操作工具类
@staticmethod
def array_to_bytes(pic,formatter="jpeg",quality=70):
'''
静态方法,将numpy数组转化二进制流
:param pic: numpy数组
:param format: 图片格式
:param quality:压缩比,压缩比越高,产生的二进制数据越短
:return:
'''
stream = BytesIO()
picture = Image.fromarray(pic)
picture.save(stream,format=formatter,quality=quality)
jepg = stream.getvalue()
stream.close()
return jepg
@staticmethod
def bytes_to_base64(byte):
'''
静态方法,bytes转base64编码
:param byte:
:return:
'''
return base64.b64encode(byte)
@staticmethod
def transport_rgb(frame):
'''
将bgr图像转化为rgb图像,或者将rgb图像转化为bgr图像
'''
return frame[...,::-1]
@staticmethod
def byte_to_package(bytes,cmd,var=1):
'''
将每一帧的图片流的二进制数据进行分包
:param byte: 二进制文件
:param cmd:命令
:return:
'''
head = [ver,len(byte),cmd]
headPack = struct.pack("!3I", *head)
senddata = headPack+byte
return senddata
@staticmethod
def mkdir(filePath):
'''
创建文件夹
'''
if not os.path.exists(filePath):
os.mkdir(filePath)
@staticmethod
def countCenter(box):
'''
计算一个矩形的中心
'''
return (int(abs(box[0][0] - box[1][0])*0.5) + box[0][0],int(abs(box[0][1] - box[1][1])*0.5) +box[0][1])
@staticmethod
def countBox(center):
'''
根据两个点计算出,x,y,c,r
'''
return (center[0][0],center[0][1],center[1][0]-center[0][0],center[1][1]-center[0][1])
@staticmethod
def getImageFileName():
return time.strftime("%Y_%m_%d_%H_%M_%S", time.localtime())+'.png'
多目标跟踪:
和单目标差不多,改用MultiTracker_create()
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Sun Jan 5 18:02:33 2020
多目标跟踪
@author: youxinlin
"""import numpy as np
import cv2
import sys
'''
if len(sys.argv) != 2:
print('Input video name is missing')
exit()
'''
print('Select multiple tracking targets')
cv2.namedWindow("tracking")
camera = cv2.VideoCapture(r"/Users/youxinlin/Desktop/video_data/complex6.MOV")
#camera = cv2.VideoCapture(0)
tracker = cv2.MultiTracker_create() #多目标跟踪
a= cv2.Tracker_c
init_once = False
ok, image=camera.read()
if not ok:
print('Failed to read video')
exit()
bbox1 = cv2.selectROI('tracking', image)
bbox2 = cv2.selectROI('tracking', image)
bbox3 = cv2.selectROI('tracking', image)
while camera.isOpened():
ok, image=camera.read()
if not ok:
print ('no image to read')
break
if not init_once:
ok = tracker.add(cv2.TrackerKCF_create(),image,bbox1)
ok = tracker.add(cv2.TrackerKCF_create( ),image, bbox2)
ok = tracker.add(cv2.TrackerKCF_create(),image, bbox3)
init_once = True
ok, boxes = tracker.update(image)
for newbox in boxes:
p1 = (int(newbox[0]), int(newbox[1]))
p2 = (int(newbox[0] + newbox[2]), int(newbox[1] + newbox[3]))
cv2.rectangle(image, p1, p2, (0,0,255))
cv2.imshow('tracking', image)
k = cv2.waitKey(1)
if k == 27 : break # esc pressed
多尺度检测的KCF、自定义所用特征的KCF
在一些场景下,不想使用默认的hog特征跟踪,或需要对比不同特征的跟踪效果,那么封装好的方法似乎不可用,需要可以自己撸一波kcf的代码,从而使用自己设定的特征。
总结
以上所述是小编给大家介绍的python实现单目标、多目标、多尺度、自定义特征的KCF跟踪算法,希望对大家有所帮助,如果大家有任何疑问请给我留言,小编会及时回复大家的。在此也非常感谢大家对脚本之家网站的支持!如果你觉得本文对你有帮助,欢迎转载,烦请注明出处,谢谢!
Copyright © 2003-2013 www.wpsshop.cn 版权所有,并保留所有权利。