246 lines
7.0 KiB
Python
246 lines
7.0 KiB
Python
import ffmpeg
|
||
import numpy as np
|
||
import matplotlib
|
||
import cv2
|
||
import os
|
||
import sys
|
||
from matplotlib import pyplot as plt
|
||
|
||
|
||
'''
|
||
TODO:
|
||
0:读取视频 √
|
||
1:获取视差 √
|
||
2:获取运动矢量 √
|
||
3:确定舒适度
|
||
4:加舒适度水印
|
||
...
|
||
'''
|
||
|
||
# 打开视频文件
|
||
|
||
|
||
def openVid():
|
||
fileName = input("video path:")
|
||
while not os.path.isfile(fileName):
|
||
print("file doesn't exist!")
|
||
fileName = input("video path:")
|
||
cap = cv2.VideoCapture(fileName)
|
||
if cap.isOpened():
|
||
return cap
|
||
else:
|
||
print("cannot open video.")
|
||
sys.exit()
|
||
|
||
|
||
# 获取视频总帧数
|
||
def getFrameCount(cap):
|
||
if cap.isOpened():
|
||
return cap.get(cv2.CAP_PROP_FRAME_COUNT)
|
||
else:
|
||
print("cannot open video.")
|
||
sys.exit()
|
||
|
||
# 获取帧速率
|
||
|
||
|
||
def getFrameRate(cap):
|
||
if cap.isOpened():
|
||
return cap.get(cv2.CAP_PROP_FPS)
|
||
else:
|
||
print("cannot open video.")
|
||
sys.exit()
|
||
|
||
# 给出左右画面,计算景深
|
||
|
||
|
||
def getDepthMap(imgL, imgR):
|
||
stereo = cv2.StereoSGBM_create(numDisparities=64, blockSize=3)
|
||
return stereo.compute(imgL, imgR)
|
||
|
||
|
||
# 给出前后两帧,计算帧间运动矢量
|
||
def getMotionVector(prvs, next):
|
||
hsv = np.zeros_like(imgR) # 将运动矢量按hsv显示,以色调h表示运动方向,以明度v表示运动位移
|
||
hsv[..., 1] = 255 # 饱和度置为最高
|
||
|
||
# 转为灰度以计算光流
|
||
prvs = cv2.cvtColor(prvs, cv2.COLOR_BGR2GRAY)
|
||
next = cv2.cvtColor(next, cv2.COLOR_BGR2GRAY)
|
||
|
||
flow = cv2.calcOpticalFlowFarneback(
|
||
prvs, next, None, 0.5, 3, 15, 3, 5, 1.2, 0) # 计算两帧间的光流,即运动矢量的直角坐标表示
|
||
mag, ang = cv2.cartToPolar(
|
||
flow[..., 0], flow[..., 1]) # 运动矢量的直角坐标表示转换为极坐标表示
|
||
hsv[..., 0] = ang*180/np.pi/2 # 角度对应色调
|
||
hsv[..., 2] = cv2.normalize(mag, None, 0, 255, cv2.NORM_MINMAX) # 位移量对应明度
|
||
return hsv
|
||
|
||
|
||
if __name__ == "__main__":
|
||
cap = openVid()
|
||
isDemo = int(input("is Demo(0/1)?"))
|
||
frameRate = getFrameRate(cap)
|
||
frameCount = getFrameCount(cap)
|
||
|
||
isSuccess, img = cap.read()
|
||
if not isSuccess:
|
||
print("video read error.")
|
||
sys.exit()
|
||
|
||
# 分割左右画面
|
||
imgL = np.split(img, 2, 1)[0]
|
||
imgR = np.split(img, 2, 1)[1]
|
||
prvs = imgR # 上一帧的右画面,用于运动矢量计算
|
||
|
||
# 每秒取10帧进行计算
|
||
for frameID in range(round(cap.get(cv2.CAP_PROP_POS_FRAMES)), round(frameCount), round(frameRate/10)):
|
||
cap.set(cv2.CAP_PROP_POS_FRAMES, frameID)
|
||
isSuccess, img = cap.read()
|
||
if not isSuccess:
|
||
print("video read error.")
|
||
sys.exit()
|
||
|
||
# 分割左右画面
|
||
imgL = np.split(img, 2, 1)[0]
|
||
imgR = np.split(img, 2, 1)[1]
|
||
|
||
next = imgR # 当前帧的右画面,用于运动矢量计算
|
||
hsv = getMotionVector(prvs, next)
|
||
|
||
# 计算深度图
|
||
disparity = getDepthMap(imgL, imgR)
|
||
|
||
# 显示计算结果
|
||
print("time: ", round(frameID/frameRate, 2))
|
||
print("AVG depth: ", round(np.mean(disparity), 2))
|
||
print("AVG motion: ", round(np.mean(hsv[..., 2]), 2))
|
||
print()
|
||
|
||
# 当为demo模式时显示当前帧画面、运动矢量图和景深图
|
||
if isDemo:
|
||
# 显示当前帧
|
||
cv2.namedWindow("img", cv2.WINDOW_NORMAL)
|
||
cv2.imshow('img', img)
|
||
|
||
# 显示当前帧的运动矢量的hsv表示
|
||
bgr = cv2.cvtColor(hsv, cv2.COLOR_HSV2BGR) # hsv转为rgb用于显示
|
||
cv2.namedWindow("MotionVector", cv2.WINDOW_NORMAL)
|
||
cv2.imshow("MotionVector", bgr)
|
||
|
||
# 显示当前帧的景深图
|
||
plt.title("DepthMap")
|
||
plt.imshow(disparity)
|
||
plt.pause(0.2)
|
||
|
||
prvs = next # 当前帧覆盖上一帧,继续计算
|
||
print("success")
|
||
|
||
|
||
# ffmpeg.input("./vid/avatar.mkv")
|
||
|
||
# Motion Vector
|
||
#cap = cv2.VideoCapture('./vid/zootopia.mkv')
|
||
# for i in range(1,10000):
|
||
# cap.read()
|
||
# params for ShiTomasi corner detection
|
||
# feature_params = dict( maxCorners = 100,
|
||
# qualityLevel = 0.3,
|
||
# minDistance = 7,
|
||
# blockSize = 7 )
|
||
# Parameters for lucas kanade optical flow
|
||
# lk_params = dict( winSize = (15,15),
|
||
# maxLevel = 2,
|
||
# criteria = (cv2.TERM_CRITERIA_EPS | cv2.TERM_CRITERIA_COUNT, 10, 0.03))
|
||
# Create some random colors
|
||
#color = np.random.randint(0,255,(100,3))
|
||
# Take first frame and find corners in it
|
||
#ret, old_frame = cap.read()
|
||
#old_gray = cv2.cvtColor(old_frame, cv2.COLOR_BGR2GRAY)
|
||
#p0 = cv2.goodFeaturesToTrack(old_gray, mask = None, **feature_params)
|
||
# Create a mask image for drawing purposes
|
||
#mask = np.zeros_like(old_frame)
|
||
# while(1):
|
||
# ret,frame = cap.read()
|
||
# frame_gray = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY)
|
||
# # calculate optical flow
|
||
# p1, st, err = cv2.calcOpticalFlowPyrLK(old_gray, frame_gray, p0, None, **lk_params)
|
||
# # Select good points
|
||
# good_new = p1[st==1]
|
||
# good_old = p0[st==1]
|
||
# # draw the tracks
|
||
# for i,(new,old) in enumerate(zip(good_new,good_old)):
|
||
# a,b = new.ravel()
|
||
# c,d = old.ravel()
|
||
# mask = cv2.line(mask, (a,b),(c,d), color[i].tolist(), 2)
|
||
# frame = cv2.circle(frame,(a,b),5,color[i].tolist(),-1)
|
||
# img = cv2.add(frame,mask)
|
||
# cv2.imshow('frame',img)
|
||
# k = cv2.waitKey(30) & 0xff
|
||
# if k == 27:
|
||
# break
|
||
# # Now update the previous frame and previous points
|
||
# old_gray = frame_gray.copy()
|
||
# p0 = good_new.reshape(-1,1,2)
|
||
# cv2.destroyAllWindows()
|
||
# cap.release()
|
||
|
||
|
||
# 中文文件名无法识别
|
||
# imgDirs = os.listdir("./pic_en")
|
||
|
||
|
||
# def read_frame_as_jpeg(in_filename, frame_num):
|
||
# out, err = (
|
||
# ffmpeg
|
||
# .input(in_filename)
|
||
# .filter('select', 'gte(n,{})'.format(frame_num))
|
||
# .output('pipe:', vframes=1, format='image2', vcodec='mjpeg')
|
||
# .run(capture_stdout=True)
|
||
# )
|
||
# return out
|
||
|
||
# ffmpeg.input("./vid/venom.mkv")
|
||
# ffmpeg.
|
||
|
||
#img = read_frame_as_jpeg("./vid/venom.mkv", 648)
|
||
# print(type(img))
|
||
#img = cv2.imdecode(img,0)
|
||
# print(img)
|
||
# cv2.imshow("img",img)
|
||
#imgL = np.split(img, 2, 1)[0]
|
||
#imgR = np.split(img, 2, 1)[1]
|
||
#stereo = cv2.StereoBM_create(numDisparities=64, blockSize=11)
|
||
#disparity = stereo.compute(imgL, imgR)
|
||
# plt.imshow(disparity)
|
||
# plt.show()
|
||
|
||
|
||
# for imgDir in imgDirs:
|
||
# dir = "./pic_en/"+imgDir
|
||
# print(dir)
|
||
# img = cv2.imread(dir)
|
||
|
||
# imgL = np.split(img, 2, 1)[0]
|
||
# imgR = np.split(img, 2, 1)[1]
|
||
# print(img.shape)
|
||
# print(imgL.shape)
|
||
# print(imgR.shape)
|
||
# cv2.imshow("img", img)
|
||
# stereo = cv2.StereoSGBM_create(numDisparities=96, blockSize=11)
|
||
# disparity = stereo.compute(imgL, imgR)
|
||
|
||
# plt.imshow(disparity)
|
||
# plt.show()
|
||
# import numpy as np
|
||
# import cv2
|
||
# from matplotlib import pyplot as plt
|
||
#
|
||
# imgL = cv2.imread('tsukuba_l.png',0)
|
||
# imgR = cv2.imread('tsukuba_r.png',0)
|
||
#
|
||
# stereo = cv2.StereoBM_create(numDisparities=16, blockSize=15)
|
||
# disparity = stereo.compute(imgL,imgR)
|
||
# plt.imshow(disparity,'gray')
|
||
# plt.show()
|