HivisionIDPhotos / src /cuny_tools.py
TheEeeeLin's picture
update files
d5d20be verified
raw
history blame
28.2 kB
import cv2
import numpy as np
from hivisionai.hycv.utils import get_box_pro
from hivisionai.hycv.vision import cover_image, draw_picture_dots
from math import fabs, sin, radians, cos
def opencv_rotate(img, angle):
h, w = img.shape[:2]
center = (w / 2, h / 2)
scale = 1.0
# 2.1获取M矩阵
"""
M矩阵
[
cosA -sinA (1-cosA)*centerX+sinA*centerY
sinA cosA -sinA*centerX+(1-cosA)*centerY
]
"""
M = cv2.getRotationMatrix2D(center, angle, scale)
# 2.2 新的宽高,radians(angle) 把角度转为弧度 sin(弧度)
new_H = int(w * fabs(sin(radians(angle))) + h * fabs(cos(radians(angle))))
new_W = int(h * fabs(sin(radians(angle))) + w * fabs(cos(radians(angle))))
# 2.3 平移
M[0, 2] += (new_W - w) / 2
M[1, 2] += (new_H - h) / 2
rotate = cv2.warpAffine(img, M, (new_W, new_H), borderValue=(0, 0, 0))
return rotate
def transformationNeck2(image:np.ndarray, per_to_side:float=0.8)->np.ndarray:
"""
透视变换脖子函数,输入图像和四个点(矩形框)
矩形框内的图像可能是不完整的(边角有透明区域)
我们将根据透视变换将矩形框内的图像拉伸成和矩形框一样的形状.
算法分为几个步骤: 选择脖子的四个点 -> 选定这四个点拉伸后的坐标 -> 透视变换 -> 覆盖原图
"""
_, _, _, a = cv2.split(image) # 这应该是一个四通道的图像
height, width = a.shape
def locate_side(image_:np.ndarray, x_:int, y_max:int) -> int:
# 寻找x=y, 且 y <= y_max 上从下往上第一个非0的点,如果没找到就返回0
y_ = 0
for y_ in range(y_max - 1, -1, -1):
if image_[y_][x_] != 0:
break
return y_
def locate_width(image_:np.ndarray, y_:int, mode, left_or_right:int=None):
# 从y=y这个水平线上寻找两边的非零点
# 增加left_or_right的原因在于为下面check_jaw服务
if mode==1: # 左往右
x_ = 0
if left_or_right is None:
left_or_right = 0
for x_ in range(left_or_right, width):
if image_[y_][x_] != 0:
break
else: # 右往左
x_ = width
if left_or_right is None:
left_or_right = width - 1
for x_ in range(left_or_right, -1, -1):
if image_[y_][x_] != 0:
break
return x_
def check_jaw(image_:np.ndarray, left_, right_):
"""
检查选择的点是否与截到下巴,如果截到了,就往下平移一个单位
"""
f= True # True代表没截到下巴
# [x, y]
for x_cell in range(left_[0] + 1, right_[0]):
if image_[left_[1]][x_cell] == 0:
f = False
break
if f is True:
return left_, right_
else:
y_ = left_[1] + 2
x_left_ = locate_width(image_, y_, mode=1, left_or_right=left_[0])
x_right_ = locate_width(image_, y_, mode=2, left_or_right=right_[0])
left_, right_ = check_jaw(image_, [x_left_, y_], [x_right_, y_])
return left_, right_
# 选择脖子的四个点,核心在于选择上面的两个点,这两个点的确定的位置应该是"宽出来的"两个点
_, _ ,_, a = cv2.split(image) # 这应该是一个四通道的图像
ret,a_thresh = cv2.threshold(a,127,255,cv2.THRESH_BINARY)
y_high, y_low, x_left, x_right = get_box_pro(image=image, model=1) # 直接返回矩阵信息
y_left_side = locate_side(image_=a_thresh, x_=x_left, y_max=y_low) # 左边的点的y轴坐标
y_right_side = locate_side(image_=a_thresh, x_=x_right, y_max=y_low) # 右边的点的y轴坐标
y = min(y_left_side, y_right_side) # 将两点的坐标保持相同
cell_left_above, cell_right_above = check_jaw(a_thresh,[x_left, y], [x_right, y])
x_left, x_right = cell_left_above[0], cell_right_above[0]
# 此时我们寻找到了脖子的"宽出来的"两个点,这两个点作为上面的两个点, 接下来寻找下面的两个点
if per_to_side >1:
assert ValueError("per_to_side 必须小于1!")
# 在后面的透视变换中我会把它拉成矩形, 在这里我先获取四个点的高和宽
height_ = 150 # 这个值应该是个变化的值,与拉伸的长度有关,但是现在先规定为150
width_ = x_right - x_left # 其实也就是 cell_right_above[1] - cell_left_above[1]
y = int((y_low - y)*per_to_side + y) # 定位y轴坐标
cell_left_below, cell_right_bellow = ([locate_width(a_thresh, y_=y, mode=1), y], [locate_width(a_thresh, y_=y, mode=2), y])
# 四个点全齐,开始透视变换
# 寻找透视变换后的四个点,只需要变换below的两个点即可
# cell_left_below_final, cell_right_bellow_final = ([cell_left_above[1], y_low], [cell_right_above[1], y_low])
# 需要变换的四个点为 cell_left_above, cell_right_above, cell_left_below, cell_right_bellow
rect = np.array([cell_left_above, cell_right_above, cell_left_below, cell_right_bellow],
dtype='float32')
# 变化后的坐标点
dst = np.array([[0, 0], [width_, 0], [0 , height_], [width_, height_]],
dtype='float32')
# 计算变换矩阵
M = cv2.getPerspectiveTransform(rect, dst)
warped = cv2.warpPerspective(image, M, (width_, height_))
final = cover_image(image=warped, background=image, mode=3, x=cell_left_above[0], y=cell_left_above[1])
# tmp = np.zeros(image.shape)
# final = cover_image(image=warped, background=tmp, mode=3, x=cell_left_above[0], y=cell_left_above[1])
# final = cover_image(image=image, background=final, mode=3, x=0, y=0)
return final
def transformationNeck(image:np.ndarray, cutNeckHeight:int, neckBelow:int,
toHeight:int,per_to_side:float=0.75) -> np.ndarray:
"""
脖子扩充算法, 其实需要输入的只是脖子扣出来的部分以及需要被扩充的高度/需要被扩充成的高度.
"""
height, width, channels = image.shape
_, _, _, a = cv2.split(image) # 这应该是一个四通道的图像
ret, a_thresh = cv2.threshold(a, 127, 255, cv2.THRESH_BINARY) # 将透明图层二值化
def locate_width(image_:np.ndarray, y_:int, mode, left_or_right:int=None):
# 从y=y这个水平线上寻找两边的非零点
# 增加left_or_right的原因在于为下面check_jaw服务
if mode==1: # 左往右
x_ = 0
if left_or_right is None:
left_or_right = 0
for x_ in range(left_or_right, width):
if image_[y_][x_] != 0:
break
else: # 右往左
x_ = width
if left_or_right is None:
left_or_right = width - 1
for x_ in range(left_or_right, -1, -1):
if image_[y_][x_] != 0:
break
return x_
def check_jaw(image_:np.ndarray, left_, right_):
"""
检查选择的点是否与截到下巴,如果截到了,就往下平移一个单位
"""
f= True # True代表没截到下巴
# [x, y]
for x_cell in range(left_[0] + 1, right_[0]):
if image_[left_[1]][x_cell] == 0:
f = False
break
if f is True:
return left_, right_
else:
y_ = left_[1] + 2
x_left_ = locate_width(image_, y_, mode=1, left_or_right=left_[0])
x_right_ = locate_width(image_, y_, mode=2, left_or_right=right_[0])
left_, right_ = check_jaw(image_, [x_left_, y_], [x_right_, y_])
return left_, right_
x_left = locate_width(image_=a_thresh, mode=1, y_=cutNeckHeight)
x_right = locate_width(image_=a_thresh, mode=2, y_=cutNeckHeight)
# 在这里我们取消了对下巴的检查,原因在于输入的imageHeight并不能改变
# cell_left_above, cell_right_above = check_jaw(a_thresh, [x_left, imageHeight], [x_right, imageHeight])
cell_left_above, cell_right_above = [x_left, cutNeckHeight], [x_right, cutNeckHeight]
toWidth = x_right - x_left # 矩形宽
# 此时我们寻找到了脖子的"宽出来的"两个点,这两个点作为上面的两个点, 接下来寻找下面的两个点
if per_to_side >1:
assert ValueError("per_to_side 必须小于1!")
y_below = int((neckBelow - cutNeckHeight) * per_to_side + cutNeckHeight) # 定位y轴坐标
cell_left_below = [locate_width(a_thresh, y_=y_below, mode=1), y_below]
cell_right_bellow = [locate_width(a_thresh, y_=y_below, mode=2), y_below]
# 四个点全齐,开始透视变换
# 需要变换的四个点为 cell_left_above, cell_right_above, cell_left_below, cell_right_bellow
rect = np.array([cell_left_above, cell_right_above, cell_left_below, cell_right_bellow],
dtype='float32')
# 变化后的坐标点
dst = np.array([[0, 0], [toWidth, 0], [0 , toHeight], [toWidth, toHeight]],
dtype='float32')
M = cv2.getPerspectiveTransform(rect, dst)
warped = cv2.warpPerspective(image, M, (toWidth, toHeight))
# 将变换后的图像覆盖到原图上
final = cover_image(image=warped, background=image, mode=3, x=cell_left_above[0], y=cell_left_above[1])
return final
def bestJunctionCheck_beta(image:np.ndarray, stepSize:int=4, if_per:bool=False):
"""
最优衔接点检测算法, 去寻找脖子的"拐点"
"""
point_k = 1
_, _, _, a = cv2.split(image) # 这应该是一个四通道的图像
height, width = a.shape
ret, a_thresh = cv2.threshold(a, 127, 255, cv2.THRESH_BINARY) # 将透明图层二值化
y_high, y_low, x_left, x_right = get_box_pro(image=image, model=1) # 直接返回矩阵信息
def scan(y_:int, max_num:int=2):
num = 0
left = False
right = False
for x_ in range(width):
if a_thresh[y_][x_] != 0:
if x_ < width // 2 and left is False:
num += 1
left = True
elif x_ > width // 2 and right is False:
num += 1
right = True
return True if num >= max_num else False
def locate_neck_above():
"""
定位脖子的尖尖脚
"""
for y_ in range( y_high - 2, height):
if scan(y_):
return y_, y_
y_high_left, y_high_right = locate_neck_above()
def locate_width_pro(image_:np.ndarray, y_:int, mode):
"""
这会是一个生成器,用于生成脖子两边的轮廓
x_, y_ 是启始点的坐标,每一次寻找都会让y_+1
mode==1说明是找左边的边,即,image_[y_][x_] == 0 且image_[y_][x_ + 1] !=0 时跳出;
否则 当image_[y_][x_] != 0 时, x_ - 1; 当image_[y_][x_] == 0 且 image_[y_][x_ + 1] ==0 时x_ + 1
mode==2说明是找右边的边,即,image_[y_][x_] == 0 且image_[y_][x_ - 1] !=0 时跳出
否则 当image_[y_][x_] != 0 时, x_ + 1; 当image_[y_][x_] == 0 且 image_[y_][x_ - 1] ==0 时x_ - 1
"""
y_ += 1
if mode == 1:
x_ = 0
while 0 <= y_ < height and 0 <= x_ < width:
while image_[y_][x_] != 0 and x_ >= 0:
x_ -= 1
while image_[y_][x_] == 0 and image_[y_][x_ + 1] == 0 and x_ < width - 2:
x_ += 1
yield [y_, x_]
y_ += 1
elif mode == 2:
x_ = width-1
while 0 <= y_ < height and 0 <= x_ < width:
while image_[y_][x_] != 0 and x_ < width - 2: x_ += 1
while image_[y_][x_] == 0 and image_[y_][x_ - 1] == 0 and x_ >= 0: x_ -= 1
yield [y_, x_]
y_ += 1
yield False
def kGenerator(image_:np.ndarray, mode):
"""
导数生成器,用来生成每一个点对应的导数
"""
y_ = y_high_left if mode == 1 else y_high_right
c_generator = locate_width_pro(image_=image_, y_=y_, mode=mode)
for cell in c_generator:
nc = locate_width_pro(image_=image_, y_=cell[0] + stepSize, mode=mode)
nextCell = next(nc)
if nextCell is False:
yield False, False
else:
k = (cell[1] - nextCell[1]) / stepSize
yield k, cell
def findPt(image_:np.ndarray, mode):
k_generator = kGenerator(image_=image_, mode=mode)
k, cell = next(k_generator)
k_next, cell_next = next(k_generator)
if k is False:
raise ValueError("无法找到拐点!")
while k_next is not False:
k_next, cell_next = next(k_generator)
if (k_next < - 1 / stepSize) or k_next > point_k:
break
cell = cell_next
# return int(cell[0] + stepSize / 2)
return cell[0]
# 先找左边的拐点:
pointY_left = findPt(image_=a_thresh, mode=1)
# 再找右边的拐点:
pointY_right = findPt(image_=a_thresh, mode=2)
point = (pointY_left + pointY_right) // 2
if if_per is True:
point = (pointY_left + pointY_right) // 2
return point / (y_low - y_high)
pointX_left = next(locate_width_pro(image_=a_thresh, y_= point - 1, mode=1))[1]
pointX_right = next(locate_width_pro(image_=a_thresh, y_=point- 1, mode=2))[1]
return [pointX_left, point], [pointX_right, point]
def bestJunctionCheck(image:np.ndarray, offset:int, stepSize:int=4):
"""
最优点检测算算法输入一张脖子图片(无论这张图片是否已经被二值化,我都认为没有被二值化),输出一个小数(脖子最上方与衔接点位置/脖子图像长度)
与beta版不同的是它新增了一个阈值限定内容.
对于脖子而言,我我们首先可以定位到上面的部分,然后根据上面的这个点向下进行遍历检测.
与beta版类似,我们使用一个stepSize来用作斜率的检测
但是对于遍历检测而言,与beta版不同的是,我们需要对遍历的地方进行一定的限制.
限制的标准是,如果当前遍历的点的横坐标和起始点横坐标的插值超过了某个阈值,则认为是越界.
"""
point_k = 1
_, _, _, a = cv2.split(image) # 这应该是一个四通道的图像
height, width = a.shape
ret, a_thresh = cv2.threshold(a, 127, 255, cv2.THRESH_BINARY) # 将透明图层二值化
# 直接返回脖子的位置信息, 修正系数为0, get_box_pro内部也封装了二值化,所以直接输入原图
y_high, y_low, _, _ = get_box_pro(image=image, model=1, correction_factor=0)
# 真正有用的只有上下y轴的两个值...
# 首先当然是确定起始点的位置,我们用同样的scan扫描函数进行行遍历.
def scan(y_:int, max_num:int=2):
num = 0
# 设定两个值,分别代表脖子的左边和右边
left = False
right = False
for x_ in range(width):
if a_thresh[y_][x_] != 0:
# 检测左边
if x_ < width // 2 and left is False:
num += 1
left = True
# 检测右边
elif x_ > width // 2 and right is False:
num += 1
right = True
return True if num >= max_num else False
def locate_neck_above():
"""
定位脖子的尖尖脚
"""
# y_high就是脖子的最高点
for y_ in range(y_high, height):
if scan(y_):
return y_
y_start = locate_neck_above() # 得到遍历的初始高度
if y_low - y_start < stepSize: assert ValueError("脖子太小!")
# 然后获取一下初始的坐标点
x_left, x_right = 0, width
for x_left_ in range(0, width):
if a_thresh[y_start][x_left_] != 0:
x_left = x_left_
break
for x_right_ in range(width -1 , -1, -1):
if a_thresh[y_start][x_right_] != 0:
x_right = x_right_
break
# 接下来我定义两个生成器,首先是脖子轮廓(向下寻找的)生成器,每进行一次next,生成器会返回y+1的脖子轮廓点
def contoursGenerator(image_:np.ndarray, y_:int, mode):
"""
这会是一个生成器,用于生成脖子两边的轮廓
y_ 是启始点的y坐标,每一次寻找都会让y_+1
mode==1说明是找左边的边,即,image_[y_][x_] == 0 且image_[y_][x_ + 1] !=0 时跳出;
否则 当image_[y_][x_] != 0 时, x_ - 1; 当image_[y_][x_] == 0 且 image_[y_][x_ + 1] ==0 时x_ + 1
mode==2说明是找右边的边,即,image_[y_][x_] == 0 且image_[y_][x_ - 1] !=0 时跳出
否则 当image_[y_][x_] != 0 时, x_ + 1; 当image_[y_][x_] == 0 且 image_[y_][x_ - 1] ==0 时x_ - 1
"""
y_ += 1
try:
if mode == 1:
x_ = 0
while 0 <= y_ < height and 0 <= x_ < width:
while image_[y_][x_] != 0 and x_ >= 0: x_ -= 1
# 这里其实会有bug,不过可以不管
while x_ < width and image_[y_][x_] == 0 and image_[y_][x_ + 1] == 0: x_ += 1
yield [y_, x_]
y_ += 1
elif mode == 2:
x_ = width-1
while 0 <= y_ < height and 0 <= x_ < width:
while x_ < width and image_[y_][x_] != 0: x_ += 1
while x_ >= 0 and image_[y_][x_] == 0 and image_[y_][x_ - 1] == 0: x_ -= 1
yield [y_, x_]
y_ += 1
# 当处理失败则返回False
except IndexError:
yield False
# 然后是斜率生成器,这个生成器依赖子轮廓生成器,每一次生成轮廓后会计算斜率,另一个点的选取和stepSize有关
def kGenerator(image_: np.ndarray, mode):
"""
导数生成器,用来生成每一个点对应的导数
"""
y_ = y_start
# 对起始点建立一个生成器, mode=1时是左边轮廓,mode=2时是右边轮廓
c_generator = contoursGenerator(image_=image_, y_=y_, mode=mode)
for cell in c_generator:
# 寻找距离当前cell距离为stepSize的轮廓点
kc = contoursGenerator(image_=image_, y_=cell[0] + stepSize, mode=mode)
kCell = next(kc)
if kCell is False:
# 寻找失败
yield False, False
else:
# 寻找成功,返回当坐标点和斜率值
# 对于左边而言,斜率必然是前一个点的坐标减去后一个点的坐标
# 对于右边而言,斜率必然是后一个点的坐标减去前一个点的坐标
k = (cell[1] - kCell[1]) / stepSize if mode == 1 else (kCell[1] - cell[1]) / stepSize
yield k, cell
# 接着开始写寻找算法,需要注意的是我们是分两边选择的
def findPt(image_:np.ndarray, mode):
x_base = x_left if mode == 1 else x_right
k_generator = kGenerator(image_=image_, mode=mode)
k, cell = k_generator.__next__()
if k is False:
raise ValueError("无法找到拐点!")
k_next, cell_next = k_generator.__next__()
while k_next is not False:
cell = cell_next
if cell[1] > x_base and mode == 2:
x_base = cell[1]
elif cell[1] < x_base and mode == 1:
x_base = cell[1]
# 跳出循环的方式一:斜率超过了某个值
if k_next > point_k:
print("K out")
break
# 跳出循环的方式二:超出阈值
elif abs(cell[1] - x_base) > offset:
print("O out")
break
k_next, cell_next = k_generator.__next__()
if abs(cell[1] - x_base) > offset:
cell[0] = cell[0] - offset - 1
return cell[0]
# 先找左边的拐点:
pointY_left = findPt(image_=a_thresh, mode=1)
# 再找右边的拐点:
pointY_right = findPt(image_=a_thresh, mode=2)
point = min(pointY_right, pointY_left)
per = (point - y_high) / (y_low - y_high)
# pointX_left = next(contoursGenerator(image_=a_thresh, y_= point- 1, mode=1))[1]
# pointX_right = next(contoursGenerator(image_=a_thresh, y_=point - 1, mode=2))[1]
# return [pointX_left, point], [pointX_right, point]
return per
def checkSharpCorner(image:np.ndarray):
_, _, _, a = cv2.split(image) # 这应该是一个四通道的图像
height, width = a.shape
ret, a_thresh = cv2.threshold(a, 127, 255, cv2.THRESH_BINARY) # 将透明图层二值化
# 直接返回脖子的位置信息, 修正系数为0, get_box_pro内部也封装了二值化,所以直接输入原图
y_high, y_low, _, _ = get_box_pro(image=image, model=1, correction_factor=0)
def scan(y_:int, max_num:int=2):
num = 0
# 设定两个值,分别代表脖子的左边和右边
left = False
right = False
for x_ in range(width):
if a_thresh[y_][x_] != 0:
# 检测左边
if x_ < width // 2 and left is False:
num += 1
left = True
# 检测右边
elif x_ > width // 2 and right is False:
num += 1
right = True
return True if num >= max_num else False
def locate_neck_above():
"""
定位脖子的尖尖脚
"""
# y_high就是脖子的最高点
for y_ in range(y_high, height):
if scan(y_):
return y_
y_start = locate_neck_above()
return y_start
def checkJaw(image:np.ndarray, y_start:int):
# 寻找"马鞍点"
_, _, _, a = cv2.split(image) # 这应该是一个四通道的图像
height, width = a.shape
ret, a_thresh = cv2.threshold(a, 127, 255, cv2.THRESH_BINARY) # 将透明图层二值化
if width <=1: raise TypeError("图像太小!")
x_left, x_right = 0, width - 1
for x_left in range(width):
if a_thresh[y_start][x_left] != 0:
while a_thresh[y_start][x_left] != 0: x_left += 1
break
for x_right in range(width-1, -1, -1):
if a_thresh[y_start][x_right] != 0:
while a_thresh[y_start][x_right] != 0: x_right -= 1
break
point_list_y = []
point_list_x = []
for x in range(x_left, x_right):
y = y_start
while a_thresh[y][x] == 0: y += 1
point_list_y.append(y)
point_list_x.append(x)
y = max(point_list_y)
x = point_list_x[point_list_y.index(y)]
return x, y
def checkHairLOrR(cloth_image_input_cut,
input_a,
neck_a,
cloth_image_input_top_y,
cutbar_top=0.4,
cutbar_bottom=0.5,
threshold=0.3):
"""
本函数用于检测衣服是否被头发遮挡,当前只考虑左右是否被遮挡,即"一刀切"
返回int
0代表没有被遮挡
1代表左边被遮挡
2代表右边被遮挡
3代表全被遮挡了
约定,输入的图像是一张灰度图,且被二值化过.
"""
def per_darkPoint(img:np.ndarray) -> int:
"""
用于遍历相加图像上的黑点.
然后返回黑点数/图像面积
"""
h, w = img.shape
sum_darkPoint = 0
for y in range(h):
for x in range(w):
if img[y][x] == 0:
sum_darkPoint += 1
return sum_darkPoint / (h * w)
if threshold < 0 or threshold > 1: raise TypeError("阈值设置必须在0和1之间!")
# 裁出cloth_image_input_cut按高度40%~50%的区域-cloth_image_input_cutbar,并转换为A矩阵,做二值化
cloth_image_input_height = cloth_image_input_cut.shape[0]
_, _, _, cloth_image_input_cutbar = cv2.split(cloth_image_input_cut[
int(cloth_image_input_height * cutbar_top):int(
cloth_image_input_height * cutbar_bottom), :])
_, cloth_image_input_cutbar = cv2.threshold(cloth_image_input_cutbar, 127, 255, cv2.THRESH_BINARY)
# 裁出input_image、neck_image的A矩阵的对应区域,并做二值化
input_a_cutbar = input_a[cloth_image_input_top_y + int(cloth_image_input_height * cutbar_top):
cloth_image_input_top_y + int(cloth_image_input_height * cutbar_bottom), :]
_, input_a_cutbar = cv2.threshold(input_a_cutbar, 127, 255, cv2.THRESH_BINARY)
neck_a_cutbar = neck_a[cloth_image_input_top_y + int(cloth_image_input_height * cutbar_top):
cloth_image_input_top_y + int(cloth_image_input_height * cutbar_bottom), :]
_, neck_a_cutbar = cv2.threshold(neck_a_cutbar, 50, 255, cv2.THRESH_BINARY)
# 将三个cutbar合到一起-result_a_cutbar
input_a_cutbar = np.uint8(255 - input_a_cutbar)
result_a_cutbar = cv2.add(input_a_cutbar, cloth_image_input_cutbar)
result_a_cutbar = cv2.add(result_a_cutbar, neck_a_cutbar)
if_mask = 0
# 我们将图像 一刀切,分为左边和右边
height, width = result_a_cutbar.shape # 一通道图像
left_image = result_a_cutbar[:, :width//2]
right_image = result_a_cutbar[:, width//2:]
if per_darkPoint(left_image) > threshold:
if_mask = 1
if per_darkPoint(right_image) > threshold:
if_mask = 3 if if_mask == 1 else 2
return if_mask
def find_black(image):
"""
找黑色点函数,遇到输入矩阵中的第一个黑点,返回它的y值
"""
height, width = image.shape[0], image.shape[1]
for i in range(height):
for j in range(width):
if image[i, j] < 127:
return i
return None
def convert_black_array(image):
height, width = image.shape[0], image.shape[1]
mask = np.zeros([height, width])
for j in range(width):
for i in range(height):
if image[i, j] > 127:
mask[i:, j] = 1
break
return mask
def checkLongHair(neck_image, head_bottom_y, neck_top_y):
"""
长发检测函数,输入为head/neck图像,通过下巴是否为最低点,来判断是否为长发
:return 0 : 短发
:return 1 : 长发
"""
jaw_y = neck_top_y + checkJaw(neck_image, y_start=checkSharpCorner(neck_image))[1]
if jaw_y >= head_bottom_y-3:
return 0
else:
return 1
def checkLongHair2(head_bottom_y, cloth_top_y):
if head_bottom_y > cloth_top_y+10:
return 1
else:
return 0
if __name__ == "__main__":
for i in range(1, 8):
img = cv2.imread(f"./neck_temp/neck_image{i}.png", cv2.IMREAD_UNCHANGED)
# new = transformationNeck(image=img, cutNeckHeight=419,neckBelow=472, toHeight=150)
# point_list = bestJunctionCheck(img, offset=5, stepSize=3)
# per = bestJunctionCheck(img, offset=5, stepSize=3)
# # 返回一个小数的形式, 接下来我将它处理为两个点
point_list = []
# y_high_, y_low_, _, _ = get_box_pro(image=img, model=1, conreection_factor=0)
# _y = y_high_ + int((y_low_ - y_high_) * per)
# _, _, _, a_ = cv2.split(img) # 这应该是一个四通道的图像
# h, w = a_.shape
# r, a_t = cv2.threshold(a_, 127, 255, cv2.THRESH_BINARY) # 将透明图层二值化
# _x = 0
# for _x in range(w):
# if a_t[_y][_x] != 0:
# break
# point_list.append([_x, _y])
# for _x in range(w - 1, -1, -1):
# if a_t[_y][_x] != 0:
# break
# point_list.append([_x, _y])
y = checkSharpCorner(img)
point = checkJaw(image=img, y_start=y)
point_list.append(point)
new = draw_picture_dots(img, point_list, pen_size=2)
cv2.imshow(f"{i}", new)
cv2.waitKey(0)