add aphla image
This commit is contained in:
parent
d5db3a3020
commit
16f61c46bb
@ -56,17 +56,18 @@ def detect(net, img, device):
|
|||||||
return bboxlist
|
return bboxlist
|
||||||
|
|
||||||
|
|
||||||
def batch_detect(net, imgs, device):
|
def batch_detect(net, images, device):
|
||||||
imgs = imgs - np.array([104, 117, 123])
|
rgb = images[:, :, :, :3]
|
||||||
imgs = imgs.transpose(0, 3, 1, 2)
|
rgb = rgb - np.array([104, 117, 123])
|
||||||
|
rgb = rgb.transpose(0, 3, 1, 2)
|
||||||
|
|
||||||
if 'cuda' in device:
|
if 'cuda' in device:
|
||||||
torch.backends.cudnn.benchmark = True
|
torch.backends.cudnn.benchmark = True
|
||||||
|
|
||||||
imgs = torch.from_numpy(imgs).float().to(device)
|
images1 = torch.from_numpy(rgb).float().to(device)
|
||||||
BB, CC, HH, WW = imgs.size()
|
BB, CC, HH, WW = images1.size()
|
||||||
with torch.no_grad():
|
with torch.no_grad():
|
||||||
olist = net(imgs)
|
olist = net(images1)
|
||||||
|
|
||||||
bboxlist = []
|
bboxlist = []
|
||||||
for i in range(len(olist) // 2):
|
for i in range(len(olist) // 2):
|
||||||
|
@ -44,7 +44,6 @@ class HumanContext:
|
|||||||
logging.info(f'face images length: {face_images_length}')
|
logging.info(f'face images length: {face_images_length}')
|
||||||
print(f'face images length: {face_images_length}')
|
print(f'face images length: {face_images_length}')
|
||||||
|
|
||||||
|
|
||||||
def __del__(self):
|
def __del__(self):
|
||||||
print(f'HumanContext: __del__')
|
print(f'HumanContext: __del__')
|
||||||
object_stop(self._asr)
|
object_stop(self._asr)
|
||||||
|
@ -30,8 +30,9 @@ class VideoRender(BaseRender):
|
|||||||
|
|
||||||
clock_time = self._play_clock.clock_time()
|
clock_time = self._play_clock.clock_time()
|
||||||
time_difference = clock_time - ps
|
time_difference = clock_time - ps
|
||||||
|
print("Video frame time", clock_time, ps, time_difference)
|
||||||
if abs(time_difference) > self._play_clock.audio_diff_threshold:
|
if abs(time_difference) > self._play_clock.audio_diff_threshold:
|
||||||
if self._diff_avg_count < 3:
|
if self._diff_avg_count < 5:
|
||||||
self._diff_avg_count += 1
|
self._diff_avg_count += 1
|
||||||
else:
|
else:
|
||||||
if time_difference < -self._play_clock.audio_diff_threshold:
|
if time_difference < -self._play_clock.audio_diff_threshold:
|
||||||
@ -62,7 +63,7 @@ class VideoRender(BaseRender):
|
|||||||
except:
|
except:
|
||||||
print('resize error')
|
print('resize error')
|
||||||
return
|
return
|
||||||
combine_frame[y1:y2, x1:x2] = res_frame
|
combine_frame[y1:y2, x1:x2, :3] = res_frame
|
||||||
|
|
||||||
image = combine_frame
|
image = combine_frame
|
||||||
# image = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)
|
# image = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)
|
||||||
|
10
ui.py
10
ui.py
@ -57,7 +57,7 @@ class App(customtkinter.CTk):
|
|||||||
# self.main_button_1.grid(row=2, column=2, padx=(20, 20), pady=(20, 20), sticky="nsew")
|
# self.main_button_1.grid(row=2, column=2, padx=(20, 20), pady=(20, 20), sticky="nsew")
|
||||||
background = os.path.join(current_file_path, 'data', 'background', 'background.webp')
|
background = os.path.join(current_file_path, 'data', 'background', 'background.webp')
|
||||||
logger.info(f'background: {background}')
|
logger.info(f'background: {background}')
|
||||||
# self._background = ImageTk.PhotoImage(read_image(background))
|
self._background = read_image(background).convert("RGBA")
|
||||||
|
|
||||||
self._init_image_canvas()
|
self._init_image_canvas()
|
||||||
|
|
||||||
@ -105,7 +105,13 @@ class App(customtkinter.CTk):
|
|||||||
image = cv2.resize(image, (int(iwidth * height / iheight), int(height)), interpolation=cv2.INTER_AREA)
|
image = cv2.resize(image, (int(iwidth * height / iheight), int(height)), interpolation=cv2.INTER_AREA)
|
||||||
img = Image.fromarray(image)
|
img = Image.fromarray(image)
|
||||||
|
|
||||||
imgtk = ImageTk.PhotoImage(image=img)
|
bg_width, bg_height = self._background.size
|
||||||
|
fg_width, fg_height = img.size
|
||||||
|
x = (bg_width - fg_width) // 2
|
||||||
|
y = (bg_height - fg_height) // 2
|
||||||
|
self._background.paste(img, (x, y), img)
|
||||||
|
|
||||||
|
imgtk = ImageTk.PhotoImage(self._background)
|
||||||
|
|
||||||
self._canvas.delete("all")
|
self._canvas.delete("all")
|
||||||
|
|
||||||
|
@ -36,7 +36,7 @@ def read_images(img_list):
|
|||||||
print(f'read image path:{img_path}')
|
print(f'read image path:{img_path}')
|
||||||
# frame = cv2.imread(img_path, cv2.IMREAD_UNCHANGED)
|
# frame = cv2.imread(img_path, cv2.IMREAD_UNCHANGED)
|
||||||
frame = Image.open(img_path)
|
frame = Image.open(img_path)
|
||||||
frame = frame.convert("RGBA")
|
# frame = frame.convert("RGBA")
|
||||||
frame = np.array(frame)
|
frame = np.array(frame)
|
||||||
frames.append(frame)
|
frames.append(frame)
|
||||||
return frames
|
return frames
|
||||||
@ -179,7 +179,7 @@ def load_avatar(path, img_size, device):
|
|||||||
face_frames = []
|
face_frames = []
|
||||||
coord_frames = []
|
coord_frames = []
|
||||||
for face, coord in face_det_results:
|
for face, coord in face_det_results:
|
||||||
resized_crop_frame = cv2.resize(face, (img_size, img_size))
|
resized_crop_frame = cv2.resize(face[:, :, :3], (img_size, img_size))
|
||||||
face_frames.append(resized_crop_frame)
|
face_frames.append(resized_crop_frame)
|
||||||
coord_frames.append(coord)
|
coord_frames.append(coord)
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user