From ac64cdcc54194c898cc08ea7f555366b4b14d3c3 Mon Sep 17 00:00:00 2001 From: zhouchengrong Date: Thu, 18 Jul 2024 10:28:26 +0800 Subject: [PATCH] =?UTF-8?q?feat=20fix=20=20=E9=87=8D=E5=86=99=E6=89=80?= =?UTF-8?q?=E6=9C=89resize=E4=BB=A3=E7=A0=81=EF=BC=8Cmmcv=E6=9B=BF?= =?UTF-8?q?=E6=8D=A2=E4=B8=BAcv?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- app/service/design/items/pipelines/split.py | 18 ++++++++++++++---- 1 file changed, 14 insertions(+), 4 deletions(-) diff --git a/app/service/design/items/pipelines/split.py b/app/service/design/items/pipelines/split.py index 1e06712..7b010ab 100644 --- a/app/service/design/items/pipelines/split.py +++ b/app/service/design/items/pipelines/split.py @@ -45,10 +45,15 @@ class Split(object): rgba_image = rgb_to_rgba((result['final_image'].shape[0], result['final_image'].shape[1]), result['final_image'], result['mask']) result_front_image = np.zeros_like(rgba_image) result_front_image[front_mask != 0] = rgba_image[front_mask != 0] + # TODO PIL resize替换为CV2 + # result_front_image_pil = Image.fromarray(cvtColor(result_front_image, COLOR_BGR2RGBA)) + # front_new_size = (int(result_front_image_pil.width * result["scale"] * result["resize_scale"][0]), int(result_front_image_pil.height * result["scale"] * result["resize_scale"][1])) + # result_front_image_pil = result_front_image_pil.resize(front_new_size, Image.LANCZOS) + front_new_size = (int(result_front_image.shape[1] * result["scale"] * result["resize_scale"][0]), int(result_front_image.shape[0] * result["scale"] * result["resize_scale"][1])) + result_front_image = cv2.resize(result_front_image, front_new_size) result_front_image_pil = Image.fromarray(cvtColor(result_front_image, COLOR_BGR2RGBA)) - front_new_size = (int(result_front_image_pil.width * result["scale"] * result["resize_scale"][0]), int(result_front_image_pil.height * result["scale"] * result["resize_scale"][1])) - result_front_image_pil = result_front_image_pil.resize(front_new_size, Image.LANCZOS) + # result['front_mask_image'] = cv2.resize(front_mask, front_new_size) # result['front_image'] = result_front_image_pil front_mask = cv2.resize(front_mask, front_new_size) @@ -57,10 +62,15 @@ class Split(object): if result["name"] in ('blouse', 'dress', 'outwear', 'tops'): result_back_image = np.zeros_like(rgba_image) result_back_image[back_mask != 0] = rgba_image[back_mask != 0] + # TODO PIL resize替换为CV2 + # result_back_image_pil = Image.fromarray(cvtColor(result_back_image, COLOR_BGR2RGBA)) + # back_new_size = (int(result_back_image_pil.width * result["scale"] * result["resize_scale"][0]), int(result_back_image_pil.height * result["scale"] * result["resize_scale"][1])) + # result_back_image_pil = result_back_image_pil.resize(back_new_size, Image.LANCZOS) + back_new_size = (int(result_back_image.shape[1] * result["scale"] * result["resize_scale"][0]), int(result_back_image.shape[0] * result["scale"] * result["resize_scale"][1])) + result_back_image = cv2.resize(result_back_image, back_new_size) result_back_image_pil = Image.fromarray(cvtColor(result_back_image, COLOR_BGR2RGBA)) - back_new_size = (int(result_back_image_pil.width * result["scale"] * result["resize_scale"][0]), int(result_back_image_pil.height * result["scale"] * result["resize_scale"][1])) - result_back_image_pil = result_back_image_pil.resize(back_new_size, Image.LANCZOS) + # result['back_mask_image'] = cv2.resize(back_mask, back_new_size) # result['back_image'] = result_back_image_pil