From d7a5ee36979dfef2be24da38c7bc59fde867d22b Mon Sep 17 00:00:00 2001 From: zhouchengrong Date: Mon, 15 Apr 2024 18:37:03 +0800 Subject: [PATCH 01/19] =?UTF-8?q?feat=20generate=20=E8=BF=81=E7=A7=BB?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- app/core/config.py | 11 ++++------- 1 file changed, 4 insertions(+), 7 deletions(-) diff --git a/app/core/config.py b/app/core/config.py index 0444760..bbb2a0b 100644 --- a/app/core/config.py +++ b/app/core/config.py @@ -19,19 +19,16 @@ class Settings(BaseSettings): LOGGING_CONFIG_FILE = os.path.join(BASE_DIR, 'logging_env.py') -DEBUG = True +DEBUG = False ENV = 0 if DEBUG: LOGS_PATH = "logs/errors.log" else: LOGS_PATH = "app/logs/errors.log" -RABBITMQ_ENV = "" - -if ENV == 1: - RABBITMQ_ENV = "dev" -elif ENV == 2: - RABBITMQ_ENV = "local" +# RABBITMQ_ENV = "" +# RABBITMQ_ENV = "dev" +RABBITMQ_ENV = "local" settings = Settings() From 6bb988bd88c3edfc60ede110cfb95dd4d9851865 Mon Sep 17 00:00:00 2001 From: zhouchengrong Date: Mon, 15 Apr 2024 18:41:06 +0800 Subject: [PATCH 02/19] =?UTF-8?q?feat=20generate=20=E8=BF=81=E7=A7=BB?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- Dockerfile | 2 +- app/core/config.py | 4 ++-- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/Dockerfile b/Dockerfile index 0e94af1..2f4ba1b 100644 --- a/Dockerfile +++ b/Dockerfile @@ -19,4 +19,4 @@ LABEL maintainer="zchengrong@yeah.net" \ name="trinity_aida" -CMD ["gunicorn", "-c", "gunicorn_config.py", "app.main:app" , "-e", "SR_RABBITMQ_QUEUES=SuperResolution" ,"-e", "GI_RABBITMQ_QUEUES=GenerateImage"] \ No newline at end of file +CMD ["gunicorn", "-c", "gunicorn_config.py", "app.main:app" , "-e", "SR_RABBITMQ_QUEUES=SuperResolution-local" ,"-e", "GI_RABBITMQ_QUEUES=GenerateImage-local"] \ No newline at end of file diff --git a/app/core/config.py b/app/core/config.py index 6bf0556..a9220f1 100644 --- a/app/core/config.py +++ b/app/core/config.py @@ -25,9 +25,9 @@ if DEBUG: else: LOGS_PATH = "app/logs/errors.log" -RABBITMQ_ENV = "" # 生产环境 +# RABBITMQ_ENV = "" # 生产环境 # RABBITMQ_ENV = "-dev" # 开发环境 -# RABBITMQ_ENV = "-local" # 本地测试环境 +RABBITMQ_ENV = "-local" # 本地测试环境 settings = Settings() From b596692b35c1d30cc5ee9180f165ffe79c2432ad Mon Sep 17 00:00:00 2001 From: zhouchengrong Date: Tue, 16 Apr 2024 15:53:08 +0800 Subject: [PATCH 03/19] =?UTF-8?q?feat=20generate=20=E5=8D=87=E7=BA=A7=20at?= =?UTF-8?q?tribute=20retrieve=20=E8=BF=81=E7=A7=BB?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- Dockerfile | 2 +- app/core/config.py | 6 +++--- 2 files changed, 4 insertions(+), 4 deletions(-) diff --git a/Dockerfile b/Dockerfile index c577312..2f4ba1b 100644 --- a/Dockerfile +++ b/Dockerfile @@ -19,4 +19,4 @@ LABEL maintainer="zchengrong@yeah.net" \ name="trinity_aida" -CMD ["gunicorn", "-c", "gunicorn_config.py", "app.main:app" , "-e", "SR_RABBITMQ_QUEUES=SuperResolution-dev" ,"-e", "GI_RABBITMQ_QUEUES=GenerateImage-dev"] \ No newline at end of file +CMD ["gunicorn", "-c", "gunicorn_config.py", "app.main:app" , "-e", "SR_RABBITMQ_QUEUES=SuperResolution-local" ,"-e", "GI_RABBITMQ_QUEUES=GenerateImage-local"] \ No newline at end of file diff --git a/app/core/config.py b/app/core/config.py index d7810db..a114fa6 100644 --- a/app/core/config.py +++ b/app/core/config.py @@ -19,7 +19,7 @@ class Settings(BaseSettings): LOGGING_CONFIG_FILE = os.path.join(BASE_DIR, 'logging_env.py') -DEBUG = True +DEBUG = False if DEBUG: LOGS_PATH = "logs/" CATEGORY_PATH = "service/attribute/config/descriptor/category/category_dis.csv" @@ -28,8 +28,8 @@ else: CATEGORY_PATH = "app/service/attribute/config/descriptor/category/category_dis.csv" # RABBITMQ_ENV = "" # 生产环境 -RABBITMQ_ENV = "-dev" # 开发环境 -# RABBITMQ_ENV = "-local" # 本地测试环境 +# RABBITMQ_ENV = "-dev" # 开发环境 +RABBITMQ_ENV = "-local" # 本地测试环境 settings = Settings() From 6759b873d520f9643ca56c1640d3de4534143a90 Mon Sep 17 00:00:00 2001 From: zhouchengrong Date: Tue, 16 Apr 2024 16:36:17 +0800 Subject: [PATCH 04/19] =?UTF-8?q?feat=20generate=20=E9=80=BB=E8=BE=91?= =?UTF-8?q?=E8=B0=83=E6=95=B4?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- app/core/config.py | 4 ++-- app/service/generate_image/service.py | 27 ++++++++++++++++++++++----- 2 files changed, 24 insertions(+), 7 deletions(-) diff --git a/app/core/config.py b/app/core/config.py index a114fa6..73e6e67 100644 --- a/app/core/config.py +++ b/app/core/config.py @@ -19,7 +19,7 @@ class Settings(BaseSettings): LOGGING_CONFIG_FILE = os.path.join(BASE_DIR, 'logging_env.py') -DEBUG = False +DEBUG = True if DEBUG: LOGS_PATH = "logs/" CATEGORY_PATH = "service/attribute/config/descriptor/category/category_dis.csv" @@ -29,7 +29,7 @@ else: # RABBITMQ_ENV = "" # 生产环境 # RABBITMQ_ENV = "-dev" # 开发环境 -RABBITMQ_ENV = "-local" # 本地测试环境 +RABBITMQ_ENV = "-local" # 本地测试环境 settings = Settings() diff --git a/app/service/generate_image/service.py b/app/service/generate_image/service.py index fed3d41..b1448b2 100644 --- a/app/service/generate_image/service.py +++ b/app/service/generate_image/service.py @@ -10,7 +10,10 @@ import json import logging import time +from io import BytesIO +import cv2 +import minio import redis import tritonclient.grpc as grpcclient import numpy as np @@ -20,7 +23,6 @@ from tritonclient.utils import np_to_triton_dtype from app.core.config import * from app.schemas.generate_image import GenerateImageModel from app.service.generate_image.utils.upload_sd_image import upload_png_sd -from app.service.utils.generate_uuid import generate_uuid logger = logging.getLogger() @@ -32,13 +34,15 @@ class GenerateImage: self.redis_client = redis.StrictRedis(host=REDIS_HOST, port=REDIS_PORT, db=REDIS_DB, decode_responses=True) self.connection = pika.BlockingConnection(pika.ConnectionParameters(**RABBITMQ_PARAMS)) self.channel = self.connection.channel() - if request_data.mode == "txt2img": - self.image = np.random.randint(0, 256, (1024, 1024, 3), dtype=np.uint8) + if request_data.mode == "img2img": + self.image = self.get_image(request_data.image_url) + self.prompt = request_data.prompt else: - self.image = request_data.image + self.image = np.random.randint(0, 256, (1024, 1024, 3), dtype=np.uint8) + self.prompt = request_data.prompt + self.tasks_id = request_data.tasks_id self.user_id = self.tasks_id[self.tasks_id.rfind('-') + 1:] - self.prompt = request_data.prompt self.mode = request_data.mode self.batch_size = 1 self.category = request_data.category @@ -49,9 +53,22 @@ class GenerateImage: self.grpc_client.close() self.connection.close() + def get_image(self, image_url): + # Get data of an object. + # Read data from response. + try: + response = self.minio_client.get_object(image_url.split('/')[0], image_url[image_url.find('/') + 1:]) + image_file = BytesIO(response.data) + image_array = np.asarray(bytearray(image_file.read()), dtype=np.uint8) + image_cv2 = cv2.imdecode(image_array, cv2.IMREAD_COLOR) + except minio.error.S3Error: + image_cv2 = np.random.randint(0, 256, (1024, 1024, 3), dtype=np.uint8) + return image_cv2 + def __call__(self, *args, **kwargs): self.generate_data = json.dumps({'status': 'PENDING', 'message': "pending", 'data': ''}) self.redis_client.set(self.tasks_id, self.generate_data) + self.redis_client.expire(self.tasks_id, 600) def callback(self, result, error): if error: From 43ddc82fb40b205fa28f8011b6c61ec678af8fe1 Mon Sep 17 00:00:00 2001 From: zhouchengrong Date: Tue, 16 Apr 2024 16:52:25 +0800 Subject: [PATCH 05/19] =?UTF-8?q?feat=20=E7=8E=AF=E5=A2=83=E6=9B=B4?= =?UTF-8?q?=E6=96=B0?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- Dockerfile | 1 + requirements.txt | Bin 708 -> 814 bytes 2 files changed, 1 insertion(+) diff --git a/Dockerfile b/Dockerfile index 2f4ba1b..d97cfae 100644 --- a/Dockerfile +++ b/Dockerfile @@ -6,6 +6,7 @@ RUN apt install -y libgl1-mesa-glx COPY ./requirements.txt /requirements.txt RUN pip install --upgrade pip RUN pip install -r requirements.txt +RUN mkdir logs RUN pip install gunicorn RUN pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu118 RUN #pip install mmcv==1.4.2 -f https://download.openmmlab.com/mmcv/dist/cu117/torch1.13/index.html diff --git a/requirements.txt b/requirements.txt index b77dd7d0cfc9f81464fad1b15bb2f83471bba06e..152908247f5c59e3953542a8479bc450084ee395 100644 GIT binary patch delta 117 zcmX@Yx{hr_2cu*GLn1>SLkf^AX0Qc9BL+P%Hk@q9)a}j7zy%afX2@j72I3M1T_BkY pRGZF_3RZ0ZRt?f)zyMNP!cYh_D;F$h0u(g^LXdetxyj9p+W{JN6c_*i delta 15 XcmZ3-c7%092jk=#rcIL Date: Tue, 16 Apr 2024 17:05:21 +0800 Subject: [PATCH 06/19] =?UTF-8?q?feat=20=E7=8E=AF=E5=A2=83=E6=9B=B4?= =?UTF-8?q?=E6=96=B0?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- Dockerfile | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/Dockerfile b/Dockerfile index d97cfae..674faec 100644 --- a/Dockerfile +++ b/Dockerfile @@ -6,7 +6,7 @@ RUN apt install -y libgl1-mesa-glx COPY ./requirements.txt /requirements.txt RUN pip install --upgrade pip RUN pip install -r requirements.txt -RUN mkdir logs +RUN mkdir app/logs RUN pip install gunicorn RUN pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu118 RUN #pip install mmcv==1.4.2 -f https://download.openmmlab.com/mmcv/dist/cu117/torch1.13/index.html From f53dc0b33114bc2d8d20213c90e9dda2be21caad Mon Sep 17 00:00:00 2001 From: zhouchengrong Date: Tue, 16 Apr 2024 17:09:53 +0800 Subject: [PATCH 07/19] =?UTF-8?q?feat=20=E7=8E=AF=E5=A2=83=E6=9B=B4?= =?UTF-8?q?=E6=96=B0?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- Dockerfile | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/Dockerfile b/Dockerfile index 674faec..0bd3e74 100644 --- a/Dockerfile +++ b/Dockerfile @@ -6,7 +6,7 @@ RUN apt install -y libgl1-mesa-glx COPY ./requirements.txt /requirements.txt RUN pip install --upgrade pip RUN pip install -r requirements.txt -RUN mkdir app/logs +RUN mkdir -p app/logs RUN pip install gunicorn RUN pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu118 RUN #pip install mmcv==1.4.2 -f https://download.openmmlab.com/mmcv/dist/cu117/torch1.13/index.html From 32ebb2232ce1d983c478f73eea9bcf5d331b83b8 Mon Sep 17 00:00:00 2001 From: zhouchengrong Date: Tue, 16 Apr 2024 17:14:02 +0800 Subject: [PATCH 08/19] =?UTF-8?q?feat=20=E7=8E=AF=E5=A2=83=E6=9B=B4?= =?UTF-8?q?=E6=96=B0?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- app/service/generate_image/service.py | 5 ----- 1 file changed, 5 deletions(-) diff --git a/app/service/generate_image/service.py b/app/service/generate_image/service.py index b1448b2..46c96a9 100644 --- a/app/service/generate_image/service.py +++ b/app/service/generate_image/service.py @@ -48,11 +48,6 @@ class GenerateImage: self.category = request_data.category self.index = 0 - def __del__(self): - self.redis_client.close() - self.grpc_client.close() - self.connection.close() - def get_image(self, image_url): # Get data of an object. # Read data from response. From c9b407b2a4340a139d6d54141fde5e33d331cb0a Mon Sep 17 00:00:00 2001 From: zhouchengrong Date: Tue, 16 Apr 2024 17:15:00 +0800 Subject: [PATCH 09/19] =?UTF-8?q?feat=20=E7=8E=AF=E5=A2=83=E6=9B=B4?= =?UTF-8?q?=E6=96=B0?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- app/service/attribute/service_att_recognition.py | 3 --- app/service/generate_image/test.py | 5 ----- app/service/super_resolution/service.py | 5 ----- 3 files changed, 13 deletions(-) diff --git a/app/service/attribute/service_att_recognition.py b/app/service/attribute/service_att_recognition.py index d307474..da71c16 100644 --- a/app/service/attribute/service_att_recognition.py +++ b/app/service/attribute/service_att_recognition.py @@ -30,9 +30,6 @@ class AttributeRecognition: self.const = const self.triton_client = httpclient.InferenceServerClient(url=f"{ATT_TRITON_URL}") - def __del__(self): - self.triton_client.close() - def get_result(self): for sketch in self.request_data: if sketch['category'] == "Tops" or sketch['category'] == "Blouse": diff --git a/app/service/generate_image/test.py b/app/service/generate_image/test.py index 0e03900..ab2dc43 100644 --- a/app/service/generate_image/test.py +++ b/app/service/generate_image/test.py @@ -64,11 +64,6 @@ class GenerateImage: pass - def __del__(self): - self.redis_client.close() - self.triton_client.close() - self.connection.close() - @staticmethod def image_grid(imgs, rows, cols): assert len(imgs) == rows * cols diff --git a/app/service/super_resolution/service.py b/app/service/super_resolution/service.py index e20eb70..95b2811 100644 --- a/app/service/super_resolution/service.py +++ b/app/service/super_resolution/service.py @@ -29,11 +29,6 @@ class SuperResolution: self.connection = pika.BlockingConnection(pika.ConnectionParameters(**RABBITMQ_PARAMS)) self.channel = self.connection.channel() - def __del__(self): - self.redis_client.close() - self.triton_client.close() - self.connection.close() - # @RunTime def read_image(self): try: From 004607a9583fffd0c27a12dbced240f6a1268359 Mon Sep 17 00:00:00 2001 From: zhouchengrong Date: Wed, 17 Apr 2024 09:41:35 +0800 Subject: [PATCH 10/19] =?UTF-8?q?feat=20=E7=8E=AF=E5=A2=83=E6=9B=B4?= =?UTF-8?q?=E6=96=B0?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- Dockerfile | 2 +- app/core/config.py | 4 ++-- docker-compose.yml | 2 +- 3 files changed, 4 insertions(+), 4 deletions(-) diff --git a/Dockerfile b/Dockerfile index 0bd3e74..c1e42d1 100644 --- a/Dockerfile +++ b/Dockerfile @@ -20,4 +20,4 @@ LABEL maintainer="zchengrong@yeah.net" \ name="trinity_aida" -CMD ["gunicorn", "-c", "gunicorn_config.py", "app.main:app" , "-e", "SR_RABBITMQ_QUEUES=SuperResolution-local" ,"-e", "GI_RABBITMQ_QUEUES=GenerateImage-local"] \ No newline at end of file +CMD ["gunicorn", "-c", "gunicorn_config.py", "app.main:app" , "-e", "SR_RABBITMQ_QUEUES=SuperResolution-dev" ,"-e", "GI_RABBITMQ_QUEUES=GenerateImage-dev"] \ No newline at end of file diff --git a/app/core/config.py b/app/core/config.py index 4889db8..b88390c 100644 --- a/app/core/config.py +++ b/app/core/config.py @@ -28,8 +28,8 @@ else: CATEGORY_PATH = "app/service/attribute/config/descriptor/category/category_dis.csv" # RABBITMQ_ENV = "" # 生产环境 -# RABBITMQ_ENV = "-dev" # 开发环境 -RABBITMQ_ENV = "-local" # 本地测试环境 +RABBITMQ_ENV = "-dev" # 开发环境 +# RABBITMQ_ENV = "-local" # 本地测试环境 settings = Settings() diff --git a/docker-compose.yml b/docker-compose.yml index 9d3b725..a550e44 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -6,4 +6,4 @@ services: volumes: - ./trinity_client_aida:/trinity ports: - - "10201:4562" \ No newline at end of file + - "10200:4562" \ No newline at end of file From 5ed53a1e7c452602be8f8a633e0b3bb39183fb8a Mon Sep 17 00:00:00 2001 From: zhouchengrong Date: Wed, 17 Apr 2024 17:37:51 +0800 Subject: [PATCH 11/19] =?UTF-8?q?feat=20generate=20image=20=E9=80=BB?= =?UTF-8?q?=E8=BE=91=E8=A1=A5=E5=85=85?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- app/service/generate_image/service.py | 94 +++++++++++++++------------ 1 file changed, 52 insertions(+), 42 deletions(-) diff --git a/app/service/generate_image/service.py b/app/service/generate_image/service.py index 46c96a9..d0bda9a 100644 --- a/app/service/generate_image/service.py +++ b/app/service/generate_image/service.py @@ -47,6 +47,9 @@ class GenerateImage: self.batch_size = 1 self.category = request_data.category self.index = 0 + self.generate_data = {'tasks_id': self.tasks_id, 'status': 'PENDING', 'message': "pending", 'data': ''} + self.redis_client.set(self.tasks_id, json.dumps(self.generate_data)) + self.redis_client.expire(self.tasks_id, 600) def get_image(self, image_url): # Get data of an object. @@ -60,24 +63,23 @@ class GenerateImage: image_cv2 = np.random.randint(0, 256, (1024, 1024, 3), dtype=np.uint8) return image_cv2 - def __call__(self, *args, **kwargs): - self.generate_data = json.dumps({'status': 'PENDING', 'message': "pending", 'data': ''}) - self.redis_client.set(self.tasks_id, self.generate_data) - self.redis_client.expire(self.tasks_id, 600) - def callback(self, result, error): if error: - generate_data = json.dumps({'status': 'FAILURE', 'message': f"{error}", 'data': f"{error}"}) - self.redis_client.set(self.tasks_id, generate_data) + self.generate_data['status'] = "FAILURE" + self.generate_data['message'] = str(error) + self.generate_data['data'] = str(error) + self.redis_client.set(self.tasks_id, json.dumps(self.generate_data)) else: image_result = result.as_numpy("generated_image")[0] image_url = upload_png_sd(image_result, user_id=self.user_id, category=f"{self.category}", object_name=f"{self.tasks_id}.png") - generate_data = json.dumps({'status': 'SUCCESS', 'message': 'success', 'data': f'{image_url}'}) - self.redis_client.set(self.tasks_id, generate_data) + self.generate_data['status'] = "SUCCESS" + self.generate_data['message'] = "success" + self.generate_data['data'] = str(image_url) + self.redis_client.set(self.tasks_id, json.dumps(self.generate_data)) def read_tasks_status(self): - status_data = json.loads(self.redis_client.get(self.tasks_id)) - return status_data + status_data = self.redis_client.get(self.tasks_id) + return json.loads(status_data), status_data def infer(self, inputs): return self.grpc_client.async_infer( @@ -87,45 +89,53 @@ class GenerateImage: ) def get_result(self): - prompts = [self.prompt] * self.batch_size - modes = [self.mode] * self.batch_size - images = [self.image.astype(np.float16)] * self.batch_size + try: + prompts = [self.prompt] * self.batch_size + modes = [self.mode] * self.batch_size + images = [self.image.astype(np.float16)] * self.batch_size - text_obj = np.array(prompts, dtype="object").reshape((-1, 1)) - mode_obj = np.array(modes, dtype="object").reshape((-1, 1)) - image_obj = np.array(images, dtype=np.float16).reshape((-1, 1024, 1024, 3)) + text_obj = np.array(prompts, dtype="object").reshape((-1, 1)) + mode_obj = np.array(modes, dtype="object").reshape((-1, 1)) + image_obj = np.array(images, dtype=np.float16).reshape((-1, 1024, 1024, 3)) - input_text = grpcclient.InferInput("prompt", text_obj.shape, np_to_triton_dtype(text_obj.dtype)) - input_image = grpcclient.InferInput("input_image", image_obj.shape, "FP16") - input_mode = grpcclient.InferInput("mode", mode_obj.shape, np_to_triton_dtype(text_obj.dtype)) + input_text = grpcclient.InferInput("prompt", text_obj.shape, np_to_triton_dtype(text_obj.dtype)) + input_image = grpcclient.InferInput("input_image", image_obj.shape, "FP16") + input_mode = grpcclient.InferInput("mode", mode_obj.shape, np_to_triton_dtype(text_obj.dtype)) - input_text.set_data_from_numpy(text_obj) - input_image.set_data_from_numpy(image_obj) - input_mode.set_data_from_numpy(mode_obj) + input_text.set_data_from_numpy(text_obj) + input_image.set_data_from_numpy(image_obj) + input_mode.set_data_from_numpy(mode_obj) - inputs = [input_text, input_image, input_mode] - ctx = self.infer(inputs) - time_out = 60 - while time_out > 0: - generate_data = self.read_tasks_status() - if generate_data['status'] in ["REVOKED", "FAILURE"]: - ctx.cancel() - self.channel.basic_publish(exchange='', routing_key=GI_RABBITMQ_QUEUES, body=json.dumps(generate_data)) - logger.info(f" [x] Sent {json.dumps(generate_data, indent=4)}") - break - elif generate_data['status'] == "SUCCESS": - self.channel.basic_publish(exchange='', routing_key=GI_RABBITMQ_QUEUES, body=json.dumps(generate_data)) - logger.info(f" [x] Sent {json.dumps(generate_data, indent=4)}") - break - time_out -= 1 - time.sleep(0.1) - return self.read_tasks_status() + inputs = [input_text, input_image, input_mode] + ctx = self.infer(inputs) + time_out = 60 + generate_data = None + while time_out > 0: + generate_data, _ = self.read_tasks_status() + if generate_data['status'] in ["REVOKED", "FAILURE"]: + ctx.cancel() + break + elif generate_data['status'] == "SUCCESS": + break + time_out -= 1 + time.sleep(0.1) + return generate_data + except Exception as e: + self.generate_data['status'] = "FAILURE" + self.generate_data['message'] = "failure" + self.generate_data['data'] = str(e) + self.redis_client.set(self.tasks_id, json.dumps(self.generate_data)) + raise Exception(str(e)) + finally: + dict_generate_data, str_generate_data = self.read_tasks_status() + self.channel.basic_publish(exchange='', routing_key=GI_RABBITMQ_QUEUES, body=str_generate_data) + logger.info(f" [x] Sent {json.dumps(dict_generate_data, indent=4)}") def infer_cancel(tasks_id): redis_client = redis.StrictRedis(host=REDIS_HOST, port=REDIS_PORT, db=REDIS_DB, decode_responses=True) - data = {'status': 'REVOKED', 'message': "revoked", 'data': 'revoked'} - generate_data = json.dumps({'status': 'REVOKED', 'message': "revoked", 'data': 'revoked'}) + data = {'tasks_id': tasks_id, 'status': 'REVOKED', 'message': "revoked", 'data': 'revoked'} + generate_data = json.dumps(data) redis_client.set(tasks_id, generate_data) return data From 36447f934e86beb841a630c98871245f6218bbc1 Mon Sep 17 00:00:00 2001 From: zhouchengrong Date: Wed, 17 Apr 2024 17:39:10 +0800 Subject: [PATCH 12/19] =?UTF-8?q?feat=20generate=20image=20=E9=80=BB?= =?UTF-8?q?=E8=BE=91=E8=A1=A5=E5=85=85?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- app/core/config.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/app/core/config.py b/app/core/config.py index 73e6e67..4889db8 100644 --- a/app/core/config.py +++ b/app/core/config.py @@ -19,7 +19,7 @@ class Settings(BaseSettings): LOGGING_CONFIG_FILE = os.path.join(BASE_DIR, 'logging_env.py') -DEBUG = True +DEBUG = False if DEBUG: LOGS_PATH = "logs/" CATEGORY_PATH = "service/attribute/config/descriptor/category/category_dis.csv" From 74481a3736c59899db31477293cdc9491f911e5e Mon Sep 17 00:00:00 2001 From: zhouchengrong Date: Thu, 18 Apr 2024 13:26:59 +0800 Subject: [PATCH 13/19] =?UTF-8?q?feat=20generate=20image=20=E9=80=BB?= =?UTF-8?q?=E8=BE=91=E8=A1=A5=E5=85=85?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- docker-compose.yml | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/docker-compose.yml b/docker-compose.yml index a550e44..1072cf9 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -1,9 +1,9 @@ version: "3" services: - trinity_aida_local: + trinity_aida_dev: build: . - container_name: trinity_aida_local + container_name: trinity_aida_dev volumes: - ./trinity_client_aida:/trinity ports: - - "10200:4562" \ No newline at end of file + - "10201:4562" \ No newline at end of file From 0460568f7608ced6a8f88f389a206c48061e70bd Mon Sep 17 00:00:00 2001 From: zhouchengrong Date: Thu, 18 Apr 2024 13:27:26 +0800 Subject: [PATCH 14/19] =?UTF-8?q?feat=20generate=20image=20=E9=80=BB?= =?UTF-8?q?=E8=BE=91=E8=A1=A5=E5=85=85?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- docker-compose.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/docker-compose.yml b/docker-compose.yml index 1072cf9..e0d001d 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -6,4 +6,4 @@ services: volumes: - ./trinity_client_aida:/trinity ports: - - "10201:4562" \ No newline at end of file + - "10200:4562" \ No newline at end of file From 64141a88a2918d4f30edbef075572fac71b1f7d3 Mon Sep 17 00:00:00 2001 From: zhouchengrong Date: Mon, 22 Apr 2024 10:52:03 +0800 Subject: [PATCH 15/19] =?UTF-8?q?feat=20=E6=96=B0=E5=A2=9Eredis=20?= =?UTF-8?q?=E8=BF=87=E6=9C=9F=E6=97=B6=E9=97=B4?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- app/service/super_resolution/service.py | 1 + 1 file changed, 1 insertion(+) diff --git a/app/service/super_resolution/service.py b/app/service/super_resolution/service.py index 95b2811..e87f1a7 100644 --- a/app/service/super_resolution/service.py +++ b/app/service/super_resolution/service.py @@ -26,6 +26,7 @@ class SuperResolution: self.sr_xn = data.sr_xn self.minio_client = Minio(MINIO_URL, access_key=MINIO_ACCESS, secret_key=MINIO_SECRET, secure=MINIO_SECURE) self.redis_client.set(self.tasks_id, json.dumps({'status': 'PENDING', 'message': "pending", 'data': ''})) + self.redis_client.expire(self.tasks_id, 600) self.connection = pika.BlockingConnection(pika.ConnectionParameters(**RABBITMQ_PARAMS)) self.channel = self.connection.channel() From 5d009735d4b8cc8afa109a6bbd6368d9d7b0b6db Mon Sep 17 00:00:00 2001 From: zhouchengrong Date: Mon, 22 Apr 2024 11:16:11 +0800 Subject: [PATCH 16/19] feat --- .gitignore | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/.gitignore b/.gitignore index 7b16d72..1bf82fb 100644 --- a/.gitignore +++ b/.gitignore @@ -125,7 +125,9 @@ seg_result/ seg_result *.png uwsgi -#*.yaml +*.yaml +*.yml +Dockerfile .conf app/logs From 0e601fdb50d5033fd8b37e17bcb14116451ebe78 Mon Sep 17 00:00:00 2001 From: zchengrong <124802516+zchengrong@users.noreply.github.com> Date: Mon, 22 Apr 2024 11:16:42 +0800 Subject: [PATCH 17/19] Delete docker-compose.yml --- docker-compose.yml | 9 --------- 1 file changed, 9 deletions(-) delete mode 100644 docker-compose.yml diff --git a/docker-compose.yml b/docker-compose.yml deleted file mode 100644 index e0d001d..0000000 --- a/docker-compose.yml +++ /dev/null @@ -1,9 +0,0 @@ -version: "3" -services: - trinity_aida_dev: - build: . - container_name: trinity_aida_dev - volumes: - - ./trinity_client_aida:/trinity - ports: - - "10200:4562" \ No newline at end of file From b51da2de17c49fd8cb5360da187d58645b049eb3 Mon Sep 17 00:00:00 2001 From: zchengrong <124802516+zchengrong@users.noreply.github.com> Date: Mon, 22 Apr 2024 11:16:51 +0800 Subject: [PATCH 18/19] Delete Dockerfile --- Dockerfile | 23 ----------------------- 1 file changed, 23 deletions(-) delete mode 100644 Dockerfile diff --git a/Dockerfile b/Dockerfile deleted file mode 100644 index c1e42d1..0000000 --- a/Dockerfile +++ /dev/null @@ -1,23 +0,0 @@ -FROM python:3.9 -ENV TZ=Asia/Shanghai -RUN apt update -RUN apt install -y vim -RUN apt install -y libgl1-mesa-glx -COPY ./requirements.txt /requirements.txt -RUN pip install --upgrade pip -RUN pip install -r requirements.txt -RUN mkdir -p app/logs -RUN pip install gunicorn -RUN pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu118 -RUN #pip install mmcv==1.4.2 -f https://download.openmmlab.com/mmcv/dist/cu117/torch1.13/index.html - -WORKDIR /app -COPY . . -ENV FLASK_APP=manage.py -LABEL maintainer="zchengrong@yeah.net" \ - description="My Python 3.9 - trinity aida " \ - version="1.0" \ - name="trinity_aida" - - -CMD ["gunicorn", "-c", "gunicorn_config.py", "app.main:app" , "-e", "SR_RABBITMQ_QUEUES=SuperResolution-dev" ,"-e", "GI_RABBITMQ_QUEUES=GenerateImage-dev"] \ No newline at end of file From 25fdcdb6b09b3f581748430810b9a5aa5fea15c5 Mon Sep 17 00:00:00 2001 From: zhouchengrong Date: Mon, 22 Apr 2024 11:44:17 +0800 Subject: [PATCH 19/19] feat --- app/core/config.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/app/core/config.py b/app/core/config.py index b88390c..4889db8 100644 --- a/app/core/config.py +++ b/app/core/config.py @@ -28,8 +28,8 @@ else: CATEGORY_PATH = "app/service/attribute/config/descriptor/category/category_dis.csv" # RABBITMQ_ENV = "" # 生产环境 -RABBITMQ_ENV = "-dev" # 开发环境 -# RABBITMQ_ENV = "-local" # 本地测试环境 +# RABBITMQ_ENV = "-dev" # 开发环境 +RABBITMQ_ENV = "-local" # 本地测试环境 settings = Settings()