From 5491c54bda681448fe93632e4898de6af82c58d4 Mon Sep 17 00:00:00 2001 From: zhouchengrong Date: Mon, 2 Dec 2024 20:31:46 +0800 Subject: [PATCH] =?UTF-8?q?feat=20=20translator=20=E5=88=87=E6=8D=A2ollama?= =?UTF-8?q?=20fix?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- app/api/api_prompt_generation.py | 4 ++-- app/service/prompt_generation/chatgpt_for_translation.py | 8 +++++++- 2 files changed, 9 insertions(+), 3 deletions(-) diff --git a/app/api/api_prompt_generation.py b/app/api/api_prompt_generation.py index 59e5779..11733e8 100644 --- a/app/api/api_prompt_generation.py +++ b/app/api/api_prompt_generation.py @@ -6,7 +6,7 @@ from fastapi import APIRouter, HTTPException from app.schemas.prompt_generation import PromptGenerationImageModel from app.schemas.response_template import ResponseModel -from app.service.prompt_generation.chatgpt_for_translation import translate_to_en +from app.service.prompt_generation.chatgpt_for_translation import translate_to_en, get_translation_from_llama3 router = APIRouter() logger = logging.getLogger() @@ -26,7 +26,7 @@ def prompt_generation(request_data: PromptGenerationImageModel): """ try: logger.info(f"prompt_generation request item is : @@@@@@:{request_data}") - data = translate_to_en("[" + request_data.text + "]") + data = get_translation_from_llama3("[" + request_data.text + "]") logger.info(f"prompt_generation response @@@@@@:{data}") except Exception as e: logger.warning(f"prompt_generation Run Exception @@@@@@:{e}") diff --git a/app/service/prompt_generation/chatgpt_for_translation.py b/app/service/prompt_generation/chatgpt_for_translation.py index 5d720b9..e541781 100644 --- a/app/service/prompt_generation/chatgpt_for_translation.py +++ b/app/service/prompt_generation/chatgpt_for_translation.py @@ -1,4 +1,6 @@ import json +import logging +import time import requests from dashscope import Generation @@ -7,6 +9,8 @@ from retry import retry from app.core.config import QWEN_API_KEY +logger = logging.getLogger(__name__) + # os.environ["http_proxy"] = "http://127.0.0.1:7890" # os.environ["https_proxy"] = "http://127.0.0.1:7890" @@ -83,6 +87,7 @@ def get_response(messages): def get_translation_from_llama3(text): + start_time = time.time() url = "http://10.1.1.240:11434/api/generate" # url = "http://10.1.1.240:1143/api/generate" @@ -98,15 +103,16 @@ def get_translation_from_llama3(text): # 将负载转换为 JSON 格式 headers = {'Content-Type': 'application/json'} response = requests.post(url, data=json.dumps(payload), headers=headers) - # 处理响应 if response.status_code == 200: # print("Response from server:") # print(response.json()) resp = json.loads(response.content).get("response") + logger.info(f"translation server runtime is {time.time() - start_time} , response is {resp}") print("input : {}, translate result : {}".format(text, resp)) return resp else: + logger.info(f"translation server runtime is {time.time() - start_time} , response is {response.content}") print(f"Request failed with status code {response.status_code}") print(response.text)