feat translator 切换ollama
fix
This commit is contained in:
@@ -6,7 +6,7 @@ from fastapi import APIRouter, HTTPException
|
|||||||
|
|
||||||
from app.schemas.prompt_generation import PromptGenerationImageModel
|
from app.schemas.prompt_generation import PromptGenerationImageModel
|
||||||
from app.schemas.response_template import ResponseModel
|
from app.schemas.response_template import ResponseModel
|
||||||
from app.service.prompt_generation.chatgpt_for_translation import translate_to_en
|
from app.service.prompt_generation.chatgpt_for_translation import translate_to_en, get_translation_from_llama3
|
||||||
|
|
||||||
router = APIRouter()
|
router = APIRouter()
|
||||||
logger = logging.getLogger()
|
logger = logging.getLogger()
|
||||||
@@ -26,7 +26,7 @@ def prompt_generation(request_data: PromptGenerationImageModel):
|
|||||||
"""
|
"""
|
||||||
try:
|
try:
|
||||||
logger.info(f"prompt_generation request item is : @@@@@@:{request_data}")
|
logger.info(f"prompt_generation request item is : @@@@@@:{request_data}")
|
||||||
data = translate_to_en("[" + request_data.text + "]")
|
data = get_translation_from_llama3("[" + request_data.text + "]")
|
||||||
logger.info(f"prompt_generation response @@@@@@:{data}")
|
logger.info(f"prompt_generation response @@@@@@:{data}")
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.warning(f"prompt_generation Run Exception @@@@@@:{e}")
|
logger.warning(f"prompt_generation Run Exception @@@@@@:{e}")
|
||||||
|
|||||||
@@ -1,4 +1,6 @@
|
|||||||
import json
|
import json
|
||||||
|
import logging
|
||||||
|
import time
|
||||||
|
|
||||||
import requests
|
import requests
|
||||||
from dashscope import Generation
|
from dashscope import Generation
|
||||||
@@ -7,6 +9,8 @@ from retry import retry
|
|||||||
|
|
||||||
from app.core.config import QWEN_API_KEY
|
from app.core.config import QWEN_API_KEY
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
# os.environ["http_proxy"] = "http://127.0.0.1:7890"
|
# os.environ["http_proxy"] = "http://127.0.0.1:7890"
|
||||||
# os.environ["https_proxy"] = "http://127.0.0.1:7890"
|
# os.environ["https_proxy"] = "http://127.0.0.1:7890"
|
||||||
@@ -83,6 +87,7 @@ def get_response(messages):
|
|||||||
|
|
||||||
|
|
||||||
def get_translation_from_llama3(text):
|
def get_translation_from_llama3(text):
|
||||||
|
start_time = time.time()
|
||||||
url = "http://10.1.1.240:11434/api/generate"
|
url = "http://10.1.1.240:11434/api/generate"
|
||||||
# url = "http://10.1.1.240:1143/api/generate"
|
# url = "http://10.1.1.240:1143/api/generate"
|
||||||
|
|
||||||
@@ -98,15 +103,16 @@ def get_translation_from_llama3(text):
|
|||||||
# 将负载转换为 JSON 格式
|
# 将负载转换为 JSON 格式
|
||||||
headers = {'Content-Type': 'application/json'}
|
headers = {'Content-Type': 'application/json'}
|
||||||
response = requests.post(url, data=json.dumps(payload), headers=headers)
|
response = requests.post(url, data=json.dumps(payload), headers=headers)
|
||||||
|
|
||||||
# 处理响应
|
# 处理响应
|
||||||
if response.status_code == 200:
|
if response.status_code == 200:
|
||||||
# print("Response from server:")
|
# print("Response from server:")
|
||||||
# print(response.json())
|
# print(response.json())
|
||||||
resp = json.loads(response.content).get("response")
|
resp = json.loads(response.content).get("response")
|
||||||
|
logger.info(f"translation server runtime is {time.time() - start_time} , response is {resp}")
|
||||||
print("input : {}, translate result : {}".format(text, resp))
|
print("input : {}, translate result : {}".format(text, resp))
|
||||||
return resp
|
return resp
|
||||||
else:
|
else:
|
||||||
|
logger.info(f"translation server runtime is {time.time() - start_time} , response is {response.content}")
|
||||||
print(f"Request failed with status code {response.status_code}")
|
print(f"Request failed with status code {response.status_code}")
|
||||||
print(response.text)
|
print(response.text)
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user