|
|
@@ -277,8 +277,6 @@ class OllamaTranslator(BaseTranslator):
|
|
|
self.client = ollama.Client()
|
|
|
self.prompttext = prompt
|
|
|
self.add_cache_impact_parameters("temperature", self.options["temperature"])
|
|
|
- if prompt:
|
|
|
- self.add_cache_impact_parameters("prompt", prompt.template)
|
|
|
|
|
|
def do_translate(self, text):
|
|
|
maxlen = max(2000, len(text) * 5)
|
|
|
@@ -320,8 +318,6 @@ class XinferenceTranslator(BaseTranslator):
|
|
|
self.client = xinference_client.RESTfulClient(self.envs["XINFERENCE_HOST"])
|
|
|
self.prompttext = prompt
|
|
|
self.add_cache_impact_parameters("temperature", self.options["temperature"])
|
|
|
- if prompt:
|
|
|
- self.add_cache_impact_parameters("prompt", prompt.template)
|
|
|
|
|
|
def do_translate(self, text):
|
|
|
maxlen = max(2000, len(text) * 5)
|
|
|
@@ -384,8 +380,6 @@ class OpenAITranslator(BaseTranslator):
|
|
|
)
|
|
|
self.prompttext = prompt
|
|
|
self.add_cache_impact_parameters("temperature", self.options["temperature"])
|
|
|
- if prompt:
|
|
|
- self.add_cache_impact_parameters("prompt", prompt.template)
|
|
|
|
|
|
def do_translate(self, text) -> str:
|
|
|
response = self.client.chat.completions.create(
|
|
|
@@ -438,8 +432,6 @@ class AzureOpenAITranslator(BaseTranslator):
|
|
|
)
|
|
|
self.prompttext = prompt
|
|
|
self.add_cache_impact_parameters("temperature", self.options["temperature"])
|
|
|
- if prompt:
|
|
|
- self.add_cache_impact_parameters("prompt", prompt.template)
|
|
|
|
|
|
def do_translate(self, text) -> str:
|
|
|
response = self.client.chat.completions.create(
|
|
|
@@ -476,8 +468,6 @@ class ModelScopeTranslator(OpenAITranslator):
|
|
|
model = self.envs["MODELSCOPE_MODEL"]
|
|
|
super().__init__(lang_in, lang_out, model, base_url=base_url, api_key=api_key)
|
|
|
self.prompttext = prompt
|
|
|
- if prompt:
|
|
|
- self.add_cache_impact_parameters("prompt", prompt.template)
|
|
|
|
|
|
|
|
|
class ZhipuTranslator(OpenAITranslator):
|
|
|
@@ -497,8 +487,6 @@ class ZhipuTranslator(OpenAITranslator):
|
|
|
model = self.envs["ZHIPU_MODEL"]
|
|
|
super().__init__(lang_in, lang_out, model, base_url=base_url, api_key=api_key)
|
|
|
self.prompttext = prompt
|
|
|
- if prompt:
|
|
|
- self.add_cache_impact_parameters("prompt", prompt.template)
|
|
|
|
|
|
def do_translate(self, text) -> str:
|
|
|
try:
|
|
|
@@ -534,8 +522,6 @@ class SiliconTranslator(OpenAITranslator):
|
|
|
model = self.envs["SILICON_MODEL"]
|
|
|
super().__init__(lang_in, lang_out, model, base_url=base_url, api_key=api_key)
|
|
|
self.prompttext = prompt
|
|
|
- if prompt:
|
|
|
- self.add_cache_impact_parameters("prompt", prompt.template)
|
|
|
|
|
|
|
|
|
class GeminiTranslator(OpenAITranslator):
|
|
|
@@ -555,8 +541,6 @@ class GeminiTranslator(OpenAITranslator):
|
|
|
model = self.envs["GEMINI_MODEL"]
|
|
|
super().__init__(lang_in, lang_out, model, base_url=base_url, api_key=api_key)
|
|
|
self.prompttext = prompt
|
|
|
- if prompt:
|
|
|
- self.add_cache_impact_parameters("prompt", prompt.template)
|
|
|
|
|
|
|
|
|
class AzureTranslator(BaseTranslator):
|
|
|
@@ -634,8 +618,6 @@ class AnythingLLMTranslator(BaseTranslator):
|
|
|
"Content-Type": "application/json",
|
|
|
}
|
|
|
self.prompttext = prompt
|
|
|
- if prompt:
|
|
|
- self.add_cache_impact_parameters("prompt", prompt.template)
|
|
|
|
|
|
def do_translate(self, text):
|
|
|
messages = self.prompt(text, self.prompttext)
|
|
|
@@ -752,8 +734,6 @@ class GorkTranslator(OpenAITranslator):
|
|
|
model = self.envs["GORK_MODEL"]
|
|
|
super().__init__(lang_in, lang_out, model, base_url=base_url, api_key=api_key)
|
|
|
self.prompttext = prompt
|
|
|
- if prompt:
|
|
|
- self.add_cache_impact_parameters("prompt", prompt.template)
|
|
|
|
|
|
|
|
|
class GroqTranslator(OpenAITranslator):
|
|
|
@@ -772,8 +752,6 @@ class GroqTranslator(OpenAITranslator):
|
|
|
model = self.envs["GROQ_MODEL"]
|
|
|
super().__init__(lang_in, lang_out, model, base_url=base_url, api_key=api_key)
|
|
|
self.prompttext = prompt
|
|
|
- if prompt:
|
|
|
- self.add_cache_impact_parameters("prompt", prompt.template)
|
|
|
|
|
|
|
|
|
class DeepseekTranslator(OpenAITranslator):
|
|
|
@@ -792,8 +770,6 @@ class DeepseekTranslator(OpenAITranslator):
|
|
|
model = self.envs["DEEPSEEK_MODEL"]
|
|
|
super().__init__(lang_in, lang_out, model, base_url=base_url, api_key=api_key)
|
|
|
self.prompttext = prompt
|
|
|
- if prompt:
|
|
|
- self.add_cache_impact_parameters("prompt", prompt.template)
|
|
|
|
|
|
|
|
|
class OpenAIlikedTranslator(OpenAITranslator):
|
|
|
@@ -822,5 +798,3 @@ class OpenAIlikedTranslator(OpenAITranslator):
|
|
|
api_key = self.envs["OPENAILIKED_API_KEY"]
|
|
|
super().__init__(lang_in, lang_out, model, base_url=base_url, api_key=api_key)
|
|
|
self.prompttext = prompt
|
|
|
- if prompt:
|
|
|
- self.add_cache_impact_parameters("prompt", prompt.template)
|