G2E prompt template adjustment

This commit is contained in:
gdw6463
2024-05-17 18:29:39 +08:00
parent cc44574ffb
commit b0b24ff6c2

View File

@ -19,7 +19,7 @@ class G2E(Blackbox):
return isinstance(data, list)
# model_name有 Qwen1.5-14B-Chat , internlm2-chat-20b
def processing(self, model_name, prompt, template, context: list) -> str:
def processing(self, model_name, prompt, template, context: list) -> str:
if context == None:
context = []
#url = 'http://120.196.116.194:48890/v1'
@ -51,14 +51,13 @@ class G2E(Blackbox):
# {"role": "system", "content": background_prompt + prompt1},
#]
prompt_template = [
{"role": "system", "content": prompt1},
{"role": "system", "content": ''}
]
messages = prompt_template + context + [
{
"role": "user",
"content": prompt + inject_prompt
"content": prompt
}
]
print("**** History with current prompt input ****")
@ -75,10 +74,10 @@ class G2E(Blackbox):
model=model_name,
messages=messages,
temperature=0.8,
top_p=0.8
#frequency_penalty=0.5,
#presence_penalty=0.8,
#stop=100
top_p=0.8,
frequency_penalty=0.5,
presence_penalty=0.8,
stop=100
)
fastchat_content = response.choices[0].message.content