G2E prompt template adjustment

This commit is contained in:
gdw6463
2024-05-17 18:29:39 +08:00
parent cc44574ffb
commit b0b24ff6c2

View File

@ -19,7 +19,7 @@ class G2E(Blackbox):
return isinstance(data, list) return isinstance(data, list)
# model_name有 Qwen1.5-14B-Chat , internlm2-chat-20b # model_name有 Qwen1.5-14B-Chat , internlm2-chat-20b
def processing(self, model_name, prompt, template, context: list) -> str: def processing(self, model_name, prompt, template, context: list) -> str:
if context == None: if context == None:
context = [] context = []
#url = 'http://120.196.116.194:48890/v1' #url = 'http://120.196.116.194:48890/v1'
@ -51,14 +51,13 @@ class G2E(Blackbox):
# {"role": "system", "content": background_prompt + prompt1}, # {"role": "system", "content": background_prompt + prompt1},
#] #]
prompt_template = [ prompt_template = [
{"role": "system", "content": prompt1}, {"role": "system", "content": ''}
] ]
messages = prompt_template + context + [ messages = prompt_template + context + [
{ {
"role": "user", "role": "user",
"content": prompt + inject_prompt "content": prompt
} }
] ]
print("**** History with current prompt input ****") print("**** History with current prompt input ****")
@ -75,10 +74,10 @@ class G2E(Blackbox):
model=model_name, model=model_name,
messages=messages, messages=messages,
temperature=0.8, temperature=0.8,
top_p=0.8 top_p=0.8,
#frequency_penalty=0.5, frequency_penalty=0.5,
#presence_penalty=0.8, presence_penalty=0.8,
#stop=100 stop=100
) )
fastchat_content = response.choices[0].message.content fastchat_content = response.choices[0].message.content