222 lines
7.5 KiB
Python
222 lines
7.5 KiB
Python
|
"""
|
|||
|
This script is an example of using the OpenAI API to create various interactions with a ChatGLM3 model.
|
|||
|
It includes functions to:
|
|||
|
|
|||
|
1. Conduct a basic chat session, asking about weather conditions in multiple cities.
|
|||
|
2. Initiate a simple chat in Chinese, asking the model to tell a short story.
|
|||
|
3. Retrieve and print embeddings for a given text input.
|
|||
|
|
|||
|
Each function demonstrates a different aspect of the API's capabilities, showcasing how to make requests
|
|||
|
and handle responses.
|
|||
|
"""
|
|||
|
|
|||
|
from openai import OpenAI
|
|||
|
|
|||
|
base_url = "http://127.0.0.1:8000/v1/"
|
|||
|
client = OpenAI(api_key="EMPTY", base_url=base_url)
|
|||
|
|
|||
|
|
|||
|
def function_chat():
|
|||
|
messages = [{"role": "user", "content": "What's the weather like in San Francisco, Tokyo, and Paris?"}]
|
|||
|
tools = [
|
|||
|
{
|
|||
|
"type": "function",
|
|||
|
"function": {
|
|||
|
"name": "get_current_weather",
|
|||
|
"description": "Get the current weather in a given location",
|
|||
|
"parameters": {
|
|||
|
"type": "object",
|
|||
|
"properties": {
|
|||
|
"location": {
|
|||
|
"type": "string",
|
|||
|
"description": "The city and state, e.g. San Francisco, CA",
|
|||
|
},
|
|||
|
"unit": {"type": "string", "enum": ["celsius", "fahrenheit"]},
|
|||
|
},
|
|||
|
"required": ["location"],
|
|||
|
},
|
|||
|
},
|
|||
|
}
|
|||
|
]
|
|||
|
|
|||
|
response = client.chat.completions.create(
|
|||
|
model="chatglm3",
|
|||
|
messages=messages,
|
|||
|
tools=tools,
|
|||
|
tool_choice="auto",
|
|||
|
)
|
|||
|
if response:
|
|||
|
content = response.choices[0].message.content
|
|||
|
print(content)
|
|||
|
else:
|
|||
|
print("Error:", response.status_code)
|
|||
|
|
|||
|
|
|||
|
def chat(text):
|
|||
|
# 定义API请求的数据
|
|||
|
data = {
|
|||
|
"model": "chatglm3-6b",
|
|||
|
"prompt": text,
|
|||
|
"temperature": 0.5, # 控制输出结果的随机性,范围从0.0到1.0,越高越随机
|
|||
|
"max_tokens": 75, # 控制输出文本的长度
|
|||
|
"top_p": 1, # 一个更复杂的参数,与temperature类似但更加精细控制
|
|||
|
"n": 1, # 要返回的最完整的文本段落数
|
|||
|
"stream": False # 是否以流的形式返回输出
|
|||
|
}
|
|||
|
# 发送API请求
|
|||
|
response = client.chat.completions.create(**data)
|
|||
|
# 打印响应结果
|
|||
|
print(response.get("choices")[0]["text"])
|
|||
|
|
|||
|
def chat2(text):
|
|||
|
messages = [
|
|||
|
{
|
|||
|
"role": "user",
|
|||
|
"content": text
|
|||
|
}
|
|||
|
]
|
|||
|
response = client.chat.completions.create(
|
|||
|
model="chatglm3-6b",
|
|||
|
prompt=messages,
|
|||
|
stream=False,
|
|||
|
max_tokens=256,
|
|||
|
temperature=0.8,
|
|||
|
presence_penalty=1.1,
|
|||
|
top_p=0.8)
|
|||
|
if response:
|
|||
|
if False:
|
|||
|
for chunk in response:
|
|||
|
print(chunk.choices[0].delta.content)
|
|||
|
else:
|
|||
|
content = response.choices[0].message.content
|
|||
|
print(content)
|
|||
|
else:
|
|||
|
print("Error:", response.status_code)
|
|||
|
|
|||
|
def simple_chat(use_stream=True):
|
|||
|
messages = [
|
|||
|
{
|
|||
|
"role": "system",
|
|||
|
"content": "You are ChatGLM3, a large language model trained by Zhipu.AI. Follow the user's "
|
|||
|
"instructions carefully. Respond using markdown.",
|
|||
|
},
|
|||
|
{
|
|||
|
"role": "user",
|
|||
|
"content": "你好,请你用生动的话语给我讲一个猫和狗的小故事吧"
|
|||
|
}
|
|||
|
]
|
|||
|
response = client.chat.completions.create(
|
|||
|
model="chatglm3-6b",
|
|||
|
messages=messages,
|
|||
|
stream=use_stream,
|
|||
|
max_tokens=256,
|
|||
|
temperature=0.8,
|
|||
|
presence_penalty=1.1,
|
|||
|
top_p=0.8)
|
|||
|
if response:
|
|||
|
if use_stream:
|
|||
|
for chunk in response:
|
|||
|
print(chunk.choices[0].delta.content)
|
|||
|
else:
|
|||
|
content = response.choices[0].message.content
|
|||
|
print(content)
|
|||
|
else:
|
|||
|
print("Error:", response.status_code)
|
|||
|
def chat3(text):
|
|||
|
history = [['你好', '你好,有什么帮到你呢?'],['你好,给我讲一个七仙女的故事,大概20字', '七个仙女下凡,来到人间,遇见了王子,经历了许多冒险和考验,最终爱情获胜']]
|
|||
|
messages=[]
|
|||
|
if history is not None:
|
|||
|
for string in history:
|
|||
|
# 打印字符串
|
|||
|
print(string)
|
|||
|
# for his in string:
|
|||
|
# print(his)
|
|||
|
i = 0
|
|||
|
for his in string:
|
|||
|
print(his)
|
|||
|
if i==0:
|
|||
|
dialogue={
|
|||
|
"role": "user",
|
|||
|
"content": his
|
|||
|
}
|
|||
|
elif i==1:
|
|||
|
dialogue={
|
|||
|
"role": "assistant",
|
|||
|
"content": his
|
|||
|
}
|
|||
|
messages.append(dialogue)
|
|||
|
i = 1
|
|||
|
current = {
|
|||
|
"role": "user",
|
|||
|
"content": text
|
|||
|
}
|
|||
|
messages.append(current)
|
|||
|
print("===============messages=========================")
|
|||
|
print(messages)
|
|||
|
print("===============messages=========================")
|
|||
|
# messages = [
|
|||
|
|
|||
|
# {
|
|||
|
# "role": "user",
|
|||
|
# "content": text
|
|||
|
# }
|
|||
|
# ]
|
|||
|
response = client.chat.completions.create(
|
|||
|
model="chatglm3-6b",
|
|||
|
messages=messages,
|
|||
|
stream=False,
|
|||
|
max_tokens=256,
|
|||
|
temperature=0.8,
|
|||
|
presence_penalty=1.1,
|
|||
|
top_p=0.8)
|
|||
|
if response:
|
|||
|
if False:
|
|||
|
for chunk in response:
|
|||
|
print(chunk.choices[0].delta.content)
|
|||
|
else:
|
|||
|
content = response.choices[0].message.content
|
|||
|
print(content)
|
|||
|
else:
|
|||
|
print("Error:", response.status_code)
|
|||
|
|
|||
|
|
|||
|
|
|||
|
def embedding():
|
|||
|
response = client.embeddings.create(
|
|||
|
model="bge-large-zh-1.5",
|
|||
|
input=["你好,给我讲一个故事,大概100字"],
|
|||
|
)
|
|||
|
embeddings = response.data[0].embedding
|
|||
|
print("嵌入完成,维度:", len(embeddings))
|
|||
|
|
|||
|
|
|||
|
if __name__ == "__main__":
|
|||
|
chat3("你好,给我讲楚汉相争的故事,大概20字")
|
|||
|
# simple_chat(use_stream=False)
|
|||
|
# simple_chat(use_stream=True)
|
|||
|
# embedding()
|
|||
|
# function_chat()
|
|||
|
|
|||
|
# curl -X POST "http://127.0.0.1:8000/v1/chat/completions" \
|
|||
|
# -H "Content-Type: application/json" \
|
|||
|
# -d "{\"model\": \"chatglm3-6b\", \"messages\": [{\"role\": \"system\", \"content\": \"You are ChatGLM3, a large language model trained by Zhipu.AI. Follow the user's instructions carefully. Respond using markdown.\"}, {\"role\": \"user\", \"content\": \"你好,给我讲一个故事,大概100字\"}], \"stream\": false, \"max_tokens\": 100, \"temperature\": 0.8, \"top_p\": 0.8}"
|
|||
|
|
|||
|
# curl -X POST "http://127.0.0.1:8000/v1/completions" \
|
|||
|
# -H 'Content-Type: application/json' \
|
|||
|
# -d '{"prompt": "请用20字内回复我.你今年多大了", "history": []}'
|
|||
|
|
|||
|
# curl -X POST "http://127.0.0.1:8000/v1/completions" \
|
|||
|
# -H 'Content-Type: application/json' \
|
|||
|
# -d '{"prompt": "请用20字内回复我.你今年多大了", "history": [{"你好","你好👋!我是人工智能助手 ChatGLM-6B,很高兴见到你,欢迎问我任何问题。"}]}'
|
|||
|
|
|||
|
|
|||
|
# curl -X POST "http://127.0.0.1:8000/v1/completions" \
|
|||
|
# -H 'Content-Type: application/json' \
|
|||
|
# -d '{"prompt": "请用20字内回复我.你今年多大了", "history": [["你好","你好👋!我是人工智能助手 ChatGLM-6B,很高兴见到你,欢迎问我任何问题。"]]}'
|
|||
|
|
|||
|
|
|||
|
# curl -X POST "http://127.0.0.1:8000/v1/completions" \
|
|||
|
# -H 'Content-Type: application/json' \
|
|||
|
# -d '{"prompt": "请用20字内回复我.你今年多大了", "history": ["你好"]}'
|
|||
|
|
|||
|
|