提交 37ceeae6 作者: soon 提交者: imClumsyPanda

提交更

上级 d35eb120
...@@ -117,6 +117,9 @@ class LocalDocQA: ...@@ -117,6 +117,9 @@ class LocalDocQA:
问题: 问题:
{question}""" {question}"""
if vs_path is None or vs_path =="":# or (not os.path.exists(vs_path))
result = self.llm.chat(query)
else:
prompt = PromptTemplate( prompt = PromptTemplate(
template=prompt_template, template=prompt_template,
input_variables=["context", "question"] input_variables=["context", "question"]
......
...@@ -19,10 +19,11 @@ llm_model_dict = { ...@@ -19,10 +19,11 @@ llm_model_dict = {
"chatglm-6b-int4-qe": "THUDM/chatglm-6b-int4-qe", "chatglm-6b-int4-qe": "THUDM/chatglm-6b-int4-qe",
"chatglm-6b-int4": "THUDM/chatglm-6b-int4", "chatglm-6b-int4": "THUDM/chatglm-6b-int4",
"chatglm-6b": "THUDM/chatglm-6b", "chatglm-6b": "THUDM/chatglm-6b",
"chatyuan": "ClueAI/ChatYuan-large-v2",
} }
# LLM model name # LLM model name
LLM_MODEL = "chatglm-6b" LLM_MODEL = "chatyuan" #"chatglm-6b"
# Use p-tuning-v2 PrefixEncoder # Use p-tuning-v2 PrefixEncoder
USE_PTUNING_V2 = False USE_PTUNING_V2 = False
......
...@@ -82,6 +82,19 @@ class ChatGLM(LLM): ...@@ -82,6 +82,19 @@ class ChatGLM(LLM):
self.history = self.history+[[None, response]] self.history = self.history+[[None, response]]
return response return response
def chat(self,
prompt: str) -> str:
response, _ = self.model.chat(
self.tokenizer,
prompt,
history=[],#self.history[-self.history_len:] if self.history_len>0 else
max_length=self.max_token,
temperature=self.temperature,
)
torch_gc()
self.history = self.history+[[None, response]]
return response
def load_model(self, def load_model(self,
model_name_or_path: str = "THUDM/chatglm-6b", model_name_or_path: str = "THUDM/chatglm-6b",
llm_device=LLM_DEVICE, llm_device=LLM_DEVICE,
......
Markdown 格式
0%
您添加了 0 到此讨论。请谨慎行事。
请先完成此评论的编辑!
注册 或者 后发表评论