description='About langchain-ChatGLM, local knowledge based ChatGLM with langchain | '
'基于本地知识库的 ChatGLM 问答')
parser.add_argument('--no-remote-model',action='store_true',default=NO_REMOTE_MODEL,help='remote in the model on '
parser.add_argument('--no-remote-model',action='store_true',default=False,help='remote in the model on loader checkpoint, if your load local model to add the ` --no-remote-model`')
'loader checkpoint, '
parser.add_argument('--model',type=str,default='chatglm-6b',help='Name of the model to load by default.')
'if your load local '
'model to add the ` '
'--no-remote-model`')
parser.add_argument('--model',type=str,default=LLM_MODEL,help='Name of the model to load by default.')
parser.add_argument('--lora',type=str,help='Name of the LoRA to apply to the model by default.')
parser.add_argument('--lora',type=str,help='Name of the LoRA to apply to the model by default.')
parser.add_argument("--model-dir",type=str,default='model/',help="Path to directory with all the models")
parser.add_argument("--model-dir",type=str,default=MODEL_DIR,help="Path to directory with all the models")
parser.add_argument("--lora-dir",type=str,default='loras/',help="Path to directory with all the loras")
parser.add_argument("--lora-dir",type=str,default=LORA_DIR,help="Path to directory with all the loras")
# Accelerate/transformers
# Accelerate/transformers
parser.add_argument('--cpu',action='store_true',help='Use the CPU to generate text. Warning: Training on CPU is extremely slow.')
parser.add_argument('--auto-devices',action='store_true',help='Automatically split the model across the available GPU(s) and CPU.')
help='Load the model with 8-bit precision.')
parser.add_argument('--gpu-memory',type=str,nargs="+",help='Maxmimum GPU memory in GiB to be allocated per GPU. Example: --gpu-memory 10 for a single GPU, --gpu-memory 10 5 for two GPUs. You can also set values in MiB like --gpu-memory 3500MiB.')