Skip to content

Specify Custom Model Path

bash
python infer_lora_chat.py --base_dir my-base-model --adapter_dir my-lora-adapter

Use Merged Model

bash
python infer_lora_chat.py --merged true --adapter_dir my-lora-adapter

Adjust Generation Parameters

bash
python infer_lora_chat.py --temperature 0.9 --top_p 0.95 --max_new_tokens 1024

Use Custom System Prompt

bash
python infer_lora_chat.py --system_prompt "You are a helpful AI assistant."

Command Line Parameters

Parameter NameTypeDefault ValueDescription
--base_dirstrqwen3-8b-baseBase model directory
--adapter_dirstrfinetune/models/qwen3-8b-qloraLoRA adapter directory
--mergedboolFalseIf True, load merged full weights from adapter_dir/merged
--system_promptstrQing's digital avatar personaModel's system prompt
--max_new_tokensint512Maximum number of new tokens to generate
--temperaturefloat0.7Sampling temperature
--top_pfloat0.9Top-p sampling parameter
--trust_remote_codeboolTrueWhether to trust remote code