Environment & Hardware Configuration
Please provide your specific setup. Use the suggested commands for your OS to verify.
- OS Version: Windows
- CPU: AMD Ryzen 5 3600 6-Core Processor
- RAM: 16 GB
- GPU: NVIDIA RTX 3060 12GB
- Python: 3.12
Model & Logic Context
- Model source: HuggingFace
- Model: Qwen3.5-VL
Failure Timing & Logs
The issue is occurring during Runtime.
Analysis & Brainstorming
I've noticed that tool_choice isn't taken into account in multimodal models such as Qwen3.5. The tools are passed and executed for every request as if tool_choice were set to “required” by default. I suspect the reason might be that the code doesn't pass this parameter when rendering the chat template.
I've made sure to set tool_choice to None. I've also tried other tool_choice like "auto".
llama-cliorllama-serverprovided byllama.cpp, and the problem (exists/does not exist) still exists.Environment & Hardware Configuration
Please provide your specific setup. Use the suggested commands for your OS to verify.
Model & Logic Context
Failure Timing & Logs
The issue is occurring during Runtime.
Analysis & Brainstorming
I've noticed that
tool_choiceisn't taken into account in multimodal models such as Qwen3.5. The tools are passed and executed for every request as iftool_choicewere set to “required” by default. I suspect the reason might be that the code doesn't pass this parameter when rendering the chat template.I've made sure to set
tool_choiceto None. I've also tried othertool_choicelike "auto".