Thank you for your contribution to the code. Due to certain circumstances, I’m unable to use online LLM APIs such as ChatGPT. Is it possible to use a local model instead—for example, vision models like Qwen2.5-VL 7B?
Please let me know if you have any suggestions.
Thank you.