ValueError: Model name Qwen/Qwen2.5-3B does not support function calling API.
llm = VllmServer( model='Qwen/Qwen2.5-3B', api_url='http://localhost:8000/v1', tensor_parallel_size=4, max_new_tokens=256, temperature=0.0, dtype='bfloat16', vllm_kwargs={ 'max_model_len': 32_750, }, ) program = FunctionCallingProgram.from_defaults( output_cls=output_class, prompt_template_str=prompt, llm=llm, verbose=True, ) program(markdown)
--enable-auto-tool-choice --tool-call-parser hermes
?pip install OpenAILike
from llama_index.llms.openai_like import OpenAILike llm = OpenAILike( model="Qwen/Qwen2.5-3B", api_base="http://localhost:8000/v1", api_key="fake", max_tokens=256, temperature=0.0, context_window=32750, is_chat_model=True, is_function_calling_model=True, additional_kwargs={"max_model_len": 32750} )