Hugging Face Backend
case "huggingface":
self.llms.append(HfLocalBackend(
model_name=config.name,
max_gpu_memory=config.max_gpu_memory,
eval_device=config.eval_device
))completed_response = model.generate(**completion_kwargs)
return completed_response, Trueif tools:
new_messages = merge_messages_with_tools(messages, tools)
completion_kwargs["messages"] = new_messages
completed_response = model.generate(**completion_kwargs)
# During processing
if tools:
if isinstance(model, HfLocalBackend):
if finished:
tool_calls = decode_hf_tool_calls(completed_response)
tool_calls = double_underscore_to_slash(tool_calls)
return LLMResponse(
response_message=None,
tool_calls=tool_calls,
finished=finished
)Last updated