You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
There are more and more open sourced small models(7b, 30b, 70b) out there that support tool calling. Is it possible that llama.cpp starts to support those? I also created my own tool calling models: https://huggingface.co/rubra-ai/Meta-Llama-3-8B-Instruct-GGUF, but it requires extra preprocess and post process steps to handle the function calling requests, so I have to create a fork tools.cpp and implemented custom logic.
I wonder if it is possible to create something like a standard tool calling template so that tool calling models can follow. Basically it needs to cover 3 things:
convert tool calling output from model to OAI json format.
convert OAI format input function definitions to proper system prompt of local model.
convert OAI format input chat messages with role tool_call or tools to a proper format that a tool calling model can support.
Possible Implementation
No response
The text was updated successfully, but these errors were encountered:
Tool calls is not just a chat template, so it's trivial to implement. The problem is that it must be done per model. I think we can base on the same infrastructure provided by #5695
Prerequisites
Feature Description
In the server mode, user should be able to use the OAI api for tool calling just like they do to do tool calling for gpt-4. Today this is not supported yet: https://github.com/ggerganov/llama.cpp/blob/master/examples/server/utils.hpp#L394
Motivation
There are more and more open sourced small models(7b, 30b, 70b) out there that support tool calling. Is it possible that llama.cpp starts to support those? I also created my own tool calling models: https://huggingface.co/rubra-ai/Meta-Llama-3-8B-Instruct-GGUF, but it requires extra preprocess and post process steps to handle the function calling requests, so I have to create a fork tools.cpp and implemented custom logic.
I wonder if it is possible to create something like a standard tool calling template so that tool calling models can follow. Basically it needs to cover 3 things:
tool_call
ortools
to a proper format that a tool calling model can support.Possible Implementation
No response
The text was updated successfully, but these errors were encountered: