A specialized Llama 3 8B model fine-tuned for function calling and tool use, optimized for generating structured API calls and handling multi-turn conversations with tool integration.
A Meta Llama 3 8B Instruct model fine-tuned specifically for function calling and tool use capabilities, quantized to GGUF format for efficient local deployment.
This skill provides a specialized version of Meta's Llama 3 8B model that has been fine-tuned on the Trelis function_calling_v3 dataset. The model excels at:
The model is available in multiple GGUF quantization levels, allowing you to balance between quality and resource usage based on your hardware constraints.
Select the appropriate quantization based on your available VRAM/RAM:
Download your chosen quantization from HuggingFace:
```bash
wget https://huggingface.co/mradermacher/Meta-Llama-3-8B-Instruct-function-calling-GGUF/resolve/main/Meta-Llama-3-8B-Instruct-function-calling.Q4_K_M.gguf
```
Set up the model in your preferred runtime environment. For llama.cpp-based tools:
```bash
./main -m Meta-Llama-3-8B-Instruct-function-calling.Q4_K_M.gguf \
-p "Your prompt here" \
-n 512 \
--temp 0.7
```
Use the Llama 3 instruction format with function definitions:
```
<|begin_of_text|><|start_header_id|>system<|end_header_id|>
You are a helpful assistant with access to the following functions:
{
"name": "get_weather",
"description": "Get current weather for a location",
"parameters": {
"location": {"type": "string", "description": "City name"},
"units": {"type": "string", "enum": ["celsius", "fahrenheit"]}
}
}
<|eot_id|><|start_header_id|>user<|end_header_id|>
What's the weather like in San Francisco?<|eot_id|><|start_header_id|>assistant<|end_header_id|>
```
The model will generate structured function calls that you should parse and execute:
```json
{
"function": "get_weather",
"arguments": {
"location": "San Francisco",
"units": "fahrenheit"
}
}
```
After executing the function, provide the result back to the model:
```
<|start_header_id|>tool<|end_header_id|>
{"temperature": 68, "condition": "partly cloudy"}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
```
Define available API endpoints as functions and let the model generate the appropriate calls based on user requests.
Expose database operations as functions and have the model translate natural language queries into structured function calls.
Define workflow steps as callable functions and let the model orchestrate multi-step processes based on high-level instructions.
Apache 2.0 - Fine-tuned from Meta Llama 3 8B Instruct base model
Model card: https://huggingface.co/mradermacher/Meta-Llama-3-8B-Instruct-function-calling-GGUF
Base model: https://huggingface.co/Trelis/Meta-Llama-3-8B-Instruct-function-calling
Training dataset: https://huggingface.co/datasets/Trelis/function_calling_v3
Leave a review
No reviews yet. Be the first to review this skill!
# Download SKILL.md from killerskills.ai/api/skills/llama-3-8b-function-calling/raw