Quantized Llama 3 70B model optimized for tool use and function calling. Available in multiple GGUF quantization formats for different hardware configurations.
This skill has safety concerns that you should review before use. Some patterns were detected that may pose a risk.Safety score: 60/100.
KillerSkills scans all public content for safety. Use caution before installing or executing flagged content.
A quantized version of Groq's Llama 3 70B model optimized for tool use and function calling. Available in multiple GGUF quantization formats to support various hardware configurations, from high-end GPUs to resource-constrained systems.
Multiple quantization levels available to balance quality vs. resource requirements:
Determine your available resources:
1. **For GPU-only inference**: Choose a quant 1-2GB smaller than your GPU VRAM
2. **For CPU+GPU inference**: Add system RAM + GPU VRAM, then choose a quant 1-2GB smaller than the total
3. **For quality priority**: Use Q5_K_M or higher
4. **For resource-constrained systems**: Use Q4_K_M or lower
Using huggingface-cli:
```bash
pip install -U "huggingface_hub[cli]"
huggingface-cli download bartowski/Llama-3-Groq-70B-Tool-Use-GGUF --include "Llama-3-Groq-70B-Tool-Use-Q4_K_M.gguf" --local-dir ./
huggingface-cli download bartowski/Llama-3-Groq-70B-Tool-Use-GGUF --include "Llama-3-Groq-70B-Tool-Use-Q8_0.gguf/*" --local-dir Llama-3-Groq-70B-Tool-Use-Q8_0
```
**In LM Studio:**
1. Open LM Studio
2. Import the downloaded GGUF file
3. Configure inference settings
4. Start chatting
**With llama.cpp:**
```bash
./main -m Llama-3-Groq-70B-Tool-Use-Q4_K_M.gguf -p "Your prompt here" -n 512
```
```
<|begin_of_text|><|start_header_id|>system<|end_header_id|>
{system_prompt}<|eot_id|><|start_header_id|>user<|end_header_id|>
{prompt}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
```
When defining tools for the model, structure them according to Llama 3's tool use format. The model is specifically trained to generate function calls in response to user queries that require external tools or APIs.
**Basic function calling:**
```
System: You are a helpful assistant with access to a weather API.
User: What's the weather like in San Francisco?
Leave a review
No reviews yet. Be the first to review this skill!
# Download SKILL.md from killerskills.ai/api/skills/llama-3-groq-70b-tool-use-gguf/raw