Run the YiXin-Agentic-Qwen3-14B model locally using optimized GGUF quantizations. This agentic AI model is built on Qwen3-14B architecture and optimized for conversational tasks with multiple quantization options for different hardware constraints.
Run the YiXin-Agentic-Qwen3-14B language model locally using optimized GGUF quantizations from mradermacher. This skill helps you download and set up the appropriate quantization level based on your hardware capabilities.
This skill guides you through selecting and downloading the appropriate quantized version of the YiXin-Agentic-Qwen3-14B model for local inference. The model is available in multiple quantization formats (from 3.7GB to 12.2GB) to balance quality and hardware requirements.
When the user requests this skill, follow these steps:
1. **Assess Hardware Requirements**
- Ask the user about their available RAM/VRAM
- Ask about their priority: speed, quality, or size
- Recommend an appropriate quantization based on their constraints:
- Limited RAM (<8GB): IQ2_XXS or IQ2_XS (4.4-4.8GB)
- Moderate RAM (8-12GB): IQ3_S or IQ3_M (6.8-7.0GB)
- Good RAM (12-16GB): Q4_K_S or Q4_K_M (8.7-9.1GB) - recommended
- High RAM (>16GB): Q5_K_M or Q6_K (10.6-12.2GB)
2. **Download the Model**
- Provide the direct download link from: `https://huggingface.co/mradermacher/YiXin-Agentic-Qwen3-14B-i1-GGUF`
- Use wget or curl to download the selected .gguf file
- Example: `wget https://huggingface.co/mradermacher/YiXin-Agentic-Qwen3-14B-i1-GGUF/resolve/main/YiXin-Agentic-Qwen3-14B.i1-Q4_K_M.gguf`
3. **Verify the Download**
- Check the file size matches the expected size
- Confirm the file has downloaded completely
4. **Provide Usage Instructions**
- Explain how to use the model with llama.cpp, ollama, or other GGUF-compatible inference engines
- Provide a basic command example for inference
- Example for llama.cpp: `./main -m YiXin-Agentic-Qwen3-14B.i1-Q4_K_M.gguf -p "Your prompt here" -n 512`
5. **Next Steps**
- Suggest creating a simple test prompt to verify the model works
- Provide information about the model's capabilities (conversational, agentic tasks)
- Link to the base model page for more details: `https://huggingface.co/YiXin-AILab/YiXin-Agentic-Qwen3-14B`
Recommend based on user needs:
User: "I need to run a local LLM for agentic tasks"
Assistant response:
1. Assess available RAM
2. Recommend Q4_K_M if they have 12GB+ RAM
3. Download the selected quantization
4. Provide setup instructions for their preferred inference tool
5. Test with a sample agentic prompt
Leave a review
No reviews yet. Be the first to review this skill!
# Download SKILL.md from killerskills.ai/api/skills/yixin-agentic-qwen3-14b/raw