Purple Squirrel AI — Models, Papers & Data
Collection
9 models, 3 papers, 3 datasets. Distributed AI, GPU video, multichain DeFi, Solana wallets. GGUF quants + LoRA + 1.3K training pairs. MIT. • 12 items • Updated
GGUF quantized versions of Purple Squirrel R1 for local inference via llama.cpp, Ollama, or LM Studio.
| File | Quant | Size | Quality | Speed | Use Case |
|---|---|---|---|---|---|
purple-squirrel-r1-f16.gguf |
F16 | 15 GB | Best | Slowest | Reference, re-quantization |
purple-squirrel-r1-Q8_0.gguf |
Q8_0 | ~8 GB | Excellent | Fast | High-quality local inference |
purple-squirrel-r1-Q5_K_M.gguf |
Q5_K_M | ~5.5 GB | Great | Faster | Balanced quality/speed |
purple-squirrel-r1-Q4_K_M.gguf |
Q4_K_M | 4.6 GB | Good | Fastest | Memory-constrained devices |
| Property | Value |
|---|---|
| Base Model | DeepSeek-R1-Distill-Llama-8B |
| Parameters | 8B |
| Architecture | Llama |
| Context Length | 4096 tokens |
| Specialization | AIDP platform ops, video analysis, blockchain |
A ready-to-use Modelfile is included in this repo.
# Download the Modelfile and a GGUF
huggingface-cli download purplesquirrelnetworks/purple-squirrel-r1-gguf \
Modelfile purple-squirrel-r1-Q5_K_M.gguf --local-dir .
# Create and run
ollama create purple-squirrel-r1 -f Modelfile
ollama run purple-squirrel-r1
To use a different quantization, edit the FROM line in the Modelfile.
./llama-cli -m purple-squirrel-r1-Q4_K_M.gguf \
-p "Explain how distributed GPU inference reduces costs" \
-n 500 -c 4096
| Resource | Link |
|---|---|
| Full Model (safetensors) | purple-squirrel-r1 |
| Multichain Edition (MLX) | purple-squirrel-r1-multichain |
| LoRA Adapters | purple-squirrel-r1-multichain-lora |
| Research Paper | AIDP Neural Cloud |
| Research Paper | AIDP Video Forge |
| Coldstar Whitepaper | coldstar-whitepaper |
| Training Data | multichain-day-training |
| Full Collection | Purple Squirrel AI |
@misc{purplesquirrel-r1-gguf-2025,
title={Purple Squirrel R1 GGUF Quantizations},
author={Karsten, Matthew},
year={2025},
publisher={Purple Squirrel Media},
howpublished={\url{https://huggingface.co/purplesquirrelnetworks/purple-squirrel-r1-gguf}},
note={GGUF quantized DeepSeek-R1-Distill-Llama-8B for local inference}
}
MIT
Built by Purple Squirrel Media | GitHub
4-bit
5-bit
8-bit
16-bit
Base model
deepseek-ai/DeepSeek-R1-Distill-Llama-8B