Instructions to use enosislabs/math-mini-1.7b-preview-16bits with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use enosislabs/math-mini-1.7b-preview-16bits with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="enosislabs/math-mini-1.7b-preview-16bits") messages = [ {"role": "user", "content": "Who are you?"}, ] pipe(messages)# Load model directly from transformers import AutoTokenizer, AutoModelForCausalLM tokenizer = AutoTokenizer.from_pretrained("enosislabs/math-mini-1.7b-preview-16bits") model = AutoModelForCausalLM.from_pretrained("enosislabs/math-mini-1.7b-preview-16bits") messages = [ {"role": "user", "content": "Who are you?"}, ] inputs = tokenizer.apply_chat_template( messages, add_generation_prompt=True, tokenize=True, return_dict=True, return_tensors="pt", ).to(model.device) outputs = model.generate(**inputs, max_new_tokens=40) print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[-1]:])) - Notebooks
- Google Colab
- Kaggle
- Local Apps
- vLLM
How to use enosislabs/math-mini-1.7b-preview-16bits with vLLM:
Install from pip and serve model
# Install vLLM from pip: pip install vllm # Start the vLLM server: vllm serve "enosislabs/math-mini-1.7b-preview-16bits" # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:8000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "enosislabs/math-mini-1.7b-preview-16bits", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }'Use Docker
docker model run hf.co/enosislabs/math-mini-1.7b-preview-16bits
- SGLang
How to use enosislabs/math-mini-1.7b-preview-16bits with SGLang:
Install from pip and serve model
# Install SGLang from pip: pip install sglang # Start the SGLang server: python3 -m sglang.launch_server \ --model-path "enosislabs/math-mini-1.7b-preview-16bits" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "enosislabs/math-mini-1.7b-preview-16bits", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }'Use Docker images
docker run --gpus all \ --shm-size 32g \ -p 30000:30000 \ -v ~/.cache/huggingface:/root/.cache/huggingface \ --env "HF_TOKEN=<secret>" \ --ipc=host \ lmsysorg/sglang:latest \ python3 -m sglang.launch_server \ --model-path "enosislabs/math-mini-1.7b-preview-16bits" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "enosislabs/math-mini-1.7b-preview-16bits", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }' - Docker Model Runner
How to use enosislabs/math-mini-1.7b-preview-16bits with Docker Model Runner:
docker model run hf.co/enosislabs/math-mini-1.7b-preview-16bits
Math Mini 1.7B (Preview)
Math Mini 1.7B (Preview) is a larger, more capable model developed by Enosis Labs as part of the "Mini Series." Building on the foundation of the 0.6B version, this 1.7B model delivers significantly improved performance, deeper reasoning, and greater accuracy in mathematical tasks. It is fine-tuned from the original Qwen/Qwen3-1.7B base model (not from Unsloth's pre-adapted versions).
Philosophy & Capabilities
The Mini Series, along with the "Enosis Math" and "Enosis Code" models, incorporates step-by-step reasoning by default, enabling more efficient, clear, and well-founded answers. All models in the Math series have been trained with carefully curated step-by-step problem-solving datasets, resulting in a greater ability to reason and explain solutions in a structured way.
Math Mini 1.7B (Preview) is optimized for:
- Basic and Intermediate Algebra: Solving equations, manipulating expressions, and handling more complex algebraic problems.
- Arithmetic & Sequential Reasoning: Calculations and breaking down problems into logical steps, with improved multi-step reasoning.
- Elementary & Intermediate Logic: Applying deduction in mathematical contexts, now with broader coverage.
- Competition Problem Solving (Introductory to Intermediate): Enhanced foundational and competition-style skills, adapted to the increased model scale.
Larger models in the "Enosis Math" series address even more advanced topics such as calculus, higher algebra, and olympiad problems. The "Code Mini" and "Enosis Code" series are oriented towards programming and algorithmic tasks, maintaining the same philosophy of explicit and efficient reasoning.
This model is a preview version and is under continuous improvement and evaluation.
Quick Start
Available in Hugging Face Transformers format and for high-throughput inference servers like vLLM.
vLLM (Inference Server)
Install vLLM:
pip install vllm
Start the vLLM server with the model (16-bit version):
vllm serve "enosislabs/math-mini-1.7b-preview-16bits"
Call the server using curl:
curl -X POST "http://localhost:8000/v1/chat/completions" \
-H "Content-Type: application/json" \
--data '{
"model": "enosislabs/math-mini-1.7b-preview-16bits",
"messages": [
{"role": "user", "content": "What is the capital of France?"}
]
}'
Transformers (Hugging Face)
Use a pipeline as a high-level helper:
from transformers import pipeline
pipe = pipeline("text-generation", model="enosislabs/math-mini-1.7b-preview-16bits")
messages = [
{"role": "user", "content": "Who are you?"},
]
pipe(messages)
Prompt Format (Qwen3 ChatML)
For best results, use the Qwen3 ChatML format. The tokenizer.apply_chat_template method handles this automatically.
<|im_start|>system
You are a helpful AI assistant. Provide a detailed step-by-step solution.
<|im_end|>
<|im_start|>user
{user_question}
<|im_end|>
<|im_start|>assistant
Acknowledgements
- Fine-tuned from the original
Qwen/Qwen3-1.7Bbase model. - Training process accelerated and optimized using Unsloth for efficiency, but not using Unsloth's pre-adapted weights.
Citation
If you use this model, please cite:
@software{enosislabs_math_mini_1.7b_preview_2025,
author = {{Enosis Labs}},
title = {{Math Mini 1.7B (Preview)}},
year = {2025},
publisher = {Hugging Face},
version = {0.1-preview},
url = {https://huggingface.co/enosislabs/math-mini-1.7b-preview-16bits}
}
- Downloads last month
- 12