Text Generation
Transformers
PyTorch
notagen
fill-mask
music-generation
symbolic-music
abc-notation
quantized
Instructions to use manoskary/NotaGenX-Quantized with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use manoskary/NotaGenX-Quantized with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="manoskary/NotaGenX-Quantized")# Load model directly from transformers import AutoModelWithLMHead model = AutoModelWithLMHead.from_pretrained("manoskary/NotaGenX-Quantized", dtype="auto") - Notebooks
- Google Colab
- Kaggle
- Local Apps
- vLLM
How to use manoskary/NotaGenX-Quantized with vLLM:
Install from pip and serve model
# Install vLLM from pip: pip install vllm # Start the vLLM server: vllm serve "manoskary/NotaGenX-Quantized" # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:8000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "manoskary/NotaGenX-Quantized", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }'Use Docker
docker model run hf.co/manoskary/NotaGenX-Quantized
- SGLang
How to use manoskary/NotaGenX-Quantized with SGLang:
Install from pip and serve model
# Install SGLang from pip: pip install sglang # Start the SGLang server: python3 -m sglang.launch_server \ --model-path "manoskary/NotaGenX-Quantized" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "manoskary/NotaGenX-Quantized", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }'Use Docker images
docker run --gpus all \ --shm-size 32g \ -p 30000:30000 \ -v ~/.cache/huggingface:/root/.cache/huggingface \ --env "HF_TOKEN=<secret>" \ --ipc=host \ lmsysorg/sglang:latest \ python3 -m sglang.launch_server \ --model-path "manoskary/NotaGenX-Quantized" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "manoskary/NotaGenX-Quantized", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }' - Docker Model Runner
How to use manoskary/NotaGenX-Quantized with Docker Model Runner:
docker model run hf.co/manoskary/NotaGenX-Quantized
NotaGenX-Quantized
This is a quantized version of the NotaGen model for symbolic music generation. The model generates music in ABC notation format and has been optimized for faster inference and reduced memory usage.
Model Description
- Base Model: sander-wood/notagen
- Quantization: INT8 dynamic quantization using PyTorch
- Size Reduction: ~75% smaller than the original model
- Performance: Faster inference with minimal quality loss
- Memory: Reduced VRAM requirements
Model Architecture
- Type: GPT-2 based transformer for symbolic music generation
- Input: Period, Composer, Instrumentation prompts
- Output: ABC notation music scores
- Patch Size: 16
- Patch Length: 1024
- Hidden Size: 1280
- Layers: 20 (encoder) + 6 (decoder)
Usage
from weavemuse.tools.notagen_tool import NotaGenTool
# Initialize the tool (will automatically use quantized model)
notagen = NotaGenTool()
# Generate music
result = notagen("Classical", "Mozart", "Piano")
print(result["abc"])
Quantization Details
This model has been quantized using PyTorch's dynamic quantization:
- Method: Dynamic INT8 quantization
- Target: Linear and embedding layers
- Preserved: Model architecture and functionality
- Testing: Validated against original model outputs
Performance Comparison
| Metric | Original | Quantized | Improvement |
|---|---|---|---|
| Model Size | ~2.3GB | ~0.6GB | 75% reduction |
| Load Time | ~15s | ~4s | 73% faster |
| Inference | Baseline | 1.2-1.5x faster | 20-50% speedup |
| VRAM Usage | ~2.1GB | ~0.8GB | 62% reduction |
Installation
pip install weavemuse
Citation
If you use this model, please cite the original NotaGen paper:
@article{notagen2024,
title={NotaGen: Symbolic Music Generation with Fine-Grained Control},
author={Wood, Sander and others},
year={2024}
}
License
MIT License - see the original model repository for full license details.
Contact
- Maintainer: manoskary
- Repository: weavemuse
- Issues: Please report issues on the main repository
- Downloads last month
- 8