sovthpaw commited on
Commit
306ce24
·
verified ·
1 Parent(s): bcb7b3c

Upload folder using huggingface_hub

Browse files
Files changed (2) hide show
  1. README.md +53 -18
  2. download_omni_senter.sh +52 -0
README.md CHANGED
@@ -21,30 +21,57 @@ Omni Senter is a trained version of Qwen2.5-Omni-3B that can:
21
 
22
  Perfect for building voice-controlled phone assistants!
23
 
24
- ## Quick Start
25
 
26
- ### Running with llama.cpp
27
 
28
  ```bash
29
- llama-server \
30
- -m /path/to/Qwen2.5-Omni-3B-Q4_K_M.gguf \
31
- --mmproj /path/to/mmproj-Qwen2.5-Omni-3B-Q8_0.gguf \
32
- --lora omni-senter-3b.gguf \
33
- --ctx-size 8192 \
34
- --n-gpu-layers -1 \
 
 
 
 
 
 
35
  --port 8107
36
  ```
37
 
38
- ### Merging with HF Model
 
 
 
 
39
 
40
- ```python
41
- from transformers import AutoModelForCausalLM
42
- from peft import PeftModel
43
 
44
- base = AutoModelForCausalLM.from_pretrained("Qwen/Qwen2.5-Omni-3B")
45
- model = PeftModel.from_pretrained(base, ".")
46
- merged = model.merge_and_unload()
47
- merged.save_pretrained("omni-senter-3b-merged")
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
48
  ```
49
 
50
  ## Input Formats
@@ -129,11 +156,19 @@ Model: <speak>I hear you! You asked me to open the browser.</speak>
129
 
130
  ## Files
131
 
132
- - `omni-senter-3b.gguf` - LoRA adapter (apply to base GGUF model)
133
- - `adapter_model.safetensors` - LoRA weights (for HF merge)
 
134
  - `adapter_config.json` - LoRA configuration
135
  - `tokenizer*` - Tokenizer files
136
 
 
 
 
 
 
 
 
137
  ## Requirements
138
 
139
  - Qwen2.5-Omni-3B GGUF model
 
21
 
22
  Perfect for building voice-controlled phone assistants!
23
 
24
+ ## Download & Setup (Recommended)
25
 
26
+ We've included a download script that automatically gets the base model and applies our LoRA adapter:
27
 
28
  ```bash
29
+ # 1. Clone this repo
30
+ git lfs install
31
+ git clone https://huggingface.co/sovthpaw/Omni-Senter-3B
32
+
33
+ # 2. Run the setup script
34
+ cd Omni-Senter-3B
35
+ chmod +x download_omni_senter.sh
36
+ ./download_omni_senter.sh
37
+
38
+ # 3. Run the model
39
+ llama-server -m ~/.cache/llama.cpp/models/omni-senter-merged.gguf \
40
+ --mmproj ~/.cache/llama.cpp/models/mmproj-Qwen2.5-Omni-3B-Q8_0.gguf \
41
  --port 8107
42
  ```
43
 
44
+ The script will:
45
+ 1. Download Qwen2.5-Omni GGUF base model (~4GB)
46
+ 2. Download mmproj for vision/audio (~1GB)
47
+ 3. Download our LoRA adapter (46MB)
48
+ 4. Apply the LoRA and create a merged model
49
 
50
+ ## Quick Start (Manual)
 
 
51
 
52
+ ### Option 1: Use LoRA with llama.cpp
53
+
54
+ ```bash
55
+ # Download base model
56
+ huggingface-cli download ggml-org/Qwen2.5-Omni-3B-GGUF Qwen2.5-Omni-3B-Q4_K_M.gguf
57
+ huggingface-cli download ggml-org/Qwen2.5-Omni-3B-GGUF mmproj-Qwen2.5-Omni-3B-Q8_0.gguf
58
+
59
+ # Download LoRA adapter from this repo
60
+ huggingface-cli download sovthpaw/Omni-Senter-3B senter-lora-500.gguf
61
+
62
+ # Apply LoRA
63
+ llama-export-lora -m Qwen2.5-Omni-3B-Q4_K_M.gguf -o omni-senter.gguf --lora senter-lora-500.gguf
64
+
65
+ # Run
66
+ llama-server -m omni-senter.gguf --mmproj mmproj-Qwen2.5-Omni-3B-Q8_0.gguf --port 8107
67
+ ```
68
+
69
+ ### Option 2: Use Full HF Model (8.8GB)
70
+
71
+ ```bash
72
+ # Download full merged model
73
+ git lfs install
74
+ git clone https://huggingface.co/sovthpaw/Omni-Senter-3B-full
75
  ```
76
 
77
  ## Input Formats
 
156
 
157
  ## Files
158
 
159
+ - `download_omni_senter.sh` - Automated download & setup script
160
+ - `senter-lora-500.gguf` - LoRA adapter (46MB) - apply to base GGUF model
161
+ - `adapter_model.safetensors` - LoRA weights for HF merge
162
  - `adapter_config.json` - LoRA configuration
163
  - `tokenizer*` - Tokenizer files
164
 
165
+ ## Full Model Download
166
+
167
+ For the fully merged model (8.8GB, no LoRA needed):
168
+ ```bash
169
+ git clone https://huggingface.co/sovthpaw/Omni-Senter-3B-full
170
+ ```
171
+
172
  ## Requirements
173
 
174
  - Qwen2.5-Omni-3B GGUF model
download_omni_senter.sh ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ # Download and setup Omni Senter 3B
3
+ # This script downloads the base Qwen2.5-Omni GGUF and applies our LoRA adapter
4
+
5
+ set -e
6
+
7
+ MODEL_DIR="$HOME/.cache/llama.cpp/models"
8
+ mkdir -p "$MODEL_DIR"
9
+
10
+ echo "=== Omni Senter 3B Setup ==="
11
+
12
+ # Download base model (Qwen2.5-Omni GGUF)
13
+ BASE_MODEL="$MODEL_DIR/Qwen2.5-Omni-3B-Q4_K_M.gguf"
14
+ if [ ! -f "$BASE_MODEL" ]; then
15
+ echo "Downloading Qwen2.5-Omni-3B GGUF model..."
16
+ huggingface-cli download ggml-org/Qwen2.5-Omni-3B-GGUF Qwen2.5-Omni-3B-Q4_K_M.gguf \
17
+ --local-dir "$MODEL_DIR" \
18
+ --token HF_TOKEN
19
+ else
20
+ echo "Base model already exists: $BASE_MODEL"
21
+ fi
22
+
23
+ # Download mmproj
24
+ MMPROJ="$MODEL_DIR/mmproj-Qwen2.5-Omni-3B-Q8_0.gguf"
25
+ if [ ! -f "$MMPROJ" ]; then
26
+ echo "Downloading mmproj..."
27
+ huggingface-cli download ggml-org/Qwen2.5-Omni-3B-GGUF mmproj-Qwen2.5-Omni-3B-Q8_0.gguf \
28
+ --local-dir "$MODEL_DIR" \
29
+ --token HF_TOKEN
30
+ else
31
+ echo "mmproj already exists: $MMPROJ"
32
+ fi
33
+
34
+ # Download LoRA adapter
35
+ LORA_DIR="$HOME/.cache/llama.cpp/omni-senter"
36
+ mkdir -p "$LORA_DIR"
37
+ echo "Downloading Omni Senter LoRA adapter..."
38
+ huggingface-cli download sovthpaw/Omni-Senter-3B senter-lora-500.gguf \
39
+ --local-dir "$LORA_DIR" \
40
+ --token HF_TOKEN
41
+
42
+ LORA_FILE="$LORA_DIR/senter-lora-500.gguf"
43
+
44
+ # Apply LoRA and create merged model
45
+ MERGED_MODEL="$MODEL_DIR/omni-senter-merged.gguf"
46
+ echo "Applying LoRA adapter..."
47
+ ./llama-export-lora -m "$BASE_MODEL" -o "$MERGED_MODEL" --lora "$LORA_FILE"
48
+
49
+ echo ""
50
+ echo "=== Setup Complete! ==="
51
+ echo "Run with:"
52
+ echo " llama-server -m $MERGED_MODEL --mmproj $MMPROJ --port 8107"