Add library name and link to Github repository (#2)
Browse files- Add library name and link to Github repository (623d0babcfb6ef616d3b4477e745560d8ce6a1f8)
Co-authored-by: Niels Rogge <nielsr@users.noreply.huggingface.co>
README.md
CHANGED
|
@@ -1,19 +1,19 @@
|
|
| 1 |
---
|
| 2 |
-
|
|
|
|
| 3 |
language:
|
| 4 |
- en
|
| 5 |
- fr
|
| 6 |
- it
|
| 7 |
- de
|
| 8 |
- es
|
| 9 |
-
|
| 10 |
-
- PleIAs/Pleias-350m-Preview
|
| 11 |
pipeline_tag: text-generation
|
| 12 |
tags:
|
| 13 |
- transformers
|
|
|
|
| 14 |
---
|
| 15 |
|
| 16 |
-
|
| 17 |
# Pleias-RAG-350m
|
| 18 |
|
| 19 |
<div align="center">
|
|
@@ -21,7 +21,7 @@ tags:
|
|
| 21 |
</div>
|
| 22 |
|
| 23 |
<p align="center">
|
| 24 |
-
<a href="https://
|
| 25 |
</p>
|
| 26 |
|
| 27 |
**Pleias-RAG-350M** is a 350 million parameters Small Reasoning Model, trained for retrieval-augmented general (RAG), search and source summarization. Along with Pleias-RAG-1B it belongs to the first generation of Pleias specialized reasoning models.
|
|
@@ -127,4 +127,6 @@ With only 350 million parameters, Pleias-RAG-350M is classified among the *phone
|
|
| 127 |
|
| 128 |
We also release an unquantized [GGUF version](https://huggingface.co/PleIAs/Pleias-RAG-350M-gguf) for deployment on CPU. Our internal performance benchmarks suggest that waiting times are currently acceptable for most either even under constrained RAM: about 20 seconds for a complex generation including reasoning traces on 8g RAM and below. Since the model is unquantized, quality of text generation should be identical to the original model.
|
| 129 |
|
| 130 |
-
Once integrated into a RAG system, Pleias-RAG-350M can also be use in a broader range of non-conversational use cases including user support or educational assistance. Through this release, we aims to make tiny model workable in production by relying systematically on an externalized memory.
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
+
base_model:
|
| 3 |
+
- PleIAs/Pleias-350m-Preview
|
| 4 |
language:
|
| 5 |
- en
|
| 6 |
- fr
|
| 7 |
- it
|
| 8 |
- de
|
| 9 |
- es
|
| 10 |
+
license: apache-2.0
|
|
|
|
| 11 |
pipeline_tag: text-generation
|
| 12 |
tags:
|
| 13 |
- transformers
|
| 14 |
+
library_name: transformers
|
| 15 |
---
|
| 16 |
|
|
|
|
| 17 |
# Pleias-RAG-350m
|
| 18 |
|
| 19 |
<div align="center">
|
|
|
|
| 21 |
</div>
|
| 22 |
|
| 23 |
<p align="center">
|
| 24 |
+
<a href="https://huggingface.co/papers/2504.18225"><b>Full model report</b></a>
|
| 25 |
</p>
|
| 26 |
|
| 27 |
**Pleias-RAG-350M** is a 350 million parameters Small Reasoning Model, trained for retrieval-augmented general (RAG), search and source summarization. Along with Pleias-RAG-1B it belongs to the first generation of Pleias specialized reasoning models.
|
|
|
|
| 127 |
|
| 128 |
We also release an unquantized [GGUF version](https://huggingface.co/PleIAs/Pleias-RAG-350M-gguf) for deployment on CPU. Our internal performance benchmarks suggest that waiting times are currently acceptable for most either even under constrained RAM: about 20 seconds for a complex generation including reasoning traces on 8g RAM and below. Since the model is unquantized, quality of text generation should be identical to the original model.
|
| 129 |
|
| 130 |
+
Once integrated into a RAG system, Pleias-RAG-350M can also be use in a broader range of non-conversational use cases including user support or educational assistance. Through this release, we aims to make tiny model workable in production by relying systematically on an externalized memory.
|
| 131 |
+
|
| 132 |
+
Github repository: https://github.com/Pleias/Pleias-RAG-Library
|