Upload complete model
Browse files
README.md
CHANGED
|
@@ -9,7 +9,7 @@ tags:
|
|
| 9 |
### CURRENTLY UPLOADING
|
| 10 |
### CURRENTLY UPLOADING
|
| 11 |
|
| 12 |
-
**See DeepSeek-V3.2 5.5bit MLX in action - [demonstration video
|
| 13 |
|
| 14 |
*q5.5bit quant typically achieves 1.141 perplexity in our testing*
|
| 15 |
| Quantization | Perplexity |
|
|
@@ -29,7 +29,7 @@ tags:
|
|
| 29 |
* sudo sysctl iogpu.wired_limit_mb=507000
|
| 30 |
* Expect ~16.6 tokens/s @ 1000 tokens
|
| 31 |
* Quantized with a modified version of [MLX](https://github.com/ml-explore/mlx) 0.28
|
| 32 |
-
* For more details see [demonstration video
|
| 33 |
|
| 34 |
## Disclaimer
|
| 35 |
|
|
|
|
| 9 |
### CURRENTLY UPLOADING
|
| 10 |
### CURRENTLY UPLOADING
|
| 11 |
|
| 12 |
+
**See DeepSeek-V3.2 5.5bit MLX in action - [demonstration video](https://youtu.be/b6RgBIROK5o)**
|
| 13 |
|
| 14 |
*q5.5bit quant typically achieves 1.141 perplexity in our testing*
|
| 15 |
| Quantization | Perplexity |
|
|
|
|
| 29 |
* sudo sysctl iogpu.wired_limit_mb=507000
|
| 30 |
* Expect ~16.6 tokens/s @ 1000 tokens
|
| 31 |
* Quantized with a modified version of [MLX](https://github.com/ml-explore/mlx) 0.28
|
| 32 |
+
* For more details see [demonstration video](https://youtu.be/b6RgBIROK5o) or visit [DeepSeek-V3.2](https://huggingface.co/deepseek-ai/DeepSeek-V3.2).
|
| 33 |
|
| 34 |
## Disclaimer
|
| 35 |
|