inferencerlabs commited on
Commit
a6322e7
·
verified ·
1 Parent(s): 4bbdeab

Upload complete model

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -9,7 +9,7 @@ tags:
9
  ### CURRENTLY UPLOADING
10
  ### CURRENTLY UPLOADING
11
 
12
- **See DeepSeek-V3.2 5.5bit MLX in action - [demonstration video - coming soon](https://youtube.com/xcreate)**
13
 
14
  *q5.5bit quant typically achieves 1.141 perplexity in our testing*
15
  | Quantization | Perplexity |
@@ -29,7 +29,7 @@ tags:
29
  * sudo sysctl iogpu.wired_limit_mb=507000
30
  * Expect ~16.6 tokens/s @ 1000 tokens
31
  * Quantized with a modified version of [MLX](https://github.com/ml-explore/mlx) 0.28
32
- * For more details see [demonstration video - coming soon](https://youtube.com/xcreate) or visit [DeepSeek-V3.2](https://huggingface.co/deepseek-ai/DeepSeek-V3.2).
33
 
34
  ## Disclaimer
35
 
 
9
  ### CURRENTLY UPLOADING
10
  ### CURRENTLY UPLOADING
11
 
12
+ **See DeepSeek-V3.2 5.5bit MLX in action - [demonstration video](https://youtu.be/b6RgBIROK5o)**
13
 
14
  *q5.5bit quant typically achieves 1.141 perplexity in our testing*
15
  | Quantization | Perplexity |
 
29
  * sudo sysctl iogpu.wired_limit_mb=507000
30
  * Expect ~16.6 tokens/s @ 1000 tokens
31
  * Quantized with a modified version of [MLX](https://github.com/ml-explore/mlx) 0.28
32
+ * For more details see [demonstration video](https://youtu.be/b6RgBIROK5o) or visit [DeepSeek-V3.2](https://huggingface.co/deepseek-ai/DeepSeek-V3.2).
33
 
34
  ## Disclaimer
35