Update README.md
Browse files
README.md
CHANGED
@@ -10,11 +10,11 @@ pipeline_tag: any-to-any
|
|
10 |
|
11 |
本模型为 https://huggingface.co/inclusionAI/Ming-Lite-Omni-1.5 模型的 NF4 量化版本,可用官方代码直接加载使用,从而能使该项目在 24GB 消费级显卡内进行推理测试和学习研究。
|
12 |
|
13 |
-
这里只是主模型部分的量化权重,其它部分如:connector / mlp / talker / transformer / vae 等,请从官方 repo 下载,其中
|
14 |
|
15 |
This model is the NF4 quantized version of the model: https://huggingface.co/inclusionAI/Ming-Lite-Omni-1.5, it can be directly loaded using the official demo code, let the project can be perform inference testing and learning research on a 24GB consumer GPU.
|
16 |
|
17 |
-
Here is only the quantized weights of the main model part. Other parts such as: connector / mlp / talker / transformer / vae, please download from the official repo, and the
|
18 |
|
19 |
以下部分引用自原模型说明内容:
|
20 |
|
|
|
10 |
|
11 |
本模型为 https://huggingface.co/inclusionAI/Ming-Lite-Omni-1.5 模型的 NF4 量化版本,可用官方代码直接加载使用,从而能使该项目在 24GB 消费级显卡内进行推理测试和学习研究。
|
12 |
|
13 |
+
这里只是主模型部分的量化权重,其它部分如:connector / mlp / talker / transformer / vae 等,请从官方 repo 下载,其中 transformer 也可以按需量化。
|
14 |
|
15 |
This model is the NF4 quantized version of the model: https://huggingface.co/inclusionAI/Ming-Lite-Omni-1.5, it can be directly loaded using the official demo code, let the project can be perform inference testing and learning research on a 24GB consumer GPU.
|
16 |
|
17 |
+
Here is only the quantized weights of the main model part. Other parts such as: connector / mlp / talker / transformer / vae, please download from the official repo, and the transformer part can also be quantized as needed.
|
18 |
|
19 |
以下部分引用自原模型说明内容:
|
20 |
|