Add/update the quantized ONNX model files and README.md for Transformers.js v3 (#2)
Browse files- Add/update the quantized ONNX model files and README.md for Transformers.js v3 (fb5dbab37b4a3a352753bc1039dd8a202394c245)
Co-authored-by: Yuichiro Tachibana <[email protected]>
- README.md +9 -7
- onnx/model_bnb4.onnx +3 -0
- onnx/model_int8.onnx +3 -0
- onnx/model_q4.onnx +3 -0
- onnx/model_q4f16.onnx +3 -0
- onnx/model_uint8.onnx +3 -0
README.md
CHANGED
@@ -6,17 +6,20 @@ pipeline_tag: feature-extraction
|
|
6 |
|
7 |
https://huggingface.co/jinaai/jina-embeddings-v2-base-en with ONNX weights to be compatible with Transformers.js.
|
8 |
|
9 |
-
|
10 |
## Usage with 🤗 Transformers.js
|
11 |
|
|
|
|
|
|
|
|
|
|
|
12 |
```js
|
13 |
-
|
14 |
-
import { pipeline, cos_sim } from '@xenova/transformers';
|
15 |
|
16 |
// Create feature extraction pipeline
|
17 |
-
const extractor = await pipeline('feature-extraction', 'Xenova/jina-embeddings-v2-base-en',
|
18 |
-
|
19 |
-
);
|
20 |
|
21 |
// Generate embeddings
|
22 |
const output = await extractor(
|
@@ -28,5 +31,4 @@ const output = await extractor(
|
|
28 |
console.log(cos_sim(output[0].data, output[1].data)); // 0.9341313949712492 (unquantized) vs. 0.9022937687830741 (quantized)
|
29 |
```
|
30 |
|
31 |
-
|
32 |
Note: Having a separate repo for ONNX weights is intended to be a temporary solution until WebML gains more traction. If you would like to make your models web-ready, we recommend converting to ONNX using [🤗 Optimum](https://huggingface.co/docs/optimum/index) and structuring your repo like this one (with ONNX weights located in a subfolder named `onnx`).
|
|
|
6 |
|
7 |
https://huggingface.co/jinaai/jina-embeddings-v2-base-en with ONNX weights to be compatible with Transformers.js.
|
8 |
|
|
|
9 |
## Usage with 🤗 Transformers.js
|
10 |
|
11 |
+
If you haven't already, you can install the [Transformers.js](https://huggingface.co/docs/transformers.js) JavaScript library from [NPM](https://www.npmjs.com/package/@huggingface/transformers) using:
|
12 |
+
```bash
|
13 |
+
npm i @huggingface/transformers
|
14 |
+
```
|
15 |
+
|
16 |
```js
|
17 |
+
import { pipeline, cos_sim } from '@huggingface/transformers';
|
|
|
18 |
|
19 |
// Create feature extraction pipeline
|
20 |
+
const extractor = await pipeline('feature-extraction', 'Xenova/jina-embeddings-v2-base-en', {
|
21 |
+
dtype: "fp32" // Options: "fp32", "fp16", "q8", "q4"
|
22 |
+
});
|
23 |
|
24 |
// Generate embeddings
|
25 |
const output = await extractor(
|
|
|
31 |
console.log(cos_sim(output[0].data, output[1].data)); // 0.9341313949712492 (unquantized) vs. 0.9022937687830741 (quantized)
|
32 |
```
|
33 |
|
|
|
34 |
Note: Having a separate repo for ONNX weights is intended to be a temporary solution until WebML gains more traction. If you would like to make your models web-ready, we recommend converting to ONNX using [🤗 Optimum](https://huggingface.co/docs/optimum/index) and structuring your repo like this one (with ONNX weights located in a subfolder named `onnx`).
|
onnx/model_bnb4.onnx
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:942109bb9c0654eca2aec01d23b7ad9f5d84ea4cc3622a059d8e29dd5df77386
|
3 |
+
size 158113971
|
onnx/model_int8.onnx
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5a01f42d114a6bb92228d1a47c4cccef79cfd931791a499134db80e177a3a93d
|
3 |
+
size 137390704
|
onnx/model_q4.onnx
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:180bd960d42efb0f4671e34776461ff54f19f4b9471dfa07162b3ff5e5c3ae80
|
3 |
+
size 165191319
|
onnx/model_q4f16.onnx
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:10f89500a7dfd28b4aeaf87b9ee03d9c8c37befeb231b27615be63e8354df3b5
|
3 |
+
size 111050734
|
onnx/model_uint8.onnx
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cdc0b9730a2cbae3edee4ff1869a45aeb6874e0a0d505f1b509b4e30964ace2c
|
3 |
+
size 137390742
|