fcogidi commited on
Commit
5bcde88
·
verified ·
1 Parent(s): 8603c75

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -1
README.md CHANGED
@@ -5,4 +5,19 @@ language:
5
  pipeline_tag: summarization
6
  library_name: transformers.js
7
  ---
8
- https://huggingface.co/JustinDu/BARTxiv with ONNX weights to be compatible with Transformers.js.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  pipeline_tag: summarization
6
  library_name: transformers.js
7
  ---
8
+ https://huggingface.co/JustinDu/BARTxiv with ONNX weights to be compatible with Transformers.js.
9
+
10
+ ## Steps to get ONNX quantized weights
11
+ 1. Use optimum-cli to convert the PyTorch weights to ONNX format.
12
+
13
+ ```bash
14
+ optimum-cli export onnx --model JustinDu/BARTxiv BARTxiv_onnx
15
+ ```
16
+
17
+ 2. Use the `quantize` script in the Transformers.js repo to quantize the ONNX weights.
18
+
19
+ ```bash
20
+ python -m scripts.quantize --input_folder BARTxiv_onnx --output_folder BARTxiv_onnx/onnx
21
+ ```
22
+
23
+ Move the `.onnx` files in the `BARTxiv_onnx` folder to `BARTxiv_onnx/onnx` (these are the FP32 weights).