Upload folder using huggingface_hub

#9
by Xenova HF staff - opened
onnx/decoder_model_merged_quantized.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e0784dfa87ac4cba563c01a081f22841e782bdbe91ce751b5cc955f6c499c4c6
3
- size 426777507
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9dade5431ef5f3690cde7449d8e881b2d596d1ee5c6472752787d7d470d67091
3
+ size 428422300
onnx/decoder_model_quantized.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f52ff6681bc2b74ef3326584cd9e74ff6949742ca61efaf5b41ada412e1b8919
3
- size 424993274
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3fff451a7f89ae5a7e547ffa3b7933e5116f56ee928b75277d2d451418e9b09
3
+ size 426638067
onnx/decoder_with_past_model_quantized.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8f079fa848c6a21ba458fee27968c7a165e0d56d4f1e81658d7b370f466e8449
3
- size 373572413
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5f013ec948fa4f8731c13af8fdcd24a83a5fa8e4891d5a651cfb3c11cba08c6
3
+ size 374966081
onnx/encodec_decode_quantized.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a5d97ab43fba9152418343efd9e61fe1eb3d00175cf8e97b124f661ac77d54b
3
- size 59796623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b2a729fd329ad5390f209234c697f56c8d8f03ee6c273959b7adbccd2d4ac66
3
+ size 59878565
onnx/text_encoder_quantized.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:77494d927d25ea09e6abc8d566281b5b9cd58f39c5f89a7da60e157b0a711607
3
- size 110027935
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:336afda84d24205409f20e71da691e5c21585e32aa86e3e38d73543ac6d481fc
3
+ size 110443015
quantize_config.json CHANGED
@@ -179,8 +179,8 @@
179
  "weight_type": "QInt8"
180
  }
181
  },
182
- "per_channel": false,
183
- "reduce_range": false
184
  },
185
  "int8": {
186
  "per_model_config": {
 
179
  "weight_type": "QInt8"
180
  }
181
  },
182
+ "per_channel": true,
183
+ "reduce_range": true
184
  },
185
  "int8": {
186
  "per_model_config": {