igor commited on
Commit
8245433
1 Parent(s): 554833e

updated README

Browse files
Files changed (1) hide show
  1. README.md +3 -8
README.md CHANGED
@@ -18,12 +18,7 @@ tags:
18
 
19
  GPT-J 6B is a transformer model trained using Ben Wang's [Mesh Transformer JAX](https://github.com/kingoflolz/mesh-transformer-jax/). "GPT-J" refers to the class of model, while "6B" represents the number of trainable parameters.
20
 
21
- This repository contains TensorRT engines with mixed precission int8 + fp32. You can find prebuilt engines for the following GPUs:
22
- * RTX 4090
23
- * RTX 3080 Ti
24
- * RTX 2080 Ti
25
-
26
- ONNX model generated by [ENOT-AutoDL](https://pypi.org/project/enot-autodl/) and build script will be published soon.
27
 
28
  ## Metrics:
29
 
@@ -62,7 +57,7 @@ ONNX model generated by [ENOT-AutoDL](https://pypi.org/project/enot-autodl/) and
62
 
63
  ## How to use
64
 
65
- Example of inference and accuracy test [published on github](https://github.com/ENOT-AutoDL/gpt-j-6B-tensorrt-int8):
66
  ```shell
67
- git clone https://github.com/ENOT-AutoDL/gpt-j-6B-tensorrt-int8
68
  ```
 
18
 
19
  GPT-J 6B is a transformer model trained using Ben Wang's [Mesh Transformer JAX](https://github.com/kingoflolz/mesh-transformer-jax/). "GPT-J" refers to the class of model, while "6B" represents the number of trainable parameters.
20
 
21
+ This repository contains GPT-J 6B onnx model suitable for building TensorRT int8+fp32 engines. Quantization of model was performed by the [ENOT-AutoDL](https://pypi.org/project/enot-autodl/) framework. Code for building of TensorRT engines and examples published on [github](https://github.com/ENOT-AutoDL/ENOT-transformers).
 
 
 
 
 
22
 
23
  ## Metrics:
24
 
 
57
 
58
  ## How to use
59
 
60
+ Example of inference and accuracy test [published on github](https://github.com/ENOT-AutoDL/ENOT-transformers):
61
  ```shell
62
+ git clone https://github.com/ENOT-AutoDL/ENOT-transformers
63
  ```