igor
commited on
Commit
•
8245433
1
Parent(s):
554833e
updated README
Browse files
README.md
CHANGED
@@ -18,12 +18,7 @@ tags:
|
|
18 |
|
19 |
GPT-J 6B is a transformer model trained using Ben Wang's [Mesh Transformer JAX](https://github.com/kingoflolz/mesh-transformer-jax/). "GPT-J" refers to the class of model, while "6B" represents the number of trainable parameters.
|
20 |
|
21 |
-
This repository contains TensorRT engines
|
22 |
-
* RTX 4090
|
23 |
-
* RTX 3080 Ti
|
24 |
-
* RTX 2080 Ti
|
25 |
-
|
26 |
-
ONNX model generated by [ENOT-AutoDL](https://pypi.org/project/enot-autodl/) and build script will be published soon.
|
27 |
|
28 |
## Metrics:
|
29 |
|
@@ -62,7 +57,7 @@ ONNX model generated by [ENOT-AutoDL](https://pypi.org/project/enot-autodl/) and
|
|
62 |
|
63 |
## How to use
|
64 |
|
65 |
-
Example of inference and accuracy test [published on github](https://github.com/ENOT-AutoDL/
|
66 |
```shell
|
67 |
-
git clone https://github.com/ENOT-AutoDL/
|
68 |
```
|
|
|
18 |
|
19 |
GPT-J 6B is a transformer model trained using Ben Wang's [Mesh Transformer JAX](https://github.com/kingoflolz/mesh-transformer-jax/). "GPT-J" refers to the class of model, while "6B" represents the number of trainable parameters.
|
20 |
|
21 |
+
This repository contains GPT-J 6B onnx model suitable for building TensorRT int8+fp32 engines. Quantization of model was performed by the [ENOT-AutoDL](https://pypi.org/project/enot-autodl/) framework. Code for building of TensorRT engines and examples published on [github](https://github.com/ENOT-AutoDL/ENOT-transformers).
|
|
|
|
|
|
|
|
|
|
|
22 |
|
23 |
## Metrics:
|
24 |
|
|
|
57 |
|
58 |
## How to use
|
59 |
|
60 |
+
Example of inference and accuracy test [published on github](https://github.com/ENOT-AutoDL/ENOT-transformers):
|
61 |
```shell
|
62 |
+
git clone https://github.com/ENOT-AutoDL/ENOT-transformers
|
63 |
```
|