RedXeol commited on
Commit
1d6f719
1 Parent(s): fa531fd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -1
README.md CHANGED
@@ -16,10 +16,13 @@ tags:
16
 
17
  <strong><span style="font-size: larger;">bertin-gpt-j-6B-alpaca-4bit-128g 🤗</span></strong>
18
 
 
 
19
  **descripción en español agregado ⬇️**
20
 
21
  This is a 4-bit GPTQ version of the [bertin-project/bertin-gpt-j-6B-alpaca]( https://huggingface.co/bertin-project/bertin-gpt-j-6B-alpaca)
22
 
 
23
  this is the result of quantizing to 4 bits using [AutoGPTQ](https://github.com/PanQiWei/AutoGPTQ).
24
 
25
  ** How to easily download and use this model in text-generation-webui** (tutorial by [TheBloke](https://huggingface.co/TheBloke))
@@ -61,7 +64,7 @@ This dataset cannot be used to create models that compete in any way with OpenAI
61
  Finetuning
62
  To fine-tune the BERTIN GPT-J-6B model we used the code available on BERTIN's fork of mesh-transformer-jax, which provides code adapt an Alpaca dataset to finetune any GPT-J-6B model. We run finetuning for 3 epochs using sequence length of 2048 on a single TPUv3-8 for 3 hours on top of BERTIN GPT-J-6B.
63
 
64
- ** español **
65
 
66
  Esta es una versión GPTQ de 4 bits del [bertin-project/bertin-gpt-j-6B-alpaca]( https://huggingface.co/bertin-project/bertin-gpt-j-6B-alpaca)
67
 
 
16
 
17
  <strong><span style="font-size: larger;">bertin-gpt-j-6B-alpaca-4bit-128g 🤗</span></strong>
18
 
19
+ ![bertin](https://aeiljuispo.cloudimg.io/v7/https://s3.amazonaws.com/moonup/production/uploads/1627122830113-5ef3829e518622264685b0cd.png?w=200&h=200&f=face)
20
+
21
  **descripción en español agregado ⬇️**
22
 
23
  This is a 4-bit GPTQ version of the [bertin-project/bertin-gpt-j-6B-alpaca]( https://huggingface.co/bertin-project/bertin-gpt-j-6B-alpaca)
24
 
25
+
26
  this is the result of quantizing to 4 bits using [AutoGPTQ](https://github.com/PanQiWei/AutoGPTQ).
27
 
28
  ** How to easily download and use this model in text-generation-webui** (tutorial by [TheBloke](https://huggingface.co/TheBloke))
 
64
  Finetuning
65
  To fine-tune the BERTIN GPT-J-6B model we used the code available on BERTIN's fork of mesh-transformer-jax, which provides code adapt an Alpaca dataset to finetune any GPT-J-6B model. We run finetuning for 3 epochs using sequence length of 2048 on a single TPUv3-8 for 3 hours on top of BERTIN GPT-J-6B.
66
 
67
+ **español**
68
 
69
  Esta es una versión GPTQ de 4 bits del [bertin-project/bertin-gpt-j-6B-alpaca]( https://huggingface.co/bertin-project/bertin-gpt-j-6B-alpaca)
70