Commit
•
5f5d133
1
Parent(s):
f75c568
Update README.md
Browse files
README.md
CHANGED
@@ -11,13 +11,9 @@ datasets:
|
|
11 |
|
12 |
HuggingFaceH4/vsft-llava-1.5-7b-hf-trl is a Vision Language Model, created by performing VSFT on the [llava-hf/llava-1.5-7b-hf](https://huggingface.co/llava-hf/llava-1.5-7b-hf) model
|
13 |
|
14 |
-
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6200d0a443eb0913fa2df7cc/q5GXv6Om4Hf2n6IB3e7DQ.png)
|
15 |
|
16 |
-
|
17 |
-
|
18 |
-
Check out also the Google Colab demo to run Llava on a free-tier Google Colab instance: [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1qsl6cd2c8gGtEW1xV5io7S8NHh-Cp1TV?usp=sharing)
|
19 |
-
|
20 |
-
Or check out our Spaces demo! [![Open in Spaces](https://huggingface.co/datasets/huggingface/badges/resolve/main/open-in-hf-spaces-md-dark.svg)](https://huggingface.co/spaces/llava-hf/llava-4bit)
|
21 |
|
22 |
|
23 |
## Model details
|
@@ -30,7 +26,7 @@ It is an auto-regressive language model, based on the transformer architecture.
|
|
30 |
The model was trained on April the 11th 2024
|
31 |
|
32 |
**Example training script**
|
33 |
-
https://github.com/huggingface/trl/blob/main/examples/scripts/vsft_llava.py
|
34 |
|
35 |
## How to use the model
|
36 |
|
@@ -117,4 +113,16 @@ model = LlavaForConditionalGeneration.from_pretrained(
|
|
117 |
|
118 |
## License
|
119 |
Llama 2 is licensed under the LLAMA 2 Community License,
|
120 |
-
Copyright (c) Meta Platforms, Inc. All Rights Reserved.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
|
12 |
HuggingFaceH4/vsft-llava-1.5-7b-hf-trl is a Vision Language Model, created by performing VSFT on the [llava-hf/llava-1.5-7b-hf](https://huggingface.co/llava-hf/llava-1.5-7b-hf) model
|
13 |
|
14 |
+
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6200d0a443eb0913fa2df7cc/q5GXv6Om4Hf2n6IB3e7DQ.png) model with 260k image and conversation pairs from the [HuggingFaceH4/llava-instruct-mix-vsft](https://huggingface.co/datasets/HuggingFaceH4/llava-instruct-mix-vsft) dataset.
|
15 |
|
16 |
+
Or check out our Spaces demo! [![Open in Spaces](https://huggingface.co/datasets/huggingface/badges/resolve/main/open-in-hf-spaces-md-dark.svg)](https://huggingface.co/spaces/HuggingFaceH4/vlm-playground)
|
|
|
|
|
|
|
|
|
17 |
|
18 |
|
19 |
## Model details
|
|
|
26 |
The model was trained on April the 11th 2024
|
27 |
|
28 |
**Example training script**
|
29 |
+
[Train a VLM yourself with our TRL example](https://github.com/huggingface/trl/blob/main/examples/scripts/vsft_llava.py)
|
30 |
|
31 |
## How to use the model
|
32 |
|
|
|
113 |
|
114 |
## License
|
115 |
Llama 2 is licensed under the LLAMA 2 Community License,
|
116 |
+
Copyright (c) Meta Platforms, Inc. All Rights Reserved.
|
117 |
+
|
118 |
+
## Citation
|
119 |
+
```
|
120 |
+
@misc{vonwerra2022trl,
|
121 |
+
author = {Edward Beeching and Kashif Rasul and Younes Belkada and Shengyi Huangand and Leandro von Werra and Lewis Tunstall},
|
122 |
+
title = {TRL: Transformer Reinforcement Learning},
|
123 |
+
year = {2020},
|
124 |
+
publisher = {GitHub},
|
125 |
+
journal = {GitHub repository},
|
126 |
+
howpublished = {\url{https://github.com/huggingface/trl}}
|
127 |
+
}
|
128 |
+
```
|