edbeeching HF staff commited on
Commit
5f5d133
1 Parent(s): f75c568

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -8
README.md CHANGED
@@ -11,13 +11,9 @@ datasets:
11
 
12
  HuggingFaceH4/vsft-llava-1.5-7b-hf-trl is a Vision Language Model, created by performing VSFT on the [llava-hf/llava-1.5-7b-hf](https://huggingface.co/llava-hf/llava-1.5-7b-hf) model
13
 
14
- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6200d0a443eb0913fa2df7cc/q5GXv6Om4Hf2n6IB3e7DQ.png)
15
 
16
- Below is the model card of Llava model 7b, which is copied from the original Llava model card that you can find [here](https://huggingface.co/liuhaotian/llava-v1.5-13b).
17
-
18
- Check out also the Google Colab demo to run Llava on a free-tier Google Colab instance: [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1qsl6cd2c8gGtEW1xV5io7S8NHh-Cp1TV?usp=sharing)
19
-
20
- Or check out our Spaces demo! [![Open in Spaces](https://huggingface.co/datasets/huggingface/badges/resolve/main/open-in-hf-spaces-md-dark.svg)](https://huggingface.co/spaces/llava-hf/llava-4bit)
21
 
22
 
23
  ## Model details
@@ -30,7 +26,7 @@ It is an auto-regressive language model, based on the transformer architecture.
30
  The model was trained on April the 11th 2024
31
 
32
  **Example training script**
33
- https://github.com/huggingface/trl/blob/main/examples/scripts/vsft_llava.py
34
 
35
  ## How to use the model
36
 
@@ -117,4 +113,16 @@ model = LlavaForConditionalGeneration.from_pretrained(
117
 
118
  ## License
119
  Llama 2 is licensed under the LLAMA 2 Community License,
120
- Copyright (c) Meta Platforms, Inc. All Rights Reserved.
 
 
 
 
 
 
 
 
 
 
 
 
 
11
 
12
  HuggingFaceH4/vsft-llava-1.5-7b-hf-trl is a Vision Language Model, created by performing VSFT on the [llava-hf/llava-1.5-7b-hf](https://huggingface.co/llava-hf/llava-1.5-7b-hf) model
13
 
14
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6200d0a443eb0913fa2df7cc/q5GXv6Om4Hf2n6IB3e7DQ.png) model with 260k image and conversation pairs from the [HuggingFaceH4/llava-instruct-mix-vsft](https://huggingface.co/datasets/HuggingFaceH4/llava-instruct-mix-vsft) dataset.
15
 
16
+ Or check out our Spaces demo! [![Open in Spaces](https://huggingface.co/datasets/huggingface/badges/resolve/main/open-in-hf-spaces-md-dark.svg)](https://huggingface.co/spaces/HuggingFaceH4/vlm-playground)
 
 
 
 
17
 
18
 
19
  ## Model details
 
26
  The model was trained on April the 11th 2024
27
 
28
  **Example training script**
29
+ [Train a VLM yourself with our TRL example](https://github.com/huggingface/trl/blob/main/examples/scripts/vsft_llava.py)
30
 
31
  ## How to use the model
32
 
 
113
 
114
  ## License
115
  Llama 2 is licensed under the LLAMA 2 Community License,
116
+ Copyright (c) Meta Platforms, Inc. All Rights Reserved.
117
+
118
+ ## Citation
119
+ ```
120
+ @misc{vonwerra2022trl,
121
+ author = {Edward Beeching and Kashif Rasul and Younes Belkada and Shengyi Huangand and Leandro von Werra and Lewis Tunstall},
122
+ title = {TRL: Transformer Reinforcement Learning},
123
+ year = {2020},
124
+ publisher = {GitHub},
125
+ journal = {GitHub repository},
126
+ howpublished = {\url{https://github.com/huggingface/trl}}
127
+ }
128
+ ```