Update README.md
Browse files
README.md
CHANGED
@@ -3,6 +3,7 @@ license: apache-2.0
|
|
3 |
datasets:
|
4 |
- liuhaotian/LLaVA-Pretrain
|
5 |
- liuhaotian/LLaVA-Instruct-150K
|
|
|
6 |
---
|
7 |
## Model
|
8 |
llava-dinov2-internlm2-7b-v1 is a LLaVA model fine-tuned from [InternLM2-Chat-7B](https://huggingface.co/internlm/internlm2-chat-7b) and [Dinov2-large](https://huggingface.co/facebook/dinov2-large) with [LLaVA-Pretrain](liuhaotian/LLaVA-Pretrain) and [LLaVA-Instruct](https://huggingface.co/datasets/liuhaotian/LLaVA-Instruct-150K) by [XTuner](https://github.com/InternLM/xtuner). I thank the help of [Zhihao Lin](https://github.com/LZHgrla) and [pppppM](https://github.com/pppppM) from the Xtuner team. I also thank the Huggingface transformers team for approving [my pull request](https://github.com/huggingface/transformers/pull/28504) so training Dinov2 in bf16 becomes possible.
|
|
|
3 |
datasets:
|
4 |
- liuhaotian/LLaVA-Pretrain
|
5 |
- liuhaotian/LLaVA-Instruct-150K
|
6 |
+
pipeline_tag: visual-question-answering
|
7 |
---
|
8 |
## Model
|
9 |
llava-dinov2-internlm2-7b-v1 is a LLaVA model fine-tuned from [InternLM2-Chat-7B](https://huggingface.co/internlm/internlm2-chat-7b) and [Dinov2-large](https://huggingface.co/facebook/dinov2-large) with [LLaVA-Pretrain](liuhaotian/LLaVA-Pretrain) and [LLaVA-Instruct](https://huggingface.co/datasets/liuhaotian/LLaVA-Instruct-150K) by [XTuner](https://github.com/InternLM/xtuner). I thank the help of [Zhihao Lin](https://github.com/LZHgrla) and [pppppM](https://github.com/pppppM) from the Xtuner team. I also thank the Huggingface transformers team for approving [my pull request](https://github.com/huggingface/transformers/pull/28504) so training Dinov2 in bf16 becomes possible.
|