LLaVA_int4 / README.md
emon-j's picture
Update README.md
c031eb3
|
raw
history blame contribute delete
No virus
570 Bytes
---
license: cc
---
# LLaVA Compress model weights to int4 using NNCF
[LLaVA](https://llava-vl.github.io) (Large Language and Vision Assistant) is large multimodal model that aims to develop a general-purpose visual assistant that can follow both language and image instructions to complete various real-world tasks.
LLaVA connects pre-trained [CLIP ViT-L/14](https://openai.com/research/clip) visual encoder and large language model like Vicuna, LLaMa v2 or MPT, using a simple projection matrix.
![vlp_matrix.png](https://llava-vl.github.io/images/llava_arch.png)