Update README.md
#1
by
Jason0214
- opened
README.md
CHANGED
@@ -16,9 +16,9 @@ Crystal-based models mimics the training recipie used for Vicuna 7B in LLaVA mul
|
|
16 |
|
17 |
| LLM Backbone | MME-P | MME-C | POPE | SciQA | TextVQA |
|
18 |
|-----------------------------------|---------|--------|-------|--------|---------|
|
19 |
-
| CrystalCoder-7B | 1359.83 | 238.92 | 86.
|
20 |
| CrystalChat-7B | 1456.53 | **308.21** | 86.96 | 67.77 | **57.84** |
|
21 |
-
| Vicuna-7B | **1481.12** | 302.85 | **87.
|
22 |
|
23 |
*Table: Comparison of different LLM backbones on visual language understanding benchmarks. All models are instruction-tuned on the general domain data (i.e. LLaVA)*
|
24 |
|
|
|
16 |
|
17 |
| LLM Backbone | MME-P | MME-C | POPE | SciQA | TextVQA |
|
18 |
|-----------------------------------|---------|--------|-------|--------|---------|
|
19 |
+
| CrystalCoder-7B | 1359.83 | 238.92 | 86.18 | 64.15 | 50.39 |
|
20 |
| CrystalChat-7B | 1456.53 | **308.21** | 86.96 | 67.77 | **57.84** |
|
21 |
+
| Vicuna-7B | **1481.12** | 302.85 | **87.17** | **67.97** | 56.49 |
|
22 |
|
23 |
*Table: Comparison of different LLM backbones on visual language understanding benchmarks. All models are instruction-tuned on the general domain data (i.e. LLaVA)*
|
24 |
|