Add link to paper
#1
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,13 +1,15 @@
|
|
| 1 |
---
|
| 2 |
-
license: apache-2.0
|
| 3 |
datasets:
|
| 4 |
- liuhaotian/LLaVA-CC3M-Pretrain-595K
|
| 5 |
library_name: transformers
|
|
|
|
| 6 |
pipeline_tag: image-text-to-text
|
| 7 |
---
|
| 8 |
|
| 9 |
# Model Card: LLaVA_MORE-llama_3_1-8B-pretrain
|
| 10 |
|
|
|
|
|
|
|
| 11 |
```LLaVA-MORE``` enhances the well-known LLaVA architecture by integrating the use of LLaMA 3.1 as the language model. We are publicly releasing the checkpoints for stages one and two for the first model with 8B parameters.
|
| 12 |
|
| 13 |
In this model space, you will find the stage one (pretrain) weights of LLaVA-MORE LLaMA 3.1 8B.
|
|
|
|
| 1 |
---
|
|
|
|
| 2 |
datasets:
|
| 3 |
- liuhaotian/LLaVA-CC3M-Pretrain-595K
|
| 4 |
library_name: transformers
|
| 5 |
+
license: apache-2.0
|
| 6 |
pipeline_tag: image-text-to-text
|
| 7 |
---
|
| 8 |
|
| 9 |
# Model Card: LLaVA_MORE-llama_3_1-8B-pretrain
|
| 10 |
|
| 11 |
+
This repository contains the model described in [LLaVA-MORE: Enhancing Visual Instruction Tuning with LLaMA 3.1](https://huggingface.co/papers/2503.15621).
|
| 12 |
+
|
| 13 |
```LLaVA-MORE``` enhances the well-known LLaVA architecture by integrating the use of LLaMA 3.1 as the language model. We are publicly releasing the checkpoints for stages one and two for the first model with 8B parameters.
|
| 14 |
|
| 15 |
In this model space, you will find the stage one (pretrain) weights of LLaVA-MORE LLaMA 3.1 8B.
|