Image-Text-to-Text
Transformers
llava_llama
text-generation

Add link to paper

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +3 -1
README.md CHANGED
@@ -1,13 +1,15 @@
1
  ---
2
- license: apache-2.0
3
  datasets:
4
  - liuhaotian/LLaVA-CC3M-Pretrain-595K
5
  library_name: transformers
 
6
  pipeline_tag: image-text-to-text
7
  ---
8
 
9
  # Model Card: LLaVA_MORE-llama_3_1-8B-pretrain
10
 
 
 
11
  ```LLaVA-MORE``` enhances the well-known LLaVA architecture by integrating the use of LLaMA 3.1 as the language model. We are publicly releasing the checkpoints for stages one and two for the first model with 8B parameters.
12
 
13
  In this model space, you will find the stage one (pretrain) weights of LLaVA-MORE LLaMA 3.1 8B.
 
1
  ---
 
2
  datasets:
3
  - liuhaotian/LLaVA-CC3M-Pretrain-595K
4
  library_name: transformers
5
+ license: apache-2.0
6
  pipeline_tag: image-text-to-text
7
  ---
8
 
9
  # Model Card: LLaVA_MORE-llama_3_1-8B-pretrain
10
 
11
+ This repository contains the model described in [LLaVA-MORE: Enhancing Visual Instruction Tuning with LLaMA 3.1](https://huggingface.co/papers/2503.15621).
12
+
13
  ```LLaVA-MORE``` enhances the well-known LLaVA architecture by integrating the use of LLaMA 3.1 as the language model. We are publicly releasing the checkpoints for stages one and two for the first model with 8B parameters.
14
 
15
  In this model space, you will find the stage one (pretrain) weights of LLaVA-MORE LLaMA 3.1 8B.