Text Generation
Transformers
PyTorch
llava
medical
tnaumann bsnelling commited on
Commit
0dd1c26
·
verified ·
1 Parent(s): 305b38d

Upload data_summary_card.md (#5)

Browse files

- Upload data_summary_card.md (aed283b3ca53d768d756612fe2dd217a631db68a)


Co-authored-by: Blake S <[email protected]>

Files changed (1) hide show
  1. data_summary_card.md +145 -0
data_summary_card.md ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+
3
+ Data Summary for microsoft_llava-med-7b-delta
4
+
5
+
6
+
7
+
8
+
9
+ ## 1. General information
10
+
11
+ **1.0.1 Version of the Summary:** 1.0
12
+
13
+
14
+
15
+ **1.0.2 Last update:** 24-Nov-2025
16
+
17
+
18
+
19
+ ## 1.1 Model Developer Identification
20
+
21
+ **1.1.1 Model Developer name and contact details:** Microsoft Corporation at One Microsoft Way, Redmond, WA 98052. Tel: 425-882-8080.
22
+
23
+
24
+
25
+ ## 1.2 Model Identification
26
+
27
+ **1.2.1 Versioned model name(s):** LLaVA-Med-7b-delta
28
+
29
+
30
+
31
+ **1.2.2 Model release date:** 09-Nov-2023
32
+
33
+
34
+
35
+ ## 1.3 Overall training data size and characteristics
36
+
37
+ ### 1.3.1 Size of dataset and characteristics
38
+
39
+ **1.3.1.A Text training data size:** Less than 1 billion tokens
40
+
41
+
42
+
43
+ **1.3.1.B Text training data content:** LLaVA-Med builds upon PMC-15M dataset, which is a large-scale parallel image-text dataset for biomedical vision-language processing. It contains 15 million figure-caption pairs extracted from biomedical research articles in PubMed Central.
44
+
45
+
46
+
47
+ **1.3.1.C Image training data size:** Less than 1 billion tokens
48
+
49
+
50
+
51
+
52
+
53
+ **1.3.1.D Image training data content:** Figure images from biomedical research articles in PubMed Central paired with captions across modalities including microscopy, radiography, histology, CT, MRI, chest X-ray, pathology figures
54
+
55
+
56
+
57
+ **1.3.1.E Audio training data size:** Not applicable
58
+
59
+
60
+
61
+ **1.3.1.F Audio training data content:** Not applicable
62
+
63
+
64
+
65
+ **1.3.1.G Video training data size:** Not applicable
66
+
67
+
68
+
69
+ **1.3.1.H Video training data content:** Not applicable
70
+
71
+
72
+
73
+ **1.3.1.I Other training data size:** The model also uses instruction-following conversational data generated by GPT-4 from captions and inline mentions; sizes include versions of 10K and 60K samples
74
+
75
+
76
+
77
+ **1.3.1.J Other training data content:** GPT-4 generated multi-round biomedical instruction-following conversations derived from figure captions and sentences mentioning figures (inline mentions) from PubMed Central articles
78
+
79
+
80
+
81
+ **1.3.2 Latest date of data acquisition/collection for model training:** 01-May-2023
82
+
83
+ **1.3.3 Is data collection ongoing to update the model with new data collection after deployment?** No
84
+
85
+
86
+
87
+ **1.3.4 Date the training dataset was first used to train the model:** 01-May-2023
88
+
89
+
90
+
91
+ **1.3.5 Rationale or purpose of data selection:** Datasets were selected to adapt a general-domain multimodal model to biomedical vision-language tasks by aligning biomedical concepts and enabling instruction-following. PMC-15M figure-caption pairs provide broad coverage of biomedical images and terminology, while GPT-4 generated conversations from captions and inline mentions create diverse, open-ended instruction data to support visual chat and VQA performance
92
+
93
+
94
+
95
+ ## 2. List of data sources
96
+
97
+ ### 2.1 Publicly available datasets
98
+
99
+ **2.1.1 Have you used publicly available datasets to train the model?** Yes
100
+
101
+
102
+
103
+ ## 2.2 Private non-publicly available datasets obtained from third parties
104
+
105
+ ### 2.2.1 Datasets commercially licensed by rights holders or their representatives
106
+
107
+ **2.2.1.A Have you concluded transactional commercial licensing agreement(s) with rights holder(s) or with their representatives?** This information cannot be provided due to unavailability of the underlying data (e.g., loss, corruption, or other access limitations)
108
+
109
+ ### 2.2.2 Private datasets obtained from other third-parties
110
+
111
+ ** 2.2.2.A Have you obtained private datasets from third parties that are not licensed as described in Section 2.2.1, such as data obtained from providers of private databases, or data intermediaries?** No
112
+
113
+
114
+
115
+ ## 2.3 Personal Information
116
+
117
+ **2.3.1 Was personal data used to train the model?** Microsoft follows all relevant laws and regulations pertaining to personal information.
118
+
119
+
120
+
121
+ ## 2.4 Synthetic data
122
+
123
+ **2.4.1 Was any synthetic AI-generated data used to train the model?** Yes
124
+
125
+
126
+
127
+ ## 3. Data processing aspects
128
+
129
+ ## 3.1 Respect of reservation of rights from text and data mining exception or limitation
130
+
131
+ **3.1.1 Does this dataset include any data protected by copyright, trademark, or patent?** Microsoft follows all required regulations and laws for processing data protected by copyright, trademark, or patent.
132
+
133
+
134
+
135
+ ### 3.2 Other information
136
+
137
+ **3.2.1 Does the dataset include information about consumer groups without revealing individual consumer identities?** Microsoft follows all required regulations and laws for protecting consumer identities.
138
+
139
+
140
+
141
+ **3.2.2 Was the dataset cleaned or modified before model training?** Yes
142
+
143
+
144
+
145
+