CodeXRyu commited on
Commit
d839810
·
verified ·
1 Parent(s): 49ce6c0

Enhanced meeting summarizer with QMSum dataset

Browse files
README.md ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ language: en
4
+ license: apache-2.0
5
+ tags:
6
+ - text-summarization
7
+ - meeting-summarization
8
+ - t5
9
+ - transformers
10
+ - qmsum
11
+ datasets:
12
+ - qmsum
13
+ metrics:
14
+ - rouge
15
+ pipeline_tag: summarization
16
+ ---
17
+
18
+ # Meeting Summarizer
19
+
20
+ This model is a fine-tuned version of `t5-small` for meeting summarization tasks.
21
+
22
+ ## Model Details
23
+ - **Base Model**: t5-small
24
+ - **Task**: Abstractive Meeting Summarization
25
+ - **Training Data**: QMSum Dataset + Enhanced Training
26
+ - **Parameters**: t5-small architecture
27
+
28
+ ## Training Configuration
29
+ - **Max Input Length**: 256 tokens
30
+ - **Max Output Length**: 64 tokens
31
+ - **Batch Size**: 16
32
+ - **Learning Rate**: 5e-05
33
+ - **Training Epochs**: 1
34
+ - **Training Samples**: N/A
35
+
36
+ ## Usage
37
+
38
+ ```python
39
+ from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
40
+
41
+ # Load model and tokenizer
42
+ tokenizer = AutoTokenizer.from_pretrained("CodeXRyu/meeting-summarizer")
43
+ model = AutoModelForSeq2SeqLM.from_pretrained("CodeXRyu/meeting-summarizer")
44
+
45
+ def generate_summary(meeting_text, max_length=150):
46
+ # Prepare input
47
+ input_text = "summarize: " + meeting_text
48
+ inputs = tokenizer(input_text, max_length=512, truncation=True, return_tensors="pt")
49
+
50
+ # Generate summary
51
+ summary_ids = model.generate(
52
+ inputs["input_ids"],
53
+ max_length=max_length,
54
+ num_beams=4,
55
+ length_penalty=2.0,
56
+ early_stopping=True
57
+ )
58
+
59
+ return tokenizer.decode(summary_ids[0], skip_special_tokens=True)
60
+
61
+ # Example usage
62
+ meeting_transcript = '''
63
+ John: Good morning team. Let's discuss our Q3 results.
64
+ Sarah: Our sales exceeded targets by 15%, reaching $2.1M in revenue.
65
+ Mike: The new marketing campaign was very effective.
66
+ John: Great work everyone. Let's plan for Q4.
67
+ '''
68
+
69
+ summary = generate_summary(meeting_transcript)
70
+ print(summary)
71
+ ```
72
+
73
+ ## Training Data
74
+ This model was trained on the QMSum dataset, which contains real meeting transcripts from multiple domains:
75
+ - Academic meetings
76
+ - Product development meetings
77
+ - Committee meetings
78
+
79
+ ## Performance
80
+ The model achieves competitive ROUGE scores on meeting summarization benchmarks.
81
+
82
+ ## Limitations
83
+ - Optimized for English meeting transcripts
84
+ - Performance may vary on very long meetings (>512 tokens input)
85
+ - Best suited for structured meeting formats with speaker labels
86
+
87
+ ## Citation
88
+ If you use this model, please cite:
89
+ ```
90
+ @misc{meeting-summarizer-codexryu,
91
+ author = {CodeXRyu},
92
+ title = {Meeting Summarizer},
93
+ year = {2025},
94
+ publisher = {Hugging Face},
95
+ url = {https://huggingface.co/CodeXRyu/meeting-summarizer}
96
+ }
97
+ ```
checkpoint-13/config.json CHANGED
@@ -9,6 +9,7 @@
9
  "decoder_start_token_id": 0,
10
  "dense_act_fn": "relu",
11
  "dropout_rate": 0.1,
 
12
  "eos_token_id": 1,
13
  "feed_forward_proj": "relu",
14
  "initializer_factor": 1.0,
@@ -53,8 +54,7 @@
53
  "prefix": "translate English to Romanian: "
54
  }
55
  },
56
- "torch_dtype": "float32",
57
- "transformers_version": "4.53.3",
58
  "use_cache": true,
59
  "vocab_size": 32128
60
  }
 
9
  "decoder_start_token_id": 0,
10
  "dense_act_fn": "relu",
11
  "dropout_rate": 0.1,
12
+ "dtype": "float32",
13
  "eos_token_id": 1,
14
  "feed_forward_proj": "relu",
15
  "initializer_factor": 1.0,
 
54
  "prefix": "translate English to Romanian: "
55
  }
56
  },
57
+ "transformers_version": "4.56.1",
 
58
  "use_cache": true,
59
  "vocab_size": 32128
60
  }
checkpoint-13/generation_config.json CHANGED
@@ -1,7 +1,9 @@
1
  {
2
  "_from_model_config": true,
3
  "decoder_start_token_id": 0,
4
- "eos_token_id": 1,
 
 
5
  "pad_token_id": 0,
6
- "transformers_version": "4.53.3"
7
  }
 
1
  {
2
  "_from_model_config": true,
3
  "decoder_start_token_id": 0,
4
+ "eos_token_id": [
5
+ 1
6
+ ],
7
  "pad_token_id": 0,
8
+ "transformers_version": "4.56.1"
9
  }
checkpoint-13/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0f11c74df923063d7933164164348d9be823169f5aceadd8cb0a28023f1ac33c
3
  size 242041896
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dee695d09e479b8a4c7d50551558815291b54f64c9a0da8779f318595db5c109
3
  size 242041896
checkpoint-13/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d75fb4a23edebce5ed875b14f08369d282ec73aee2ef4eabb12be446a8d91028
3
  size 484160331
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8780b109dfc27662eb489f645bfc3ad832cc24826668c1a8f9e819f58db915f
3
  size 484160331
checkpoint-13/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c3ceacd01edce49487ee660bb5920ae9feaf059c5e1b776cdf398943de0cb848
3
  size 14455
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b5761b78333a5edd4121621235e678648f1ce38c9d2b47fab279860f3f0fe5c
3
  size 14455
checkpoint-13/tokenizer.json CHANGED
@@ -1,19 +1,7 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 64,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
- "padding": {
10
- "strategy": "BatchLongest",
11
- "direction": "Right",
12
- "pad_to_multiple_of": null,
13
- "pad_id": 0,
14
- "pad_type_id": 0,
15
- "pad_token": "<pad>"
16
- },
17
  "added_tokens": [
18
  {
19
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
checkpoint-13/trainer_state.json CHANGED
@@ -3,17 +3,25 @@
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
  "epoch": 1.0,
6
- "eval_steps": 50,
7
  "global_step": 13,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
- "log_history": [],
12
- "logging_steps": 50,
 
 
 
 
 
 
 
 
13
  "max_steps": 13,
14
  "num_input_tokens_seen": 0,
15
  "num_train_epochs": 1,
16
- "save_steps": 50,
17
  "stateful_callbacks": {
18
  "TrainerControl": {
19
  "args": {
@@ -26,8 +34,8 @@
26
  "attributes": {}
27
  }
28
  },
29
- "total_flos": 6767090073600.0,
30
- "train_batch_size": 8,
31
  "trial_name": null,
32
  "trial_params": null
33
  }
 
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
  "epoch": 1.0,
6
+ "eval_steps": 20,
7
  "global_step": 13,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.7692307692307693,
14
+ "grad_norm": 4.401995658874512,
15
+ "learning_rate": 4.5e-05,
16
+ "loss": 3.8386,
17
+ "step": 10
18
+ }
19
+ ],
20
+ "logging_steps": 10,
21
  "max_steps": 13,
22
  "num_input_tokens_seen": 0,
23
  "num_train_epochs": 1,
24
+ "save_steps": 500,
25
  "stateful_callbacks": {
26
  "TrainerControl": {
27
  "args": {
 
34
  "attributes": {}
35
  }
36
  },
37
+ "total_flos": 13534180147200.0,
38
+ "train_batch_size": 16,
39
  "trial_name": null,
40
  "trial_params": null
41
  }
checkpoint-13/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:27571a09d091e90ec5d81bea87b6f6338619054c1522938a6dde9820cca267f4
3
- size 5841
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8d2ff0b4552b3f5fe2fa68c6574a4535d4620fe5a9b086dedbf0ccefd6ed20d
3
+ size 5905
config.json CHANGED
@@ -9,6 +9,7 @@
9
  "decoder_start_token_id": 0,
10
  "dense_act_fn": "relu",
11
  "dropout_rate": 0.1,
 
12
  "eos_token_id": 1,
13
  "feed_forward_proj": "relu",
14
  "initializer_factor": 1.0,
@@ -53,8 +54,7 @@
53
  "prefix": "translate English to Romanian: "
54
  }
55
  },
56
- "torch_dtype": "float32",
57
- "transformers_version": "4.53.3",
58
  "use_cache": true,
59
  "vocab_size": 32128
60
  }
 
9
  "decoder_start_token_id": 0,
10
  "dense_act_fn": "relu",
11
  "dropout_rate": 0.1,
12
+ "dtype": "float32",
13
  "eos_token_id": 1,
14
  "feed_forward_proj": "relu",
15
  "initializer_factor": 1.0,
 
54
  "prefix": "translate English to Romanian: "
55
  }
56
  },
57
+ "transformers_version": "4.56.1",
 
58
  "use_cache": true,
59
  "vocab_size": 32128
60
  }
generation_config.json CHANGED
@@ -1,7 +1,8 @@
1
  {
2
- "_from_model_config": true,
3
  "decoder_start_token_id": 0,
4
- "eos_token_id": 1,
 
 
5
  "pad_token_id": 0,
6
- "transformers_version": "4.53.3"
7
  }
 
1
  {
 
2
  "decoder_start_token_id": 0,
3
+ "eos_token_id": [
4
+ 1
5
+ ],
6
  "pad_token_id": 0,
7
+ "transformers_version": "4.56.1"
8
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0f11c74df923063d7933164164348d9be823169f5aceadd8cb0a28023f1ac33c
3
  size 242041896
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dee695d09e479b8a4c7d50551558815291b54f64c9a0da8779f318595db5c109
3
  size 242041896
tokenizer.json CHANGED
@@ -2,18 +2,11 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 64,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
- "padding": {
10
- "strategy": "BatchLongest",
11
- "direction": "Right",
12
- "pad_to_multiple_of": null,
13
- "pad_id": 0,
14
- "pad_type_id": 0,
15
- "pad_token": "<pad>"
16
- },
17
  "added_tokens": [
18
  {
19
  "id": 0,
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 256,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
+ "padding": null,
 
 
 
 
 
 
 
10
  "added_tokens": [
11
  {
12
  "id": 0,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:27571a09d091e90ec5d81bea87b6f6338619054c1522938a6dde9820cca267f4
3
- size 5841
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8d2ff0b4552b3f5fe2fa68c6574a4535d4620fe5a9b086dedbf0ccefd6ed20d
3
+ size 5905
training_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_name": "t5-small",
3
+ "max_input_length": 256,
4
+ "max_target_length": 64,
5
+ "batch_size": 16,
6
+ "learning_rate": 5e-05,
7
+ "num_epochs": 1,
8
+ "train_size": 200,
9
+ "val_size": 40,
10
+ "test_size": 20
11
+ }