faresfawzi commited on
Commit
0098913
·
verified ·
1 Parent(s): c3d970c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -10
README.md CHANGED
@@ -12,7 +12,7 @@ tags:
12
  license: apache-2.0
13
  ---
14
 
15
- # Model Card for faresfawzi/ToolACE-2-8B-SCRIBE
16
 
17
  ## Abstract
18
  Language models can be used to provide interactive, personalized student feedback in educational settings. However, real-world deployment faces three key challenges: privacy concerns, limited computational resources, and the need for pedagogically valid responses. These constraints require small, open-source models that can run locally and reliably ground their outputs in correct information. We introduce SCRIBE, a framework for multi-hop, tool-augmented reasoning designed to generate valid responses to student questions about feedback reports. SCRIBE combines domain-specific tools with a self-reflective inference pipeline that supports iterative reasoning, tool use, and error recovery. We distil these capabilities into 3B and 8B models via two-stage LoRA fine-tuning on synthetic GPT-4o-generated data. Evaluation with a human-aligned GPT-Judge and a user study with 108 students shows that 8B-SCRIBE models achieve comparable or superior quality to much larger models in key dimensions such as relevance and actionability, while being perceived on par with GPT-4o and Llama-3.3 70B by students. These findings demonstrate the viability of SCRIBE for low-resource, privacy-sensitive educational applications.
@@ -21,8 +21,8 @@ Language models can be used to provide interactive, personalized student feedbac
21
 
22
  ## Model Description
23
 
24
- **ToolACE-2-8B-SCRIBE** is a fine-tuned large language model for **interactive educational feedback**.
25
- It builds on **Team-ACE/ToolACE-2.5-Llama-3.1-8B** and incorporates the **SCRIBE framework**: structured chain reasoning with multi-hop tool calling and self-reflection, enabling small models to deliver **pedagogically valid, actionable, and context-grounded explanations** to student questions.
26
 
27
  - **Developed by:** EPFL (Machine Learning for Education Lab)
28
  - **Paper:** *SCRIBE: Structured Chain Reasoning for Interactive Behavior Explanations using Tool Calling*
@@ -30,7 +30,7 @@ It builds on **Team-ACE/ToolACE-2.5-Llama-3.1-8B** and incorporates the **SCRIBE
30
  - **Model type:** Tool-augmented 8B LLM fine-tuned with two-stage LoRA
31
  - **Languages:** English
32
  - **License:** Apache 2.0
33
- - **Finetuned from:** `Team-ACE/ToolACE-2.5-Llama-3.1-8B`
34
 
35
  ---
36
 
@@ -71,12 +71,6 @@ The model is designed to:
71
 
72
  If you use this model, please cite:
73
 
74
- **APA**
75
- Fawzi, F., Swamy, V., Glandorf, D., Nazaretsky, T., & Käser, T. (2025).
76
- *SCRIBE: Structured Chain Reasoning for Interactive Behavior Explanations using Tool Calling*. EPFL.
77
-
78
- Liu, W., Huang, X., Zeng, X., Hao, X., Yu, S., Li, D., Wang, S., Gan, W., Liu, Z., Yu, Y., Wang, Z., Wang, Y., Ning, W., Hou, Y., Wang, B., Wu, C., Xinzhi, W., Liu, Y., Wang, Y., Tang, D., Tu, D., Shang, L., Jiang, X., Tang, R., Lian, D., Liu, Q., & Chen, E. (2025). *ToolACE: Winning the Points of LLM Function Calling*. In ICLR 2025.
79
-
80
  **BibTeX**
81
  ```bibtex
82
  @inproceedings{2025-EMNLP-Scribe,
 
12
  license: apache-2.0
13
  ---
14
 
15
+ # Model Card for faresfawzi/Qwen3-8B-SCRIBE
16
 
17
  ## Abstract
18
  Language models can be used to provide interactive, personalized student feedback in educational settings. However, real-world deployment faces three key challenges: privacy concerns, limited computational resources, and the need for pedagogically valid responses. These constraints require small, open-source models that can run locally and reliably ground their outputs in correct information. We introduce SCRIBE, a framework for multi-hop, tool-augmented reasoning designed to generate valid responses to student questions about feedback reports. SCRIBE combines domain-specific tools with a self-reflective inference pipeline that supports iterative reasoning, tool use, and error recovery. We distil these capabilities into 3B and 8B models via two-stage LoRA fine-tuning on synthetic GPT-4o-generated data. Evaluation with a human-aligned GPT-Judge and a user study with 108 students shows that 8B-SCRIBE models achieve comparable or superior quality to much larger models in key dimensions such as relevance and actionability, while being perceived on par with GPT-4o and Llama-3.3 70B by students. These findings demonstrate the viability of SCRIBE for low-resource, privacy-sensitive educational applications.
 
21
 
22
  ## Model Description
23
 
24
+ **Qwen3-8B-SCRIBE** is a fine-tuned large language model for **interactive educational feedback**.
25
+ It builds on **Qwen/Qwen3-8B** and incorporates the **SCRIBE framework**: structured chain reasoning with multi-hop tool calling and self-reflection, enabling small models to deliver **pedagogically valid, actionable, and context-grounded explanations** to student questions.
26
 
27
  - **Developed by:** EPFL (Machine Learning for Education Lab)
28
  - **Paper:** *SCRIBE: Structured Chain Reasoning for Interactive Behavior Explanations using Tool Calling*
 
30
  - **Model type:** Tool-augmented 8B LLM fine-tuned with two-stage LoRA
31
  - **Languages:** English
32
  - **License:** Apache 2.0
33
+ - **Finetuned from:** `Qwen/Qwen3-8B`
34
 
35
  ---
36
 
 
71
 
72
  If you use this model, please cite:
73
 
 
 
 
 
 
 
74
  **BibTeX**
75
  ```bibtex
76
  @inproceedings{2025-EMNLP-Scribe,