--- language: en tags: - question-answering - squad - transformers datasets: - squad metrics: - exact_match - f1 model-index: - name: HariomSahu/distilbert-base-uncased-squadv1-adam-lin-e526 results: - task: type: question-answering name: Question Answering dataset: name: SQuAD type: squad metrics: - type: exact_match value: 76.35761589403974 - type: f1 value: N/A --- # distilbert-base-uncased fine-tuned on SQuAD This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the SQuAD dataset. ## Training Details ### Training Hyperparameters - **Model**: distilbert-base-uncased - **Dataset**: SQuAD - **Optimizer**: adamw - **Learning Rate Scheduler**: linear - **Learning Rate**: 2e-05 - **Batch Size**: 16 per device - **Total Batch Size**: 64 - **Epochs**: 5 (with early stopping) - **Weight Decay**: 0.01 - **Warmup Ratio**: 0.1 - **Max Gradient Norm**: 1.0 ### Early Stopping - **Patience**: 3 - **Metric**: exact_match - **Best Epoch**: 3 ## Usage ```python from transformers import AutoTokenizer, AutoModelForQuestionAnswering tokenizer = AutoTokenizer.from_pretrained("HariomSahu/distilbert-base-uncased-squadv1-adam-lin-e526") model = AutoModelForQuestionAnswering.from_pretrained("HariomSahu/distilbert-base-uncased-squadv1-adam-lin-e526") # Example usage question = "What is the capital of France?" context = "France is a country in Europe. Its capital city is Paris." inputs = tokenizer(question, context, return_tensors="pt") outputs = model(**inputs) # Get answer start_scores, end_scores = outputs.start_logits, outputs.end_logits start_index = start_scores.argmax() end_index = end_scores.argmax() answer = tokenizer.decode(inputs["input_ids"][0][start_index:end_index+1]) print(f"Answer: {answer}") ``` ## Evaluation Results The model achieved the following results on the evaluation set: - **Exact Match**: 76.3103 - **F1 Score**: 84.5885 ## Training Configuration Hash Config Hash: e5265f15 This hash can be used to reproduce the exact training configuration.