rabah2026 commited on
Commit
80cad1a
·
verified ·
1 Parent(s): ac0d78a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -18,7 +18,7 @@ license: apache-2.0
18
  ## Dataset Description:
19
  Ayah-Corpus is a large-scale, multi-reciter Arabic speech dataset meticulously curated for Automatic Speech Recognition (ASR) tasks. It consists of high-quality audio recordings of Quranic verses (Ayahs) paired with their corresponding exact transcriptions. The audio is sourced from two primary repositories: Al-Quran.cloud and EveryAyah.com.
20
 
21
- This dataset is specifically designed to facilitate the development of ASR models for Quranic Arabic, which features a distinct vocabulary, phonetic structure, and recitation style (Tajweed) compared to Modern Standard Arabic or colloquial dialects. All audio files have been standardized to a 16kHz sampling rate to be compatible with most modern ASR pipelines.
22
 
23
  ## Dataset Structure:
24
  The dataset is divided into train, validation, and test splits, ensuring a strict separation of reciters between the sets to evaluate model generalization to unseen voices.
 
18
  ## Dataset Description:
19
  Ayah-Corpus is a large-scale, multi-reciter Arabic speech dataset meticulously curated for Automatic Speech Recognition (ASR) tasks. It consists of high-quality audio recordings of Quranic verses (Ayahs) paired with their corresponding exact transcriptions. The audio is sourced from two primary repositories: Al-Quran.cloud and EveryAyah.com.
20
 
21
+ This dataset is specifically designed to facilitate the development of ASR models for Quranic Arabic, which features a distinct vocabulary, phonetic structure, and recitation style (Tajweed) compared to Modern Standard Arabic or colloquial dialects. All audio files have been standardized to a **16kHz** sampling rate to be compatible with most modern ASR pipelines.
22
 
23
  ## Dataset Structure:
24
  The dataset is divided into train, validation, and test splits, ensuring a strict separation of reciters between the sets to evaluate model generalization to unseen voices.