fawzanaramam commited on
Commit
4cf9d04
·
verified ·
1 Parent(s): 4ba4139

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +59 -38
README.md CHANGED
@@ -4,59 +4,80 @@ language:
4
  license: apache-2.0
5
  base_model: openai/whisper-medium
6
  tags:
7
- - generated_from_trainer
 
 
 
 
8
  datasets:
9
  - fawzanaramam/the-amma-juz
10
  model-index:
11
  - name: Whisper Medium Finetuned on Amma Juz of Quran
12
- results: []
 
 
 
 
 
 
 
 
 
 
 
13
  ---
14
 
15
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
- should probably proofread and complete it, then remove this comment. -->
17
-
18
  # Whisper Medium Finetuned on Amma Juz of Quran
19
 
20
- This model is a fine-tuned version of [openai/whisper-medium](https://huggingface.co/openai/whisper-medium) on the The Truth Amma Juz dataset.
21
- It achieves the following results on the evaluation set:
22
- - eval_loss: 0.0032
23
- - eval_wer: 0.5102
24
- - eval_runtime: 47.9061
25
- - eval_samples_per_second: 2.087
26
- - eval_steps_per_second: 0.271
27
- - epoch: 0.6653
28
- - step: 950
29
 
30
- ## Model description
31
 
32
- More information needed
 
 
 
 
 
33
 
34
- ## Intended uses & limitations
35
 
36
- More information needed
37
 
38
- ## Training and evaluation data
 
 
 
39
 
40
- More information needed
 
 
 
41
 
42
- ## Training procedure
43
 
44
- ### Training hyperparameters
45
 
 
 
 
46
  The following hyperparameters were used during training:
47
- - learning_rate: 1e-05
48
- - train_batch_size: 16
49
- - eval_batch_size: 8
50
- - seed: 42
51
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
- - lr_scheduler_type: linear
53
- - lr_scheduler_warmup_steps: 10
54
- - num_epochs: 3.0
55
- - mixed_precision_training: Native AMP
56
-
57
- ### Framework versions
58
-
59
- - Transformers 4.41.1
60
- - Pytorch 2.2.1+cu121
61
- - Datasets 2.19.1
62
- - Tokenizers 0.19.1
 
4
  license: apache-2.0
5
  base_model: openai/whisper-medium
6
  tags:
7
+ - fine-tuned
8
+ - Quran
9
+ - automatic-speech-recognition
10
+ - arabic
11
+ - whisper
12
  datasets:
13
  - fawzanaramam/the-amma-juz
14
  model-index:
15
  - name: Whisper Medium Finetuned on Amma Juz of Quran
16
+ results:
17
+ - task:
18
+ type: automatic-speech-recognition
19
+ name: Speech Recognition
20
+ dataset:
21
+ name: The Amma Juz Dataset
22
+ type: fawzanaramam/the-amma-juz
23
+ metrics:
24
+ - type: eval_loss
25
+ value: 0.0032
26
+ - type: eval_wer
27
+ value: 0.5102
28
  ---
29
 
 
 
 
30
  # Whisper Medium Finetuned on Amma Juz of Quran
31
 
32
+ This model is a fine-tuned version of [openai/whisper-medium](https://huggingface.co/openai/whisper-medium), tailored for transcribing Arabic audio with a focus on Quranic recitation from the *Amma Juz* dataset. It is optimized for high accuracy and minimal word error rates in Quranic transcription tasks.
33
+
34
+ ## Model Description
35
+
36
+ Whisper Medium is a transformer-based automatic speech recognition (ASR) model developed by OpenAI. This fine-tuned version leverages the *Amma Juz* dataset to enhance performance in recognizing Quranic recitations. The model is particularly effective for Arabic speech transcription in religious contexts, while retaining Whisper's general-purpose ASR capabilities.
 
 
 
 
37
 
38
+ ## Performance Metrics
39
 
40
+ On the evaluation set, the model achieved:
41
+ - **Evaluation Loss**: 0.0032
42
+ - **Word Error Rate (WER)**: 0.5102%
43
+ - **Evaluation Runtime**: 47.9061 seconds
44
+ - **Evaluation Samples per Second**: 2.087
45
+ - **Evaluation Steps per Second**: 0.271
46
 
47
+ These metrics demonstrate the model's superior accuracy and efficiency, making it suitable for applications requiring high-quality Quranic transcription.
48
 
49
+ ## Intended Uses & Limitations
50
 
51
+ ### Intended Uses
52
+ - **Speech-to-text transcription** of Quranic recitation in Arabic, specifically from the *Amma Juz*.
53
+ - Research and development of tools for Quranic education and learning.
54
+ - Projects focused on Arabic ASR in religious and educational domains.
55
 
56
+ ### Limitations
57
+ - The model is fine-tuned on Quranic recitations and may not generalize well to non-Quranic Arabic speech or casual conversations.
58
+ - Variations in recitation style, audio quality, or heavy accents may impact transcription accuracy.
59
+ - For optimal performance, use clean and high-quality audio inputs.
60
 
61
+ ## Training and Evaluation Data
62
 
63
+ The model was trained using the *Amma Juz* dataset, which includes Quranic audio recordings and corresponding transcripts. The dataset was carefully curated to ensure the integrity and accuracy of Quranic content.
64
 
65
+ ## Training Procedure
66
+
67
+ ### Training Hyperparameters
68
  The following hyperparameters were used during training:
69
+ - **Learning Rate**: 1e-05
70
+ - **Training Batch Size**: 16
71
+ - **Evaluation Batch Size**: 8
72
+ - **Seed**: 42
73
+ - **Optimizer**: Adam (betas=(0.9, 0.999), epsilon=1e-08)
74
+ - **Learning Rate Scheduler**: Linear
75
+ - **Warmup Steps**: 10
76
+ - **Number of Epochs**: 3.0
77
+ - **Mixed Precision Training**: Native AMP
78
+
79
+ ### Framework Versions
80
+ - **Transformers**: 4.41.1
81
+ - **PyTorch**: 2.2.1+cu121
82
+ - **Datasets**: 2.19.1
83
+ - **Tokenizers**: 0.19.1