Melody Training with Segment-Based Tilt Contour for Quranic Tarannum (original) (raw)

Dynamic Time Warping Features Extraction Design for Quranic Syllable-based Harakaat Assessment

International Journal of Advanced Computer Science and Applications

The use of technological speech recognition systems with a variety of approaches and techniques has grown exponentially in varieties of human-machine interaction applications. The assessment for Qur'anic recitation errors based on syllables utterance is used to meet the Tajweed rules which generally consist of Harakaat (prolonging). The digital transformation of Quranic voice signals with identification of Tajweed-based recitation errors of Harakaat is the main research work in this paper. The study focused on speech processing implemented using the representation of Quranic Recitation Speech Signals (QRSS) in the best digital format based on Al-Quran syllables and feature extraction design to reveal similarities or differences in recitation (based on Al-Quran syllables) between experts and student. The method of Dynamic Time Warping (DTW) is used as Short Time Frequency Transform (STFT) of QRSS syllable feature for Harakaat measurement. Findings from this paper include an approach based on human-guidance threshold classification that is used specifically to evaluate Harakaat based on the syllables of the Qur'an. The threshold classification performance obtained for Harakaat is above 80% in the training and testing stages. The results of the analysis at the end of the experiment have concluded that the threshold classification method for Minimum Path Cost (MPC) feature parameters can be used as an important feature to evaluate the rules of Tajwid Harakaat embedded in syllables.

Quranic Reciter Recognition: A Machine Learning Approach

Advances in Science, Technology and Engineering Systems Journal

Recitation and listening of the Holy Quran with Tajweed is an essential activity as a Muslim and is a part of the faith. In this article, we use a machine learning approach for the Quran Reciter recognition. We use the database of Twelve Qari who recites the last Ten Surah of Quran. The twelve Qari thus represents the 12-class problem. Two approaches are used for audio representation, firstly, the audio is analyzed in the frequency domain, and secondly, the audio is treated as images through Spectrogram. The Mel Frequency Cepstral Coefficients (MFCC) and Pitch are used as the features for model learning in the first case. In the second case of audio as images, Auto-correlograms are used to extract features. In both cases, the features are learned with the classical machine learning which includes the Naïve Bayes, J48, and the Random Forest. These classifiers are selected due to their overall good performance in the state-of-the-art. It is observed that classifiers can efficiently learn the separation between classes, when the audio is represented by the MFCC, and the Pitch features. In such a case, we get 88% recognition accuracy with the Naïve Bayes and the Random Forest showing that Qari can be effectively recognized from the recitation of the Quranic verses.

Quranic Verse Recitation Feature Extraction Using Mel-Frequency Cepstral Coefficient (MFCC)

Each person's voice is different. Thus, the Quran sound, which had been recited by most of recitors will probably tend to differ a lot from one person to another. Although those Quranic sentence were particularly taken from the same verse, but the way of the sentence in Al-Quran been recited or delivered may be different. It may produce the difference sounds for the different recitors. Those same combinations of letters may be pronounced differently due to the use of harakates. This paper explores the viability of Mel-Frequency Cepstral Coefficient (MFCC) technique to extract features from Quranic verse recitation. Features extraction is crucial to prepare data for classification process. MFCC is one of the most popular feature extraction techniques used in speech recognition, whereby it is based on the frequency domain of Mel scale for human ear scale. MFCCs consist of preprocessing, framing, windowing, DFT, Mel Filterbank, Logarithm and Inverse DFT.

Strategies for Implementing an Optimal ASR System for Quranic Recitation Recognition

With the help of automatic speech recognition (ASR) techniques, computers become capable of recognizing speech. The Quran is the speech of Allah (The God); it is the Holy book for all Muslims in the world; it is written and recited in Classical Arabic language, the language in which it was revealed by Allah to the Prophet Muhammad. Knowing how to pronounce correctly the Quranic sounds and correct mistakes occurred in reading is one of the most important topics in Quranic ASR applications, which assist self-learning, memorizing and checking the Holy Quran recitations. This paper presents a practical framework for development and implementation of an optimal ASR system for Quranic sounds recognition. The system uses the statistical approach of Hidden Markov Models (HMMs) for modeling the Quranic sounds and the Cambridge HTK tools as a development environment. Since sounds duration is regarded as a distinguishing factor in Quranic recitation and discrimination between certain Quranic sounds relies heavily on their durations, we have proposed and tested various strategies for modeling the Quranic sounds' durations in order to increase the ability in distinguishing them properly and thus enhancing their overall recognition accuracy. Experiments have been carried out on a particular Quranic Corpus containing ten male speakers and more than eight hours of speech collected from recitations of the Holy Quran. The implemented system reached (99%) as average recognition rate; which reflects its robustness and performance.

Computer speech recognition to text for recite Holy Quran

IOP Conference Series: Materials Science and Engineering

Memorizing Holy Quran or Tahfidz is important to worship for Muslim around the world. This research proposed a solution in memorizing and learning Holy Quran easily. To help in remembering the sentence of Holy Quran, Fisher-Yates Shuffle had implemented for randomization of the letter of the Holy Quran. In this research, the sound of Holy Quran had recorded and it was converted into Arabic text to recognize the character of text. Jaro-Winkler was used for text matching algorithm, and Google Speech API help to define speech recognition. The result showed that Fisher-Yates Shuffle Algorithm was successfully applied in randomization with 15 times of experiments. And also, Jaro-Winkler Distance algorithm had performed well as text matching between text from speech recognition and Holy Quran text. The result showed that the percentage of accuracy was around 91% and an average of matching time was 1.9 ms.

Using deep learning for automatically determining correct application of basic quranic recitation rules

2018

Quranic Recitation Rules (Ahkam Al-Tajweed) are the articulation rules that should be applied properly when reciting the Holy Quran. Most of the current automatic Quran recitation systems focus on the basic aspects of recitation, which are concerned with the correct pronunciation of words and neglect the advanced Ahkam Al-Tajweed that are related to the rhythmic and melodious way of recitation such as where to stop and how to “stretch” or “merge” certain letters. The only existing works on the latter parts are limited in terms of the rules they consider or the parts of Quran they cover. This paper comes to fill these gaps. It addresses the problem of identifying the correct usage of Ahkam Al-Tajweed in the entire Quran. Specifically, we focus on eight Ahkam Al-Tajweed faced by early learners of recitation. Popular audio processing techniques for feature extraction (such as LPC, MFCC and WPD) and classification (KNN, SVM, RF, etc.) are tested on an in-house dataset. Moreover, we stud...

On Computational Transcription and Analysis of Oral and Semi-Oral Chant Traditions

2012

Variation is considered a universal principle in music. In terms of semiotics, variation in music is omnipresent and distinguishes music from language (Middleton, 1990). In oral music traditions, variation is introduced to the music due to the absence of a concrete notation. In this paper we investigate melodic stability and variation in cadences as they occur in oral and semi-oral traditions. Creating a new framework for transcription, we have quantized and compared cadences found in Torah trope, strophic melodies from the Dutch folk song collection Onder de groene linde and Qur’an recitation. We have developed computational methods to analyze similarity and variation in melodic formulas in cadences as they occur in recorded examples of the beforementioned oral/semi-oral traditions. Concentrating on cadences, we have investigated melodic, durational and contour similarities in cadences within individual songs/chants, within chant types and between chant types. Using computational m...

Quranic Verse Recitation Recognition Module for Support in j-QAF Learning: A Review

International Journal of …, 2008

Each person's voice is different. Thus, the Quran sound, which had been recited by most of recitors will probably tend to differ a lot from one person to another. Although those Quranic sentence were particularly taken from the same verse, but the way of the sentence in Al-Quran been recited or delivered may be different. It may produce the difference sounds for the different recitors. Those same combinations of letters may be pronounced differently due to the use of harakates. This paper seeks to provide a comprehensive review of Quran Arabic verse recitation recognition focusing on the techniques used, the advantages, and drawbacks. Areas with potential of further expansion are identified for future research for support in j-QAF learning.

Towards Building a Speech Recognition System for Quranic Recitations: A Pilot Study Involving Female Reciters

Jordan Journal of Electrical Engineering, 2022

This paper is the first step in an effort toward building automatic speech recognition (ASR) system for Quranic recitations that caters specifically to female reciters. To function properly, ASR systems require a huge amount of data for training. Surprisingly, the data readily available for Quranic recitations suffer from major limitations. Specifically, the currently available audio recordings of Quran recitations have massive volume, but they are mostly done by male reciters (who have dedicated most of their lives to perfecting their recitation skills) using professional and expensive equipment. Such proficiency in the training data (along with the fact that the reciters come from a specific demographic group; adult males) will most likely lead to some bias in the resulting model and limit their ability to process input from other groups, such as non-/semi-professionals, females or children. This work aims at empirically exploring this shortcoming. To do so, we create a first-of-its-kind (to the best of our knowledge) benchmark dataset called the Quran recitations by females and males (QRFAM) dataset. QRFAM is a relatively big dataset of audio recordings made by male and female reciters from different age groups and proficiency levels. After creating the dataset, we experiment on it by building ASR systems based on one of the most popular open-source ASR models, which is the celebrated DeepSpeech model from Mozilla. The speaker-independent end-to-end models, that we produce, are evaluated using word error rate (WER). Despite DeepSpeech's known flexibility and prowess (which is shown when trained and tested on recitations from the same group), the models trained on the recitations of one group could not recognize most of the recitations done by the other groups in the testing phase. This shows that there is still a long way to go in order to produce an ASR system that can be used by anyone and the first step is to build and expand the resources needed for this such as QRFAM. Hopefully, our work will be the first step in this direction and it will inspire the community to take more interest in this problem.