Browsing by Author "Haboussi, Samia"
Now showing 1 - 2 of 2
- Results Per Page
- Sort Options
Item ArabAlg: A new Dataset for Arabic Speech Command Recognition for Machine Learning Applications(University of Bahrain, 2024) Oukas, Nourredine; Haboussi, Samia; Maiza, Chafik; Benslimane, NassimAutomatic Speech Recognition (ASR) systems have witnessed significant advancements in recent years, thanks to the emergence of deep learning techniques and the availability of large speech datasets in various languages. With the increasing demand for Arabic voice-enabled technologies, the availability of a high-quality and representative dataset for the Arabic language becomes crucial. This paper presents the development of a new dataset called ArabAlg, specifically designed for Arabic Speech Command Recognition (ASCR), to support the integration of Arabic voice recognition systems into smart devices in the Internet of Things (IoT). This research focuses on collecting and annotating a diverse range of Arabic speech commands, encompassing various domains and applications. The dataset construction process involves recording and preprocessing several utterances from native Arabic speakers. To ensure precision and reliability, quality control measures are implemented during data collection and annotation. The resulting dataset provides a valuable resource for training and evaluating ASCR systems tailored for Arabic speakers using Machine Learning and Deep Learning.Item Arabic speech recognition using deep learning and common voice dataset(IEEE, 2022) Oukas, Nourredine; Zerrouki, Taha; Haboussi, Samia; Djettou, HalimaSpeech recognition is critical in creating a natural voice interface for human-to-human communication with modern digital life equipment. Smart homes, vehicles, autonomous devices in the Internet of Things, and others need to recognize various spoken languages. Meanwhile, the Arabic language has a shortage of speech recognition systems. This study comes to develop an Arabic speech-to-text tool for Arabic language. Our solution uses DeepSpeech model which is a deep learning approach and uses a data set from the Common Voice Mozilla project. The results showed a 24.3 percent Word Error Rate and a 17.6 percent character error rate. So, the proposed model reduces the Word Error Rate by 11.7% compared to Bakheet's Wav2Vec model
