Voice-Based Sinhala Document Maker Application

Abstract

This research paper introduces "Word Sri," an inventive voice-based web application designed to enhance written Sinhala communication and originality. The application incorporates features such as a plagiarism checker, grammar checker, voice-activated punctuation, and intuitive voice commands in an effort to accommodate a diverse user base, including individuals with disabilities, language learners, writers, journalists, students, and professionals. The first aspect of the paper explores the creation of a user-friendly Sinhala Grammar Checker with Fix Sentences & Punctuation, addressing the dearth of effective language tools for grammar, punctuation, and orthography correction. Using a comprehensive analysis of Sinhala language components, literature, and a corpus of text data, the model provides students, academicians, and professionals with real-time error feedback and correction. The second aspect highlights the distinctive characteristics of the "Word Sri" application, emphasizing its usefulness for voice-based Sinhala content creation. Notably, the application includes a cutting-edge Sinhala plagiarism analyzer that uses machine learning algorithms, synonym and paraphrase detection to improve the originality of content. This research represents a significant advancement in Sinhala language technology, as it provides a unified platform for efficient voice-based typing, accurate grammar checks, and plagiarism detection, thereby facilitating effective written communication and originality in Sinhala for global users.

Country : Sri Lanka

1 D.D.D.Dissanayaka2 J.M.O.K.Jayasundara3 Dr. Dilshan De Silva

  1. Department of Computer Science and Software Engineering, Sri Lanka Institute of Information Technology, Colombo, Sri Lanka
  2. Department of Computer Science and Software Engineering, Sri Lanka Institute of Information Technology, Colombo, Sri Lanka
  3. Department of Computer Science and Software Engineering, Sri Lanka Institute of Information Technology, Colombo, Sri Lanka

IRJIET, Volume 8, Issue 1, January 2024 pp. 38-44

doi.org/10.47001/IRJIET/2024.801005

References

  1. S. Lanka, “7 th INTERNATIONAL CONFERENCE ON HUMANITIES AND SOCIAL SCIENCES Faculty of Arts.”
  2. J. Singh, K. Steele, and L. Singh, “Combining the Best of Online and Face-to-Face Learning: Hybrid and Blended Learning Approach for COVID-19, Post Vaccine, & Post-Pandemic World,” J. Educ. Technol. Syst., vol. 50, no. 2, pp. 140–171, Dec. 2021, doi: 10.1177/00472395211047865.
  3. Azir M.Y.M, Lorensuhewa S.A.S, and Kalyani M.A.L, “Sinhala speech recognition using hidden Markov based model and deep neural network based model for number sequences.”
  4. Anaharris-ms eric-urban, “What is the Speech service? - Azure AI services | Microsoft Learn,” Jul. 18, 2023. https://learn.microsoft.com/en-us/azure/ai-services/speech-service/overview (accessed Aug. 23, 2023).
  5. Y. Wijeratne, N. de Silva, and Y. Shanmugarajah, “Natural Language Processing for Government: Problems and Potential,” Build. Chatbots with Python, no. August, pp. 29–61, 2019.
  6. Y. Kim, M. Reza, J. McGrenere, and D. Yoon, “Designers characterize naturalness in voice user interfaces: Their goals, practices, and challenges,” Conf. Hum. Factors Comput. Syst. - Proc., 2021, doi: 10.1145/3411764.3445579.
  7. A.Mukhamadiyev, I. Khujayarov, O. Djuraev, and J. Cho, “Automatic Speech Recognition Method Based on Deep Learning Approaches for Uzbek Language,” Sensors, vol. 22, no. 10, 2022, doi: 10.3390/s22103683.
  8. A.Valledor, A. Olmedo, C. J. Hellín, A. Tayebi, S. Otón-Tortosa, and J. Gómez, “The Eclectic Approach in English Language Teaching Applications: A Qualitative Synthesis of the Literature,” Sustain., vol. 15, no. 15, 2023, doi: 10.3390/su151511978.
  9. A.Mehrish, N. Majumder, R. Bharadwaj, R. Mihalcea, and S. Poria, “A review of deep learning techniques for speech processing,” Inf. Fusion, vol. 99, 2023, doi: 10.1016/j.inffus.2023.101869.
  10. S. S. Mukrimaa et al., “No 主観的健康感を中心とした在宅高齢者における 健康関連指標に関する共分散構造分析 Title,” J. Penelit. Pendidik. Guru Sekol. Dasar, vol. 6, no. August, p. 128, 2016.
  11. R. Tardi, “THE RIGHT TO INCLUSIVE EDUCATION FOR CHILDREN WITH DISABILITIES: good Practices in the CEECIS Region and Recommendations to the Albanian government,” World Vis., no. March, 2012, [Online]. Available: https://www.wvi.org/sites/default/files/Raport_Rajonal_3_Eng_Web.pdf.
  12. L. Senevirathne, P. Demotte, B. Karunanayake, U. Munasinghe, and S. Ranathunga, “Sentiment Analysis for Sinhala Language using Deep Learning Techniques,” 2020, [Online]. Available: http://arxiv.org/abs/2011.07280.
  13. E. Innovation, Innovating Education and Educating for Innovation. 2016.
  14. “International Conference on Innovations in Info-business and Technology 2022 Digital Transformation & Innovation for a Sustainable Society,” no. February, 2022.
  15. N. de Silva, “Survey on Publicly Available Sinhala Natural Language Processing Tools and Research,” 2019, [Online]. Available: http://arxiv.org/abs/1906.02358.
  16. P. F. Jason, “Personal Desktop Voice Assistant - Research Paper,” vol. 12, no. 3, pp. 126–135, 2023, doi: 10.17148/IJARCCE.2023.12324.
  17. M. Munasinghe, “Sustainable Sri Lanka 2030 Vision and Strategic Path i Sustainable Sri Lanka 2030 Vision and Strategic Path Presidential Expert Committee (PEC) Sustainable Sri Lanka 2030 Vision and Strategic Path,” 2019.
  18. F. Rakotomalala, H. N. Randriatsarafara, A. R. Hajalalaina, and N. M. V. Ravonimanantsoa, “Voice User Interface: Literature review, challenges and future directions,” Syst. THEORY, Control Comput. J., vol. 1, no. 2, pp. 65–89, Dec. 2021, doi: 10.52846/STCCJ.2021.1.2.26.
  19. E. Cooper, “Text-to-Speech Synthesis Using Found Data for Low-Resource Languages,” ICASSP, IEEE Int. Conf. Acoust. Speech Signal Process. - Proc., vol. 2021-June, pp. 6593–6597, 2019.
  20. M. Labied and A. Belangour, “Automatic Speech Recognition Features Extraction Techniques: A Multi-criteria Comparison,” Int. J. Adv. Comput. Sci. Appl., vol. 12, no. 8, pp. 177–182, 2021, doi: 10.14569/IJACSA.2021.0120821.
  21. M. Ravanelli et al., “SpeechBrain: A General-Purpose Speech Toolkit,” Jun. 2021, Accessed: Aug. 23, 2023. [Online]. Available: https://huggingface.co/speechbrain/asr-crdnn-commonvoice-it.
  22. A.Chatterjee, “Automated Information Organization,” Elem. Inf. Organ. Dissem., pp. 497–526, 2017, doi: 10.1016/B978-0-08-102025-8.00025-9.
  23. M. Wankhade, A. C. S. Rao, and C. Kulkarni, “A survey on sentiment analysis methods, applications, and challenges,” Artif. Intell. Rev. 2022 557, vol. 55, no. 7, pp. 5731–5780, Feb. 2022, doi: 10.1007/S10462-022-10144-1.
  24. M. K. Ngueajio and G. Washington, “Hey ASR System! Why Aren’t You More Inclusive?: Automatic Speech Recognition Systems’ Bias and Proposed Bias Mitigation Techniques. A Literature Review,” Lect. Notes Comput. Sci. (including Subser. Lect. Notes Artif. Intell. Lect. Notes Bioinformatics), vol. 13518 LNCS, pp. 421–440, 2022, doi: 10.1007/978-3-031-21707-4_30.
  25. AWS, “Security and compliance - Overview of Amazon Web Services,” 2022. https://docs.aws.amazon.com/whitepapers/latest/aws-overview/security-and-compliance.html (accessed Aug. 23, 2023).
  26. D. Vazhenina and K. Markov, “End-to-end noisy speech recognition using fourier and hilbert spectrum features,” Electron., vol. 9, no. 7, pp. 1–18, Jul. 2020, doi: 10.3390/ELECTRONICS9071157.