Detection of errors in the Indonesian standard mushaf based on pixels to support accelerated verification

(1) Tri Wahyu Widyaningsih Mail (Universitas Gunadarma, Indonesia)
(2) * Sarifuddin Madenda Mail (Universitas Gunadarma, Indonesia)
(3) Ravi Ahmad Salim Mail (Universitas Gunadarma, Indonesia)
(4) Nurma Nugraha Mail (Universitas Gunadarma, Indonesia)
*corresponding author

Abstract


One effort to maintain the validity of the Al-Qur'an manuscript is the analysis and verification of the manuscript by experts (Pentashih). Currently, manuscript verification without translation takes 30 working days. Therefore, to support Pentashih in reviewing the manuscript, technology is needed to expedite the Pentashih process and prevent analysis errors caused by Pentashih fatigue. This study conducts a writing analysis of the target manuscript by referring to the template manuscript, implementing image preprocessing stages, applying SSIM for analysis, and employing the pixel-matching method. This method examines the manuscript's writing by comparing two block images at the pixel level. Block images are produced by preprocessing the manuscript images before image-matching analysis is performed. Image preprocessing comprises: cropping the outer frame, cropping the inner frame, segmenting the page into row images, adjusting margins, aligning image sizes, segmenting rows into block images, and aligning positions between block images. Pixel value differences are calculated at the same positions across each column and row of the template and target block images. Block image positions with pixel values ≥ 200 occur in 5 consecutive columns, adjacent rows with a distance = 1, and an SSIM value ≥ 0.9, both images meet the mismatch criteria. These findings indicate that the proposed approach provides an efficient and accurate solution for automating the verification of the Indonesian Standard Mushaf.

Keywords


Tashih, Pixel Matching, Writing Error Analysis, Text Image, Image Matching

   

DOI

https://doi.org/10.26555/ijain.v11i4.1820
      

Article metrics

Abstract views : 173 | PDF views : 45

   

Cite

   

Full Text

Download

References


[1] J. Li, “Research on Multimedia Information Processing System based on Data Mining Technology,” in 2023 3rd International Symposium on Computer Technology and Information Science (ISCTIS), IEEE, Jul. 2023, pp. 877–883. doi: 10.1109/ISCTIS58954.2023.10213025.

[2] K. Maharana, S. Mondal, and B. Nemade, “A review: Data pre-processing and data augmentation techniques,” Glob. Transitions Proc., vol. 3, no. 1, pp. 91–99, 2022, doi: 10.1016/j.gltp.2022.04.020.

[3] A. Tuppad and S. D. Patil, “Data Pre-processing Issues in Medical Data Classification,” in 2023 International Conference on Network, Multimedia and Information Technology (NMITCON), IEEE, Sep. 2023, pp. 1–6. doi: 10.1109/NMITCON58196.2023.10275855.

[4] X. He, Y. Chen, Y. Fan, X. Kong, and Z. Cai, “A Reversible Multimedia Representation Method and Its Applications in Multimedia Processing,” IEEE Access, vol. 11, no. May, pp. 62436–62448, 2023, doi: 10.1109/ACCESS.2023.3288020.

[5] Z. Abidin, “Mengetahui Spesifikasi Tiga Mushaf Al Quran Standar Indonesia,” 2020. Available at: https://tafsiralquran.id/mengetahui-spesifikasi-tiga-Mushaf-al-quran-standar-indonesia/

[6] F. Hoerulloh, D. T. Murdiansyah, and K. Lhaksmana, “Juz Classification of English-Translated Qur’anic Verses Using Support Vector Machine,” in 2025 IEEE International Conference on Artificial Intelligence for Learning and Optimization (ICoAILO), 2025, pp. 372–377. doi: 10.1109/ICoAILO66760.2025.11156072.

[7] LPMQ, “Lajnah Pentashihan Mushaf Al-Qur’an Kementerian Agama RI,” 2023. Available at: https://lajnah.kemenag.go.id/info-lpmq/#:~:text=Lajnah Pentashihan Mushaf Al-Qur’an (LPMQ) adalah,hasil kajian LPMQ%3B dan pengelolaan

[8] F. Lukman, The Official Indonesian Qurʾān Translation: The History and Politics of Al-Qur’an dan Terjemahnya. Cambridge: Open Book Publishers, 2022. doi: 10.11647/OBP.0289.

[9] M. Hanafi et al., “Evaluasi Kepatuhan Penerbit Al-Qur’an terhadap Mekanisme Penerbitan, Pentashihan,dan Peredaran Mushaf Al-Qur’an di Indonesia,” Suhuf, vol. 14, no. 2, pp. 245–263, 2021, doi: 10.22548/shf.v14i2.691.

[10] T. S. Arulananth, P. Chinnasamy, J. C. Babu, A. Kiran, J. Hemalatha, and M. Abbas, “Edge detection using fast pixel based matching and contours mapping algorithms,” PLoS One, vol. 18, no. 8, p. e0289823, 2023, doi: 10.1371/journal.pone.0289823.

[11] Z. Ding, S. Qian, Y. Li, and Z. Li, “An image matching method based on the analysis of grey correlation degree and feature points,” in NAECON 2014 - IEEE National Aerospace and Electronics Conference, IEEE, Jun. 2014, pp. 157–162. doi: 10.1109/NAECON.2014.7045795.

[12] T. Lu, “Improving Image Matching Accuracy with Keypoint Features and Feature Fusion,” in 2021 IEEE 3rd International Conference on Frontiers Technology of Information and Computer (ICFTIC), IEEE, Nov. 2021, pp. 513–520. doi: 10.1109/ICFTIC54370.2021.9647145.

[13] R. Gonzalez and R. E. Woods, Digital Image Processing 3rd Edition, 3rd ed. 2008. [Online]. Available at: https://sde.uoc.ac.in/sites/default/files/sde_videos/Digital Image Processing 3rd ed. - R. Gonzalez, R. Woods-ilovepdf-compressed.pdf

[14] I. Kesen, J. F. Lotz, I. Ziegler, P. Rust, and D. Elliott, “Multilingual Pretraining for Pixel Language Models,” in Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, Stroudsburg, PA, USA: Association for Computational Linguistics, 2025, pp. 29582–29599. doi: 10.18653/v1/2025.emnlp-main.1504.

[15] R. C. Gonzalez, R. E. Woods, and P. Hall, Digital Image Processing.Pearson Education India, 2001. Available at: Google Scholar.

[16] Q. Huang, X. Guo, Y. Wang, H. Sun, and L. Yang, “A survey of feature matching methods,” IET Image Process., vol. 18, no. 6, pp. 1385–1410, May 2024, doi: 10.1049/ipr2.13032.

[17] S. Verykokou and C. Ioannidis, “Image Matching: A Comprehensive Overview of Conventional and Learning-Based Methods,” Encyclopedia, vol. 5, no. 1, p. 4, Jan. 2025, doi: 10.3390/encyclopedia5010004.

[18] J. Mukherjee, “A Deep Neural Network Based Holistic Approach for Optical Character Recognition of Handwritten Documents,” SN Comput. Sci., vol. 5, no. 4, p. 347, 2024, doi: 10.1007/s42979-024-02713-8.

[19] R. Szeliski, Computer Vision, 2nd ed. in Texts in Computer Science. Cham: Springer International Publishing, 2022. doi: 10.1007/978-3-030-34372-9.

[20] S. Ji, C. Zeng, Y. Zhang, and Y. Duan, “An evaluation of conventional and deep learning‐based image‐matching methods on diverse datasets,” Photogramm. Rec., vol. 38, no. 182, pp. 137–159, 2023, doi: 10.1111/phor.12445.

[21] W. Dang et al., “A Feature Matching Method based on the Convolutional Neural Network.,” J. Imaging Sci. Technol., vol. 67, no. 3, 2023, doi: 10.2352/J.ImagingSci.Technol.2023.67.3.030402.

[22] L. C. Kiew, A. J. M. Muzahid, and S. F. Kamarulzaman, “Vehicle Route Tracking System based on Vehicle Registration Number Recognition using Template Matching Algorithm,” in 2021 International Conference on Software Engineering & Computer Systems and 4th International Conference on Computational Science and Information Management (ICSECS-ICOCSIM), 2021, pp. 249–254. doi: 10.1109/ICSECS52883.2021.00052

[23] J. Xu, C. Yang, and J. Chen, “A customized template matching classification system,” J. Korean Phys. Soc., vol. 85, no. 10, pp. 867–882, 2024, doi: 10.1007/s40042-024-01182-9.

[24] Z. Osman, L. Hamandi, R. Zantout, and F. N. Sibai, “Automatic processing of Arabic text,” in 2009 International Conference on Innovations in Information Technology (IIT), IEEE, Dec. 2009, pp. 140–144. doi: 10.1109/IIT.2009.5413793.

[25] A. K. J. Saudagar, H. V. Mohammed, K. Iqbal, and Y. J. Gyani, “Efficient Arabic text extraction and recognition using thinning and dataset comparison technique,” in 2015 International Conference on Communication, Information & Computing Technology (ICCICT), IEEE, Jan. 2015, pp. 1–5. doi: 10.1109/ICCICT.2015.7045725.

[26] T. Faisal and S. AlMaadeed, “Enabling indexing and retrieval of historical Arabic manuscripts through template matching based word spotting,” in 2017 1st International Workshop on Arabic Script Analysis and Recognition (ASAR), IEEE, Apr. 2017, pp. 57–63. doi: 10.1109/ASAR.2017.8067760.

[27] N. Zaghden, R. Mullot, and A. M. Alimi, “Characterization of ancient document images composed by Arabic and Latin scripts,” in 2011 International Conference on Innovations in Information Technology, IEEE, Apr. 2011, pp. 124–127. doi: 10.1109/INNOVATIONS.2011.5893801.

[28] S. Alghyaline, “Arabic Optical Character Recognition: A Review,” Comput. Model. Eng. Sci., vol. 135, no. 3, pp. 1825–1861, 2023, doi: 10.32604/cmes.2022.024555.

[29] F. A. Abdul-Aziz, Q. M. Hussein, and M. S. Al-Taei, “Arabic Printed Word Recognition For Encrypted Document Management,” 2023, Thess, Tikrit University. available at: Google Scholar.

[30] R. Patgiri, S. Nayak, T. Akutota, and B. Paul, “Machine Learning: A Dark Side of Cancer Computing,” Int’l Conf. Bioinforma. Comput. Biol., pp. 92–98, 2019, doi: 10.48550/arXiv.1903.07167.

[31] L. Myllyaho, M. Raatikainen, T. Männistö, J. K. Nurminen, and T. Mikkonen, “On misbehaviour and fault tolerance in machine learning systems,” J. Syst. Softw., vol. 183, no. January, p. 111096, Jan. 2022, doi: 10.1016/j.jss.2021.111096.

[32] Yuhandri, S. Madenda, E. P. Wibowo, and - Karmilasari, “Object Feature Extraction of Songket Image Using Chain Code Algorithm,” Int. J. Adv. Sci. Eng. Inf. Technol., vol. 7, no. 1, p. 235, Feb. 2017, doi: 10.18517/ijaseit.7.1.1479.

[33] C. Wibawa and D. T. Anggraeni, “Comparison Of Image Segmentation Method In Image Character Extraction Preprocessing Using Optical Character Recoginiton,” J. Tek. Inform., vol. 4, no. 3, pp. 583–589, Jun. 2023, doi: 10.52436/1.jutif.2023.4.3.956.

[34] I. Rabaev and M. Litvak, “Recent advances in text line segmentation and baseline detection in historical document images: a systematic review,” Int. J. Doc. Anal. Recognit., May 2025, doi: 10.1007/s10032-025-00526-w.

[35] Y. Deng, X. Hu, D. Teng, B. Li, C. Zhang, and W. Hu, “Dynamic adjustment of hyperparameters for anchor-based detection of objects with large image size differences,” Pattern Recognit. Lett., vol. 167, pp. 196–203, 2023, doi: 10.1016/j.patrec.2023.02.019.

[36] K. Lee, S. Lee, and K. M. Lee, “Auto-regressive transformation for image alignment,” pp. 1–16, 2025, doi: 10.48550/arXiv.2505.04864.

[37] W. Song and I. Deng, “A Hybrid Architecture Combining CNN, LSTM, and Attention Mechanisms for Automatic Speech Recognition,” in 2025 11th International Conference on Computing and Artificial Intelligence (ICCAI), 2025, pp. 285–292. doi: 10.1109/ICCAI66501.2025.00052.

[38] M. R. Raigonda and Shweta, “Signature Verification System Using SSIM In Image Processing,” J. Sci. Res. Technol., vol. 2, no. 1, pp. 5–11, Jan. 2024, doi: 10.61808/jsrt79.

[39] Z. Tang et al., “A Fast Image Stitching Algorithm Based on Texture Classification and Improved SIFT,” IEEE Access, vol. 12, no. August, pp. 124183–124208, 2024, doi: 10.1109/ACCESS.2024.3443111.

[40] V. Mudeng, M. Kim, and S. Choe, “Prospects of Structural Similarity Index for Medical Image Analysis,” Appl. Sci., vol. 12, no. 8, p. 3754, Apr. 2022, doi: 10.3390/app12083754.

[41] L. K. Singh, M. Khanna, H. Garg, and R. Singh, “Emperor penguin optimization algorithm-and bacterial foraging optimization algorithm-based novel feature selection approach for glaucoma classification from fundus images,” Soft Comput., vol. 28, no. 3, pp. 2431–2467, 2024, doi: 10.1007/s00500-023-08449-6.




Creative Commons License
This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.

___________________________________________________________
International Journal of Advances in Intelligent Informatics
ISSN 2442-6571  (print) | 2548-3161 (online)
Organized by UAD and ASCEE Computer Society
Published by Universitas Ahmad Dahlan
W: http://ijain.org
E: info@ijain.org (paper handling issues)
 andri.pranolo.id@ieee.org (publication issues)

View IJAIN Stats

This work is licensed under a Creative Commons Attribution-ShareAlike 4.0