Artificial Intelligence in English Language Assessment: Opportunities and Ethical Concerns – A Balochistan‑Focused Analysis
DOI:
https://doi.org/10.5281/zenodo.19762632Keywords:
artificial intelligence, English language assessment, automated essay scoring, algorithmic bias, accent discrimination, Balochi, Pashto, Brahvi, Saraiki, Punjabi, data privacy, teacher oversight, offline AI, edge computing, internet shutdowns,Balochistan.Abstract
Automated Essay Scoring (AES) softwares and AI-based Speaking assessment software are being utilized in classroom and in testing Centres worldwide during integrating AI in learning and teaching. This paper outlines opportunities and ethical issues of AI integration in English language assessment, emphasizing on algorithmic bias, data privacy and oversighting role of English teacher. Based on recent empirical studies, such as those that show that there are racial and linguistic prejudices in the grading of GPT-4o essays that were so serious. Indeed, still there is a continuous discrimination due to diverse accent in automated speech recognition systems. The review has also revealed that the current AI assessment technologies, although they seem to be efficient, consistent, and scalable are fundamentally incapable of evaluating language impartially, amongst heterogeneous learners. Especially, this article pays special attention on the learners whose native languages are Balochi, Pashto, Saraiki and Punjabi limiting the scope to Pakistan generally and Balochistan particularly. This article also considers and addresses the critical infrastructural issues of Balochistan (sporadic electricity, poor internet connection, internet blackouts caused by security) and suggests a tangible, offline-first, edge-computing implementation paradigm, which can allow complex AI evaluation even in the most isolated and security-impacted districts of Balochistan. Then this article recommends a human-focused, hybrid assessment model where AI is used as a supportive tool and not as a substitute to teacher judgment that is based on transparency standards and periodic bias audits and transparent policies regarding data governance. The conclusion provides context-specific practical recommendations to educators, institutions, and policymakers who wish to use AI assessment tools in a responsible manner in the context of Balochistan during English language teaching.
Downloads
References
1. Ahmad, B., Zhao, Z., Jile, X., Gultaj, H., Khan, N., & Yunxian, Y. (2024). Exploring the influence of internet technology adoption on the technical efficiency of food production: insight from wheat farmers. Frontiers in Sustainable Food Systems, 8. https://doi.org/10.3389/fsufs.2024.1385935
2. Aldino, A. A., Maheshi, B., Li, Y., Zhou, Y., Tsai, Y., Gašević, D., & Chen, G. (2026). Enhancing learner-centered feedback with AI: teachers’ practices and perceptions. Assessment & Evaluation in Higher Education, 1. https://doi.org/10.1080/02602938.2026.2638920
3. Bandodkar, G., Agarwal, S., Sughosh, A. K., Singh, S., & Choi, T. (2024). “Allot?” is “A Lot!” Towards Developing More Generalized Speech Recognition System for Accessible Communication. Proceedings of the AAAI Conference on Artificial Intelligence, 38(21), 23327. https://doi.org/10.1609/aaai.v38i21.30381
4. Barrot, J. S. (2025). Trinka: Facilitating academic writing through an intelligent writing evaluation system. Assessing Writing, 65, 100953. https://doi.org/10.1016/j.asw.2025.100953
5. Batool, K., Zhao, Z., Atif, F., & Dilanchiev, A. (2022). Nexus Between Energy Poverty and Technological Innovations: A Pathway for Addressing Energy Sustainability. Frontiers in Environmental Science, 10. https://doi.org/10.3389/fenvs.2022.888080
6. Chen, Q. (2025). Students’ Perceptions of AI-Powered Feedback in English Writing: Benefits and Challenges in Higher Education. International Journal of Changes in Education. https://doi.org/10.47852/bonviewijce52025580
7. Ding, L., & Zou, D. (2024). Automated writing evaluation systems: A systematic review of Grammarly, Pigai, and Criterion with a perspective on future directions in the age of generative artificial intelligence. Education and Information Technologies, 29(11), 14151. https://doi.org/10.1007/s10639-023-12402-3
8. Fletcher, T., & Hayes-Birchler, A. (2023). Is remote measurement a better assessment of internet censorship than expert analysis? Analyzing tradeoffs for international donors and advocacy organizations of current data and methodologies. Data & Policy, 5. https://doi.org/10.1017/dap.2023.5
9. Gong, K. (2023). Challenges and opportunities for spoken English learning and instruction brought by automated speech scoring in large-scale speaking tests: a mixed-method investigation into the washback of SpeechRater in TOEFL iBT. Asian-Pacific Journal of Second and Foreign Language Education, 8(1). https://doi.org/10.1186/s40862-023-00197-2
10. Gupta, S., Unnam, A., Yadav, K. S., & Aggarwal, V. (2024). Towards Building a Language-Independent Speech Scoring Assessment. Proceedings of the AAAI Conference on Artificial Intelligence, 38(21), 23200. https://doi.org/10.1609/aaai.v38i21.30366
11. Handley, Z., & Wang, H. (2023). What Do the Measures of Utterance Fluency Employed in Automatic Speech Evaluation (ASE) Tell Us About Oral Proficiency? Language Assessment Quarterly, 21(1), 3. https://doi.org/10.1080/15434303.2023.2283839
12. Ifenthaler, D. (2022). Automated Essay Scoring Systems. In Handbook of Open, Distance and Digital Education (p. 1). https://doi.org/10.1007/978-981-19-0351-9_59-1
13. Javed, U., & Nabi, I. (2022). Heterogeneous fragility in Pakistan. In Routledge eBooks (p. 141). Informa. https://doi.org/10.4324/9781003297697-5
14. Khalifa, M., & Albadawy, M. (2024). Using artificial intelligence in academic writing and research: An essential productivity tool. Computer Methods and Programs in Biomedicine Update, 5, 100145. https://doi.org/10.1016/j.cmpbup.2024.100145
15. Kim, J., Chapman, M., Willner, L. S., Kemp, J. A., & Kim, A. A. (2026). Educator perspectives on automated writing scoring and feedback for young language learners: Applying a fairness and justice lens. Assessing Writing, 69, 101050. https://doi.org/10.1016/j.asw.2026.101050
16. Kim, J., Yu, S., Detrick, R., & Li, N. (2024). Exploring students’ perspectives on Generative AI-assisted academic writing. Education and Information Technologies. https://doi.org/10.1007/s10639-024-12878-7
17. Landa-Blanco, M. (2026). Artificial intelligence in education: applications and limitations for teachers in low- and middle-income countries. Frontiers in Education, 10. https://doi.org/10.3389/feduc.2025.1681836
18. Li, Y., Shan, Z., Raković, M., Guan, Q., Gašević, D., & Chen, G. (2025). When AI explains in natural language: Unveiling the impact of generative AI explanations on educators’ grading and feedback practices. Education and Information Technologies, 30(17), 24931. https://doi.org/10.1007/s10639-025-13741-z
19. Pack, A., Barrett, A., & Escalante, J. (2024). Large language models and automated essay scoring of English language learner writing: Insights into validity and reliability. Computers and Education Artificial Intelligence, 6, 100234. https://doi.org/10.1016/j.caeai.2024.100234
20. Payande, I., & Charkameh, H. (2025). From .com to .gov: The internet’s inevitable nationalist turn. Internet Policy Review, 14(3). https://doi.org/10.14763/2025.3.2029
21. Sajja, R., Sermet, Y., Fodale, B., & Demir, İ. (2026). Evaluating AI-powered learning assistants in engineering higher education with implications for student engagement, ethics, and policy. Scientific Reports, 16(1). https://doi.org/10.1038/s41598-026-39237-5
22. Saleem, T., Saleem, A., & Aslam, D. M. (2025). Integrating AI in Pakistani ESL classrooms: Teachers’ practices, perspectives, and impact on student performance. PLoS ONE, 20(9). https://doi.org/10.1371/journal.pone.0333352
23. Sharma, P., Sharma, R. K., Singh, K., Maity, M., & Chakravarty, S. (2023). Dolphin: A Cellular Voice Based Internet Shutdown Resistance System. Proceedings on Privacy Enhancing Technologies, 2023(1), 589. https://doi.org/10.56553/popets-2023-0034
24. Uto, M., & Aramaki, K. (2024). Linking essay-writing tests using many-facet models and neural automated essay scoring. Behavior Research Methods, 56(8), 8450. https://doi.org/10.3758/s13428-024-02485-2
25. Yang, K., Raković, M., Li, Y., Guan, Q., Gašević, D., & Chen, G. (2024). Unveiling the Tapestry of Automated Essay Scoring: A Comprehensive Investigation of Accuracy, Fairness, and Generalizability. Proceedings of the AAAI Conference on Artificial Intelligence, 38(20), 22466. https://doi.org/10.1609/aaai.v38i20.30254
26. Zellou, G., & Holliday, N. (2024). Linguistic analysis of human-computer interaction. Frontiers in Computer Science, 6. https://doi.org/10.3389/fcomp.2024.1384252
27. Zhang, S. (2021). Review of automated writing evaluation systems. Journal of China Computer-Assisted Language Learning, 1(1), 170. https://doi.org/10.1515/jccall-2021-2007
28. Zheng, A. (2024). Dissecting bias of ChatGPT in college major recommendations. Information Technology and Management. https://doi.org/10.1007/s10799-024-00430-5
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2026 Journal of Quranic and Social Studies

This work is licensed under a Creative Commons Attribution 4.0 International License.


















