On the Relationship between Digital Literacy and Dynamic Assessment of Educational strategies
محورهای موضوعی : Journal of Teaching English Language StudiesMohammad Raza Raeisi 1 , Davood Mashhadi Heidar 2 , Mohammad Reza khodareza 3
1 - Department of English, Tonekabon Branch, Islamic Azad University, Tonekabon, Iran
2 - Department of English, Tonekabon Branch, Islamic Azad University, Tonekabon Branch, Iran
3 - Assistant Professor of TEFL, Department of English Language, Islamic Azad University, Tonekabon Branch, Tonekabon, Iran.
کلید واژه: Digital literacy, Dynamic assessment, Reading comprehension skill,
چکیده مقاله :
Given the emergence of innovative devices necessitates language practitioners to delve into their academic settings This study investigated the role of dynamic assessment and digital literacy in education. To that end, three intact classes consisting of ninety-six Iranian EFL learners participated in the study. The learners received instructions on utilizing appropriate digital tools based on educational strategies within the classroom. The study established a connection between digital literacy employed during dynamic assessment episodes, as described by Brown (2006), and the interactive model of reading comprehension proposed by Grabe (2008). To verify a dynamic assessment episodes model through the research project, the researchers chose a teacher who served as a facilitator for a thorough assessment of each component. After the treatment period, all participants were assessed based on their performance during three months. The results of the study showed that the group exhibited improved performance after the treatment, as indicated by significant differences between their pretest and posttest scores. All the same, an effective digital strategy is essential for creating an optimal education system in the context of the ongoing pandemic. Finally, the findings highlight the importance of addressing digital literacy in the language education and administrating dynamic assessment in education.
Given the emergence of innovative devices necessitates language practitioners to delve into their academic settings This study investigated the role of dynamic assessment and digital literacy in education. To that end, three intact classes consisting of ninety-six Iranian EFL learners participated in the study. The learners received instructions on utilizing appropriate digital tools based on educational strategies within the classroom. The study established a connection between digital literacy employed during dynamic assessment episodes, as described by Brown (2006), and the interactive model of reading comprehension proposed by Grabe (2008). To verify a dynamic assessment episodes model through the research project, the researchers chose a teacher who served as a facilitator for a thorough assessment of each component. After the treatment period, all participants were assessed based on their performance during three months. The results of the study showed that the group exhibited improved performance after the treatment, as indicated by significant differences between their pretest and posttest scores. All the same, an effective digital strategy is essential for creating an optimal education system in the context of the ongoing pandemic. Finally, the findings highlight the importance of addressing digital literacy in the language education and administrating dynamic assessment in education.
Anderson, J. R. (1984). The development of self-recognition: A review. Dev. Psychobiol, 17, 35–49.
Amaro, A. C., Oliveira, L., & Veloso, A. I. (2017). Intergenerational and collaborative use of tablets: «in-medium» and «in-room» communication, learning and interaction. Observatories (OBS), 11(1), 83–94. https://doi.org/10.15847/obsOBS1102017995
Andrade, H., Cizek, G., & Bennett, R. (2019). The handbook of formative assessment in the discipline. Routledge.
Astiandani, F.R & Anam, S. (2021). EFL teacher’ perceptions toward the implementation of online formative assessment amidst the covid-19 pandemic. Journal of English Language Teaching. 8(2), 269-277. https://doi.org/10.26858/eltww.v8i2.21326
Bhatt, I. (2017). Assignments as controversies: Digital literacy and writing in classroom practice. Routledge.
Blikstad-Balas, M. (2012). DL in upper secondary School-What do students do their laptops for during teacher instruction? Nordic Journal of DL, 7(2), 81-96.
Brown, G. T. (2006). Teachers' conceptions of assessment: Validation of an abridged version. Psychological Reports, 99(1), 166-170. https://doi.org/10.2466/pr0.99.1.166-170
Buckingham, D. (2007). Digital media literacies: Rethinking media education in the age of the Internet. Research in Comparative and International Education, 2, 42-55.
Chastain, K. (1988 Ed). Developing second-language skills: Theory and practice. Harcourt Brace Jovanovich, Inc. Chicago: Rand McNally College Pub. Co., c1976.
Cohen, V. L., & Cowen, J., E. (2008). Literacy for children in an information age: teaching reading, writing and thinking. Thomson Higher Education.
Dardanou, M., & Kofoed, T. (2019). It is not only about the tools! Professional digital competence. In G. Colette & I. Palaeologus (Eds.), early learning in the digital age (pp. 61–76). London: SAGE Publications.
Dinther, M. V., Dochy, F., & Segers, M. (2015). The contribution of assessment experiences to student teacher' self-efficacy in competence-based education. Teaching and Teacher Education, 49, 45-55. http://dx.doi.org/10.1016/j.tate.2015.02.013
Esfandiari, R. (2020). Iranian EFL teachers’ DL in academic settings: Teacher professionalism in the digital Age. Journal of Foreign Language Research, 10.22059/JFLR.2019.266987.556
European Commission. (2006). Recommendation on key competences for lifelong learning. Council of 18 December 2006 on Key Competences for Lifelong Learning, 2006/962/EC, L. 394/15. Retrieved December 7, 2017, from http://eur-lex.europa.eu/legal-content/en/TXT/?uri=CELEX
Falloon, G. (2020). From digital literacy to digital competence: the teacher digital competency (TDC) framework. Educational Technology Research and Development, 68(5), 2449-2472. https://doi.org/10.1007/s11423-020-09767-4
Fulcher, G. (2012). Digital Lfor the language reading comprehension. Language Assessment Quarterly, 9(2), 113–132.
Garrett, M. F. (1990) Sentence processing. In Osherson, D.N. and Lasnik, H. (Eds.). An Invitation to Cognitive Science (Volume 1): Language. Cambridge MA: MIT Press
Gillen, J., & Barton, D. (2010). Digital literacies: Research briefing for the TLRP-TEL (Teaching and Learning Research Programme-Technology Enhanced Learning). London: London Knowledge Lab, Institute of Education. Retrieved June 3, 2010, from the TLRP TEL website: http://www.tlrp.org/docs/DigitalLiteracies.pdf
Grabe, W. (2008). Reading in a second language: Moving from theory to practice. New York: Cambridge University Press. https://doi.org/10.1017/CBO9781139150484
Gupta, R., Seetharaman, A., & Maddulety, K. (2020). Critical success factors influencing the adoption of digitalization for teaching and learning by business schools. Education and Information Technologies, 25, 1-22. https://doi.org/10.1007/s10639-020-10246-9.
Hashemian, M., & Fadaei, B. (2013). Fostering EFL learners’ Autonomy in Light of Portfolio Assessment: Exploring the Potential Impact of Gender. Iranian Journal of Language Teaching Research, 1(2), 135-151. https://files.eric.ed.gov/fulltext/EJ1127429.pdf
Inbar-Lourie, O., & Levi, T. (2020). Assessment literacy as praxis mediating teacher knowledge of assessment-for-learning practices. In Toward a Reconceptualization of Second Language Classroom Assessment (pp.241-259). Springer, Cham.
Katz, I. R. (2007). Testing information literacy in digital environments: ETS’s iSkills assessment. Information Technology and Libraries, 26(3), 3-12. https://doi.org/10.6017/ital.v26i3.3271
Kao, Y. T. (2020). A comparison study of DA and nonDA on EFL Chinese learners' speaking performance: Transfer of learning English Teaching and Learning, 1-21. doi:10.1007/s42321-019-00042-1
Khlaisang, J. & Koraneekij, P. (2019). Open online assessment management system platform and instrument to enhance the information, media, and ICT literacy skills of 21st century learners. International Journal of Emerging Technologies in Learning. 14(7), 111-127. https://doi:org/10.3991/ijet.v14i07.9953.
Jakeman, V., & McDowell, C. (2008). New insight into IELTS. Cambridge University Press.
Lee, L., Gundersen, E. (2011). Select readings. Teacher-approved readings for today’s students. Oxford University Press.
LIopis Nebot, M. A; Esteve-Mon, F. V & Segura, J. A. (2021). Diagnostic and educational self-assessment of the digital competence of university EFL learners. Nordic Journal of DL. 16(3), 115-131. doi. 10.18261/issn.1891-943x-2021-03-04-03
Lukitasari, M., Murtafiah, W., Ramdiah, S., Hasan, R., & Sukri, A. (2022). Constructing DL instrument and its effect on college students' learning outcomes. International Journal of Instruction, 15(2), 171-188. https://doi.org/10.29333/iji.2022.15210a
McDougall, J., Readman, M. & Wilkinson, P. (2018). The uses of (digital) literacy, Learning, Media and Technology, 43:3, 263-279, doi: 10.1080/17439884.2018.1462206
Makhachashvili, R., & Semenist, I. (2021). Student satisfaction with digital hybrid learning In European and oriental languages programs: Survey study of regional universities of Ukraine. In proceedings of the 19th international conference E-society, pp. 133– 143.
Martin, A., & Grudziecki, J. (2006). DigEuLit: Concepts and tools for DL development. Innovation in Teaching and Learning in Information and Computer Sciences, 5(4), 249–267. https://doi.org/10.11120/ital.2006.05040249
Martin, A. (2005). DigEduLit – a European Framework for DL: A Progress Report, Journal of e-Literacy, 2, 130-136.
Mehri Kamrood, A., Davouidi, M., Ghaniabadi, S., & Amirian, S. M. R. (2019). Diagnosing EFL learners’ development through online computerized DA. Computer Assisted Language Learning, 34(1), 1-30. https://doi. 10.1080.10958822101645181.
Mellati, M., & Khademi, M. (2018). Exploring EFL learners’ assessment literacy: Impact on EFL learners’ writing achievements and implications for teacher development. Australian Journal of Teacher Education, 43 (6)1-18. https://doi.org/10.14221/ajte.2018v43n6.1
Mirra, N., Morrel, E., & Filipiak, D. (2018). From digital consumption to digital invention: Toward a new critical, theory and practice of multiliteracies. Theory into Practice, 57, 12-19. http://doi.org/10.1080/00405841.2017.1390336
Murray, M. C., & Pérez, J. (2014). Unraveling the DL paradox: How higher education fails at the fourth literacy. Issues in Informing Science and Information Technology, 11, 85-100. http://iisit.org/Vol11/IISITv11p085
Norton, B. (2013). Identity, literacy, and English language teaching. Iranian Journal of Language Teaching Research (IJLTR), 1(2), 85-98.
Panadero, E., Brown, G. T. L., & Strijbos, J. W. (2016). The future of student self-assessment: A review of known unknowns and potential directions. Educational Psychology Review, 28(4), 803-830. doi:10.1007/s10648-015-9350-2
Pearson, P. D. (1976). A psycholinguistic model of reading. Language Arts, 53(3), 309-314.
Petersen, D. B., Tonn, P., Spencer, T. D., & Foster, M. E. (2020). The Classification Accuracy of a DA of Inferential Word Learning for Bilingual English/Spanish-Speaking School-Age Children. Language, Speech, and Hearing Services in Schools, 51(1), 144-164. https://doi:10.1044/2019_lshss-18-0129
Pill, J., & Harding, L. (2013). Defining the language DL gap: Evidence from a parliamentary inquiry. Language Testing, 30(3), 381-402. http://dx.doi.org/10.1177/0265532213480337
Scarino, A. (2013). Language DL as self -awareness: Understanding the role of interpretation in assessment and in teacher learning. Australia Language Testing 30(3), 309-327. doi: 10.1177/0265532213480128
Scriven, M. (1967). The methodology of evaluation. In W. R., Tylor R. M. Gagne, & M. Scriven (Eds.). Perspectives of curriculum evaluation (pp.39-83). Chicago: Rand McNally.
Shabani, K. (2018). Group DA of L2 Learners' writing abilities. International Journal of Language Literature Research 6(1) 129-149.
Spante, M.., Hashemi, S. S., Lundin, M., & Algers, A. (2018). Digital competence and DL in higher education research: Systematic review of concept use. Cogent Education, 5(1), 1519143. https://doi.org/10.1080/2331186X.2018.1519143
Stiggins, R., & Chappius, J. C. (2005). Using student-involved reading comprehension assessment to close achievement gaps. Theory into practice, 44(1), 11-18.
Storto, A. (2021). Fingerprints: Towards a multisensory approach to meaning in digital media. Nordic Journal of DL. 16(3), 132-143. http://doi.org/ 10.18201
Yousofi, N., Velayati, S., & Ebadi, S. (2024). Implementing group DA to enhance Iranian high school students’ grammar ability. Journal of English Language Teaching and Learning, (), -. doi: 10.22034/elt.2024.59652.2592
Research Paper | Volume 9, Issue 4 Autumn, 2024 |
|
Accepted: September 2024 Published: December 2024 |
Research Article |
On the Relationship between Digital Literacy and Dynamic Assessment of Educational strategies
Mohammad Reza Raeisi1 Davood Mashhadi Heidar*2 Mohammad Reza Khodareza3
1Ph.D. Candidate, Department of English, Tonekabon Branch, Islamic Azad University, Tonekabon, Iran 2Assistant Professor, Department of English, Tonekabon Branch, Islamic Azad University, Tonekabon, Iran Email: davoodm_tarbiatmodares@yahoo.com 3Assistant Professor, Department of English, Tonekabon Branch, Islamic Azad University, Tonekabon, Iran m.khodareza@gmail.com
|
ABSTRACT Given the emergence of innovative devices necessitates language practitioners to delve into their academic settings This study investigated the role of dynamic assessment and digital literacy in education. To that end, three intact classes consisting of ninety-six Iranian EFL learners participated in the study. The learners received instructions on utilizing appropriate digital tools based on educational strategies within the classroom. The study established a connection between digital literacy employed during dynamic assessment episodes, as described by Brown (2006), and the interactive model of reading comprehension proposed by Grabe (2008). To verify a dynamic assessment episodes model through the research project, the researchers chose a teacher who served as a facilitator for a thorough assessment of each component. After the treatment period, all participants were assessed based on their performance during three months. The results of the study showed that the group exhibited improved performance after the treatment, as indicated by significant differences between their pretest and posttest scores. All the same, an effective digital strategy is essential for creating an optimal education system in the context of the ongoing pandemic. Finally, the findings highlight the importance of addressing digital literacy in the language education and administrating dynamic assessment in education.
Key Words: Digital literacy, dynamic assessment, reading comprehension skill
|
1. INTRODUCTION
Digital Literacy (DL) and Dynamic Assessment (DA) are crucial within the field of social studies (Etemadi, & Abbasian, 2023). Taking this, numerous definitions have been proposed to capture the essence of the concepts, multisensory approaches involved in the digital media. The multisensory approaches were developed into a “TikTok entitled as a video-sharing and creation platform in Beijing” (Storto, 2021, p. 138). Among these definitions, Makhachashvili and Semenist (2021) have identified required attention to teacher perception and beliefs in EFL assessment innovation. Moreover, Inbar-Lourie and Levi (2020) highlighted subjects such as information literacy, media literacy, and technological proficiency in the context of language learning. These intersecting activities have brought attention to the intricate, intersecting, conflicting, and not fully understood connection between DL innovation as applied in design and innovation within the broader scope of innovation studies.
In a study conducted by Yousofi, Velayati, and Ebadi in 2024, they explored the use of group DA as a means to improve the grammar skills of high school students in Iran. The research was published in the Journal of English Language Teaching and Learning. The study aimed to investigate the impact of group DA, a growing trend in education in general, and language education in particular (Khodabakhsh, Abbasian, & Rashtchi, 2018) regarding developing EFL learners‟ level of language awareness (LA) on students' writing ability. While significant research has been conducted in the domain of DL as evident from studies conducted by Astiandani and Anam (2021), Hashemian and Fadaei (2013), and Norton (2013), none of these studies have explored potential methods to enhance the DL capabilities of English as EFL learners. The available literature on DL in the context of testing and assessment is relatively limited, with only a few studies conducted by Inbar-Lourie and Levi (2020), and Petersen et al. (2020). Moreover, studies have been developed to highlight the contrasting approaches in analyzing contemporary DL paradigms between instructional practices and the requirements of academic teaching.
In the same manner, Yu (2022) Sustained roles of students, digital literacy, learning achievements, and motivation in online learning environments during the COVID-19 pandemic. Indeed, sustaining the notion of contemporary literacies pertains to everyday learning and practices in EFL is primarily limited to the realm of reading comprehension interaction and education. Several researchers have conducted assessments to demonstrate the DL of EFL learners (Yu, 2022). Lukitasari et al. (2022) have examined the efficacy of digital tools on EFL learners' educational achievements, specifically focusing on their reading comprehension skills and knowledge. Previous investigations have also indicated that these skills have been adopted from other researchers without undergoing thorough validation testing in terms of their conceptual and visual effectiveness. Additionally, DL involves the ability to access and navigate the digital world using various skills and technologies, such as Google Chrome, WhatsApp, Telegram, Webinar, Adobe Connect, and Big Blue Button (Kao, 2020; LIopis Nebot, Esteve-Mon & Segura, 2021; Muawanah, Marini, & Sarifah, 2024).
2. REVIEW OF LITERATURE
In light of the advancements in technology, there has been a shift towards exploring the subject of DigEduLit model (Bekrizadeh, Panahi & Jamalvandi, 2024; Martin & Grudziecki, 2006). This model is often seen as a designated physical space where students engage in standardized educational activities under supervision, following a routine (Katz, 2007). This research contributes to existing literature by demonstrating how new digital networks have transformed the dynamics within a conventional classroom, isolating more traditional teaching methods. Within this context, three types of information essential for real-world knowledge creation have been identified: semantic-association information, syntactic information, and symbol-sound information (Pearson, 1976). However, different scholars have explored DL using different resources to encode and access texts, facilitating meaning negotiation in socially recognizable ways (Cizek, Andrade & Bennett, 2019; Martin, 2005) and likewise other researchers studied reading proficiency and cognitive reading strategies through online dynamic assessment (ODA) in English (Imron & Anwar, 2024). Importantly, Mirra, Morrel, and Filipiak (2018) have supported this strategy, wherein students utilize a range of digital media to practice and extract useful tools to improve their performance in assessments. These tools encompass "digital competencies," which primarily involve cooperation, social interaction, and the subsequent aspects of collaboration. (Makhachashvili & Semenist, 2021; Mellati & Khademi, 2018) as well as other researchers studied the impact of mobile-assisted hybrid dynamic assessment on arabic language leaners’ reading comprehension performance (Mustiah, Dayat & Sadek, 2024).
2.1. Salience of Digital Literacy
A similar discovery has also been documented, highlighting the importance of DL and its role in discussions about evaluating and addressing it (Murray & Perez, 2014, p. 86). This finding is sometimes referred to as the knowledge of the expanding value of the DL paradox, which encompasses the combination of computer information media or digital compatibility with literacy. However, students faced a challenge with DL as they were deeply engrossed in traditional paper-based assessments and failed to recognize the connection between DL and psycholinguistic aspects in education, especially among EFL learners. In other words, this issue manifests itself through real problems and has significant consequences for students.
2.2. Digital Literacy Model
In addition to a theoretical perspective and framework, various definitions have been proposed to assess the performance of students in education. For instance, Spante et al. (2018) defined DL as an individual's efforts to adapt their lifestyle, learning, and work to a digital environment, and to evaluate their digital skills through DA. Building upon this model, Amaro et al. (2017) and Shabani (2018) further expanded on the theoretical framework of DA, which holds significant relevance in this particular study.
Given the significance of DL and reading comprehension abilities in the EFL setting, this research endeavors to explore the views of EFL teachers regarding their students' reading skills. Prior studies have not investigated the role of learners' DL in the Iranian context within language institutes. As a result, this study aligns with previous investigations on these factors, emphasizing the influence of teacher assessment, specifically focusing on DL. Taking inspiration from the study conducted by Panadero et al. (2016) and Falloon (2020), which discovered that online assessment using Web 2.0 tools significantly enhanced the skills of EFL learners in Iran, this research seeks to validate these findings and expand them to include DL. Furthermore, building on the research carried out by van Dinther et al. (2015), which demonstrated that electronic-based assessment methods can greatly improve EFL learners' reading abilities, this study further explores the potential advantages of utilizing such an approach. According to Cizek, Andrade and Bennett (2019), DL involves the utilization of digital technologies to encode and access texts, enabling the creation, communication, and negotiation of meanings in socially recognizable manners in assessment literacy (Tavassoli & Sorat, 2023)
Additionally, Bekrizadeh, Panahi and Jamalvandi (2024) suggest that DL encompass a broad perspective that involves the capacity of EFL learners to employ digital tools to navigate, control, manage, and integrate digital resources. This enables the production of innovative literacies through collaboration with others, reflecting the specific contexts of their lives. On a similar note, numerous researchers (Fulcher, 2012; Pill & Harding, 2013; Scarino, 2013; Stiggins & Chappius, 2005) argue for the crucial role between the two categories, i.e., DL and Da in education. Conversely, some researchers believe that learners perceive English as a challenging language to use in their daily lives (Gupta, Seetharaman & Maddulety, 2020; Scriven, 1967). These researchers aim to trace the origins, purposes, and contexts of the definitions that have shaped and been utilized in this area of study.
When discussing the concept of DL, Gillan and Barton (2010) present it as an intriguing and captivating idea, encompassing both its definition and its practical applications. They emphasize the remarkable opportunities that digital tools offer for global collaboration among users of technology. Mirra et al. (2018) introduce a significant article that focuses on the impact of "media literacy word moves to production element," which aims to encourage learners to actively engage with media and utilize media tools for further learning. These tools encompass various digital skills, including collaboration, social interaction, and other aspects of cooperation.
Furthermore, this study also places great importance on assessing DL, which is considered the guiding principle of the research. This assessment serves as a means of effectively communicating the significance of the conducted research to different readers, including academic assessors, teachers, and individuals interested in the development of literacy within the school context. Thus, two research questions guided this study:
RQ1: Is there any significant relationship between DL and DA at different educational levels?
RQ1: Can DL help to the improvement of Iranian EFL learners’ reading comprehension skill through DA across three different time periods?
3. METHODOLOGY
3.1. Participants
These participants were chosen from a pool of 174 students studying at the Not-for-Profit University of Mohhades Nouri (NPUMN) in Mazandaran, Iran. After piloting, the sample of this study incorporated 96 Iranian male and female learners. They were within the age range of 18-28. Students had already been assigned to three separate classes according to the education schedule of the university, each consisting 32 participants. These classes were selected as two experimental groups (i.e., DA of DL) and one control group (i.e., without treatment group). It was crucial to mention that all of the selected students were at the pre-intermediate level of English proficiency, as determined by their scores on the Oxford English Reading Passage assessment (Pre-intermediate, Lee & Gundersen's Select Readings, 2011). All of the participants were Iranian EFL learners with a Persian background, having a general 3-credit course to pass as a curriculum syllabus.
All of them were nonnative speakers of English, but English was their foreign language. Thus, a decision was made to only homogenize them in terms of reading comprehension literacy. Accordingly, ninety-six students, who attained scores well within two standard deviations below or above the mean scores on the first 10 English reading size test, were selected as the participants.
3.2. Design
The ongoing investigation utilized a combination of mixed-method research design. The main objective of the study was to understand how the variables being studied are related to each other. To accomplish this, participants were randomly placed into groups to receive treatment. A key feature that sets crossover designs apart from other types of experiments is that the same group of participants were used throughout the study. Each student's response to the treatment they received was assessed and recorded. Essentially, the goal of this mixed-method research design was to determine the nature and strength of the connection between the two variables.
3.3. Instruments and Materials
The research made use of materials from the Reading Series (Lee & Gundersen's Select Readings, 2011) published by TOEIC®-style Final, Oxford University Press. These materials included chosen readings suitable for intermediate level learners and were employed to evaluate the language proficiency of the participants. Furthermore, the Select Reading (Extra Assessment) test was utilized to measure the learners' accomplishments and ensure their uniformity. The test consisted of 45 multiple-choice questions, with 12 passage reading questions, 6 lexicon questions, and 3 pronoun questions distributed among 3 passages. Participants had 45 minutes to answer these questions. The Eurostat's Digital Skills Indicator was also identified as the most suitable tool to adapt for assessing DL, as it met the minimum requirements. This study utilized a standardized measurement, consisting of 181 items, which was developed and validated by Khlaisang and Koraneekij (2019). The measurement focused on three crucial digital skills for the 21st century: information literacy (49 items), media literacy (63 items), and information and communication technology literacy (69 items). Confirmatory factor analysis was employed to validate the questionnaire, and its reliability was demonstrated through Cronbach's alpha coefficient.
3.3.1. First strategy: Designing the DigEduLit Model
DigEduLit Model strategies were developed for a specific group of students based on their “knowledge and skills in psycholinguistic models” (Anderson, 1984, p. 186 as cited in Chastain, 1988). These strategies were then compared to a skills model. To understand the role of DA on DL practices, researchers needed to compare the perspectives and practices of students proficient and non-proficient in assessment. Thus, at the start of the study, a group of 96 EFL learners (selected from a larger sample) studying Pre-Select Reading books were chosen. The researchers then administered a DL Eurostat Skills to these students over 10 sessions. The major difference between less proficient and highly proficient students was the approach to completing reading tasks rather than just focusing on planning and preparing for the main iChecker self-assessment tool in terms of DL (Cohen & Cowen, 2008).
3.3.2. Second strategy: Selecting the sample of mediators
In order to validate the DigEduLit model created during the research project, it was crucial to select a teacher to act as a guide for a comprehensive evaluation of each element. The teacher's responsibility was to assess the importance of the tasks and the DL characteristics involved in them. The purpose of this selection process was to eliminate any errors or irrelevant questions and to adjust or correct paragraphs as necessary Bhatt (2017). To integrate the Zone of Proximal Development (ZPD) concept into digital reading literacy and enable feedback and interventions, the researchers developed a framework that serves as the foundation for evaluating the evaluation process. Additionally, a standardized reading comprehension test (at the pre-intermediate level) was used as a benchmark for the teacher's involvement in two DL management scenarios (Heidari & Tabatabaee-Yazdi, 2021). This test provided guidelines for various sections of the reading passage and included multiple examples of this particular reading style (Buckingham, 2007).
3.4. Data Collection Procedure
In order to obtain a more accurate assessment of students' performance on the two reading tasks over three different time periods, the researcher calculated their reading scores separately and conducted t-tests to compare them. aiming to enhance their effectiveness in digital contexts. At the beginning (Month 1 [T1]), Month 2 (T2] and the end (Month 3 [T3]) of the semester term of the academic year, students were tested on their English decoding, vocabulary, and word awareness and reading comprehension with paper-and-pencil tasks and on EDU with a self-assessment reading task with reading times being compared on passages with unknown versus known words and passages with the pre-reading text (Guikema & Menke, 2014),
For this study, the researcher initially employed the Oxford Placement Test (Allan, 2004) to select a total of 175 pre-intermediate-level Iranian EFL learners from a private language institute in Nur, Iran. The participants consisted of 115 female learners and 60 male learners, divided into four intact reading comprehension classes. These classes were categorized as follows: a) male experimental group (25 learners), b) female experimental group (71 learners), c) male control group (35 learners), and d) female control group (44 learners). The researcher randomly assigned the aforementioned reading comprehension classes to the male and female experimental and control groups.
Secondly, the researcher utilized the Big Blue Button learning management system and Google Forms to administer the reading comprehension pretest to the experimental groups. In contrast, the control groups in the in-person classes were given a paper-and-pencil version of the test. The participants were given 50 minutes to answer the questions on this test.
Thirdly, the experimental groups received the integrated DL and DA treatment during twelve online sessions over a span of six weeks. These sessions lasted for 90 minutes and occurred twice a week. Specifically, the researcher divided the twelve sessions into three categories: four sessions based on the Learning-Management-System, four sessions based on WhatsApp, and four sessions based on Telegram. Throughout all twelve sessions, the learners attended Big Blue Button sessions.
In the first session of the Learning-Management-System-based sessions, the researcher spent 20 minutes using the microphone, camera, and screen-sharing features of the system to inform the learners about various aspects of Big Blue Button learning management system and Google Forms. Following this, the researcher implemented a test-teach-retest pattern as part of the treatment. Firstly, the learners were given a 150-word reading comprehension text through the file-sharing feature, and they were asked to read it and answer comprehension questions using Google Forms within 20 minutes. Secondly, the researcher provided scaffolding techniques, based on the evaluation of the EFL learners' ZPDs, using the camera feature and whiteboard feature to offer negotiated assistance for 30 minutes. Finally, the learners were given the same reading comprehension text again and asked to read it and answer a different set of comprehension questions within 20 minutes.
4. RESULTS
4.1. Results for the Reading Ability Test
To address the research questions, we utilized Mauchly's Test of Sphericity. We included DL of learners, based on educational strategies, as predictors in a model to determine L2 learners' reading comprehension scores. These scores were the main focus. Following the recommendation of Jakeman and McDowell (2008), we initially conducted Mauchly's Test of Sphericity to investigate if the relationship between DL changed significantly at different educational levels in Iranian EFL learners' reading comprehension scores. Pre-reading tests were administered at the beginning and end of the study to evaluate the learners' reading ability. Posttest comprehension scores showed variations between the two groups (ME, M = 14.56, SD = 2.56; MC, M = 15.68, SD = 1.60). Furthermore, an analysis of covariance (ANCOVA) was carried out to assess the impact of two interventions on enhancing the comprehension skills of male participants. The pretest scores of male participants' comprehension skill was used as a covariate in this analysis. The results of the analysis can be found in Table 10.
Table 1.
Assessment of ANCOVA for Male Group Comparison
Source | Type III Sum of Squares | Df | Mean Square | F | Sig. | Partial Eta Squared |
Corrected Model | 97.456b | 2 | 48.728 | 16.709 | .000 | .370 |
Intercept | 91.653 | 1 | 91.653 | 31.428 | .000 | .355 |
pretest | 78.976 | 1 | 78.976 | 27.081 | .000 | .322 |
group | 1.974 | 1 | 1.974 | .677 | .414 | .012 |
Error | 166.227 | 57 | 2.916 |
|
|
|
Total | 14156.500 | 60 |
|
|
|
|
Corrected Total | 263.683 | 59 |
|
|
|
|
a. gender = male | ||||||
b. R Squared = .370 (Adjusted R Squared = .347) |
The results of the performance outcomes for all participants in the experimental group are depicted in Table 1 over three distinct time periods. This analysis will thoroughly explore the statistical measures related to the experimental group, offering valuable insights and aiding in the comprehension of the experimental data. Through the use of clear and straightforward language, our goal is to improve understanding and provide a fresh perspective on the topic at hand.
4.2. Results for Pretest Scores for Experimental Groups
The results of the Mauchly's Test of Sphericity for DL are shown in Table 2. The findings indicate that the Mauchly's W values for the tests were .981 and 1.817 for Chi-square. Additionally, the table also presents the hypotheses and characteristics of Mauchly Sphericity. The p-values obtained from the Mauchly's Test of Sphericity for the scores in Table 2 were higher than the critical value of .05, suggesting that the distribution of scores demonstrated normality.
Table 2.
Overview of Statistical Measures for Experimental Groups
step | Mean | Std. Deviation | N |
Pretest- exp | 12.6265 | 2.3162 | 96 |
Treatment-exp | 13.3750 | 2.6046 | 96 |
posttest -exp | 14.3542 | 2.5658 | 96 |
The probability value in Table 2 confirms that the assumption of similarity in variance between each measurement time pair is accepted. The Tests of Within-Subjects Effects in RM analysis were used for this study, as indicated in the table. The probability value for DL suggests that students mostly benefited from the repeated time pair and established educational strategies. The values of Mauchly's W (.981) and approximate chi-square (1.817) along with the estimated marginal means at a certain level support this finding. The tests in Table 2 also show that most students preferred to use 'digital technology' (1.817). Around 96 students were randomly selected from four strata, including young and old individuals (18-28 years), who were both more active and less active in DL.
4.2.1. Tests of Within-Subjects Effects
The data regarding the impact of the DL on individuals is presented in Table 2. The statistical significance obtained from this analysis confirms that the assumption of equal variance between each set of measurements is valid. These findings are crucial for utilizing the "In-subject Effect Testing" table for further analysis. Table 2 indicates that there was no significant improvement in the reading comprehension skills of EFL learners after being taught using traditional methods. The average score before the teaching was 12.62 with a standard deviation of 2.31, while the average score after was 13.37 with a standard deviation of 2.60. These results suggest that there was no substantial enhancement in reading comprehension. Furthermore, a paired samples t-test was conducted on the control group with high self-efficacy to compare their scores before and after the teaching. The resulting p-value was .488, which is above the significance level of .05. Therefore, it can be concluded that there was no notable improvement in the reading comprehension skills of high self-efficacious learners in the control group. The mean difference was -.38, standard deviation was 3.16, standard error mean was .481, t-value was 11.23, and the degrees of freedom were 19. In other words, the control group, which received traditional instruction, did not show significant progress in their reading comprehension skills. Table 18 also indicates variations in comprehension scores of the experimental group across three time periods: pretest (M=12.62, SD=2.31), treatment (M=13.37, SD=2.60), and posttest (M=14.35, SD=2.56). To further evaluate the effectiveness of interventions aimed at improving participants' comprehension, a repeated measure analysis was conducted, with results displayed in Table 3
4.3. Results for First and Second Research Questions
The results from Table 3 clearly show a significant difference in reading comprehension scores across three different time periods in the experimental group. Statistical analysis indicates a substantial variance, with an F-value of 13.33 and a p-value of .000 (p < .05). The effect size, as denoted by the partial eta squared value of .123, further reinforces the belief that utilizing DA reading comprehension language learning can greatly improve comprehension skills over these time periods, thus validating the initial hypothesis
Table 3.
Evaluation of Repeated Measurement Analysis for all Participants
Source |
| Type III Sum of Squares |
|
| df | Mean Square | F | Sig. | Partial Eta Squared | |||||||||||
factor1 | Sphericity Assumed |
| 139.465 |
|
| 2 | 69.733 | 13.335 | .000 | .123 | ||||||||||
Greenhouse-Geisser |
| 139.465 |
|
| 1.484 | 93.974 | 13.335 | .000 | .123 | |||||||||||
Huynh-Feldt |
| 139.465 |
|
| 1.502 | 92.846 | 13.335 | .000 | .123 | |||||||||||
Lower-bound |
| 139.465 |
|
| 1.000 | 139.465 | 13.335 | .000 | .123 | |||||||||||
Error(factor1) | Sphericity Assumed |
| 993.535 |
|
| 190 | 5.229 |
|
|
| ||||||||||
Greenhouse-Geisser |
| 993.535 |
|
| 140.987 | 7.047 |
|
|
| |||||||||||
Huynh-Feldt |
| 993.535 |
|
| 142.701 | 6.962 |
|
|
| |||||||||||
Lower-bound |
| 993.535 |
|
| 95.000 | 10.458 |
|
|
|
In contrast, Table 3 reveals no significant differentiation in the performances of the groups (F (1,172) = 0.447, p = .505 >.05, partial eta squared = .003). Consequently, it was determined that the use of DA reading comprehension language learning did not result in a notable enhancement in the participants' comprehension skills.
Figure 1: Comparing the Average Scores Before and After the Test Among Participants
The researchers conducted an analysis in the table below, with a significance level of five percent (α= 0.05) for all tests. To begin the analysis, various Multivariate Tests were performed and listed in the table along with their corresponding names and statistics.
Table 4.
Statistics on Coefficients Analyzing Pretest and Posttest Scores
Correlation | Unstandardized Coefficients | Standardized Coefficients |
|
| 95% Confidence Interval | ||
B | Std. Error | Beta | Lower | Upper | |||
Constant | 5.323 | 1.124 |
| 4.734 | <.001 | 3.090 | 7.555 |
Pretest | .714 | .087 | .644 | 8.165 | <.001 | .540 | .887 |
Recent studies have indicated that having strong DL skills can have a positive impact on academic performance, particularly in the area of reading comprehension. For English as a Foreign Language (EFL) students, being able to effectively utilize digital resources can lead to the adoption of various learning strategies that can improve their reading comprehension abilities. These strategies may involve engaging in online reading exercises, using interactive multimedia materials, and utilizing digital tools to enhance vocabulary and analyze texts. It is worth noting that the connection between DL and reading comprehension scores can be affected by multiple factors, such as the quality of available digital resources, the level of instructional support provided, and the unique characteristics of individual learners. To gain a better understanding of how DL influences reading comprehension among Iranian EFL students, further research specific to this population is necessary.
5. DISCUSSION
The purpose of this research was to investigate a method for identifying challenges that happen during the performance to help to evaluate the reading skills of Iranian EFL learners. One key question that remains unanswered is whether improving EFL DL through DA is linked to psycholinguistic factors. Teachers are curious about how the shift from functional to positional levels occurs, a concept not clearly defined in the Garret model (Garret, 1990). The Garret mental model addresses this by incorporating conceptual structure and inferential processes, connecting functional level representation to positional level representation by assigning frame elements to the terminal string. The study revealed that the selected group showed improvement compared to the initial stage (pretest) in various stages. A major finding was the emphasis on function over forms used in the samples and the importance of understanding key instructions through a quantitative analysis. Furthermore, the research delves into an internal factor that provides a more appropriate form of assessment compared to Anderson's skills model (1986). Similarly, the study assesses the validity of its results in light of current literature on the performance of EFL learners in a digital learning setting.
According to the results, one potential explanation is that the digital evaluation may not produce satisfactory results for teachers assessing reading comprehension strategies. This could be due to the anxiety and stress experienced during the final assessment. Although this form of evaluation allows EFL learners to give feedback and engage in interactive educational activities, it falls short of meeting the expectations of EFL teachers in terms of achieving desirable outcomes. This interpretation is backed by a 2006 report from the European Commission, which proposes that involving an EFL teacher as a mediator could be advantageous in helping students navigate classroom assessments. This method would enable teachers to effectively utilize digital tools, interact with technology, and successfully achieve their goals. The findings of the study align with those of McDougall, Readman, and Wilkinson (2018), who highlight the positive impact of digital learning on EFL learners' reading comprehension abilities, ultimately enhancing educational quality.
The research findings are in line with Esfandiari's previous study in 2020, which emphasized the increasing integration of advanced digital technologies, including DL, in the field of applied linguistics. This integration has enabled language instructors to deliver more impactful lessons to language learners. Based on the psycholinguistic model, the process of DL can be categorized into seven distinct reading processes: integrated, non-sequential skills, meaning centrality, alternative, activity, inexactness, and primacy of function over forms. These processes are utilized for selecting appropriate mediators. The significance of this model for DL lies in its focus on language and is influenced by the literacy levels of instructors and instructional tools. However, despite numerous studies on DL, a clear theoretical foundation is still lacking.
Researchers like Dardanou and Kofoed (2019) and Spante et al. (2018) have highlighted the importance of merging various DL into a cohesive principle and evaluating them through quantitative studies. This study builds upon the existing theoretical model, contributing to the establishment of a framework for the DL discourse. Moreover, a study by Readman and Wilkinson (2018) explored changes in DL and observed differences in performance between participants based on the psycholinguistic model, skills model, and SQ3R process.
5. CONCLUSION
To address the findings of the study, the initial question raised was whether integrating DL instruction and DA could lead to an improvement in the reading comprehension abilities of Iranian EFL learners. Several implications for educational practice follow from this study. DL can provide unique insights into EFL reading comprehension improvement. These insights can inform our understanding of why some learners with sufficient lexical skills nevertheless show ample reading comprehension skills. Future studies will be conducted at the same site to further explore the impact of DL on improving other skills among EFL learners. In conclusion, this study highlights the importance of utilizing DA as an educational strategy to assess the DL skills of Iranian EFL learners. The results indicate that digital education has led to improved mean scores among learners, suggesting the potential effectiveness of online platforms for educational purposes. However, it is important for institutions to have a well-established digital framework to ensure effective assessment. These interconnected issues played a notable role in shaping outcomes. Given these findings, it is recommended that both teachers and students dedicate greater attention to DL methodologies and incorporate psycholinguistic strategies into educational practices. These aspects deserve heightened focus to ensure more effective learning experiences.
REFERENCES
Anderson, J. R. (1984). The development of self‐recognition: A review. Developmental Psychobiology: The Journal of the International Society for Developmental Psychobiology, 17(1), 35-49. https://doi.org/10.1002/dev.420170104
Amaro, A. C., Oliveira, L., & Veloso, A. I. (2017). Intergenerational and collaborative use of tablets: «in-medium» and «in-room» communication, learning and interaction. Observatories (OBS), 11(1), 83–94. https://doi.org/10.15847/obsOBS1102017995
Astiandani, F.R & Anam, S. (2021). EFL teacher’ perceptions toward the implementation of online formative assessment amidst the covid-19 pandemic. Journal of English Language Teaching. 8(2), 269-277. https://doi.org/10.26858/eltww.v8i2.21326
Bekrizadeh, H., Panahi, M. & Jamalvandi, B. (2024). Investigating the impact of digital literacy on electronic learning of Payam Noor University librarians with the mediating role of technology acceptance. Library and Information Science Research, 14(1), 91-110. doi: 10.22067/infosci.2024.85425.1185
Bhatt, I. (2017). Assignments as controversies: Digital literacy and writing in classroom practice. Routledge.
Blikstad-Balas, M. (2015). Digital literacy in upper secondary school–what do students use their laptops for during teacher instruction? Nordic Journal of digital literacy, 10 (Jubileumsnummer), 122-137.https://doi.org/10.18261/ISSN1891-943X-2015-Jubileumsnummer-0
Brown, G. T. (2006). Teachers' conceptions of assessment: Validation of an abridged version. Psychological Reports, 99(1), 166-170. https://doi.org/10.2466/pr0.99.1.166-170
Buckingham, D. (2007). Digital Media Literacies: rethinking media education in the age of the Internet. Research in comparative and international education, 2(1), 43-55. https://doi.org/10.2304/rcie.2007.2.1.43
Chastain, K. (1988). Developing second language skills: Theory and practice. USA: Harcourt Brace Jovanovich, Inc.
Cizek, G. J., Andrade, H. L., & Bennett, R. E. (2019). Handbook of formative assessment in the disciplines. Routledge.
Cohen, V. L., & Cowen, J., E. (2008). Literacy for children in an information age: teaching reading, writing and thinking. Thomson Higher Education.
Dardanou, M., & Kofoed, T. (2019). It is not only about the tool’s professional digital competence. In G. Colette & I. Palaeologus (Eds.), early learning in the digital age (pp. 61–76). London: SAGE Publications.
Esfandiari, R. (2020). Iranian EFL teachers’ DL in academic settings: Teacher professionalism in the digital Age. Journal of Foreign Language Research, 10.22059/JFLR.2019.266987.556
Etemadi, S. H. and Abbasian, G. (2023). Dynamic Assessment and EFL Learners' Writing Journey: Focus on DA Modalities and Writing Revision Types. Teaching English Language, 17(1), 53-79. doi: 10.22132/tel.2022.162923
European Commission. (2006). Recommendation on key competences for lifelong learning. Council of 18 December 2006 on Key Competences for Lifelong Learning, 2006/962/EC, L. 394/15. Retrieved December 7, 2017, from http://eur-lex.europa.eu/legalcontent/en/TXT/?uri=CELEX:32006H0962&qid=1496720114366
Falloon, G. (2020). From digital literacy to digital competence: the teacher digital competency (TDC) framework. Educational Technology Research and Development, 68(5), 2449-2472. https://doi.org/10.1007/s11423-020-09767-4
Fulcher, G. (2012). Digital literacy for the language reading comprehension. Language Assessment Quarterly, 9(2), 113-132. doi:10.1080/15434303.2011.642041
Garrett, M. F. (1990). Sentence processing. In D. N. Osherson & H. Lasnik (Eds.), Language: An invitation to cognitive science, Vol. 1, pp. (133–175). The MIT Press.
Gillen, J., & Barton, D. (2010). Digital literacies: Research briefing for the TLRP-TEL (Teaching and Learning Research Programme-Technology Enhanced Learning). London: London Knowledge Lab, Institute of Education. Retrieved June 3, 2010, from the TLRP TEL website: http://www.tlrp.org/docs/DigitalLiteracies.pdf
Grabe, W. (2008). Reading in a second language: Moving from theory to practice. New York: Cambridge University Press.
Gupta, R., Seetharaman, A., & Maddulety, K. (2020). Critical success factors influencing the adoption of digitalisation for teaching and learning by business schools. Education and Information Technologies, 25(5), 3481-3502. https://doi.org/10.1007/s10639-020-10246-9.
Hashemian, M., & Fadaei, B. (2013). Fostering EFL learners’ Autonomy in Light of Portfolio Assessment: Exploring the Potential Impact of Gender. Iranian Journal of Language Teaching Research, 1(2), 135-151. https://files.eric.ed.gov/fulltext/EJ1127429.pdf
Heidari, N., & Tabatabaee-Yazdi, M. (2021). Digital literacy skills among Iranian EFL teachers and students. Journal of Research in Techno-based Language Education, 1(1), 22-34. 10.22034/JRTLE.2021.133013
Inbar-Lourie, O., & Levi, T. (2020). Assessment literacy as praxis: Mediating teacher knowledge of assessment-for-learning practices. Toward a Reconceptualization of Second Language Classroom Assessment: Praxis and Researcher-teacher Partnership, 241-259.
Imron, A., & Anwar, K. (2024). Reading proficiency and cognitive reading strategies through online dynamic assessment (ODA) in English for economy. English Review: Journal of English Education, 12(3), 949-958. https://doi.org/10.25134/ erjee. v12i3.10370
Katz, I. R. (2007). Testing information literacy in digital environments: ETS’s iSkills assessment. Information Technology and Libraries, 26(3), 3-12. https://doi.org/10.6017/ital.v26i3.3271
Khlaisang, J. & Koraneekij, P. (2019). Open online assessment management system platform and instrument to enhance the information, media, and ICT literacy skills of 21st century learners. International Journal of Emerging Technologies in Learning. 14(7), 111-127. https://doi:org/10.3991/ijet.v14i07.9953.
Khodabakhsh, S., Abbasian, G., R., & Rashtchi, M. (2018). Incorporation of dynamic assessment models into developing language awareness and metacognitive strategy use in writing classes. Journal of Modern Research in English Language Studies, 5(4), 55-79.
Jakeman, V., & McDowell, C. (2008). New insight into IELTS. Cambridge University Press.
Lee, L., Gundersen, E., & Bernard, J. (2011). Select readings. Oxford: Oxford University Press.
LIopis Nebot, M. A; Esteve-Mon, F. V & Segura, J. A. (2021). Diagnostic and educational self-assessment of the digital competence of university EFL learners. Nordic Journal of DL. 16(3), 115-131. doi. 10.18261/issn.1891-943x-2021-03-04-03
Lukitasari, M., Murtafiah, W., Ramdiah, S., Hasan, R., & Sukri, A. (2022). Constructing DL instrument and its effect on college students' learning outcomes. International Journal of Instruction, 15(2), 171-188. https://doi.org/10.29333/iji.2022.15210a
McDougall, J., Readman, M. & Wilkinson, P. (2018). The uses of (digital) literacy, Learning, Media and Technology, 43:3, 263-279, doi: 10.1080/17439884.2018.1462206
Makhachashvili, R., & Semenist, I. (2021). Student satisfaction with digital hybrid learning In European and oriental languages programs: Survey study of regional universities of Ukraine. In proceedings of the 19th international conference E-society, pp. 133– 143.
Martin, A., & Grudziecki, J. (2006). DigEuLit: Concepts and tools for DL development. Innovation in Teaching and Learning in Information and Computer Sciences, 5(4), 249–267. https://doi.org/10.11120/ital.2006.05040249
Mehri Kamrood, A., Davouidi, M., Ghaniabadi, S., & Amirian, S. M. R. (2019). Diagnosing EFL learners’ development through online computerized DA. Computer Assisted Language Learning, 34(1), 1-30. https://doi. 10.1080.10958822101645181.
Mellati, M., & Khademi, M. (2018). Exploring EFL learners’ assessment literacy: Impact on EFL learners’ writing achievements and implications for teacher development. Australian Journal of Teacher Education, 43 (6)1-18. https://doi.org/10.14221/ajte.2018v43n6.1
Mirra, N., Morrel, E., & Filipiak, D. (2018). From digital consumption to digital invention: Toward a new critical, theory and practice of multiliteracies. Theory into Practice, 57, 12-19. http://doi.org/10.1080/00405841.2017.1390336
Muawanah, U., Marini, A., & Sarifah, I. (2024). The interconnection between digital literacy, artificial intelligence, and the use of E-learning applications in enhancing the sustainability of Regional Languages: Evidence from Indonesia. Elsevier, Social Sciences and Humanities Open, 10, 1-17. https://doi.org/10.1016/j.ssaho.2024.101169
Murray, M. C., & Pérez, J. (2014). Unraveling the DL paradox: How higher education fails at the fourth literacy. Issues in Informing Science and Information Technology, 11, 85-100. http://iisit.org/Vol11/IISITv11p085
Mustiah, M., Dayat, D., & Sadek, N. (2024). The Impact of mobile-assisted hybrid dynamic assessment on Arabic language leaners’ reading comprehension performance. Journal of Languages and Language Teaching, 12(1), 524-534. https://doi.org/10.33394/jollt.v12i1.9158
Norton, B. (2013). Identity, literacy, and English language teaching. Iranian Journal of Language Teaching Research, 1(2), 85-98. doi:10.18806/tesl. v28i1.1057
Panadero, E., Brown, G. T. L., & Strijbos, J. W. (2016). The future of student self-assessment: A review of known unknowns and potential directions. Educational Psychology Review, 28(4), 803-830. doi:10.1007/s10648-015-9350-2
Pearson, P. D. (1976). A psycholinguistic model of reading. Language Arts, 53(3), 309-314.
Petersen, D. B., Tonn, P., Spencer, T. D., & Foster, M. E. (2020). The Classification Accuracy of a DA of Inferential Word Learning for Bilingual English/Spanish-Speaking School-Age Children. Language, Speech, and Hearing Services in Schools, 51(1), 144-164. https://doi:10.1044/2019_lshss-18-0129
Pill, J., & Harding, L. (2013). Defining the language DL gap: Evidence from a parliamentary inquiry. Language Testing, 30(3), 381-402. http://dx.doi.org/10.1177/0265532213480337
Scarino, A. (2013). Language DL as self -awareness: Understanding the role of interpretation in assessment and in teacher learning. Australia Language Testing 30(3), 309-327. doi: 10.1177/0265532213480128
Scriven, M. (1967). The methodology of evaluation. In W. R., Tylor R. M. Gagne, & M. Scriven (Eds.). Perspectives of curriculum evaluation (pp.39-83). Chicago: Rand McNally.
Shabani, K. (2018). Group DA of L2 Learners' writing abilities. International Journal of Language Literature Research 6(1) 129-149. doi: 10.30466/IJLTR.2018.20494
Spante, M.., Hashemi, S. S., Lundin, M., & Algers, A. (2018). Digital competence and digital literacy in higher education research: Systematic review of concept use. Cogent Education, 5(1), 1519143. https://doi.org/10.1080/2331186X.2018.1519143
Stiggins, R., & Chappius, J. C. (2005). Using student-involved reading comprehension assessment to close achievement gaps. Theory into practice, 44(1), 11-18. doi:10.1207/s15430421tip4401_3
Storto, A. (2021). Fingerprints: Towards a multisensory approach to meaning in digital media. Nordic Journal of Digital Literacy, 16 (3-4), 132-143. http://doi.org/ 10.18201
Tavassoli, K. and Sorat, Z. (2023). Iranian EFL Teachers’ Oral/Aural Skills Language Assessment Literacy: Instrument Development and Validation. International Journal of Language Testing, 13(2), 56-76. doi: 10.22034/ijlt.2023.376466.1219
van Dinther, M., Dochy, F., & Segers, M. (2015). The contribution of assessment experiences to student teachers' self-efficacy in competence-based education. Teaching and Teacher Education, 49, 45-55. http://dx.doi.org/10.1016/j.tate.2015.02.013
Yousefi, N., Velayati, S., & Ebadi, S. (2024). Implementing group dynamic assessment to enhance Iranian high school students’ grammar ability. Journal of English Language Teaching and Learning, 16(33), 341-358. doi: 10.22034/elt.2024.59652.2592
Yu, Z. (2022). Sustaining student roles, digital literacy, learning achievements, and motivation in online learning environments during the COVID-19 pandemic. Sustainability, 14(8), 4388. https://doi.org/10.3390/su14084388