Introduction
This study examined the relevance of alternative assessment (AA) in the context of Ethiopian higher education. This related specifically to a communicative English Skills course (CESC) offered at Ethiopian universities. Assessment has been an essential area in English language teaching (ELT) for more than half a century, and it is an area of dynamic academic investigation (Al-Mahrooqi, & Denman, 2018; Davies, 2013; Monib et al., 2020; Phuong & Nguyen, 2019). In particular, as Abbas (2012) and Mekonnen (2014) contend, studies have identified that two opposing modes of assessment, traditional assessment (TA) and AA, have influenced ELT assessment in higher education.
In respect to CESC, TA refers to conventional or standardized methods of testing, and usually entails multiple-choice, true/false, matching, and gap-filling test items. For its part, AA involves a more informal and continuous process for evaluating students’ language ability and students are asked to carry out real-life activities that resemble the authentic application of what they have learned (Agustina, 2011). Tran (2012) defines AA as evaluation processes that can be employed in an instructional setting, and which can be readily applied to learning tasks in a particular classroom situation (also see Monib et al., 2020). The field of teaching English as a foreign language (EFL), in particular, has witnessed a paradigm shift in assessment since the 1970’s from the mere use of TA to more alternative forms of assessment (Asabe, 2017). This alternative form of assessment has resulted in the active involvement of learners in the form of peer-assessment and self-assessment to determine student performance with a view to making it as more authentic and holistic as possible (Al-Mahrooqi, & Denman, 2018; Brown, 2012; Shrestha, 2014).
In relation to the application of AA in the Ethiopian higher education context, both continuous assessment (CA) and the active learning method (ALM) became officially operative after 1994, following the implementation of the Ethiopian Educational and Training Policy (Ministry of Education, 1994). Particularly, Ethiopian universities practice standardized cumulative and more judgmental methods to assessments (Teferra et al., 2018). This is largely because standardized tests and examinations are regarded as the measurement of the outcome of a lesson that students are taught during a certain period of time (Abiy, 2013; Al-Mahrooqi, & Denman, 2018; Mekonnen, 2014; Monib et al., 2020; Motuma, 2015; Phuong & Nguyen, 2019). However, Abiy (2013) asserts that this form of assessment is insufficient for ascertaining students’ language progress. For this reason, currently, there is an inclination to use AA to gauge what learners know and can do in a more dynamic approach (Abiy, 2013; Mekonnen, 2014; Temesgen, 2017). This shift is influenced by the underlining assumption that there is a correlation between assessment and instruction within the Ethiopian higher education context (Abiy, 2013).
In CESC, the term communicative incorporates three inseparable theories: the theory of communicative course design, the theory of active learning method (ALM), and the theory of alternative assessment (AA) as used by Al-Mamari et al. (2018) and Motuma (2018) (see Figure 1).
Figure 1: Alignment of the communicative language teaching components
The communicative course design, in the context of CESC, underlines communicative language teaching and assessment practices. This includes content-based instruction (organizing activities around certain contents or subject-matters); theme-based instruction (linking the contents to the language skills); experiential learning (stimulating concrete learners’ experiences); episode hypothesis-based instruction (using series methods following a story line); and task-based learning (organizing activities around authentic and integrated tasks) (Ansarey, 2012; Brown, 2010; Davies, 2013; Smith, 2000; Ur, 2010). This is because CESC focuses on a whole set of real-world and well-integrated skills, and interactive and practical tasks that require the communicative language teaching (CLT) approach and assessment methods to help students develop English language competence.
Implementing AA in CESC at Ethiopian Universities
The purpose of the current study was to investigate the implementation of alternative assessment (AA) in CESC as part of teaching English as a foreign language (TEFL) at three Ethiopian universities (Addis Ababa University, Wollaga University, and Ambo University). The purpose of the implementation of AA strategies in CESC, in particular, should be seen in relation to the quality of teaching CESC at Ethiopian universities, in general, and at the three universities, in particular. This is because AA is a central element of a successful teaching-learning process in CESC (Herdiawan, 2018). Therefore, the types of AA strategies students are subjected to determine their success in CESC (Teelken, 2018). Scholars, such as Herdiawan (2018), Rojas Serrano (2017) and Teelken (2018), maintain that aligning AA tools and strategies with teaching-learning strategies and with the nature of the objectives of CESC, can help instructors and students achieve quality education. To this end, much effort is needed to promote the implementation of AA to enhance the quality of assessment and the quality of teaching and learning in CESC (Teelken, 2018).
To support the importance and the relevance of AA in CESC, Shrestha (2014) argues that the assessment practices of CESC are different from the assessment practices of most specific language courses, including productive and receptive language skills. This is because the complex components of the language domains in CESC require EFL instructors to use various types of AA tools and strategies to assess students’ performance (Al-Mahrooqi & Denman, 2018; Bachelor, 2017; Herdiawan, 2018; Minda, 2019). In other words, EFL instructors are not expected to assess CESC using discrete-item tests in the current context of CESC (Minda, 2015; Ministry of Education,, 2017; Motuma, 2014; Shrestha, 2014). For example, in the 1950’s and 1960’s, multiple-choice and true/false items were implemented concurrently with teacher-centered approaches (Gil & Lucas, 2013; Herdiawan, 2018; Rojas Serrano, 2017). In spite of this, the goal of CESC is to help students develop knowledge, interest, and skills in integrated English language domains through ALMs (Ministry of Education, 2017). Therefore, AA is an indispensable technique to assess both students’ knowledge and their performance in CESC (Bachelor, 2017; Minda, 2019).
Based on the above, this study set out to answer the following research questions:
What components of AA do EFL instructors currently employ to assess CESC Addis Ababa, Wollaga, and Ambo Universities?
What are instructors’ responses to AA strategies used in CESC at these three universities?
How do English major students respond to their instructors’ AA in CESC at these three universities?
Research Methodology
This study employed a convergent, parallel, mixed research design (Creswell & Creswell, 2018; Demir & Pismek, 2018; Shannon-Baker, 2016). The purpose of this design was to comprehensively address the research questions as stated above. Within this research design, a mixed-methods approach was utilized. The latter allowed for the integration of quantitative and qualitative data (Creswell & Creswell, 2018; Demir & Pismek, 2018), which were the types of data that were collected in this study.
Population
The population for this study included the following entities and groups: universities, university instructors, and university students. In this context, Ethiopia has 45 universities clustered into four generations for managerial purposes. Cluster refers to the geographical location of the universities in the country, and the generation is about the time at which these universities were established according to the Ethiopian Ministry of Education (2004). First-generation universities (ten universities) were established before 2006; second-generation universities (ten universities) were instituted from 2007 to 2009; third-generation universities (ten universities) were set up between 2010 and 2014; and the recently-established universities are categorized as fourth-generation universities.
As highlighted below, only three universities were selected for this study. In all, there were 128 English language instructors and 953 English language students at the three universities when the study was conducted.
Sampling
Universities
Before the study was conducted, an ethical clearance was granted by the UNISA’s College of Human Sciences Research Ethics Committee. The three government universities, Addis Ababa University, Wollaga University, and Ambo University were purposefully selected based on the cluster and the generation they belong to. The three selected universities belong to the first, second, and third generations, respectively.
English language instructors
As pointed out above, at the time the study was conducted, there were 128 English language instructors teaching CESC at the three universities (see Table 1). All of them were included in the study based on the maximum variation sampling technique. This was done to represent diverse variations within these instructors (Creswell, 2014; Creswell & Creswell, 2018; Palinkas et al., 2015). Before data were collected, copies of a consent letter were sent out to participants to invite them to take part in the study.
Table 1: Instructors’ background information
Students
In all, 241 students from the three Ethiopian universities were selected for this study. These were first-year, English major students. They were selected through a stratified sampling technique using Kothari’s (2004) formula as shown below:
Therefore, based on the actual size of the population of students at each of the three universities (see Table 2) during the study period, 241 students were selected from the population of 953 students using a stratified sampling technique. This type of sampling technique maintains the proportionality of the sample size to be selected from each university and gives equal chance for every individual in the population. Here, too, before data was collected, copies of a consent letter were sent out to participants to invite them to take part in the study. Table 2 indicates how the appropriate sample size for each university was arrived at.
Table 2: Sample of students at each of the three Ethiopian universities
Data collection
Data was collected through questionnaires, which were electronically sent out to participants in the first semester of 2020. The questionnaires consisted of statements that were scored in terms of a five-item Likert scale by ticking or circling either a number 1, 2, 3, 4, or 5 against each statement. The numbers were allocated the following values: 5 = always; 4 = often; 3 = sometimes; 2 = rarely and 1 = never.
Data analysis
The data obtained through questionnaires were subjected to descriptive statistical analysis and represented in percentages, weighted means, standard deviations, and rank orders based on the recommendations of Demir and Pismek (2018), and Shannon-Baker (2016). Data sets were categorized into the following themes: multi-assessor strategies; multiple AA tools; and comprehensive assessment; and progressive, continuous, and relevant assessment. As responses were collected from two groups of respondents from each university, namely, instructors and students, the Mann-Whitney U Test was employed to identify the difference between the responses of instructors and students in relation to the implementation of AA in CESC.
Similarly, the Kruskal-Wallis Test was used to determine the difference in the three universities’ participants’ responses to the aspects mentioned above. The Mann-Whitney U Test, which is an alternative to the t-test, and the Kruskal-Wallis Test, which is an alternative to an ANOVA, are useful when the difference between the sample sizes of two independent populations is large, when the test is nonparametric, and when the type of data is ordinal (Demir & Pismek, 2018), as was the case with the current study. In order to triangulate the data, the Mann-Whitney U Test was supplemented by a t-test where the number of the two groups was equal or nearly equal. The data were analyzed using SPSS® Statistics 26.0. For all statistical tests, alpha was pre-set at = 0.05 as a conventional practice to accommodate the fact that a given finding has a 5% chance of not being true (Everitt & Skrondal, 2010).
Results
Implementation of multi-assessor strategy in CESC
To determine the most frequently used AA strategies in each language domain in CESC, both EFL instructors and the students responded to questionnaires sent to them at the three universities. Based on the responses of the two sets of participants from the three universities, the mean scores for each statement were computed as summarized in Tables 3 and 4 (also see Appendices A and B for more information on CESC’s language domains).
Table 3: Instructors’ responses on the use of multi-assessor strategy in CESC by mean
The matrix in Table 3 presents the mean values of the implementation of multi-assessor strategy (rows) compared to the attention given to each component of CESC (columns). Thus, along with the rows of the table, the weighted mean 4.0 shows that instructors always assessed student learning in CESC for their own sake or using TA rather than using peer- and self-assessment, as well as invited guest assessment strategy in all the six language domains in CESC. Even within the instructor assessment, the mean values 2.4 and 3.1 indicate that listening and writing skills were given less attention in comparison to grammar (4.9), reading (4.8) and vocabulary (4.7).
The rows in the table demonstrate that instructors did not invite any guest to assess any language domain in CESC during the period of the study at the three universities. However, the analysis of the instructional materials showed that the invited guest assessment strategy is frequently recommended to measure students’ learning outcomes in CESC. Similarly, the weighted mean values 1.67, 2.15, and 2.67 imply that instructors sometimes employed peer-assessment and self-assessments and rarely integrated any two language skills in their assessment in CESC. In conjunction with the columns of the matrix, the weighted mean values 2.72 and 2.62 show that instructors employed a multi-assessor strategy more frequently to assess grammar and vocabulary knowledge respectively compared to listening (1.36), writing (1.6) and speaking (1.7) in CESC. Generally, the overall weighted mean value 2.1 implies that the instructors’ practice in implementing a multi-assessor strategy in all language domains in CESC is insignificant. Nonetheless, the CESC module suggests a variety of AA strategies to match assessment strategies with the nature of the learning objectives in CESC.
Table 4: Students’ responses to the implementation of multi-assessor strategy by mean values
In Table 4, the rows illustrate the implementation of an AA strategy to assess the entire language domains, and the columns show how a language domain in CESC is assessed by a variety of AA strategies. Concerning students’ responses to the use of multi-assessor strategies, the mean value 3.55 in this table indicates that instructors often assessed student work in CESC for themselves. Moreover, the mean value 2.73 shows that slightly more than half (55%) of the 119 students pointed out that their instructors hardly ordered them to assess their own work in CESC. Students also disclosed that they were rarely assessed by their peers (2.35) or by means of a combination of instructor, peer- and self-assessment strategies (1.59).
Implementation of multiple AA tools in CESC
As displayed in Table 5, instructors seldom used inter-student interviews (0.67) and audiotape recordings (1.0) to assess language domains in CESC at the three universities. The weighted means 3.37, 3.33, and 2.89 also exhibit that instructors gave relatively more emphasis to reading, vocabulary, and grammar knowledge respectively, than to listening (2.00), writing (2.18) and speaking (2.4) in implementing AA tools. To develop a stronger argument for this issue, Table 6 presents the analysis of the students’ responses to the implementation of AA tools in each language domain in CESC.
Table 5: Instructors’ responses to the implementation of multiple AA tools in each language domain in CESC by mean and weighted means
As illustrated in Table 6, instructors used homework/assignments, and very little individual and group work, projects, and peer-teaching activities to decide only 8% of their grades. On the other hand, the mean values in the brackets indicate that vocabulary (3.74), grammar (3.64) knowledge, and reading (3.53) received relatively higher attention than listening (2.88), speaking (2.90), and writing (3.38) at the three universities.
Table 6: Students’ responses to the implementation of AA multi-assessment tools in each language domain in CESC by mean and weighted means
Use of comprehensive and progressive assessment strategies in CESC
The symbol “f” refers to the frequency of the items as displayed in Tables 7 and 8. With respect to the comprehensive use of AA in CESC, the mean values in Table 7 demonstrate that most (82.6%) of the instructors often assessed students’ grammar (4.13), vocabulary (4.17) and their reading skills (4.3), using test items in CESC. Evidently, they rarely (2.24) used a variety of AA techniques to assess the intended learning objectives in CESC. Similarly, they infrequently assessed students’ motivation (2.62), writing (2.7), speaking (2.74) and listening (2.86), as well as students’ ability in integrating the four language skills (3.25) in CESC. Moreover, they hardly checked students’ attitudes towards the nature of CESC (2.49) at the three universities.
Table 7: Instructors’ responses on the comprehensive use of AA in CESC by percentages and means
In Table 8, the mean values 4.16, 4.25, and 4.13 show that 185 (85%) of the students thought that their instructors often assessed their reading skills, grammar, and vocabulary knowledge respectively at the three universities. Similarly, students expressed the view that their instructors sometimes assessed their speaking skills (2.65) and checked their interests in CESC (2.54) and in AA tools (2.51) concerning the teaching-learning process in CESC. Students complained that the instructors rarely checked their students’ motivation towards the CESC (1.85) at the three universities. On the whole, instructors did not comprehensively apply their assessment strategies in CESC (2.96) at the three universities.
Table 8: Students’ responses to the comprehensiveness of AA in CESC by percentages and means
The mean value, 3.62, in Table 9 shows that instructors at the three universities often checked their students’ individual learning differences in CESC. They also occasionally ascertained the suitability of their assessment tools (2.72) to the CESC instructional activities. However, instructors intermittently used AA to determine students’ language ability (3.02), their learning progress through feedback (3.05), and their problem areas for remedial consideration (2.91) at the three universities. To that end, they occasionally gave interventions to fill students’ learning gaps (2.58) in CESC. They also occasionally determined the suitability of an AA tool to the learners’ learning outcomes (2.78), individual learners’ learning styles (2.86), and to the language objectives (2.91) in CESC. However, instructors reported that they rarely (2.24) ascertained the appropriateness of their assessment tools to both students’ individual learning styles and students’ levels of language proficiency in CESC.
Table 9: Instructors’ responses on the progressiveness of AA in CESC by percentages, means and weighted means
The mean values in Table 10 represent students’ opinions about the fact that their instructors occasionally used AA to determine their learning gaps (2.72), to close gaps with tutorial sessions (2.67), and to give them feedback based on the assessment results (2.65). Students felt that their instructors rarely gave tutorials (2.56) to assist them according to their assessment results. Notably, students complained that instructors sometimes assessed what they had not taught (2.63), which means instructors occasionally set assessment items in relation to the intended learning objectives in the CESC module (2.64). This finding implies that the instructors rarely assessed what they taught in CESC, which, according to Rahman et al. (2019), can have a negative backwash effect in teaching CESC. Also known as washback effect, this particular effect refers to student readiness for a test and a mismatch, or a proper match of course content aspects, curriculum aspects, learning materials, and instructor activities (Alebachew & Minda, 2019; Kuang, 2020). In the current study, the negative backwash effect related to the mismatch between instructors’ assessment practices and CESC’s recommended assessment methods.
Table 10: Students’ responses to the progressiveness, continuity, and relevance of AA in CESC by percentages, means and weighted means
As shown in Table 11, reading, grammar, and vocabulary items 522 (82.6%) generally dominated the assessment items at the three universities. Reading activities received greater attention at Ambo and Wollaga universities, followed by grammar and vocabulary items. At the Addis Ababa University, however, grammar, 66 (28.45%) was given more attention than reading in CESC. Vocabulary was the third-most assessed language domain in CESC at the three universities. However, listening 9(1.4%), writing 45 (7.1%), and speaking (8.9%) were overlooked by instructors at the three universities.
Table 11: Frequency of language domain items in the CESC assessments at the three universities
Table 12 illustrates the difference between instructors’ and students’ views concerning the implementation of AA components in CESC as represented by the Mann-Whitney U Test. The results of the Mann-Whitney U Test indicate that there is no difference between instructors’ and students’ views regarding the implementation of AA in CESC. In other words, both the instructors and the students had similar views about the implementation of multi-assessor strategies (z = 0.54), the multiple AA tools (z = 0.69) and the use of comprehensive assessment (z =0.64), and progressive, continuous, and relevant AA in CESC (z =0.51) at the three universities. The overall result of the Mann-Whitney U Test (z = 0.46) confirms the same finding that there is no difference between the three universities in implementing AA in CESC. A Kruska-Wallis Test was also conducted to determine the difference between the opinions of the instructors and the students in relation to the three universities, as indicated in Table 12.
Table 12: Mann-Whitney U Test results concerning the implementation of AA components in CESC (The standard deviations ranged for 1.12-2.21 for instructors and from 1.13-2.31 for students)
Table 13: Kruskal-Wallis Test results regarding the difference among the three universities
As indicated in Table 13, there were 398 respondents in total, 194 from Addis Ababa University, 89 from Ambo University, and 106 from Wollaga University. The mean rank square of all the observations was 3035.07. The summary result of Kruskal–Wallis Test (t =0.257 with p = 0.579 at p-value of < 0.05) reveals that there were no statistically significant differences among the three sample universities in implementing the components of AA in CESC. In terms of the three research questions for this study, this means that: (a) instructors at the three Ethiopian universities employed more of traditional assessment strategies and less of alternative assessment (AA) strategies; (b) despite the instructors’ claims that they (instructors) implemented a multi-assessor strategy in CESC all the time, that was, nonetheless, not the case; and (c) students negated instructors’ assertions that they (instructors) employed AA methods.
Discussion
The present study investigated the implementation of the components of AA in CESC at three Ethiopian universities. In this instance, it set out to answer three research questions:
What components of AA do EFL instructors currently employ to assess CESC at Addis Ababa, Wollaga, and Ambo universities?
What are instructors’ responses to AA strategies used in CESC at these three universities?
How do English major students respond to their instructors’ AA in CESC at these three universities?
The implementation of AA components was investigated as a function of who assesses and, what and how to assess student learning in CESC. As regards a multi-assessor strategy, the mean and the weighted mean indicated that an instructor-based assessment strategy (M = 4.0) dominated the process of evaluating student learning. More than half of the students from the three universities also indicated that their instructors hardly ever used student peer assessment. In addition, instructors never used an invited guest assessment strategy to assess any of the six language domains. This particular result invalidates instructors’ claims that they (instructors) always implemented a multi-assessor strategy. In contrast, CESC emphasizes the relevance of the invited guest, peer- and self-assessment strategies. Moreover, researchers in the field argues for the use of a variety of assessors to assess the language objectives in CESC in an integrative manner (Al-Mamari et al., 2018). Other researchers contend that the use of multi-assessor strategies facilitates the alignment between AA components and the nature of CESC in the context of CLT (Bourchaib, 2017; Kaboula & Elias, 2015; Kibbe, 2017). Compared with other related contexts, the instructors at the three universities neglected the use of an integrated assessment approach (Herdiawan, 2018; Kaboula & Elias, 2015; Palacio et al., 2016). In particular, this result is inconsistent with the sentiments echoed by Nasab (2015), Rojas Serrano (2017), Temesgen (2017), and Wubshet & Menuta (2015) that instructors should employ a multi-assessor strategy in all language domains of CESC.
Pertaining to implementing multiple AA tools in CESC, the current study has shown that most (95%) of the instructors relied on traditional assessment (TA), especially testing, for measuring student performance and for grading students in CESC at the three universities. In fact, according to students, only few instructors employed individual and group work, projects, and peer-teaching. According to Davies (2013), an excessive use of testing or TA to assess the intended learning outcomes can be counterproductive, particularly when there is a mismatch between assessment strategies and course objectives (Al-Mamari et al., 2018; Davies, 2013; Temesgen, 2017; Wubshet & Menuta, 2015). The study has also discovered that instructors at the three universities only occasionally assessed students’ individual learning differences and the suitability of the assessment tools to CESC’s instructional activities. In doing so, they failed to provide sufficient interventions and remedial tasks to address learning gaps students had in CESC. Rojas Serrano(2017) and Shrestha (2014) point out that this kind of assessment is fruitless.
Concerning AA comprehensiveness, the study has revealed that the assessment applied to CESC by instructors from the three universities was non-comprehensive, focused on discrete language items, and was disproportionate to the six language domains offered by CESC. The study has also demonstrated that the assessment used in CESC at the three universities lacked progressiveness, continuity, and relevance. For instance, the collective assessment at the three universities was targeted more toward reading, grammar and vocabulary items ,and less toward the other three language domains. This is a view confirmed by students at the three universities, who indicated that their instructors often assessed their reading skills, grammar, and vocabulary knowledge. To this effect, reading activities received greater attention at AU and WU, followed by grammar and vocabulary items. At AAU, however, grammar (28.45%) received more attention than reading in CESC. Vocabulary assessment was the third-most emphasized language domain in CESC at the three universities. By contrast, listening (1.4%), writing (7.1%), and speaking (8.9%) skills were overlooked by instructors at all the three universities. According to Al-Mamari et al. (2018), Bourchaib (2017), and Kibbe (2017), incomplete assessment, point-by-point tests, and inconsistent assessment in language domains for CESC are the mainstay of TA. Nonetheless, they can create negative washback effects for CESC at the three universities. Moreover, the assessment employed by most instructors did not follow CESC’s assessment guidelines. This mismatch was caused by instructors’ attempt to assess the intended learning outcomes in CESC by using traditional testing strategies. However, Coombe and Hubley (2011) and Molla (2018) argue that instructors cannot assess the intended learning outcomes in CESC by simply following TA.
Conclusion
Based on the results of this study, the inference to make is that most of the instructors at the three Ethiopian universities misunderstand the conceptual difference between CA and AA, and the difference between TA items and AA components. There were inconsistencies between what they said and what they did in assessing the intended learning outcomes in CESC at the three universities. One observation from this study is that instructors were more optimistic and confident about their own assessment practices in CESC than the way, in which students perceived those assessment practices. Additionally, instructors’ assessment tasks were mainly dominated by traditional test items and specifically characterized by an inclination to assign student grades in CESC.
These assessment inconsistencies imply two things. First, most of the instructors at the three universities preferred TA methods to AA methods and found the former more convenient to use in CESC than the latter. Second, the instructors denied the apparently existing mismatch between their assessment practices and CESC’s recommended assessment methods. In this sense, most of the instructors’ assessment practices at the three universities negated the intended learning outcomes spelt out in CESC. Nevertheless, the efforts of those instructors, however few they could have been, who tried to use some elements of AA in CESC are worth commending, particularly for the purpose of assessment as learning. They systematically prompted the students to apply their prior knowledge and helped students develop a sense of ownership of their learning.
Future research needs to consider the value of comprehensive and progressive assessment strategies when evaluating student performance in university courses, such as CESC, which are offered to EFL students. It should also acknowledge that TA has its own inherent weaknesses, and that AA needs to be employed to complement it. Importantly, future research has to ensure that any form of assessment is fully aligned to the stated learning outcomes of any university-level course.
References
Abbas, Z. (2012). Difficulties in using methods of alternative assessment in teaching from Iraqi instructors point of view. Al-Fatih Journal, 8(48), 200-221. https://www.iasj.net/iasj/download/43ba8bcdf35fe9c4
Abiy, Y. (2013). High school English language teachers’ and students’ perceptions, attitudes and actual practices of continuous assessment. Academic Journal of Educational Research and Reviews, 8(16), 1489-1498.https://academicjournals.org/journal/ERR/article-full-text-pdf/79601868173
Agustina, R. (2011). Portfolio as alternative assessment in language testing. Eksplanasi, 6(1), 1-6. https://espace.library.uq.edu.au/view/UQ:134406b/UQ134406b.pdf
Al-Mahrooqi, R., & Denman, C. H. (2018). Alternative assessment: Alternative forms of assessment. In J. I. Liontas (Ed.), The TESOL encyclopedia of English language teaching (pp. 4851–4856). Wiley. https://doi.org/10.1002/9781118784235.eelt0325
Al-Mamari, M. R. H., Al-Mekhlafi. A. M, & Al-Barwani, T. A. (2018). Is communicative language teaching being tested communicatively? An analysis of English tests in Oman. English Language Teaching, 11(11), 83-91. http://doi.org/10.5539/elt.v11n11p83
Alebachew, M. G., & Minda, H. M. (2019). Washback effect of EFL teacher-made test on teaching-learning process of communicative English skills course at Ambo University. International Journal Online of Humanities, 5(6), 1-29.https://doi.org/10.24113/ijohmn.v5i6.146
Ansarey, D. (2012). Communicative language teaching in EFL contexts: Teachers’ attitude and perception in Bangladesh. ASA University Review, 6(1), 61-78.
Asabe, M. B. (2007). Effects of continuous assessment on academic achievement of NCE chemistry students in Kaduna state [Unpublished masters dissertation] Hawasa University.
Bachelor, R. B. (2017). Alternative assessment and student perceptions in the world language classroom. The TFLTA Journal, 6, 30-44. https://files.eric.ed.gov/fulltext/ED577103.pdf
Bourchaib, B. (2017). Exploring teachers’ assessment practices and skills. International Journal of Assessment Tools in Education, 4(1), 1-18. https://doi.org/10.21449/ijate.254581
Brown, J. D. (2010). Language assessment: Principles and classroom practice. Longman.
Brown, J. D. (2012). Choosing the right type of assessment. In C. Coombe, P. Davidson. B. O’Sullivan, S. Stoynoff (Eds.), The Cambridge guide to language assessment (pp. 133-139). Cambridge University Press.
Coombe, C., & Hubley, N. (2011). Fundamentals of language assessment manual. https://www.scribd.com/doc/78467061/Fundamentals-of-Language-Assessment-Manual-by-Coombe-and-Hubley
Creswell, J. W. (2014). Research design: Quantitative, qualitative and mixed methods approaches. Sage.
Creswell, J. W., & Creswell, J. D. (2018). Research design: Qualitative and quantitative, and mixed research approaches. Sage.
Davies, A. (2013). Fifty years of language assessment, In A. J. Kunnan (Ed.), The companion to language assessment (pp. 3-21). Wiley.
Demir, S. B., & Pismek, N. (2018). A convergent parallel mixed methods study of controversial issues in social study classrooms.Educational Sciences: Theory and Practice, 18(1), 119-149. http://doi.org/10.12738/estp.2018.1.0298
Everitt, B. S., & Skrondal, A. (2010). The Cambridge dictionary of statistics. Cambridge University Press.
Gil, D., & Lucas, D. (2013). Using alternative assessment in business and foreign language classes. Journal of International Education Research, 9(4), 359-370. https://doi.org/10.19030/jier.v9i4.8088
Herdiawan, R. D. (2018). Authentic assessment in EFL teaching and learning. Eternal English Teaching Journal, 9(1), 1-10. https://doi.org/10.26877/eternal.v9i1.2406
Kabouha, R., & Elyas, T. (2015). Aligning teaching and assessment to course objectives: The case of preparatory year English program at King Abdulaziz University. International Journal of Applied and English Literature, 4(5), 83-91. https://doi.org/10.7575/aiac.ijalel.v.4n.5p.82
Kibbe, C. T. (2017). The history of communicative language teaching (CLT) and its use in the classroom. United States Military Academy during Academic Year 2017. https://www.usma.edu/cfe/Literature/Kibbe
Kothari (2004). Research methodology. New Age.
Kuang, Q. (2020). A review of the washback of English language tests on classroom teaching. English Language Teaching, 13(9), 10-17. https://doi.org/10.5539/elt.v13n9p10
Mekonnen, G. T. (2014). EFL classroom assessment: Teachers practice and teaching techniques adjustment in Ethiopia. Educational Research Reviews. 9(20), 1071-1089. https://doi.org/10.5897/ERR2014.1857
Minda, H. M. (2015). A study on the magnitude and causes of turnover among academic staffs of Ethiopian higher institutions: Evidence from Ambo University. Science, Technology and Arts Research Journal, 4(3), 285-293. https://doi.org/10.4314/star.v4i3.41
Minda, M. H. (2019). A study on the practices of continuous assessment in communicative English skills: The case of Ambo University.Research Journal of English Language and Literature, 7(4), 54-69. https://www.doi.org/10.33329/rj3lql.74.70
Ministry of Education (2017). Higher education inspective evaluation results feedback report presented to Ambo University. Ethiopia.
Molla, T. (2018). Higher education in Ethiopia: Structural inequalities and policy responses. Springer.
Monib, W. K., Karimi, A. Q., & Nijat, N. (2020). Effects of alternative assessment in EFL classroom: A systematic review. American International Journal of Education and Linguistics Research, 3(2), 7-18. https://doi.org/10.46545/aijelr.v3i2.195
Motuma, H. (2014). An assessment of the students’ reading ability in higher institutions: The case of Asela College of Teacher Education.Journal of Science and Sustainable Development, 1(2), 76-95.
Motuma, H. (2015). The changing role of the university TEFL teacher in improving the students’ communicative English Skills: From passive technician to reflective practitioner. Journal of Science and Sustainable Development, 3(1), 133-155.
Motuma, H. (2018). Content validity of EFL teacher-made assessment: The cause of Communicative English skills Course at Ambo University. East African Journal of Social Sciences and Humanities, 3(1), 41-62.
Nasab, F. G. (2015). Alternative versus traditional assessment. Journal of Applied Linguistics and Language Research, 2(6), 165-178. http://www.jallr.com/index.php/JALLR/article/download/136/pdf136
Palacio, M., Gaviria, S., & Brown, J. D. (2016). Aligning English language testing with curriculum. Profile Issues in Teachers’ Professional Development, 18(2), 63-77. https://doi.org/10.15446/profile.v18n2.53302
Palinkas, L. A., Howitz, S. M., Green, G. A., Wisdom, J. P., Duan, N., & Hoagwood, K. (2015). Purposeful sampling for qualitative data collection and analysis in mixed method implementation research. Administration and Policy in Mental Health, 42(5), 533-544. https://doi.org/10.1007/s10488-013-0528-y
Phuong, H. Y., & Nguyen, T. V. S. (2019). Using alternative assessment to improve EFL learners’ learning achievement: From theory to practice. Nova Science.
Rojas Serrano, J. R. (2017). Making sense of alternative assessment in a qualitative evaluation system. Profile Issues in Teachers’ Professional Development, 19(2), 73-85. https://doi.org/10.15446/profile.v19n2.57178
Shannon-Baker, P. (2016). Making paradigms meaningful in mixed methods research. Journal of Mixed Methods Research, 10(4), 319-334. https://doi.org/10.1177%2F1558689815575861
Shrestha, P. (2014). Alternative assessment approaches in primary English language classrooms. Journal of NELTA, 18(1-2), 148-163. https://doi.org/10.3126/nelta.v18i1-2.10337
Smith, M. K. (2000). What is curriculum? Exploring theory and practice. The Encyclopaedia of Informal Education. http://infed.org/mobi/curriculum-theory-and-practice
Teelken, C. (2018). Teaching assessment and perceived quality of teaching: A longitudinal study among academics in three European countries. European Journal of Higher Education, 8(4), 282-399. https://doi.org/10.1080/21568235.2018.1490661
Teferra, T. Asgedom, A., Oumer, J., W/hanna, T. AkliluDalelo, & Assefa, Berhannu. (2018). Ethiopian education development roadmap (2018–30): An integrated executive summary, Ministry of Education Strategy Centre (ESC), Ethiopia. https://planipolis.iiep.unesco.org/sites/default/files/ressources/ethiopia_education_development_roadmap_2018-2030.pdf
Temesgen, T. (2017). Assessing the practices of one one-to-five cooperative language learning of grade nine students EFL speaking classes. Ethiopian Journal of Language, Culture and Communication, 2(2), 1-21.
Tran, T. H. (2012). Second language assessment for classroom teachers [Presentation]. MIDTESOL, 2021, Ames, Iowa. http://files.eric.ed.gov/fulltext/ED545800.pdf
Ur, P. (2010). A course in language teaching: Practice and theory. Cambridge University Press.
Wubshet, H., & Menuta, F. (2015). Investigating the practices of alternative assessment in English classrooms: The cases of selected grade nine English teachers’ assessment practices. International Journal of Scientific Research in Education, 8(4), 159-171. http://ijsre.com.ng/assets/vol.%2c-8(4)-wubshet
menuta.pdf