Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
International Journal of Applied Linguistics & English Literature ISSN 2200-3592 (Print), ISSN 2200-3452 (Online) Vol. 2 No. 3; May 2013 Copyright © Australian International Academic Centre, Australia The Effect of Peer Assessment on Oral Presentation in an EFL Context Saedeh Ahangari (Ph.D.) Department of English Language, Tabriz Branch, Islamic Azad University, Tabriz, Iran. Behzad Rassekh-Alqol (Corresponding Author) Department of English Language, Tabriz Branch, Islamic Azad University, Tabriz, Iran. Tel: +98-914-867-9761 E-mail: Behzad_rassekh@yahoo.com Leila Ali Akbari Hamed Member of Scientific Association of English, Tabriz branch, Islamic Azad University, Tabriz, Iran Received: 11-02-2013 Accepted: 16-03-2013 Published: 01-05-2013 doi:10.7575/aiac.ijalel.v.2n.3p.45 URL: http://dx.doi.org/10.7575/aiac.ijalel.v.2n.3p.45 Abstract The purpose of this study was to examine the effect of peer assessment on oral presentation of Iranian English nonmajor students. To this end, 52 students of Azad and State universities of Tabriz were selected as participants. Their ages ranged from 25 to 38. The peer assessment was incorporated into the experimental group’s course to explore whether and to what extent their oral presentation skills may enhance. Data were obtained through a Likert scale questionnaire of peer assessment. Data were analyzed using SPSS Software. The results specified a statistically significant difference among the groups. The findings of the present study also suggest that, when assessment criteria are definitely established, peer assessment empowers students to evaluate the performance of their peers in a manner comparable to those of the teachers. Keywords: Peer assessment, oral presentation, ELT, feedback, EFL 1. Introduction An essential feature of education is assessment and the significance and popularity of student-oriented learning demand alternative techniques of assessment to evaluate teaching and learning. A large number of novel approaches of assessment have hence been suggested which meant to develop the integration of learning and assessment by enhancing the engagement of students in the assessment tasks (Sluijsmans, Brand-Gruwel, vanMerrienboer, & Bastiens, 2003). According to these approaches, learners are regarded as active agents who have equal responsibilities, collaborate, think, and carry out a continuous conversation with the teacher and/or peers. Furthermore, assessment does not totally grade students’ activities, rather it is applied to check out learners’ achievement and promote their learning tasks by involving them in the assessment tasks. Learner engagement in the process of assessment has been enormously rising in tertiary educational settings in various fields such as medicine, science, engineering, business and writing (Falchikov, 1995a; Freeman, 1995). Not long ago, studies have proposed the use of assessment as a learning method (Purchase, 2000; Rust, Price, & O'Donovan, 2003; Smyth, 2004). Few studies of peer assessment appear to have involved EFL students being required to evaluate their peers' proficiency of spoken English, and to judge peers’ assessment of oral production contrasting nonlinguistic criteria in both written and oral tasks. The researchers of this study thought that in an EFL context, the inclusion of peer assessment of oral presentation skills would benefit students by enhancing their understanding of the type and diversity of errors made and this would also serve to indicate the potential obstacles to communicate effectively which are posed by these errors. If students could be successfully trained to judge their peers' oral presentation, this would have a positive impact on selfcorrection. The present study sought to add to the growing body of knowledge relating to the application of peer assessment in EFL contexts. 2. Review of Literature A technical shift which has recently taken place in tertiary education has been from a teacher-centered perspective to a perspective in which student engagement is more central (Boud, 1995). Peer assessment is the process in which students assess their peer(s) (Falchikov, 1995b; Freeman, 1995). Eisner (in Boud, 1995) acknowledged the characteristics of the new assessment in education as follows: • Assessment tasks need to reveal the tasks that learners will face in the settings outside schools, not just those restricted to the schools themselves; • Assessment tasks should not be constrained to the solutions that students articulate, but also demonstrate how students perform solving a problem; IJALEL 2 (3):45-53, 2013 46 • Assessment tasks should mirror the values of the intellectual society from which tasks are originated; • Assessment tasks need not be confined to a single performance; • Assessment tasks should have more than one conventional solution to a problem and more than one acceptable response to a request; • Assessment tasks should have curricular application, but not be limited to the curriculum as taught; • Assessment tasks should allow the learner to select a form of picture that he or she chooses to display what has been learned (pp.93-95). Somervell (1993) found that peer assessment involves learners in making evaluations about other students' work or the performance. Whereas many similarities exist in the oral assessment activities which are described in literature, Joughin (1998) witnessed three specific sources of variation; the exclusion or inclusion of a written paper integrated with the oral presentation to constitute the overall score, individual or group presentations, and the person/s evaluating the assessments. In addition, he also recognized four particular advantages of oral assessment: developing good learning; authenticity; promoting student strengths; and fighting plagiarism. Sambell & McDowell (1998), for instance, examined the students' perceptions towards various aspects of novel assessment. Many students who were discussing innovative assessment believed that success is, more or less, affected by continuous application and hard work, and this achievement did not depend on sheer luck or a last-minute burst of effort. Most learners thought that clarity and honesty were essential requirements of a valid and fair assessment approach. Students were optimistic about the impacts of alternative assessment on their learning. The main advantages of peer assessment are that it develops evaluative and critical abilities, offers opportunities for skill development, integrates knowledge, and motivates student collaboration. Sometimes, there are weaker points in organizational matters, such as the cheating, time constraints and stress. Finally, there could be a misfit between marks without feedback and learning (McDowell, 1995). Boud (1992, 1995) presented an assessment plan to provide an analytical and comprehensive record of learning in settings where learners have heavy responsibility for their performance. The major instruction is a handout which provides the students with the headings. As effective tools, Assessment schedules are used to enable learners to reflect on their success, to arrange a wide variety of their learning, and to investigate the practical implications for further learning (Boud, 1992; Boud & Knights, 1994). It is suggested that getting to know the criteria of a rubric and observing the peers' practice and work, results in a profound understanding of the quality of one’s own performance (Falchikov, 1995a; Freeman, 1995). Self and peer assessment training might have noticeable influences on the development of the skills concerning the content, if the training is integrated in the current syllabus (Mehrens, Popham, & Ryan, 1998). From this perspective, the assessment skill is not drilled as a skill in isolation, but is closely connected to course content. Learners will acquire how to negotiate about criteria for an acceptable presentation if the training of defining criteria is embedded in a teacher's course on presentation skills. Understanding these criteria paves the way for the students to promote their own presentations, therefore, students’ development of their presentation skills will be supported by the assessment training. Cheng & Warren (2005) have found that students adopt highly favorable attitudes towards peer assessment activities, and that such task can develop self-confidence and intrinsic motivation in the learner (Falchikov & Goldfinch, 2000). Consequently, students actually get pleasure from judging each other’s practice, and derive substantial benefit from conducting this task. As a matter of fact, Topping (1998) claims that students assess their peers in a manner which is almost comparable to their instructors, which implies it is helpful for teachers to deeply think about applying peer assessment activities in classrooms. Yet, learners must be provided with the correct instruments to do the task, and need to be trained appropriately before carrying out any formal peer evaluation activities (Cheng & Warren, 2005). A few studies have been conducted in which learners assessed their peers in EFL contexts: In a research by Jafarpur (1991), Iranian students evaluated their peers' listening and speaking skills. He found low agreement between peer and teacher- assessment. Moreover, he noted that it was more difficult for less able students to self- or peer-assess when compared to more able ones. He focused on learner training and guidance in interpreting the scoring criteria. He maintained that if the EFL learners are assumed to shoulder the burden of responsibility in study programs, this will require some modifications in testing procedures. Miller and Ng (1994) led a study that engaged learners evaluating and allocating scores to their peers’ oral performance along with expressing their perceptions towards participating in peer assessment tasks. The findings demonstrated that students were able to actually judge their group members’ oral language skill under definite conditions, specifically if the group was homogeneous, the student assessors were highly proficient language learners, the students were engaged in the preparation of the test, they had former experience with each other’s oral language skill, and the testing environment was friendly (Miller and Ng, 1994). The outcomes also illustrated a comparatively high level of agreement between teacher- and peer assessments. On the subject of perceptions, the researchers observed that their learners, all together, took a hostile attitude towards peer assessment, citing causes such as the eccentricity of the task, the partiality of the task, the inequity of the whole practice, embarrassment in front of classmates, and feelings of being inexpert, unqualified and not proficient enough in English to evaluate their peers’ oral production. IJALEL 2 (3):45-53, 2013 47 Forde’s (1996) research investigated the stability with which Chinese speakers of English in Hong Kong, signed up in a training course for IELTS, and graded the oral skills of their peers. The study demonstrated that the majority of the students were able to perfectly judge their peers’ oral proficiency, when compared with the evaluations given by the IELTS examiner. Yamashiro (1999) carried out a detailed exploration of promoting public speaking skill among Japanese university students. He developed a validation scale and the validity of this rating scale based on peer, teacher, and self-rating was confirmed. Patri (2002) studied the conformity between teacher-, self- and peer assessment of students’ oral skills with and without the use of peer feedback. He noted that there was a higher correlation between teacher and peer assessments when the procedure was boosted by peer feedback, but most of proficient students were not able to assess each other’s performances in the same way as the teacher had done, irrespective of whether or not peer feedback was integrated into the practice. Cheng and Warren (2005) examined the reliability and potential advantages of integrating peer assessment into English language programs. They compared the learners’ perceptions towards evaluating both the English language proficiency and other features of their peers’ performance. In addition, they made a comparison between peer and teacher assessments. The results proposed that learners adopted a less optimistic attitude towards judging their peers’ language proficiency, but they did not rate their peers’ language proficiency very differently from the other rating criteria. Saito (2008) investigated the influences of training on peer assessment and comments received concerning oral productions in EFL classrooms in Japan. Analyses of comments indicated that the treatment groups outperformed in both quantity and quality of comments. Accompanied by a meta-analytic summary, it was maintained that peer assessment was a vigorous method wherein training on skill features might be sufficient to attain a definite level of correspondence with the criterion variable (teacher), but training may increase student comments and decrease mismatching ratings. The above-cited studies reveal that careful planning and good management as well as learner involvement in language programs can produce good results. The question now remains unanswered as to why it is not a more prevalent aspect in all EFL teaching programs. Part of the issue may address teacher indecision as to how learners will respond: will it be regarded as handing over the undertaking or as a positive step in learner engagement? In addition, as it was previously mentioned, it is comparatively easy to accomplish student involvement as a stage of a writing course, but what about speaking and listening? This study set out to deal with some of the above mentioned areas of concern. In this study, students in a speaking skills proficiency course were given the opportunity to assess each other’s oral presentation skills. 2.1. Research Questions and Null Hypotheses This study focused on how peer assessment could improve the oral performance of the learners in an EFL context. In attempting to answer this general question, this study specifically explored the following questions: 1. Does peer assessment improve the oral presentation skills of EFL learners? 2. Can peer assessment enable students to judge the performance of their peers in a manner comparable to those of the teachers? To probe the above questions, the following null hypotheses were formulated: H01: Peer assessment does not improve the oral presentation skills of EFL learners. H02: Peer assessment cannot enable students to judge the performance of their peers in a manner comparable to those of the teachers. 3. Method 3.1 Participants The participants in the study were 52 (18 male and 34 female) university students from the State and Azad universities in Eastern Azerbaijan, Tabriz. They were graduates of Science (41), Engineering (5), Accounting (2), and Physical Education (4). 36 participants had Masters Degrees and 16 were PhD students, and their ages ranged from 25 to 38. These students were selected according to their performances on the TOEFL (Sharpe, 2001). Out of a total of 70 students, 52 obtained 400 – 450 on the TOEFL (which roughly corresponds to Intermediate High on ACTFL scale) which are typical scores for English non-major students in Iran. One of the groups was randomly selected as the experimental group and the other one as the control group. An independent t-test on the mean scores of the two groups on the TOEFL confirmed the homogeneity of the groups. 3.2 Instrumentation The instrumentation used to collect data included: (1) In order to sample the subjects, the TOEFL test (Sharpe, 2001) was used for subject selection. (2) The topics for presentation were chosen from Barron’s TOEFL Handbook (Lougheed, 2004). The treatment was built into the researchers' syllabus. IJALEL 2 (3):45-53, 2013 48 (3) The instrumentation also included a peer assessment questionnaire (Appendix 1) adopted from Patri (2002). The reliability of the instrument was reported to be .70 (Saito, 2008). Fourteen questions were included in the questionnaire which was split into four categories. Category 1 comprised questions 1–6 that covered organization and content of presentation. Category 2 (questions 7–9) was created on the use of language. Category 3 (questions 10–12) was related to manner. Lastly, Category 4 contained questions 13 and 14 which were covering interaction with the audience. The rating was done on the basis of a 5-point Likert scale in which the categories were: 1 – poor; 2 – unsatisfactory; 3 – satisfactory; 4 – good; 5 – excellent. (4) The same questionnaire was used for teacher assessment with slight adjustments; changes were restricted to phrasing in the rubric. 3.3 Procedure To evaluate the effect of peer assessment, the control and experimental groups were set up, each consisting of 26 students. An independent t-test on the TOEFL (Sharpe, 2001) scores of the two groups confirmed the homogeneity of the groups (Table 1). Participants took part in a 28- hour course with two hours of instruction each week. The experimental group met on Sundays, while the control group’s classes were held on Tuesdays. The course curriculum included making oral presentations with a focus on organization and content of the talk, language use, manner and interaction with the audience. The experiment conducted was built into the teachers’ syllabus. In week 1, the teachers held an instruction session for both groups and elaborated on the essential aspects of a good oral presentation. These characteristicss (which covered the 14 items of the peer assessment questionnaire) were explained thoroughly in order to clearly establish the criteria set by the researchers. Items 8–14 – which referred to pronunciation, fluency, and speaker’s eye contact, confidence, verbal and non-verbal interaction – had to be explained in more detail. Then the researchers described the characteristics of a poor, an average, and a good presentation (Freeman, 1995). Both classes were then divided into halves. Half the class was given topic (a): ‘How do movies or television influence people’s behavior?’ and the other half was given topic (b): ‘Television has destroyed communication among friends and family.’ to prepare for their presentation in the following week. In week 2, the week after the training session, participants from both the control and experimental groups made their presentations. Half the class presented topic (a), and the other half topic (b). 3.4 Control group: Weeks 2-14 The participants in the control group presented their talks. After each presentation which took about three minutes, the researchers assessed their performance according to the teacher assessment questionnaire and explained the basis for his comments. Then they drew the students’ attention to the major themes of each presentation and asked them to carefully follow the rules of a good presentation. The participants were given two new topics to prepare for their presentations in the following weeks. 3.5 Experimental group: Weeks 2-6 Participants made their oral presentations on the two topics. The participants in the experimental group were divided into groups of five or six depending on the class size. Participants were assigned to different groups every week to encourage maximum interaction and to expose students to different points of view. Each group was asked to assess their peers. Group members were given an appropriate number (equal to the number of students making presentation in the class) of peer assessment questionnaires. Each student from the class made a three minute presentation on the topic assigned by the researchers. The researchers assessed the participants during the presentation using the teacherassessment form. During the presentation, peers noted their comments and filled in the questionnaires. The researchers spent about five minutes with each group discussing the comments made by the peers. The participants compared their assessments at the end of each performance. When there were differences, the researchers explained the basis for their judgments. For example, ‘this student has provided a lot of detail and the detail is appropriate and diverse. That is why I gave him a “5”.’ In week 2, only 8 students made their presentations and the others assessed their peers. The teacher spent more time with groups to clarify and explain the rating process item by item. In the following weeks, the number of students who managed to make their oral presentation was increasing. As the students got more familiar with the assessment criteria, the teacher spent less time to discuss the process of peer assessment. Hence, more and more students had the opportunity to be assessed by their peers under the teacher’s guidance. During 5 weeks, each student did 56 peer assessments and made 3 oral presentations. At the end of each session, the researchers collected the peer assessment questionnaires. The participants were given two new topics to prepare for their presentations in the following week. Weeks 2- 6 provided the experimental group with practice in peer assessment. 3.6 Experimental group: Weeks 7-14 After the session in week 6, the researchers did not make any comments. The whole process of rating was done by the students. The participants were given two new topics to prepare for their presentations in the following week. At the end of the course, the participants from both groups gave an oral presentation each. Their performance was assessed by the researchers. 49 IJALEL 2 (3):45-53, 2013 3.7 Data analysis Data obtained from the experimental group in week 6 were used to correlate teacher and peer assessments. Pearson correlation coefficient was calculated between teacher–peer (T–P) assessments for the mean peer rating of students with the teacher’s rating for each student. Also a one-sample t-test was run on the mean scores obtained through peer assessment and teacher assessment to examine to what extent peer assessment was efficient to enable students to make sound judgments of their peers’ oral performance. At the end of the treatment, an independent t-test was run on the final scores to explore the difference between the scores of the experimental and control groups. 4. Results In order to test the homogeneity of both groups, an independent t-test was run on their TOEFL scores. Table 1 shows that there was no significant difference (t=.o7, p<.o5) between control and experimental groups with respect to their TOEFL scores. Table 1. Independent T-Test Displaying the TOEFL Scores The data recorded in week 6 included teacher- and peer assessments (TA and PA) obtained from the experimental group. In Table 2 Pearson correlation coefficients were calculated. The high correlation between the experimental group students’ peer assessments and the teacher ratings (r=.87) which was significant at level (p<.01) shows that peer assessment was in high agreement with the teacher assessment. This suggests that the students were able to make judgments of their peers’ oral presentations comparable to those made by the teacher. Table 2. Correlation between teacher and peer assessment in week 6 A one-sample t-test was also run on the data obtained from the experimental group in week 6 to examine to what extent peer assessment was efficient to enable students to make sound judgments of their peers’ oral performance. The researchers randomly selected one of the participants (whose rating on teacher assessment was 3.42) and recorded how other students assessed the same student’s oral presentation. Tables 3a and 3b reassure that students are able to assess their peers reasonably. Table 3a. Descriptive statistics for Experimental Group’s Peer Assessments N PA 25 Mean Std. Deviation Std. Error Mean 3.6464 .50155 .10031 As it is observed peer assessment mean is 3.64 which is a little higher than 3.42 assessed by the teacher. The small amount of the SD (=.5) shows that the assessments done by the students are more congruent, and they have grasped the main points of peer assessment. 50 IJALEL 2 (3):45-53, 2013 Table 3b. One-Sample T-Test on Teacher and Peer Assessment Table 3b shows that the difference (p<.01) was not significant and thus there was no difference between teacher assessment and peer assessments. In order to find out whether the treatment had an effect on the oral performance of the experimental group, an independent t-test was carried out on the final scores of both groups. The results indicated that there was a significant difference between the ratings the students obtained at the level of p < .05 (see Table 4) suggesting that peer assessment had a significantly positive effect on oral presentation of students receiving the treatment. Table 4. Independent T-Test Displaying the Results of Posttest 5. Discussion The results of this research demonstrate that under certain circumstances students are able to convincingly assess each other’s oral language ability. As the level of our students correlate to TOFEL scores ranging from 400-450, we can say that they have an intermediate level of language ability. We can call our students ‘sophisticated language learners’ in that they were actively involved in pursuing their language studies further, and they demonstrated their ability to cope with a complex concept in a second language testing. To be able to assess their peers’ oral skills, students must be able to access some implicit or explicit criteria. The researchers in this study gave the students a set of assessing criteria. We believe that the students, being homogenous, have an internal set of rules leading the use - and wrong use- of English. These ‘rules’ are used whenever they hear English being spoken by another speaker, or when they themselves use English. Accordingly, the students were able to institute criteria of satisfactory language use. We do not know what the results would be if, say, an Iranian speaker of English were to try to assess a Spanish speaker of English. As Spaventa and Williamson (1989) point out: “...students who come from a Japanese background and students who come from a European background have different abilities in the traditional language skill areas”. In order for the learners to assess oral production fairly and reliably, they must either be aware of the assessees’ oral skills (Miller and Ng, 1994) or be trained in evaluating and have experience as assessors (Patri, 2002; Saito, 2008). By conducting the oral assessments half way through the course, we gave the students enough time (6 weeks) to become accustomed to their peers’ oral abilities before we asked them to assess each other. In this way, my students were better prepared for the task of judging their peers. The peers were familiar with each other and therefore the assessments were conducted under friendly circumstances. This ensured that the participants were able to perform to the best of their ability. Jones (1985, p.13) explains the importance of an examiner/ examinee relationship and concludes, “... that ease of elicitation in an oral interview is often affected by the social relationship between the examiners and the examinee. The factors can include age, race, social class, education and profession.” As all our students had similar social profiles, we feel this was an important factor in their ability to assess each other fairly. it can be concluded from the results that a) they illustrated an analogous level of assessment to that of the teacher; b) the students were open in participating in this peer assessment practice; c) the procedure did not lead to lowering the standards; and d) the students promoted their understanding of and perception towards assessment by involving in this study. For most second language learners, speaking itself is a multifaceted (and occasionally distressing) task since it necessitates them to focus concurrently on content, diction, pronunciation, body language, eye contact, and so on. This sophistication is limited to oral tasks compared to writing or listening tasks. With low-achieving earners, this could be even more demanding. It is essential to point out here that the majority of the participants in the present study had obtained 400 - 450 on TOEFL. Previous studies on peer assessments with EFL (English as a Foreign Language) students had found that it was difficult for them to make sound evaluations of their peers’ speaking and learning abilities (see Jafarpur, 1991). IJALEL 2 (3):45-53, 2013 51 The teacher and peer assessments correlations are all together substantially high. This observation lends strong support to Hypothesis 2 that peer assessment enhances learners’ ability to make judgments on their peers’ oral presentation skills comparable to those of the teacher. The participants in this study demonstrated that if the criteria rules were established definitely (Patri, 2002; Saito, 2008; Yamashiro, 1999), learners could rate their peers acceptably and improve their judgments so that they could acquire a better understanding of their own skills. This explains why the experimental group outperformed the control group. 6. Conclusions and Implications Peer assessment recently has been regularly practiced as an alternative assessment technique, predominantly in higher education. In the method of peer assessment, learners are able to witness their peers and arrive at a better understanding of how their peers learn. Sluijsmans et al. (1999) have reviewed a body of research articles concerning peer assessment and have maintained that the process of peer assessment involves students in making judgments. In addition to rating or scoring, peer assessment is also an integral segment of learning experiences. During the course of the assessment, students learn to shoulder high levels of responsibility and to concentrate on learning itself. Peer evaluation also offers the learners a situation in which they can perceive the role of their teachers and appreciate the nature of assessment. It supports students to learn about learning and, consequently, increases students’ metacognitive understanding about their own learning. This research investigated the impact of peer assessment on learners’ oral presentation improvement. The results of this study indicated that when assessment criteria are clearly set, peer assessment will enable students to make judgments of their peers comparable to those of the teacher. Considering that both the control and experimental groups consisted of students with similar English proficiency – all students got 400 – 450 on TOEFL, and shared similar background – it can be assumed that peer assessment had assisted in achieving higher correlation with teacher assessments. If this, in fact, is the case, then teacher assessment could be complemented with peer assessment at a lower cost in the context of oral skills. If peers can be engaged in the task of assessment, teachers’ time could be exploited more efficiently on subjects associated with promoting their teaching methods. This research can provide comfort with language teaching practitioners to proceed with the use of peer assessment, which is formerly recommended by many researchers and is prevalent in many language teaching classes principally in the south-east Asian countries. The findings of this study may have implications for SLA theory as well as for material development and syllabus designing. References Boud, D. (1992). The use of self-assessment schedules in negotiated learning. Studies in Higher Education, 17,185– 200. Boud, D. (1995). Enhancing learning through self-assessment. London: Kogan Page. Boud, D., & Knights, S. (1994). Designing courses to promote reflective practice. Research and Development in Higher Education, 16, 229–234. Cheng, W., & Warren, M. (1997). Having second thoughts: student perceptions before and after a peer assessment exercise. Studies in Higher Education, 22, 233–239. Falchikov, N. (1995a). Peer feedback marking: development peer assessment. Innovations in Education and Training International, 32, 175 – 87. Falchikov, N. (1995b). Improving feedback to and from students. In Knight, P. (Ed.). Assessment for Learning in Higher Education. London: Kogan Page, 157 – 66. Falchikov, N., & Goldfinch, J. (2000). Student peer assessment in higher education: A meta-analysis comparing peer and teacher marks. Review of Educational Research, 70, 287–322. Forde, K. (1996). The effects of gender and proficiency on oral self- and peer-assessments. English Language Studies Working Papers, City University of Hong Kong, 1, 34 – 47. Freeman, M. (1995). Peer assessment by groups of group work. Assessment and Evaluation in Higher Education, 20, 289 – 300. Jafarpur, A. (1991). Can naive EFL learners estimate their own proficiency? Evaluation and Research in Education, 5, 145 – 57. Jones, N. (1995). Business writing, Chinese students and communicative language teaching. TESOL Journal, 4, 12 – 15. Joughin, G. (1998). Dimensions of oral assessment. Assessment & Evaluation in Higher Education, 23, 367-378. Lougheed, L. (2004). Barron’s Computer-Based TOEFL Essays. NY: Barron’s Educational Series, Inc. McDowell, L. (1995). The impact of innovative assessment on student learning. Innovations in Education and Training International, 32, 302–313. Mehrens, W.A., Popham, W.J., & Ryan, J.M. (1998). How to prepare students for performance assessments. Educational Measurement: Issues and Practice, 17, 18-22. 52 IJALEL 2 (3):45-53, 2013 Miller, L. and Ng, R. (1994). Peer assessment of oral language proficiency. Perspectives: working papers of the department of English, City Polytechnic of Hong Kong, 6, 41 – 56. Patri, M. (2002). The influence of peer feedback on self- and peer-assessment of oral skills. Language Testing, 19, 109 – 31. Purchase, H. C. (2000). Learning about interface design through peer assessment. Assessment and Evaluation in Higher Education, 27, 341-352. Rust, C., Price, M., & O'Donovan, B. (2003). Improving students' learning by developing their understanding of assessment criteria and processes. Assessment and Evaluation in Higher Education, 28, 147-164. Sambell, K., & McDowell, L. (1998). The value of self and peer assessment to the developing lifelong learner. In C. Rust (Ed.), Improving student learning — improving students as learners. Oxford, UK: Oxford Centre for Staff and Learning Development. Saito, H. (2008). EFL classroom peer assessment: Training effects on rating and commenting. Language Testing, 25. 553. Sharpe, P. J. (2001). Barron’s TOEFL Test of English as a Foreign Language (10th ed.). NY: Barron’s Educational Series, Inc. Sluijsmans, D. M. A., Brand-Gruwel, S., vanMerrienboer, J. J. G., & Bastiens, T. J. (2003). The training of peer assessment skills to promote the development of reflection skills in teacher education. Studies in Educational Evaluation, 29, 23-42. Smyth, K. (2004). The benefits of students learning about critical evaluation rather than being summatively judged. Assessment and Evaluation in Higher Education, 29, 369-378. Spaventa, L. J. & Williamson, J. S. (1989). Participatory placement: A case study. In M. Pennington (Ed.), Building Better English Language Programs. NAFSA: Washington, D.C. Topping, K. (1998). Peer assessment between students in colleges and universities. Review of Educational Research, 68, 249 – 76. Underhill, N. (1987). Testing Spoken Language. Cambridge: Cambridge University Press. Yamashiro, A. (1999). Using structural equation modeling to validate a rating scale. Presented at the 21st Language Testing Research Colloquium, Tsukuba, Japan. Appendix 1 Peer assessment questionnaire Name: Topic: Date: Rate your classmate by using the scale: Poor Unsatisfactory 1 Satisfactory 2 3 Good 4 Excellent 5 A. Introduction 1. Topic sentence – suitable? 2. Topic sentence – fascinating? 3. The opinion on the issue – well-argued? 1 2 1 3 2 4 3 5 4 5 1 2 3 4 5 4. Details supporting the main points – enough? 1 2 3 4 5 5. Details supporting the main points – related? 1 2 3 4 5 1 2 3 4 5 1 2 3 4 5 B. Body C. Conclusion 6. The main points – brief? D. Language Use 7. Grammar – precise? 8. Fluency 1 2 3 4 5 9. Pronunciation – words noticeably pronounced? 1 2 3 4 5 E. Manner 10. Confidence (not nervous) 1 2 3 4 5 53 IJALEL 2 (3):45-53, 2013 11. Confidence (looked very little at his/her notes) 1 2 3 4 5 12. Eye contact 1 2 3 4 5 F. Interaction 13. Non-verbal interaction with the audience (facial 1 2 3 4 5 4 5 expressions, gestures) 14. Verbal interaction (involving the audience by 1 2 3 asking questions and encouraging them to answer) Appendix 2 Teacher assessment questionnaire Name: Topic: Date: Rate the student by using the scale: Poor Unsatisfactory 1 Satisfactory 2 3 Good Excellent 4 5 A. Introduction 1. Topic sentence – appropriate? 1 2 3 4 5 2. Topic sentence – interesting? 1 2 3 4 5 3. The opinion on the issue – clearly stated? 1 2 3 4 5 4. Details supporting the main points – sufficient? 1 2 3 4 5 5. Details supporting the main points – relevant? 1 2 3 4 5 1 2 3 4 5 7. Grammar – accurate? 1 2 3 4 5 8. Fluency 1 2 3 4 5 B. Body C. Conclusion 6. The main points – summarized? D. Language Use 9. Pronunciation – words clearly pronounced? 1 2 3 4 5 E. Manner 10. Confidence (not nervous) 11. Confidence (depended very little on his/her notes) 1 1 12. Eye contact 2 2 1 3 4 3 2 5 4 3 5 4 5 F. Interaction 13. Non-verbal interaction with the audience (facial 1 2 3 1 2 3 4 5 expressions, gestures) 14. Verbal interaction (involving the audience during the talk by asking questions and encouraging them to respond) 4 5