Academia.eduAcademia.edu

Teachers’ perceptions of student evaluations of teaching

2014, Educational Research and Evaluation

https://doi.org/10.1080/13803611.2014.932698

Evaluation of teaching in higher education has drawn much attention due to the need for greater accountability and improvement in student learning. Our review of literature on Student Evaluation of Teaching (SET) surveys suggests that considerable controversy and criticism have surrounded its use, fairness, and validity. Yet, many universities in Hong Kong are depending primarily on student feedback in SET surveys to provide evidence for their improvement of teaching and learning, and for their decisions made on teachers' pay-rise and tenure. If universities are to continue to use SET surveys as the main source of mechanism for measuring teaching effectiveness, teachers will have to understand and be aware of its purposes. But how much do we know about teachers' perceptions of teaching and learning and, consequently, their opinions of using SET surveys as the primary standardized evaluation mechanism of teaching for promotion and tenure and for teaching effectiveness?

This article was downloaded by: [University of Hong Kong Libraries] On: 25 June 2014, At: 18:12 Publisher: Routledge Informa Ltd Registered in England and Wales Registered Number: 1072954 Registered office: Mortimer House, 37-41 Mortimer Street, London W1T 3JH, UK Educational Research and Evaluation: An International Journal on Theory and Practice Publication details, including instructions for authors and subscription information: http://www.tandfonline.com/loi/nere20 Teachers’ perceptions of student evaluations of teaching a a a Cecilia K.Y. Chan , Lillian Y.Y. Luk & Min Zeng a Centre for the Enhancement of Teaching and Learning, The University of Hong Kong, Hong Kong Published online: 23 Jun 2014. To cite this article: Cecilia K.Y. Chan, Lillian Y.Y. Luk & Min Zeng (2014): Teachers’ perceptions of student evaluations of teaching, Educational Research and Evaluation: An International Journal on Theory and Practice, DOI: 10.1080/13803611.2014.932698 To link to this article: http://dx.doi.org/10.1080/13803611.2014.932698 PLEASE SCROLL DOWN FOR ARTICLE Taylor & Francis makes every effort to ensure the accuracy of all the information (the “Content”) contained in the publications on our platform. However, Taylor & Francis, our agents, and our licensors make no representations or warranties whatsoever as to the accuracy, completeness, or suitability for any purpose of the Content. Any opinions and views expressed in this publication are the opinions and views of the authors, and are not the views of or endorsed by Taylor & Francis. The accuracy of the Content should not be relied upon and should be independently verified with primary sources of information. Taylor and Francis shall not be liable for any losses, actions, claims, proceedings, demands, costs, expenses, damages, and other liabilities whatsoever or howsoever caused arising directly or indirectly in connection with, in relation to or arising out of the use of the Content. This article may be used for research, teaching, and private study purposes. Any substantial or systematic reproduction, redistribution, reselling, loan, sub-licensing, systematic supply, or distribution in any form to anyone is expressly forbidden. Terms & Conditions of access and use can be found at http://www.tandfonline.com/page/termsand-conditions Educational Research and Evaluation, 2014 http://dx.doi.org/10.1080/13803611.2014.932698 Teachers’ perceptions of student evaluations of teaching Cecilia K.Y. Chan*, Lillian Y.Y. Luk and Min Zeng Downloaded by [University of Hong Kong Libraries] at 18:12 25 June 2014 Centre for the Enhancement of Teaching and Learning, The University of Hong Kong, Hong Kong (Received 21 January 2014; final version received 26 May 2014) Evaluation of teaching in higher education has drawn much attention due to the need for greater accountability and improvement in student learning. Our review of literature on Student Evaluation of Teaching (SET) surveys suggests that considerable controversy and criticism have surrounded its use, fairness, and validity. Yet, many universities in Hong Kong are depending primarily on student feedback in SET surveys to provide evidence for their improvement of teaching and learning, and for their decisions made on teachers’ pay-rise and tenure. If universities are to continue to use SET surveys as the main source of mechanism for measuring teaching effectiveness, teachers will have to understand and be aware of its purposes. But how much do we know about teachers’ perceptions of teaching and learning and, consequently, their opinions of using SET surveys as the primary standardized evaluation mechanism of teaching for promotion and tenure and for teaching effectiveness? Keywords: student evaluation; teaching effectiveness; higher education; SET survey; teacher perception of teaching and learning Introduction The evaluation of teaching effectiveness has become a major issue in higher education in recent years, due to the need for greater accountability and improvement in the quality of teaching and, more important, student learning (Leckey & Neill, 2001). However, the quality of teaching has not led to an improvement in academic quality but a degree of performance in peer reviews or dramaturgical compliance (Barrow, 1999) – an “implementation gap” between policies on quality and actual outcomes as described by Newton (2000). There are various mechanisms to evaluate teaching effectiveness, such as online questionnaires, class observation, peer review, surveys, and analysis of student drop-out rates (Slade & McConville, 2006). Student feedback using Student Evaluation of Teaching (SET) surveys is one of the several mechanisms commonly used to measure and improve teaching and learning. The survey usually employs the use of questionnaire items to evaluate teacher effectiveness and various areas of the course. Many universities in Hong Kong have been found to be overly reliant on SET surveys (Kember, Leung, & Kwan, 2002) to gather evaluation information for their course improvement. In particular, these universities tend to rely on SET survey results to provide benchmarking for teaching evaluation on *Corresponding author. Email: [email protected] © 2014 Taylor & Francis Downloaded by [University of Hong Kong Libraries] at 18:12 25 June 2014 2 C.K.Y. Chan et al. promotion and tenure decision-making in the absence of any standardized mechanism to evaluate good teaching (particularly where research is still viewed as the most important factor for promotion and tenure in research-intensive universities). Researchers such as Surgenor (2013) and Wachtel (1998) have suggested that faculty members have both positive and negative attitudes towards the purpose, validity, usefulness, and consequences of SET. Despite the inconsistencies in opinions towards the use of SET surveys, students’ rating of teaching effectiveness has long been playing a significant role in the evaluation of teaching in higher education institutions in Hong Kong. The present study examines science teachers’ perceptions of teaching evaluation using a SET survey mechanism at a research-intensive university in Hong Kong. In particular, the science discipline of this Hong Kong research-intensive university is of interest as this university is consistently ranked among the top-five institutions in Asia, and its science faculty has a large number of professors who have conducted world-renowned discipline research. For this university, the teaching and learning culture was not evident as teaching was not previously rewarded and there was no structural agreement on the evaluation criteria for teaching. In fact, only in recent years have a teaching and learning philosophy and university educational aims emerged. The SET instrument which the university has adopted incorporates different teaching and learning aspects including: (a) student learning information – the overall evaluation of the course including the use of English as the Medium of instruction; (b) the teacher – the overall evaluation of the respective teacher (s) who has/have delivered the main part of the course; and (c) the pedagogy – the evaluation of the delivery of the course including lecture/tutorial/laboratory in order to achieve the intended learning outcomes. Given the strong research culture of the university and its faculty, the aim of this research was to determine teachers’ perceptions of teaching and learning and, consequently, their opinions of using SET surveys as the primary and standardized mechanism for evaluating the effectiveness of teaching and learning, and for promotion and tenure. This study will also contribute to the knowledge about teachers’ interpretations of SET survey results in the Asian context and how they perceive the effectiveness and use of SET surveys. Such knowledge would be valuable for helping faculty members and policy makers evaluate, construct, and refine their current practices of SET surveys and other evaluation mechanisms for teaching and learning. Review of literature Student Evaluation of Teaching (SET) survey Reviews of the existing literature (e.g., Aleamoni, 1999; Marsh, 2007; Spooren, Brockx, & Mortelmans, 2013; Wachtel, 1998) have revealed inconsistent findings leading to an ongoing debate on the reliability and validity of SET surveys. Arguments against using SET surveys often raise issues related to (a) the definition of effective teaching; (b) other variables influencing student ratings, which may not be related to teaching ability; and (c) negative reactions from faculty members (Anderson, 2006; Scott, Stone, & Dinham, 2000). First, teaching effectiveness is a multifaceted construct (Marsh, 1984). Marsh (1984), Braskamp and Ory (1994), and Centra (1993), for example, identified six major components of effective teaching that are often included in SET surveys: course planning, communication skills, teacher–student interaction, course difficulty, grading, and student self-rated learning. However, there continues to be an absence of agreement on the definition of effective teaching (Casey, Gentile, & Bigger, 1997; Stehle, Spinath, & Downloaded by [University of Hong Kong Libraries] at 18:12 25 June 2014 Educational Research and Evaluation 3 Kadmon, 2012). Murphy, MacLaren, and Flynn (2009) have pointed out that it is difficult to measure quality of teaching because of its complex nature. Second, student ratings have been found to be influenced by other factors such as course characteristics, student characteristics, and teacher characteristics (Wachtel, 1998). For example, Marsh (1980, 2007) reported that students with a greater prior interest in the subject and students who were expecting higher grades tended to give higher teacher ratings. Apart from prior subject interest and expected grades, students’ perceptions of their teachers’ personality have been found to be related to their assessment of teacher effectiveness (Clayson, 2013; Tomasco, 1980). For instance, Shevlin, Banyard, Davies, and Griffiths (2000) identified “charisma” as a trait which influences students’ rating of their teacher. In a subsequent study conducted by Adamson, O’Kane, and Shevlin (2005), it was found that students’ perceptions of how entertaining a teacher was positively related to their ratings of teaching effectiveness. In fact, researchers such as Moore and Kuol (2005) doubted students’ ability to assess teacher performance due to their limited knowledge on teaching, and they expressed their concern towards student evaluation that is based on personal feelings and expectations, leading to students’ inability to make an objective judgement. Third, some teachers may try to get more favourable ratings by increasing students’ grades and lowering the standard of the course, causing grade inflation (Simpson & Siguaw, 2000). Thus, some teachers argue that SET surveys fail to serve their educational purpose and even violate the academic freedom and rights of teachers (Slade & McConville, 2006). Despite the criticisms directed towards using SET surveys, supporters of its use argue that effective use of feedback data (Hendry, Lyon, & Henderson-Smart, 2007; Kember et al., 2002) does provide useful information for teachers which helps improve student learning (Ballantyne, Borthwick, & Packer, 2000). A recent study conducted by Winchester and Winchester (2014) suggests that the use of weekly formative SET surveys helps teachers improve on their teaching through reflective practice, which in turn increases their summative SET score towards the end of the academic year. Aleamoni (1999) clarified 16 myths regarding SET surveys and proposed that students’ feedback does provide accurate information when data are collected using well-designed instruments. Lemos, Queiros, Teixeira, and Menezes (2011) demonstrated the reliability and validity of a students’ evaluation of teaching questionnaire, which was developed based on learning theory. Contrary to the findings in previous studies, Beran and Violato (2005) found that course and student characteristics have little influence on the rating of teachers and suggested that the ratings do measure teaching performance. It is also interesting to note that an analysis of more than 30,000 open-ended SET comments from students in a recent study conducted by Tucker (2014) revealed that the majority of the students provided meaningful feedback which was neither abusive nor offensive. However, inconsistencies across the different contexts of studies and research findings make it difficult for researchers and practitioners to agree on the reliability and validity of SET surveys. Although extensive research has been carried out on students’ feedback on the effectiveness of teaching, limited studies have examined teachers’ perceptions of SET surveys (Arthur, 2009). Teachers’ perceptions of the SET survey A review of the literature reveals that there is a lack of consensus in teachers’ attitudes towards the use of SET surveys (Beran & Rokosh, 2009; Nasser & Fresko, 2002). A national project conducted in New Zealand by Stein and colleagues (2012) and a study conducted by Surgenor (2013) at an Irish university have revealed that academics show both Downloaded by [University of Hong Kong Libraries] at 18:12 25 June 2014 4 C.K.Y. Chan et al. positive and negative attitudes towards the purpose, validity, usefulness, and consequences of the surveys. With regard to its purposes, teachers in these two studies mentioned that the surveys provided them with feedback that would help them to identify their own strengths and weaknesses in teaching. However, teachers in Surgenor’s (2013) study were concerned about the possible use of the SET survey as a means of surveillance, monitoring their performance for future reward or penalty. While more opportunities than challenges were identified with regard to the purpose of SET surveys, more barriers than opportunities have been identified with regard to the validity, usefulness, and consequences of SET (Surgenor, 2013). In terms of validity, some teachers in Surgenor’s (2013) study expressed concern towards sampling bias, which may result in questionable comments: extremely positive comments from dedicated students and extremely negative comments from discontented students. These teachers also noted that the survey may involve comments from both students who attend classes regularly and students who do not. Stein et al. (2012) found that some teachers were sceptical about the reliability of students’ feedback as they believe that students tend to be biased towards popular teachers. Despite doubts about the validity and reliability of SET surveys, a small proportion of participants in Surgenor’s (2013) study suggested that using the survey as a uniform evaluation tool across different faculties will raise the level of objectivity in the evaluation of teaching. In terms of the usefulness of the survey data, teachers were aware of the possible impact of extraneous factors on the SET survey ratings, and of the possible ways to increase their scores (e.g., increasing grades and lowering course standards). This finding coincides with that of Arthur’s (2009) study, in which teachers interviewed revealed those factors they believed to influence teaching evaluations including relationships between teacher and students, student expectations, course type, and students’ course experience. In terms of the consequences of SET surveys, teachers in Surgenor’s (2013) study acknowledged developmental opportunities that may emerge as a result of using SET surveys. For instance, they believed that using the survey will increase awareness of evaluation, which will provide them with an overview of their teaching and promote a reflective culture in teaching practice. However, they cast doubt on the degree of confidentiality and the use of data. Concerns about the confidentiality and the use of data were also reported in other studies (e.g., Beran & Rokosh, 2009; Nasser & Fresko, 2002). For example, there seems to be a general agreement that teachers should receive students’ feedback on their teaching (Nasser & Fresko, 2002) and that such data should not be made available to other staff members, students, or the public (Beran & Rokosh, 2009). On the other hand, there are conflicting views on the usage of the survey scores. Teachers interviewed by Balam and Shannon (2010) reported that the survey was used to improve their teaching; whereas in Nasser and Fresko’s (2002) study, only a few teachers mentioned that they made changes to their teaching in response to the SET data. Arthur (2009) found that whether or not teachers make any changes or improvements in response to the student feedback depends on a number of factors (e.g., the number of students reporting the same experience, teachers’ perceived importance of teaching or relevance of a certain topic, and the teaching/ research culture of the university). Arthur suggested that the decision to make any changes or improvements in teaching is dependent on teachers’ interpretations of the SET data; thus, this study aimed to understand teachers’ perceptions of SET surveys will provide information on how learning can be enhanced. Overall, teachers seem to have both positive and negative attitudes towards SET surveys. Nasser and Fresko (2002) point out that teachers who receive higher ratings Educational Research and Evaluation 5 Downloaded by [University of Hong Kong Libraries] at 18:12 25 June 2014 tend to display stronger beliefs and support for using the survey. According to Kwan (1999), science and engineering teachers tend to receive significantly lower ratings than language and business teachers do; therefore, teachers in different academic disciplines may differ in their attitudes towards SET surveys. Given the disciplinary differences, teachers’ perceived importance of teaching, and the teaching and research culture of the university, the present study aimed to investigate university science teachers’ perceptions of the use of a SET survey as the primary standardized evaluation mechanism of teaching and learning for promotion and tenure and for teaching and learning effectiveness. Methodology The objectives of this study are to: (1) investigate teachers’ perceptions to teaching and learning in the science discipline at a research-intensive university in Hong Kong; (2) analyse teachers’ current perceptions and opinions in the science discipline towards the use of a SET survey as the primary standardized evaluation mechanism of teaching and learning for promotion and tenure and for teaching and learning effectiveness at a research-intensive Hong Kong University. Quantitative and qualitative data were collected using questionnaires from teachers in the Science discipline at a research-intensive university in Hong Kong. The questionnaire included 17 close-ended items and one open-ended question. Responses to 13 close-ended items were elicited using a 5-point categorical scale (from strongly agree to strongly disagree), and 4 close-ended items were elicited using a “yes-and-no” scale. With the closed-ended items, teachers were asked about their perceived importance of teaching versus research, and their interest in teaching versus research. Related questions such as teachers’ awareness of the university’s teaching and learning support centre and their academic development activities were also included. Good teaching practices for student learning such as out-of-class contact with students, placing science teaching in a societal context, and enhancing course material in response to feedback were also incorporated into the questionnaire, to identify if the teachers realized good and effective teaching which reflected on teachers’ perceptions of teaching and learning. Only one general open-ended question was proposed, which was: “Comments (please feel free to tell us anything else you feel may be relevant to teaching and learning in the Faculty)”. The questionnaire was initially piloted with five science teachers. Certain items were revised based on the pilot to provide a more straightforward understanding and to ensure the questions were more related to the objectives according to the teachers’ feedback. Subject-matter experts were also consulted to help improve the content validity of the questionnaire items. From the qualitative comments collected, we identified a large number of comments directly related to the SET. The study received ethical approval from the university, and participant anonymity was assured. An information sheet was provided to the teachers for their inform consent and to avoid any further conflicts of interest within the faculty. An outside researcher (researcher from the teaching and learning centre) was invited to conduct the research study in order to ensure that teachers would not feel uncomfortable or threatened if the study was conducted by the faculty management. Participants from the faculty were invited to complete the survey using a purposive sampling method. To ensure sufficient questionnaire participation, both online and paper Downloaded by [University of Hong Kong Libraries] at 18:12 25 June 2014 6 C.K.Y. Chan et al. versions were distributed: An email was sent to the teachers to invite them to complete an online version of the questionnaire; a paper version of the questionnaire was also delivered to their letter boxes at their department. Personalized follow-up emails were sent to teachers to improve the response rate, and two research assistants were employed to help distribute the paper questionnaires directly to the teachers at their offices. Sixty-two out of 95 science teachers at the Hong Kong university participated in the study (65% response rate). Demographic information about the participants is provided in Table 1. The questions and the results of the questionnaire are summarized in Table 2 and are grouped together based on Objectives 1 and 2. For the teachers’ perceptions of teaching and learning (i.e., Objective 1), the quantitative data are analysed by calculating the percentage of responses for each category. The percentage of agreement (or disagreement) on the questionnaire items was calculated by adding up the responses Strongly agree (or Strongly disagree) and Agree (or Disagree). For the perception of the SET (i.e., Objective 2), in addition to the percentage of responses, the Spearman’s rank correlation coefficient was also calculated to investigate correlations between the different factors that influence teachers’ perceptions of SET. With the use of the cutting and sorting technique (Ryan & Bernard, 2003), key words and phrases in the qualitative data were identified and coded to reflect key concepts in the teachers’ comments. The codes were then sorted into categories (i.e., themes), which were named “perception of teaching versus research”, “perception of SET”, “alternative mechanism for teaching evaluation”, “curriculum reform for student learning”, and “challenges using English as a medium of instruction”. The themes from comments were compared with the quantitative data to determine any patterns (differences/similarities) in findings. Results Analysis of the quantitative data suggests that teachers had diverse attitudes towards teaching and research. Table 2 shows that 44% of the participants indicated that they enjoyed research more than teaching. The perceived time spent on research varies among participants, such that 37% indicated that they spent more time on research than teaching, 36% disagreed, and 27% held a neutral opinion. Despite participants’ diverse opinions in terms of their perceived time spent on research, most of them (56%) believed that research is a priority due to the performance review and tenure track system, and 44.3% believed that teaching is of less importance than research in their performance review. Although a generally positive attitude towards research was observed among the participants, most of them (67%) indicated that they had not engaged in any research activity related to teaching and learning. In fact, Table 1. Demographic information of participants. Background information Position Professor Associate Professor Assistant Professor Teaching Consultant Status Tenured Non-tenured % 19.7 39.3 26.2 14.8 52.5 47.5 Background information % Department Biological Sciences Chemistry Earth Sciences Mathematics Physics Statistics and Actuarial Science 19.7 9.8 19.7 18.0 14.8 18.0 Educational Research and Evaluation 7 Table 2. Teachers’ perception of teaching and research. Downloaded by [University of Hong Kong Libraries] at 18:12 25 June 2014 Frequency (%) Items on teachers’ perception on teaching & learning n I “enjoy” research more than I “enjoy” teaching. I spend more time on research than teaching. Due to the performance and tenure-track system, research is a priority for me. In my performance review, teaching is regarded less important than research. I am aware of the full scope of services provided by the teaching and learning centre such as support on course development, consultancy on learning and teaching, and educational research collaboration. I consider out-of-class contact with students as important. It is important to place science teaching in a societal context. I have engaged in a research activity related to teaching and learning. Have you ever applied for a Teaching Development Grant (TDG)? I have previously attended the teaching and learning centre seminars/workshops on learning and teaching. I update my course material each year to reflect advances in the subject. I update my course material each year in response to SET feedback. Agree Neutral Disagree 62 44 62 37 61 56 36 27 28 21 36 16 61 44 28 28 61 56 30 15 58 93 58 69 7 19 0 12 n Yes 61 33 No 67 62 29 61 64 71 36 57 95 5 57 83 18 71% of them indicated that they had never applied for teaching-related research or a development grant even though all of them would have applied for external or internal research grants in their research areas. This culture of teaching and learning is also captured in the qualitative data, where one of the teachers criticized the performance system: The system does not encourage good teaching. Teachers are only rewarded by research performance, i.e., IF (impact factor) journal papers and UGC grants. The current HR policy is killing high-quality teaching. The management style is focused on elementary numbers (SET) no matter how useless, rather than tackles the much harder job of actually finding out what goes on in the classroom. The quantitative data in Table 3 indicated the teachers’ perceptions towards the SET survey, as well as their opinions on the reasons for low SET scores. Of the participants (n = 62), 39% believed that the SET score reflected their teaching quality and ability, 40% disagreed, and 21% held a neutral attitude. In spite of the mixed attitudes toward SET, 83% of the participants used SET feedback for the refinement of their course materials. The major perceptions expressed by the participants (n = 62) for low SET scores were: the difficulty in subject matter (76%), students’ low interests in the subject matter (79%), and teachers’ low motivation (54%). Only 38% of the participants regarded students’ lack of interest in new learning approaches as the reason for low SET scores. Table 3 also presents teachers’ opinions on the reasons for low SET scores, based on whether they believed that the scores reflect their teaching quality and ability. Among those who believed that SET scores reflect their teaching quality and ability (n = 24), Downloaded by [University of Hong Kong Libraries] at 18:12 25 June 2014 8 C.K.Y. Chan et al. most agreed that the difficulty of the subject matter (75%) and student’s lack of interest in the subject matter (71%) were reasons for low SET scores. Most teachers (78%) agreed that teachers with low teaching motivation is a reason for low SET scores, and only 30% attributed this to students’ negative attitudes towards new learning approaches. Among those who had doubts on SET scores as a reflection of their teaching quality and ability (n = 25), 80% perceived that the difficulty of the subject matter would influence SET scores, 80% regarded students’ low interest in the subject matter as a reason for low SET scores, whereas only 44% agreed that students’ negative attitudes towards new learning approaches would result in low SET scores. Overall, the results indicate that regardless of whether teachers believed that SET scores reflect their teaching quality and ability, there is a general agreement that courses obtain low SET scores because of students’ lack of interest in the subject matter and the subject matter being too difficult. Compared to students’ interest and subject difficulty, fewer teachers agreed that a low SET score is a result of students’ negative attitudes towards new learning approaches. Although fewer teachers agreed that students’ acceptance of new learning approaches was a reason for a low SET score, qualitative data collected from the openended comments provide rich details on the relationship between SET score and students’ acceptance of new learning approaches, as perceived by the teachers. For example, one teacher believed that SET score is proportionally related to the way students perceive new learning approaches. Some courses with a heavy field-trip component will surely be welcome by the students and thus the SET score will be high. Others with more traditional or difficult components will be less welcome by the students and thus no matter how good the teacher was the score will just be around 70. When comparing to the former courses then the latter will be below average. Another teacher believed that “the students prefer to be spoon-fed and get easy concrete material, and many faculties focus this way to get high SET scores”. One teacher worried that new learning approaches may penalize their SET scores and careers, even though teachers agree that an authentic approach should be adapted for better student learning. I feel obliged to the students to adopt an industry-focused case-study approach for one of my courses, but it is becoming too “expensive” in harm to my career from SET to continue this. I have no choice but to switch to a grinning monkey style of teaching with all the material handed on a plate to the students. Teachers were aware that by lowering the standard of the course particularly on students’ requests, SET scores may rise, although they were also aware that this is a poor method of student learning. If the students keep on indicating that the course content was too difficult and the teacher has to lower the standard of the course we will eventually end up with substandard graduates who are not comparable to graduates of other international universities. While fewer teachers agreed on students’ acceptance of new learning approaches as a reason for low SET scores, there are more teachers who believed that low teaching motivation leads to low SET scores. This is especially true among those who believed that SET scores reflected their teaching quality and ability than those who disagreed. This finding suggests that teachers who believe in the reliability of SET in reflecting their teaching Attitude towards SET, opinions on the reasons for low SET scores, and use of SET feedback. SET n Disagree Neutral Agree All % 25 40 13 21 24 39 62 100 SM n 25 13 24 62 SI Disagree Neutral Agree (%) (%) (%) n 4 15 8 8 16 15 17 16 80 69 75 76 25 13 24 62 NA Disagree Neutral Agree (%) (%) (%) n 8 0 13 8 12 8 17 13 80 92 71 79 25 13 23 61 TM Disagree Neutral Agree (%) (%) (%) n 20 8 17 16 36 54 52 46 44 39 30 38 25 13 23 61 SETF Disagree Neutral Agree (%) (%) (%) n 36 23.1 13 25 28 31 9 21 36 46 78 54 22 12 23 57 Yes No (%) (%) 82 75 87 83 18 25 13 18 Notes: SET = “SET scores reflect my teaching quality and ability”; SM = “Subject matter is too difficult”; SI = “Students are not interested”; NA = “Students do not like new learning approaches”; TM = “Teachers are not motivated”; SETF = “I update my course material each year in response to SET feedback”. The items on SET, SM, SI, NA, and TM were assessed on a 5-point Likert scale with 1= Strongly disagree, 5 = Strongly agree, and 3 = Neutral. The percentage of agreement (or disagreement) on the questionnaire items was calculated by adding the responses Strongly agree (or Strongly disagree) and Agree (or Disagree). The participants were invited to choose “Yes” or “No” in response to SETF. Table 4. SET SM SI NA TM Correlations between opinion on SET and the reasons for low SET scores. SET SM SI NA TM 1.00 .001 −.13 −.16 .37** 1.00 .30* .17 .23 1.00 .37** −.11 1.00 −.09 1.00 Notes: SET = “SET scores reflect my teaching quality and ability”; SM = “Subject matter is too difficult”; SI = “Students are not interested”; NA = “Students do not like new learning approaches”; TM = “Teachers are not motivated”. **p < .01. *p < .05. Educational Research and Evaluation Downloaded by [University of Hong Kong Libraries] at 18:12 25 June 2014 Table 3. 9 Downloaded by [University of Hong Kong Libraries] at 18:12 25 June 2014 10 C.K.Y. Chan et al. are more likely to associate low SET scores with internal factors (e.g., teaching motivation) than those who do not. This finding is further supported by a significant correlation found between participants’ opinion on teachers’ motivation as a factor influencing SET scores, and SET scores as a measurement of teaching quality and ability (r = .37, p < .01, see Table 4), denoting that participants who believed that SET scores reflected their teaching quality and ability also believed teachers’ insufficient motivation was a reason for low SET scores. No statistically significant correlations were found between participants’ attitudes towards SET scores as a measurement of teaching quality and ability, and their opinions on the other possible reasons for low SET scores such as the perceived difficulties of the subject matters, the anticipation of students’ interests in the subject matters, and the anticipation of students’ acceptance of new learning approaches. Neither was a statistically significant correlation found between these factors and teacher’s motivation. However, there were statistically significant correlations identified among the three factors as shown in Table 4. A positive correlation was found between (a) perceived subject matter difficulty and perceived students’ subject matter interest (r = .30, p < .05) and (b) perceived students’ subject matter interest and perceived students’ acceptance of new learning approaches (r = .37, p < .01). Among the participants who believed that SET scores reflected their teaching quality and ability (n = 23), 87% updated their course material each year in response to SET feedback (see Table 3). For those who did not believe that SET scores reflected their teaching quality and ability (n = 22), 82% updated their course material each year in response to SET feedback. In other words, most teachers in this study took SET feedback seriously in informing their amendment/improvement of course material regardless of whether they believe that SET scores reflect their teaching quality and ability. The majority of teachers seemed to wish to improve their teaching quality, and some participants indicated that the university SET did not seem to be the best evaluation mechanism for identifying their teaching and learning issues: The Faculty of Science has refused, in writing, to implement mandatory peer assessment of teaching. The current SET in the Faculty of Science is not supported by any other assessment such as peer-observation of teaching. Discussion Previous research has shown a lack of consensus in teachers’ opinions on the use of SET. The results in this study indicate that within this science faculty, teachers generally perceived teaching and learning as less important than research. This is likely due to the heavy research weightings that are expected of the academics for their yearly performance reviews and for tenure and promotion. This perception of teaching and learning seems to persist even though more than half of the participants are already tenured and are either at professor or associate professor positions (at this university, only professor or associate professor positions may be considered for tenure). Participants also seemed to be aware of those features of an effective teacher, as almost all of them realized that out-of-class contact with students is an important factor, and they seemed to place science teaching in a societal context. Most of them have also reported that they update their courses regularly in response to feedback and curriculum changes. They also seem to be aware of the scope of support services provided by the university’s teaching and learning centre, and more than half of them have attended seminars and workshops on teaching and learning. In Downloaded by [University of Hong Kong Libraries] at 18:12 25 June 2014 Educational Research and Evaluation 11 relation to the perception of using SET surveys as the primary evaluation mechanism for promotion and tenure and for course effectiveness, diverse attitudes among the teachers were revealed. Science teachers who participated in our study generally believed that students’ lack of interest in the subject matter and students’ perceived difficulty of the subject matter were the two reasons for a low SET score, suggesting that teachers were aware that students’ subject interest and perceived subject difficulty could influence students’ rating of teaching effectiveness. Marsh (2007) supports the view that students with greater interest in a subject tend to give higher teacher ratings. Qualitative data from this study also reveal that teachers believe they can increase their SET survey score by adopting more traditional teaching approaches and by lowering the difficulty of their course content as requested by students. Teachers believe that students’ preferences of the teaching approaches and the standards of the curriculum are influencing their practice indicating this is because the university has been putting a lot of emphasis on the SET survey scores and this emphasis may cause harm to the teachers’ careers if students do not like the teachers’ approaches. A significant correlation was found between the three reasons for teachers’ opinions for low SET survey scores (i.e., subject matter difficulty, students’ interest in subject matter, and students’ positive attitudes towards new learning approaches). It was noted from the correlation that: (a) the more difficult the subject matter, the more likely students’ interest towards the subject decreases; (b) as students’ interest towards the subject decreases, so does their attitude towards the new learning approaches. However, these correlations are only moderate, suggesting the possibility of other extraneous factors, such as curriculum design, student grades, and course standards in influencing SET score (Surgenor, 2013). These other factors have not been investigated in the present study. The findings of this study reveal that teachers who believe in the reliability of SET in reflecting their teaching are more likely to associate low SET scores with teaching motivation than teachers who do not hold this belief. These findings indicate that factors that influence the SET scores are related to course, students, and teachers characteristics; and it is only teacher perception of their own teaching motivation that can be totally controlled by teachers – an internal factor – whereas the subject matter difficulty, students’ interest in subject matter, and students’ positive attitudes towards new learning approaches are external factors that are related to the course and student characteristics. These characteristics related to student motivation and engagement can only be marginally controlled by teachers. There has been a notable concern among the participants about using SET surveys as a dominant evaluation for tenure and promotion and as the only measurement of teaching in our study. Some teachers have expressed strong, negative opinions about using SET surveys as the primary mechanism for evaluating teaching. Some felt that the SET survey is “dumbling down” the university system and is a “threat” to improving teaching and learning quality, and this is consistent with earlier literature findings (e.g., Stein et al., 2012; Surgenor, 2013; Wong & Moni, 2014). In their opinions, SET feedback is possibly affected by the course and student characteristics such as difficulty of the subject, student interest in the subject, and student preferences in teaching and learning approaches. When SET surveys are used as the only measurement, teachers are likely to be forced to consider these factors and adjust their teaching unintentionally in order to obtain better student ratings. The teaching would then be solely driven by students’ preferences and interests, and might not necessarily align with the curriculum’s learning outcomes or instructional approaches advocated by the university. This shows that using SET surveys may have an opposite effect of demotivating teachers from trying new teaching approaches rather than encouraging for teaching improvement. Thus, our findings suggest that the inclusion of Downloaded by [University of Hong Kong Libraries] at 18:12 25 June 2014 12 C.K.Y. Chan et al. other direct evidence of quality teaching (see Australian University Teaching Criteria and Standards Framework, n.d.), such as peer evaluation of teaching (Lomas & Nicholls, 2005), in addition to SET surveys is a worthwhile consideration. Doing so may help triangulate SET surveys and provide teachers with information from various perspectives for the improvement of teaching. Most important, it may help balance the wash-back effect of SET on teaching. In fact, Stark-Wroblewski, Ahlering, and Brill (2007) proposed that other measures of student learning should be documented for promotion and tenure review in addition to SET surveys, in order to present a more comprehensive view on teaching effectiveness. Despite diverse opinions on the effectiveness of SET surveys as a measurement of teaching quality and ability and despite concerns on the use of survey results, there seems to be a dominant acceptance of the SET feedback as useful information for updating course materials among the participants. It implies that SET surveys could still benefit the teachers, especially when they are well designed. Previous researchers, after an extensive review of SET surveys, have suggested that reliable and valid student evaluation of teaching can be achieved if measurement and theoretical issues are carefully handled (Ginns, Marsh, Behnia, Cheng, & Scalas, 2009). First, a solid theoretical basis is required to start with. Second, as effective teaching is usually a hypothetical construct with a multidimensional structure, its measurement needs to be clearly defined, and it requires a robust factor structure. By attending to both of these issues, the survey can be reliable, valid, and less affected by variables such as expected student course grades, student interest in subjects, class size, and workload. It would then be able to provide university teachers with accurate data that help enhance teaching and learning. To further strengthen SET surveys as a tool to facilitate reflective practices (Winchester & Winchester, 2014), training can also be provided to faculty members to help ensure proper interpretation of SET data (Boysen, Kelly, Raesly, & Casner, 2013). Conclusions On the whole, the findings of this study remained consistent with that of previous research. While we have observed varied attitudes towards the effectiveness of SET surveys as a measurement of teaching among the participants, there was much concern about the aim and use of SET surveys, particularly for promotion and tenure decision-making. The findings suggest that it is important to ensure that students, teachers, and other stakeholders fully understand the aims and purposes of SET surveys and are aware that the survey is not simply a mechanism for benchmarking teaching, but that it serves and focuses on improving teaching and learning outcomes (as opposed to improving only the students’ perceptions of the teacher). SET surveys should not be the only data source for evaluating teaching and learning; rather, it should be part of an overall strategic plan that provides reliable triangulated evidence from different perspectives for the improvement of teaching and learning. The opinions on possible reasons for low SET scores among the participants are also similar to what has been documented in the previous literature. This study has also identified several correlation patterns among the variables. However, only limited numbers of variables have been included in the discussion. Awareness of the factors affecting the SET scores as well as the relationship among them can help inform judgement on the extent to which the scores could be used for any strategic decision-making. Thus, future studies may include more factors that will provide information required to facilitate the understanding of our observed correlation patterns between participants’ opinions on the Downloaded by [University of Hong Kong Libraries] at 18:12 25 June 2014 Educational Research and Evaluation 13 SET scores and their views on the effect of course, student, and teaching characteristics on student rating. In addition, it may also be worth surveying teachers from different disciplines for better controlled comparisons. Hence, it would be worth conducting a followup study by administering a modified questionnaire that is designed to examine academic staff members’ perceptions of the SET and to observe whether or not the current findings can be replicated. In order to measure and benchmark quality teaching, university management, university quality agencies, and academic teachers need to understand the definition of quality, and a structured, clear guideline needs to be discussed and developed in order to enhance student learning and not just to satisfy bureaucracy. Despite the limitations of our study, it has enriched the existing literature by providing insights into how Hong Kong science teachers – who work in a research-intensive context – perceive SET. This provides us with evidence to report alternative mechanisms, and for introducing ideas of how the process of SET can be improved in order to advance our understanding of SET in the context of Hong Kong. Notes on contributors Cecilia Ka Yuk Chan is an Assistant Professor at the Centre for the Enhancement of Teaching and Learning (CETL) at the University of Hong Kong (HKU). Prior to her appointment in HKU, she was an Assistant Professor and Programme Coordinator in the Faculty of Engineering at the Dublin Institute of Technology, Ireland. Cecilia has many practical experiences in relation to learning and teaching particularly in improving student retention rates. She has been involved in numerous researches and the development of engineering education. Her research fields are transferable skills in higher education, assessment and active learning, women in engineering, student retention, recruitment, and motivation. Cecilia holds a PhD in the field of Digital Signal Processing and Machine Vision in Electrical Engineering and also a Postgraduate Diploma and an MA in Higher Education. Lillian Yun Yung Luk is a senior research assistant at the Centre for Enhancement in Teaching and Learning, and she is also pursuing a PhD with the Faculty of Education at the University of Hong Kong. Her PhD research project is on students’ learning through internship experience. Her research interests include student motivation, experiential learning, assessment, and the development of transferable skills. Dr. Lily Min Zeng is a lecturer at the Centre for Academic Development at Victoria University of Wellington (VUW). Prior to her appointment as lecturer at VUW, she served as a lecturer at the Centre for the Enhancement of Teaching and Learning at the University of Hong Kong, where she oversaw the collection and analysis of research postgraduates’ research experience data at the university level, coordinated postgraduate supervision workshops for new supervisors of the research students, and coordinated certificate courses for tutors and lab demonstrators. She has over 13 years of working experience at tertiary level in Mainland China and Hong Kong. Her main research interests are conceptual change and teaching and learning, with a particular emphasis on these issues in a cross-cultural context. References Adamson, G., O’Kane, D., & Shevlin, M. (2005). Students’ ratings of teaching effectiveness: A laughing matter? Psychological Reports, 96, 225–226. Aleamoni, L. M. (1999). Student rating myths versus research facts from 1924 to 1998. Journal of Personnel Evaluation in Education, 13, 153–166. Anderson, G. (2006). Assuring quality/resisting quality assurance: Academics’ responses to “quality” in some Australian universities. Quality in Higher Education, 12, 161–173. Arthur, L. (2009). From performativity to professionalism: Lecturers’ responses to student feedback. Teaching in Higher Education, 14, 441–454. Downloaded by [University of Hong Kong Libraries] at 18:12 25 June 2014 14 C.K.Y. Chan et al. Australian University Teaching Criteria and Standards Framework. (n.d.). Overview of the Australian university teaching criteria and standards framework. Retrieved from http://uniteachingcriteria. edu.au/wp-content/uploads/2013/12/Framework_indicative-standards-arrange-by-criteria.pdf Balam, E. M., & Shannon, D. M. (2010). Student ratings of college teaching: A comparison of faculty and their students. Assessment & Evaluation in Higher Education, 35, 209–221. Ballantyne, R., Borthwick, J., & Packer, J. (2000). Beyond student evaluation of teaching: Identifying and addressing academic staff development needs. Assessment & Evaluation in Higher Education, 25, 221–236. Barrow, M. (1999). Quality-management systems and dramaturgical compliance. Quality in Higher Education, 5, 27–36. Beran, T., & Rokosh, J. L. (2009). Instructors’ perspectives on the utility of student ratings of instruction. Instructional Science, 37, 171–184. Beran, T., & Violato, C. (2005). Ratings of university teacher instruction: How much do student and course characteristics really matter? Assessment & Evaluation in Higher Education, 30, 593–601. Boysen, G. A., Kelly, T. J., Raesly, H. N., & Casner, R. W. (2013). The (mis)interpretation of teaching evaluations by college faculty and administrators. Assessment & Evaluation in Higher Education. Advance online publication. doi:10.1080/02602938.2013.860950 Braskamp, L. A., & Ory, J. C. (1994). Assessing faculty work: Enhancing individual and institutional performance. San Francisco, CA: Jossey-Bass. Casey, R. J., Gentile, P., & Bigger, S. W. (1997). Teaching appraisal in higher education: An Australian perspective. Higher Education, 34, 459–482. Centra, J. A. (1993). Reflective faculty evaluation: Enhancing teaching and determining faculty effectiveness. San Francisco, CA: Jossey-Bass. Clayson, D. E. (2013). Initial impressions and the student evaluation of teaching. Journal of Education for Business, 88, 26–35. Ginns, P., Marsh, H. W., Behnia, M., Cheng, J. H. S., & Scalas, L. F. (2009). Using postgraduate students’ evaluations of research experience to benchmark departments and faculties: Issues and challenges. British Journal of Educational Psychology, 79, 577–598. Hendry, G. D., Lyon, P. M., & Henderson-Smart, C. (2007). Teachers’ approaches to teaching and responses to student evaluation in a problem-based medical program. Assessment & Evaluation in Higher Education, 32, 143–157. Kember, D., Leung, D. Y. P., & Kwan, K. P. (2002). Does the use of student feedback questionnaires improve the overall quality of teaching? Assessment & Evaluation in Higher Education, 27, 411– 425. Kwan, K.-P. (1999). How fair are student ratings in assessing the teaching performance of university teachers?. Assessment & Evaluation in Higher Education, 24, 181–195. Leckey, J., & Neill, N. (2001). Quantifying quality: The importance of student feedback. Quality in Higher Education, 7, 19–32. Lemos, M. S., Queiros, C., Teixeira, P. M., & Menezes, I. (2011). Development and validation of a theoretically based, multidimensional questionnaire of students evaluation of university teaching. Assessment & Evaluation in Higher Education, 36, 843–864. Lomas, L., & Nicholls, G. (2005). Enhancing teaching quality through peer review of teaching. Quality in Higher Education, 11, 137–149. Marsh, H. W. (1980). The influence of student, course, and instructor characteristics in evaluations of university teaching. American Educational Research Journal, 17, 219–237. Marsh, H. W. (1984). Students’ evaluations of university teaching: Dimensionality, reliability, validity, potential biases, and utility. Journal of Educational Psychology, 76, 707–754. Marsh, H. W. (2007). Students’ evaluations of university teaching: Dimensionality, reliability, validity, potential biases and usefulness. In R. Perry & J. Smart (Eds.), The scholarship of teaching and learning in higher education: An evidence-based perspective (pp. 319–383). Dordrecht, The Netherlands: Springer. Moore, S., & Kuol, N. (2005). Students evaluating teachers: Exploring the importance of faculty reaction to feedback on teaching. Teaching in Higher Education, 10, 57–73. Murphy, T., MacLaren, I., & Flynn, S. (2009). Toward a summative system for the assessment of teaching quality in higher education. International Journal of Teaching and Learning in Higher Education, 20, 226–236. Nasser, F., & Fresko, B. (2002). Faculty views of student evaluation of college teaching. Assessment & Evaluation in Higher Education, 27, 187–198. Downloaded by [University of Hong Kong Libraries] at 18:12 25 June 2014 Educational Research and Evaluation 15 Newton, J. (2000). Feeding the beast or improving quality? Academics’ perception of quality assurance and quality monitoring. Quality in Higher Education, 6, 153–162. Ryan, G. W., & Bernard, H. R. (2003). Techniques to identify themes. Field Methods, 15, 85–109. Scott, C., Stone, B., & Dinham, S. (2000, April). International patterns of teacher discontent. Paper presented at the Annual Meeting of the American Educational Research Association, New Orleans, LA. Shevlin, M., Banyard, P., Davies, M., & Griffiths, M. (2000). The validity of student evaluation of teaching in higher education: Love me, love my lectures? Assessment & Evaluation in Higher Education, 25, 397–405. Simpson, P. M., & Siguaw, J. A. (2000). Student evaluations of teaching: An exploratory study of the faculty response. Journal of Marketing Education, 22, 199–213. Slade, P., & McConville, C. (2006). The validity of student evaluations of teaching. International Journal for Educational Integrity, 2(2), 43–59. Spooren, P., Brockx, B., & Mortelmans, D. (2013). On the validity of student evaluation of teaching: The state of the art. Review of Educational Research, 83, 598–642. Stark-Wroblewski, K., Ahlering, R. F., & Brill, F. M. (2007). Toward a more comprehensive approach to evaluating teaching effectiveness: Supplementing student evaluations of teaching with pre-post learning measures. Assessment & Evaluation in Higher Education, 32, 403–415. Stehle, S., Spinath, B., & Kadmon, M. (2012). Measuring teaching effectiveness: Correspondence between students’ evaluations of teaching and different measures of student learning. Research in Higher Education, 53, 888–904. Stein, S. J., Spiller, D., Terry, S., Harris, T., Deaker, L., & Kennedy, J. (2012). Unlocking the impact of tertiary teachers‘ perceptions of student evaluation of teaching. Wellington, New Zealand: Ako Aotearoa National Centre for Tertiary Teaching Excellence. Surgenor, P. W. G. (2013). Obstacles and opportunities: Addressing the growing pains of summative student evaluation of teaching. Assessment & Evaluation in Higher Education, 38, 363–376. Tomasco, A. T. (1980). Student perceptions of instructional and personality characteristics of faculty: A canonical analysis. Teaching of Psychology, 7, 79–82. Tucker, B. (2014). Student evaluation surveys: Anonymous comments that offend or are unprofessional. Higher Education. Advance online publication. doi:10.1007/s10734-014-9716-2 Wachtel, H. K. (1998). Student evaluation of college teaching effectiveness: A brief review. Assessment & Evaluation in Higher Education, 23, 191–212. Winchester, T. M., & Winchester, M. K. (2014). A longitudinal investigation of the impact of faculty reflective practices on students’ evaluations of teaching. British Journal of Educational Technology, 45, 112–124. Wong, W. Y., & Moni, K. (2014). Teachers’ perceptions of and responses to student evaluation of teaching: Purposes and uses in clinical education. Assessment & Evaluation in Higher Education, 39, 397–411. View publication stats