Microsoft Word - kay.doc Australasian Journal of Educational Technology 2009, 25(2), 235-249 A strategic assessment of audience response systems used in higher education Robin H. Kay and Ann LeSage University of Ontario Institute of Technology An audience response system (ARS) permits students to respond to multiple choice questions using a remote control device. All responses are instantly displayed, usually in chart form, and subsequently reviewed and discussed by the instructor and the class. This paper offers a comprehensive review of teaching strategies used with ARS and includes a discussion of general, motivational, assessment based, and learning based approaches. Several promising strategies have been identified, particularly collecting formative assessment feedback and peer based instruction. More systematic, detailed research in a broader range of contexts is recommended. Overview An audience response system (ARS) allows students to respond to multiple choice questions using a remote control device. After students click in their responses, the results are instantly amalgamated and displayed in chart form. More often than not, responses are anonymous, but can be linked to specific students for evaluation purposes. The key advantage of using an ARS is that it can give feedback to both students and instructors on how well the entire class understands concepts presented. Once this feedback is obtained, a teacher can modify the course of instruction, or students can work out misconceptions via peer or classroom discussion. ARSs were first introduced at Stanford and Cornell Universities in the mid 1960s, but did not become commercially available until 1992 (Abrahamson, 2006; Judson & Sawada, 2002). In 1999, a new generation of more affordable ARSs was created, with widespread use emerging in 2003. Today, numerous colleges and universities use ARSs (Abrahamson, 2006). The purpose of the following review is to provide a current, comprehensive analysis of strategies used with ARSs in higher education. Method Previous literature reviews At least four previous literature reviews have been completed on ARSs (Caldwell, 2007; Fies & Marshall, 2006; Judson & Sawada, 2002; Simpson & Oliver, 2007). Judson & Sawada (2002) provided a summary of ARS use prior to 1998; however their review included only eight peer reviewed references. Moreover, because the widespread use of ARS began in earnest after 2003, Judson & Sawada`s review (2002) is now dated. 236 Australasian Journal of Educational Technology, 2009, 25(2) Fies & Marshall (2006) completed an extensive analysis of methods used to assess ARSs, however they examined only 16 peer reviewed studies, of which only two were published after 2004. Therefore, some of their findings and conclusions are incomplete. For example, they noted that few studies reported the use of ARSs for formative assessment purposes, however, since 2004, 16 new studies have been completed where formative assessment was examined. The authors also claimed that ARSs were used primarily for individual, not collaborative interactions, a practice that has changed markedly since 2004. A more recent review by Simpson & Oliver (2007) analysed more than 40 papers. Only 17 of the cited articles were from peer reviewed journals, with the majority of the results founded on five references. Finally, Caldwell (2007) analysed 25 peer reviewed articles. Most papers examined were published after 2000 with 10 studies completed after 2004. The principal foci of Caldwell’s work were to discern the primary users of ARSs, determine the rationale for using ARSs, explore questioning techniques utilised during ARS lessons, and identify best practices associated with ARS classroom use. However, fewer details were offered concerning specific teaching strategies used, and their impact on student learning. Current study Data In order to address some of the limitations of previous research reviews on ARSs, a comprehensive search of peer reviewed journals and edited books was completed. Conference papers or reports were not included. The reasons for not including conference papers or reports were threefold. First, it was assumed that peer reviewed journal articles, on balance, were more rigorous than conference papers in terms of method and quality of data. Journal articles typically report on finished studies, whereas many conference papers are works in progress. Second, the sheer volume of conference papers made it unrealistic to include this type of publication. A recent search of AACE conference proceedings revealed 395 conference papers on ARSs. Third, finding a consistent, objective method to assess the quality, validity and volume of conference papers is challenging. Some conferences have a strict review process, but the details for review criteria are not as transparent and consistent as they are with journal articles. All methodological choices come with inherent biases, and this paper is no different. By not including conference papers, it is recognised that new and innovative strategies for using ARSs will be missed (e.g. peer assessment from Richards, Braiding & Vaughan, 2006; or the impact of multiculturalism on ARS use from Albon & Jewels, 2007). The articles selected for this review focus solely on strategies used with ARSs in the higher education domain. Twenty-six labels (see Appendix A in Kay, 2008a) were used to search for relevant articles. A total of 52 papers and chapters were analysed. Given that previous literature reviews included no more than 25 peer reviewed papers, one can be reasonably confident that this review accurately reflects the state of current research on ARSs. Of the 52 studies analysed, 50 were conducted between 2000 and 2007, with 37 articles published since 2004. Data analysis Each of the studies included in this review was analysed based on the following categories: rationale or theory for using ARS, context of use, benefits and challenges associated with using ARS, strategies and pedagogical use, and individual differences Kay and LeSage 237 in the use of ARS. See Appendix B (see Kay, 2008b) for a detailed description of the coding of variables used in this study. See Appendix C (Kay, 2008c) for a description of all the articles reviewed for this study. Results and discussion Context of using ARS Before presenting a detailed analysis of strategies used with ARSs, it is informative to understand the context in which ARSs have been used to date. Seventeen studies were in the domain of science, six in medicine, four in mathematics or computer science, three in business, and three in engineering. The remaining 19 studies covered a variety of subject areas including anatomy, biology, earth science, economics, law, pharmacology, philosophy, and social science. Sample size ranged from 14 to 1542 students. Eighty percent of the 30 studies that did report sample size examined over 80 students with an overall mean 241, suggesting that ARSs were typically used in large classrooms. In summary, the conclusions from the current review reflect the attitudes and learning efforts of large classes of undergraduate students who were studying mathematics or science based subject areas. Strategies and ARS While there is considerable evidence to suggest that higher education students are very positive toward the use of ARSs (e.g., Caldwell, 2007; Fies & Marshall, 2006; Simpson & Oliver, 2007), the integration of an ARS into a classroom does not guarantee improved student learning (Draper & Brown, 2004; Van Dijk, Van Den Berg & Van Keulen, 2001). It is the implementation of pedagogical strategies in combination with the technology that ultimately influences student success (Reay, Bao, Li, Warnakulasooriya & Baugh, 2005; Simpson & Oliver, 2007; Stuart, Brown & Draper, 2004). While limited research has been done examining and comparing specific instructional strategies used with ARSs, a number of techniques have been employed over the past eight years (see Table 1 for a summary). Four categories of strategies were examined in the current review. General strategies are those referring to the preparation and process involved in using an ARS. Motivational strategies involve explicit attempts to involve or engage students. Assessment strategies refer to the use of an ARS to guide instruction or evaluate progress. Finally, learning based strategies include specific techniques designed to increase learning performance. General strategies for using an ARS Explain why an ARS is being used As stated earlier, higher education students appear to embrace the use of ARSs in the classroom. However, a number of students have resisted the use of ARSs, especially during the early stages of adoption. Some objections have included concerns over the extra effort required to discuss answers presented by an ARS (Trees & Jackson, 2007), wanting responses to be anonymous (Abrahamson, 2006), feeling uncomfortable when responding incorrectly, particularly when a majority of the class responds correctly (Carnaghan & Webb, 2006), becoming overly distracted by the use of an ARS (Siau, 238 Australasian Journal of Educational Technology, 2009, 25(2) Sheng, & Nah, 2006) and general resistance to a new method of learning (Beatty, 2004; Fagan, Crouch & Mazur, 2002). Table 1: Summary of strategies for using ARS Strategy Description References General strategies Explain Explain to class why ARS is being used Beatty, 2004; Caldwell, 2007; Dufresne & Gerace, 2004; Trees & Jackson, 2007. Preparation Planning required to develop effective questions Allen & Tanner, 2005; Boyle, 2006; Beatty, 2004; Boyle, 2006; Caldwell, 2007; Poulis et al., 1998; Beatty et al., 2006; McCabe, 2006; Stuart et al., 2004. Type Type of questions that works with ARS Beatty, 2004; Beatty et al., 2006; Brewer, 2004; Caldwell, 2007; Crouch & Mazur, 2001; Cutts, 2006; Dufresne & Gerace, 2004; Fies & Marshall, 2006; McCabe, 2006; Horowitz, 2006; Kennedy & Cutts, 2005; Miller et al., 2006; Poulis et al., 1998. Format Format in which questions are offered Caldwell, 2007; Cutts, 2006; Horowitz, 2006; McCabe, 2006; Robertson, 2000; Simpson & Oliver, 2007; Uhari et al., 2003. Motivational strategies Attendance Students go to class more Caldwell , 2007; Burnstein & Lederman, 2001; Greer & Heaney, 2004. Engagement Students are more engaged in class Bergtrom, 2006; Caldwell, 2007; Draper & Brown, 2004; Latessa & Mouw, 2005; Preszler et al., 2007; Siau, Sheng, & Nah, 2006; Simpson & Oliver, 2007. Participation Students participate with peers more in class to solve problems Bullock et al., 2002; Caldwell, 2007; Draper & Brown, 2004; Greer & Heaney, 2004; Jones et al., 2001; Siau, Sheng, & Nah, 2006; Stuart et al., 2004; Uhari et al., 2003; Van Dijk et al., 2001. Assessment strategies Formative Assessment is done that improves student understanding and quality of teaching Beatty, 2004; Bergtrom, 2006; Brewer, 2004; Bullock et al., 2002; Caldwell, 2007; Draper & Brown, 2004; Dufresne & Gerace, 2004; Elliott, 2003; Greer & Heaney, 2004; Hatch et al., 2005; Jackson et al., 2005; Siau, Sheng, & Nah, 2006; Simpson & Oliver, 2007; Stuart et al., 2004. Contingent teaching Adjust teaching method based on feedback from the class Beatty, 2004; Brewer, 2004; Draper & Brown, 2004; Elliott, 2003; Greer & Heaney, 2004; Jackson et al., 2005; Kennedy & Cutts, 2005; Poulis et al., 1998; Simpson & Oliver, 2007. Summative Used ARS for graded tests Draper et al., 2002; Fies & Marshall, 2006; Simpson & Oliver, 2007. Learning based strategies Attention Students are more focused in class Bergtrom, 2006; Burnstein & Lederman, 2001; Caldwell, 2007; d'Inverno, et al., 2003; Draper & Brown, 2004; Elliott, 2003; Elliott, 2003; Jackson et al., 2005; Jones et al., 2001; Latessa & Mouw, 2005; Siau, Sheng, & Nah, 2006; Slain et al., 2004. Interaction Students interact more with peers to discuss ideas Beatty, 2004; Bergtrom, 2006; Caldwell, 2007; Elliott, 2003; Freeman et al., 2007; Kennedy et al., 2006; Sharma et al., 2005; Siau, Sheng, & Nah, 2006; Slain et al., 2004; Stuart et al., 2004; Trees & Jackson, 2007; Van Dijk et al., 2001. Kay and LeSage 239 Peer based instruction Promote peer discussion and resolution of problems Brewer, 2004; Bullock et al., 2002; Burnstein & Lederman, 2001; Caldwell, 2007; Crouch & Mazur, 2001; Draper & Brown, 2004; Jones et al., 2001; Kennedy & Cutts, 2005; Miller et al., 2006; Nicol & Boyle, 2003. Student preparation Students read materials ahead of class Beatty, 2004; Bergtrom, 2006; Bergtrom, 2006; d'Inverno, et al., 2003; El-Rady, 2006; Uhari et al., 2003. Class based discussion Promote class discussion and resolution of problems Beatty et al., 2006; Caldwell, 2007; d'Inverno, et al., 2003; Nicol & Boyle, 2003; Reay et al., 2005; Sharma et al., 2005. Case studies Present and solve case studies Jones et al., 2001. Experiments Present and solve experiments Draper et al., 2002; Simpson & Oliver, 2007. Because of the potential problems that could arise when introducing an ARS, a number of researchers recommend that instructors should explain why this new technology is being used. Beatty (2004) suggests that students will be more comfortable when they understand the rationale for a specific classroom practice. Caldwell (2007) and Tree & Jackson (2007) add that if teachers expect to garner full student support for this new method, they need to explain why ARSs are being used and what they expect to gain by using the technology. Finally, although ARSs are reportedly easy to use (d'Inverno et al., 2003; Elliott, 2003; Hinde & Hunt, 2006; Jones, Connolly, Gear & Read, 2001; Pradhan, Sparano, & Ananth, 2005; Sharma, Khachan, Chan & O'Byrne, 2005; Siau et al., 2006), it might be wise to have practice questions to allow students to become familiar with using the technology (Caldwell, 2007). To date, the need for explanation and practice has not been systematically studied; however, it seems to be a reasonable approach for addressing some of the challenges noted earlier. For example, if students understand why they are using ARS they may be more accepting of this new way of learning, more willing to put forth the extra effort required to discuss and solve problems in class, and less sensitive when they respond incorrectly to the questions presented. Question preparation While setting up and using an ARS is a relatively simple and quick task, creating effective multiple choice questions is a challenging and time consuming process (Allen & Tanner, 2005; Boyle, 2006). A number of recommendations have been tendered with respect to preparing effective questions. It has been argued that every question should have an explicit pedagogical purpose (Beatty, 2004; Beatty, Leonard, Gerace & Dufresne, 2006; Caldwell, 2007; Poulis, Massen, Robens & Gilbert 1998). In addition, McCabe (2006) notes that questions should be thoughtfully linked together. Boyle (2006) adds that even when questions have been prepared and used, there is need for continual refinement. Stuart et al. (2004) suggest that spontaneous questions based on student feedback work well, however most research supports a more systematic carefully planned approach (Allen & Tanner, 2005; Beatty, 2004; Beatty et al., 2006; Caldwell, 2007; McCabe, 2006; Poulis et al., 1998). Current research, though, is lacking with respect to concrete evidence on the impact of various question preparation strategies. In addition, there are few classroom-ready collections of ARS questions available in most fields. Thus it is incumbent upon the instructor to develop original questions and many find this task extremely time consuming (Allen & Tanner, 2005; Beatty et al., 2006; El-Rady, 2006; Fagan et al., 2002; Freeman, Comerton-Forder, Pickering & Blayney, 2007; Paschal, 2002). 240 Australasian Journal of Educational Technology, 2009, 25(2) Question type Allen & Turner (2005) maintain that the cognitive benefits of ARSs are only as strong as the questions asked. Beatty et al. (2006) add that the critical challenge is to create questions that cultivate productive classroom interaction and discourse. A wide range of suggestions have been offered regarding the most effective type of ARS questions, including those that: • allow students to apply knowledge recently acquired (Poulis et al., 1998); • are higher level and require students to compare situations or data, make predictions, and explore causal relationships (Beatty et al., 2006; Poulis et al., 1998); • are ill-defined and vague with respect to possible answers so that students are required to process and debate (Beatty, 2004; Beatty et al., 2006; Caldwell, 2007; Dufresne & Gerace, 2004; Miller, Santana-Vega & Terrell, 2006); • focus on process and reasoning as opposed to factual content (Beatty, 2004; Beatty et al., 2006; Crouch & Mazur, 2001; Fies & Marshall, 2006); • identify and help resolve misconceptions (Beatty et al., 2006; Brewer, 2004; Crouch & Mazur, 2001; Cutts, 2006; Horowitz, 2006 ); and • support a comprehensive review of a specific set of concepts (McCabe, 2006; Kennedy & Cutts, 2005). It should be noted that the above characteristics are not mutually exclusive. For example, questions could be designed to be ill-defined, require students to make predictions, but also involve several responses that isolate misconceptions. It is also worth mentioning that the type of questions created may be partially dependent on the subject being taught (McCabe, 2006). For example, more technical, application type questions may be suited to applied science, mathematics or engineering, while more open ended, less concrete questions may be more appropriate in philosophy or social sciences. Some good examples of ARS questions are available at the A2L Library (2008), Chemistry Concept Tests (2008), and the Cornell Mathematics Database (2008). For a more detailed discussion of ARS questioning, consult Beatty (2004), Caldwell (2007), and Beatty et al. (2006). Question format Once a question is developed, decisions have to be made about the number of questions to ask, the number of options to provide within a multiple choice question, and how long to take when working with a specific question. Most researchers agree that questions should be sprinkled judiciously throughout a lecture at the rate of two to five questions per 50 minute time period (Allen & Tanner, 2005; Burton, 2006; Caldwell, 2007; Preszler, Dawe, Shuster, & Shuster, 2007; Robertson, 2000). The main reason given for limiting the number of questions is to maintain student interest and enthusiasm (Allen & Tanner, 2005; Burton, 2006; Preszler et al., 2007; Robertson, 2000). Given that thoughtful ARS questions take 5-10 minutes to display, discuss, and resolve (Cutts, 2006), it would be challenging to present more than one question every 10-15 minutes. Finally, because many teachers are concerned about reduced content coverage when using an ARS, limiting the number questions asked is probably a reasonable strategy (e.g., Beatty, 2004; Beatty et al., 2006; Caldwell, 2007; Cutts, 2006; Draper & Brown, 2004). Many researchers suggest that no more than four or five options be offered when asking an ARS question (Caldwell, 2007; Cutts, 2006; Robertson, 2000; Uhari, Renko & Soini, 2003). This suggestion is based on the experiences of the researchers, yet no specific reasons are given for this restriction. Several researchers have suggested that Kay and LeSage 241 to accurately monitor whole class understanding, an “I don’t know” option be included as a potential response (McCabe, 2006; Simpson & Oliver, 2007). Little research has been published on the optimum amount of time to take when asking an ARS question. Beatty et al. (2006) anecdotally reported that when the noise level drops in a class, it is time to stop peer discussion and move on. Time allotted for discussion of an ARS question may be related to subject area, level of thinking required, question difficulty level, and the pedagogical goals of the instructor. To date, limited data has been collected comparing the efficacy of different question formats. Motivational strategies for using an ARS Attendance In an effort to address attendance issues ARSs have been introduced at some universities. Several studies point out that attendance does improve when an ARS is used, provided it is associated with a portion of the course grade. Specifically, several researchers observed dramatic increases in attendance when 15% of a student’s grade was linked to ARS participation (Burnstein & Lederman, 2001; Greer & Heaney, 2004). However, Caldwell (2007) reported that allotting just five percent of student’s grade to ARS participation was sufficient motivation to improve regular classroom attendance. Only two studies have reported increased attendance in ARS classes when grades were not used as a motivator (El-Rady, 2006; Preszler et al., 2007). However, the strategy of using an ARS to increase attendance is not universally popular and sometimes has unexpected consequences. For example, Greer & Heaney (2004) added that students were displeased about being forced to attend class in order to gain academic credit for ARS participation. An unfortunate consequence of attendance monitoring practice is that 20% to 58% of students observed their peers bringing multiple remote devices to class to record attendance for missing classmates (Caldwell, 2007). Ideally, students should want to attend class for intrinsic, not extrinsic reasons. ARSs should provide an inherent learning incentive so that students want to attend. Attaching a grade to ARS monitored attendance may foster resistance and undermine the goal of developing a student centered environment. Participation Many instructors and learning theorists would agree that students need to participate actively in the learning process. One of the main reasons an ARS is used is to increase participation. Substantial evidence indicates that using an ARS increases student participation when compared to participation rates in classrooms where an ARS was not used (Bullock et al., 2002; Caldwell, 2007; Draper & Brown, 2004; Greer & Heaney, 2004; Jones et al., 2001; Siau, Sheng, & Nah, 2006; Stuart et al., 2004; Uhari et al.2003; Van Dijk et al., 2001). For example, one study observed that “shy” students participated more in classrooms using ARSs (Greer & Heaney, 2004). Bullock et al. (2002) added that when a portion of students’ grades were assigned to ARS use, participation markedly increased. Another study reported that ARSs were more effective when case studies were employed (Jones et al., 2001). Still, other researchers have noted that students were more involved when ARSs were used in groups as opposed to individually (Jones et al., 2001; Van Dijk et al., 2001). Engagement An implicit strategy for using ARSs is the engagement value and if students are engaged, it is argued they are more likely to actively construct knowledge. In general, 242 Australasian Journal of Educational Technology, 2009, 25(2) students in ARS based classes report being more interested or engaged in concepts presented and discussed (Bergtrom, 2006; Preszler et al., 2007; Simpson & Oliver, 2007). It is important to note, though, that detailed data on the reasons for why students are engaged has not been collected to date. For example, students may be more engaged because they are actively involved in the learning process. An alternative explanation might involve the novelty of the technology - it may simply be fun to use a remote control device in class and observe other students’ responses. More comprehensive, qualitative research is required to explore plausible explanations for increased student engagement with ARS use. Assessment based strategies for using an ARS Formative assessment There are two main forms of classroom assessment: summative and formative. Summative assessment is used for tests or assignments that count toward a student’s final grade. Formative assessment is used to determine student understanding of concepts without grades in order to identify misconceptions and alter classroom instruction accordingly (Bransford, Brown & Cocking, 1999). Without a tool like an ARS, it is challenging to calibrate overall student understanding of concepts presented in class. Regular use of an ARS can offer feedback to both instructors and students as to how well concepts are being understood. Experienced teachers can quickly modify explanations or mode of instruction and students can gauge and discuss their understanding of concepts as they are presented. There is considerable evidence to suggest that ARSs help provide effective formative assessment (Beatty, 2004; Bergtrom, 2006; Brewer, 2004; Bullock et al., 2002; Caldwell, 2007; Draper & Brown, 2004; Dufresne & Gerace, 2004; Elliott, 2003; Greer & Heaney, 2004; Hatch, Jensen, & Moore, 2005; Jackson et al., 2005; Siau et al, 2006; Simpson & Oliver, 2007; Stuart et al., 2004). Contingent teaching With contingent teaching, the material that is presented and discussed in class is largely dependent on student feedback from the ARS. The instructor is presenting ideas, gathering formative assessment data, and adjusting content and teaching strategies based on how well students understand the concepts (Brewer, 2004; Cutts, 2006; Draper & Brown, 2004; Elliott, 2003; Greer & Heaney, 2004; Hinde & Hunt, 2006; Jackson et al., 2005; Kennedy & Cutts, 2005; Poulis et al., 1998; Stuart et al., 2004). There is some evidence to suggest that this approach has been successful (Brewer, 2004; Greer & Heaney, 2004), although Abrahamson (2006) speculates that success may be dependent on an instructor’s experience level and ability to instantly address problems and misconceptions. Summative assessment Fies & Marshall (2006) reported most higher education instructors regularly employ a summative assessment strategy when using an ARS. However, limited evidence exists to support this claim - most studies report the formative use of ARSs (e.g., Abrahamson, 2006; Beatty, 2004; Caldwell, 2007; Elliott, 2003; Jackson et al., 2005). It has been suggested that summative assessment encourages rote learning and cannot be used to shape instruction in a dynamic fashion (Dufresne & Gerace, 2004). In fact, understanding how well students understand concepts in traditional, lecture based classes is often a mystery until after the first exam (Bullock et al., 2002). There is some evidence to indicate that higher education students do not enjoy using ARS for grades (Caldwell, 2007). While using ARS for formal testing situations might be attractive to an instructor in terms of quick, easy marking, it may not be the most appropriate Kay and LeSage 243 pedagogical choice. More research is needed examining the benefits and challenges of using ARSs for summative assessment. Learning based strategies for using an ARS Attention One of the fundamental conditions for effective learning is that students are focused and paying attention to the content presented. Some evidence indicates that in classroom situations, student attention wanes after approximately 20 minutes (d'Inverno et al., 2003; Jackson et al., 2005). Given that the length of a traditional lecture is between 50 minutes and three hours, it is inevitable that students will be unable to concentrate on the lesson content for the entire duration of the class. Presenting ARS questions at 20 minute intervals may be one technique for segmenting a long lecture by requiring students to shift their attention and actively participate in the learning process. This hypothesis is supported by numerous studies which have reported that higher education students are more attentive when an ARS is used during lectures (Bergtrom, 2006; Burnstein & Lederman, 2001; Caldwell, 2007; d'Inverno et al., 2003; Draper & Brown, 2004; Elliott, 2003; Jackson et al., 2005; Jones et al., 2001; Latessa & Mouw, 2005; Siau et al., 2006; Slain, Abate, Hidges, Stamatakis & Wolak., 2004). It should be noted, though, that a direct link between increased attention in ARS based classrooms and learning performance has not been established. Interaction It could be argued that participation is a necessary, but not sufficient component for learning. The quality of participatory effort is perhaps more important. Students need to be interacting with each other, the instructor and new concepts being introduced. They must also be “cognitively” engaged (Van Dijk et al., 2001). Numerous studies suggest that frequent and positive interaction occurs with ARSs (Beatty, 2004; Bergtrom, 2006; Caldwell, 2007; Elliott, 2003; Freeman et al., 2007; Kennedy, Cutts & Draper, 2006; Sharma et al., 2005; Siau et al., 2006; Slain et al., 2004; Stuart et al., 2004; Trees & Jackson, 2007). When an ARS is used, researchers have reported greater articulation of student thinking (Beatty, 2004), more probing questions and an increased focus on student needs (Beatty, 2004; Siau, et al., 2006), effective peer to peer discussions (Bergtrom, 2006; Caldwell, 2007; Kennedy et al., 2006), and active learning (Elliott, 2003; Kennedy et al., 2006; Slain et al., 2004; Stuart et al., 2004). Peer based learning One of the most common and successful strategies used with ARS is peer based instruction, which involves displaying a higher level question that could identify a misconception, asking students to click in a response, giving students time to discuss and defend their answers with two to four peers, taking a re-vote on the original question, and having the instructor provide a brief summary (Brewer, 2004; Bullock et al., 2002; Burnstein & Lederman, 2001; Crouch & Mazur, 2001; Cutts, 2006; Draper & Brown, 2004; Hinde & Hunt, 2006; Jones et al., 2001; Kennedy & Cutts, 2005; Miller et al., 2006; Nicol & Boyle, 2003). Crouch & Mazur (2001) provided 10 years of evidence suggesting that peer based instruction provides significant gains in student learning performance. Nicol & Boyle (2003) added that peer instruction is central to the development of student conceptual understanding. One of the main concerns about using an ARS on a regular basis is coverage of content. Abundant research suggests that teachers, and occasionally students, feel that less content is addressed when using an ARS as opposed to a more traditional lecture 244 Australasian Journal of Educational Technology, 2009, 25(2) format (Beatty, 2004; Beatty et al., 2006; Burnstein & Lederman, 2006; Caldwell, 2007; d'Inverno, et al., 2003; Burton, 2006; Cutts, 2006; Draper & Brown, 2004; Fagan et al., 2002; Freeman et al., 2007; Hatch et al., 2005; Sharma et al., 2005; Siau et al., 2006; Slain et al., 2004; Steinert & Snell, 1999; Stuart et al., 2004). Responding to and discussing knowledge centred questions that identify and target misconceptions can take considerably more time than simply presenting material in a lecture. Some researchers have noted, though, that what is covered in a traditional lecture may not be understood as well as concepts learned with an ARS (Beatty et al., 2006; Caldwell, 2007). However, the fact remains that curriculum coverage is a reality that many lecturers have to face. One way to compensate for material not covered in class is to require students to do more reading and class preparation outside of the lecture (Bergtrom, 2006; Bullock et al., 2002; Burnstein & Lederman, 2001; Caldwell, 2007; Slain et al., 2004). This strategy is discussed next. Require student preparation before class A number of researchers have suggested that a good strategy for addressing reduced content coverage is to require students to read materials prior to class so that the class time can be devoted to refining and extending student thinking and knowledge (Beatty, 2004; Bergtrom, 2006; d'Inverno, et al., 2003). D'Inverno et al. (2003) add that providing notes to augment a lecture was a popular instructional practice. Student attitudes toward the extra effort required for class preparation has not been measured, nor has the impact of this strategy on the quality of learning. Classroom discussion A variation of the peer instruction approach augments the role of the entire class. A multiple choice question is displayed, students are immediately asked to discuss possible solutions with their peers before voting, responses are collected and displayed, and then volunteers are asked to explain the answers that they selected. At some point, the instructor stops the class discussion and summarises the results (Beatty et al., 2006; d'Inverno, et al., 2003; Nicol & Boyle, 2003; Reay et al., 2005; Sharma et al., 2005). One research study suggested that students prefer peer based instruction to classroom discussion (Nicol & Boyle, 2003) - students wanted to think about their answers on their own, before discussing it with peers. A discussion first approach tended to mute conversation with some students simply conceding to the most dominant student because they did not have sufficient time to formulate a response. Other researchers have suggested that the classroom discussion approach may be better suited to smaller classes (d'Inverno, et al., 2003). Occasionally, significant problems emerge when ARS questions are presented for classroom discussion. Some students may dominate group discussions (Nicol & Boyle, 2003) or discussion of different viewpoints may precipitate student confusion (Nicol & Boyle, 2003, Reay et al., 2005). Occasionally, students feel that ARS use distracts them from the lesson (Draper & Brown, 2004) or view class discussion as intimidating and a source of anxiety (Nicol & Boyle, 2003). While these problems have not been reported extensively, instructors and researchers need more information about creating effective discussion that is focussed, non-threatening, and efficient. Case studies Jones et al. (2003) used ARSs successfully with university students by incorporating case study questions in a peer based instruction format. The case studies appeared to make classes far more animated and students talked considerably more with peers instead of working on their own. This approach may be limited to the subject area Kay and LeSage 245 being covered. For example, it may be more difficult to come up with reasonable case studies for mathematics or physics. Clearly more research is needed to gain a comprehensive picture of the effectiveness of case studies. Experiments While this approach has not been tested rigorously, conducting psychology experiments in class and requiring students to predict outcomes has been examined by Simpson & Oliver (2007). Clearly more research is needed to determine the validity of this approach, but, again, success may be partially dependent on subject area. Science, for example, may be a good fit for an experimental approach, whereas history or English may not. Summary of strategies and future research An analysis of 52 peer reviewed articles and chapters has revealed helpful information regarding strategies used with ARSs. Promising ARS general strategies reported included explaining why an ARS is being used, asking practice questions, taking time to develop effective, meaningful questions, developing questions that stimulate class discussion, focussing on higher level as opposed to factual questions, asking only two to five questions per hour, and limiting the number of options in a multiple choice question to four or five. Effective ARS motivational strategies involved using ARSs to increase participation and engagement. Promising ARS assessment strategies included collecting formative feedback and contingency teaching. Finally, with respect to learning strategies, using ARS to increase attention levels, improving the quality of classroom interaction, peer based instruction, and requiring students to read extra materials before an ARS class appeared to work well. Preliminary research suggests that using ARSs with experiments and case studies is effective. Nonetheless, a number of significant ARS strategy based problems have yet to be examined. Overall, the results of the current review are limited to math and science based subject areas. More research is needed on the use of ARS in broader range of subject areas. Regarding ARS general strategies, formal evaluation comparing the efficacy of specific question types and format is needed. With respect to ARS motivational strategies, it is unclear why students are more engaged when ARSs are used. Anecdotal explanations have included the novelty effect of using a new technology, the fun of clicking in answers, and increased commitment to learning concepts being presented, but more in depth research is needed to explore and confirm plausible explanations for increased student engagement. Concerning ARS assessment strategies, it appears that formative assessment is a productive technique; yet more research is necessary focussing on the benefits and challenges of using ARS for summative assessment. Finally, when examining learning based strategies, learning performance needs to be assessed to determine the effectiveness of certain strategies such as the impact of increased attention levels, extra student preparation before class, experiments, and case studies. References A2L Item Library (2008). A2L- Assessing to learn physics. [viewed 28 Nov 2008] http://a2l.physics.umass.edu/library/ Abrahamson, L. (2006). A brief history of networked classrooms: Effects, cases, pedagogy, and implications. In D. A. Banks (Ed.), Audience response systems in higher education (pp. 1-25). Hershey, PA: Information Science Publishing. 246 Australasian Journal of Educational Technology, 2009, 25(2) Albon, R. & Jewels, T. (2007). The impact of audience response systems in a multicultural Asian context. In ICT: Providing choices for learners and learning. Proceedings ascilite Singapore 2007. http://www.ascilite.org.au/conferences/singapore07/procs/albon.pdf Allen, D. & Tanner, K. (2005). Infusing active learning into the large-enrolment Biology class: Seven strategies, from the simple to complex. Cell Biology Education, 4(4), 262-268. http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=1305885 Beatty, I. (2004). Transforming student learning with classroom communication systems. EDUCAUSE Research Bulletin, 2004(3), 1-13. [viewed 3 Nov 2007] http://www.educause.edu/ir/library/pdf/ERB0403.pdf Beatty, I. D., Leonard, W. J., Gerace, W. J. & Dufresne, R. J. (2006). Question driven instruction: Teaching science (well) with an audience response system. In D. A. Banks (Ed.), Audience response systems in higher education (pp. 96-115). Hershey, PA: Information Science Publishing. Bergtrom, G. (2006). Clicker sets as learning objects. Interdisciplinary Journal of Knowledge and Learning Objects, 2. [viewed 3 Nov 2007] http://ijklo.org/Volume2/v2p105-110Bergtrom.pdf Boyle, J. (2006). Eight years of asking questions. In D. A. Banks (Ed.), Audience response systems in higher education (pp. 289-304). Hershey, PA: Information Science Publishing. Bransford, J. D., Brown, A. L. & Cocking, R. R. (Eds.) (1999). How people learn: Brain, mind, experience, and school. Washington, DC: National Academy Press. [verified 8 May 2009] http://www.nap.edu/openbook.php?record_id=6160 Brewer, C. A. (2004). Near real-time assessment of student learning and understanding in biology courses. BioScience, 54(11), 1034-1039. Bullock, D. W., LaBella, V. P., Clinghan, T., Ding, Z., Stewart, G. & Thibado, P. M. (2002). Enhancing the student-instructor interaction frequency. The Physics Teacher, 40, 30-36. Burnstein, R. A. & Lederman, L. M. (2001). Using wireless keypads in lecture classes. The Physics Teacher, 39(1), 8-11. Burnstein, R. A. & Lederman, L. M. (2006). The use and evolution of an audience response system. In D. A. Banks (Ed.), Audience response systems in higher education (pp. 40-52). Hershey, PA: Information Science Publishing. Burton, K. (2006). The trial of an audience response system to facilitate problem-based learning in legal education. In D. A. Banks (Ed.), Audience response systems in higher education (pp. 265- 276). Hershey, PA: Information Science Publishing. Caldwell, J. E. (2007). Clickers in the large classroom: Current research and best-practice tips. Life Sciences Education, 6(1), 9-20. http://www.lifescied.org/cgi/reprint/6/1/9.pdf Carnaghan, C. & Webb, A. (2007). Investigating the effects of group response systems on student satisfaction, learning, and engagement in accounting education. Issues in Accounting Education, 22(3), 391-409. Chemistry Concept Tests (2008). [viewed 28 Nov 2008] http://jchemed.chem.wisc.edu/JCEDLib/QBank/collection/ConcepTests/ Cornell Mathematics Database (2008). [viewed 28 Nov 2008] http://www.math.cornell.edu/~GoodQuestions/GQbysubject_pdfversion.pdf Kay and LeSage 247 Crouch, C. H. & Mazur, E. (2001). Peer instruction: Ten years of experience and results. American Journal of Physics, 69(9), 970-977. Cutts, Q. (2006). Practical lessons from four years of using an ARS in every lecture of a large class. In D. A. Banks (Ed.), Audience response systems in higher education (pp. 65-79). Hershey, PA: Information Science Publishing. D’Inverno, R., Davis, H. & White, S. (2003). Using a personal response system for promoting student interaction. Teaching Mathematics and Its Applications, 22(4), 163-169. Draper, S. W. & Brown, M. I. (2004). Increasing interactivity in lectures using an electronic voting system. Journal of Computer Assisted Learning, 20(2), 81-94. Dufresne, R. J. & Gerace, W. J. (2004). Assessing-to-learn: Formative assessment in physics instruction. The Physics Teacher, 42, 428-433. Elliott, C. (2003). Using a personal response system in economics teaching. International Review of Economics Education, 1(1). [viewed 3 Nov 2007] http://www.economicsnetwork.ac.uk/iree/i1/elliott.htm El-Rady, J. (2006). To click or not to click: That's the question. Innovate Journal of Online Education, 2(4). [viewed 3 Nov 2007] http://www.innovateonline.info/index.php?view=article&id=171 Fagan, A. P., Crouch, C. H. & Mazur, E. (2002). Peer Instruction: Results from a range of classrooms. The Physics Teacher, 40(4), 206-209. Fies, C. & Marshall, J. (2006). Classroom response systems: A review of the literature. Journal of Science Education and Technology, 15(1), 101-109. Freeman, M., Bell, A., Comerton-Forder, C., Pickering, J. & Blayney, P. (2007). Factors affecting educational innovation with in class electronic response systems. Australasian Journal of Educational Technology, 23(2), 149-170. http://www.ascilite.org.au/ajet/ajet23/freeman.html Greer, L. & Heaney, P. J. (2004. Real-time analysis of student comprehension: an assessment of electronic student response technology in an introductory earth science course. Journal of Geoscience Education, 52(4), 345-351. Hatch, J., Jensen, M. & Moore, R. (2005). Manna from heaven or clickers from hell. Journal of College Science Teaching, 34(7), 36-39. Hinde, K. & Hunt, A. (2006). Using the personal response system to enhance student learning: Some evidence from teaching economics. In D. A. Banks (Ed.), Audience response systems in higher education (pp. 140-154). Hershey, PA: Information Science Publishing. Horowitz, H. M. (2006). ARS evolution: Reflections and recommendations. In D. A. Banks (Ed.), Audience response systems in higher education (pp. 53-63). Hershey, PA: Information Science Publishing. Jackson, M., Ganger, A. C., Bridge, P. D. & Ginsburg, K. (2005). Wireless handheld computers in the undergraduate medical curriculum. Medical Education Online, 10(5). [viewed 3 Nov 2007] http://www.med-ed-online.org/pdf/t0000062.pdf Jones, C., Connolly, M., Gear, A. & Read, M. (2001). Group integrative learning with group process support technology. British Journal of Educational Technology, 32(5), 571-581. Judson, E. & Sawada, D. (2002). Learning from past and present: Electronic response systems in college lecture halls. Journal of Computers in Mathematics and Science Teaching, 21(2), 167-181. 248 Australasian Journal of Educational Technology, 2009, 25(2) Kay, R. H. (2008a). Appendix A - Labels used to describe audience response systems. http://faculty.uoit.ca/kay/papers/arsrev/AppendixA_Labels.pdf Kay, R. H. (2008b). Appendix B – Coding of research papers reviewed for ARS strategy paper. http://faculty.uoit.ca/kay/papers/arsrev/AppendixB_Coding.pdf Kay, R. H. (2008c). Appendix C – List of studies reviewed for ARS strategy paper. http://faculty.uoit.ca/kay/papers/arsrev/AppendixC_ListOfPapers.pdf Kennedy, G. E. & Cutts, Q. I. (2005). The association between students’ use of electronic voting systems and their learning outcomes. Journal of Computer Assisted Learning, 21(4), 260-268. Kennedy, G. E., Cutts, Q. & Draper, S. W. (2006). Evaluating electronic voting systems in lectures: Two innovative methods. In D. A. Banks (Ed.), Audience response systems in higher education (pp. 155-174). Hershey, PA: Information Science Publishing. Latessa, R. & Mouw, D. (2005). Use of audience response system to augment interactive learning. Family Medicine, 37(1), 12-14. [viewed 3 Nov 2007] http://www.stfm.org/fmhub/fm2005/January/Robyn12.pdf McCabe, M. (2006). Live assessment by questioning in an interactive classroom. In D. A. Banks (Ed.), Audience response systems in higher education (pp. 276-288). Hershey, PA: Information Science Publishing. Miller, R. L., Santana-Vega, E. & Terrell, M. S. (2006). Can good questions and peer discussion improve calculus instruction? PRIMUS, 16(3), 1-9. [verified 8 May 2009; preprint] http://www.math.cornell.edu/~maria/mathfest_eduction/preprint.pdf Nicol, D. J. & Boyle, J. T. (2003). Peer instruction versus class-wide discussion in large classes: A comparison of two interaction methods in the wired classroom. Studies in Higher Education, 28(4), 457-473. Paschal, C. B. (2002). Formative assessment in physiology teaching using a wireless classroom communication system. Advances in Physiology Education, 26(4), 299-308. Poulis, J., Massen, C., Robens, E. & Gilbert, M. (1998). Physics lecturing with audience paced feedback. American Journal of Physics, 66(5), 439-441. Pradhan, A., Sparano, D., Ananth, C. V. (2005). The influence of an audience response system on knowledge retention: An application to resident education. American Journal of Obstetrics and Gynecology, 193(5), 1827-1830. Preszler, R. W., Dawe, A., Shuster, C. B. & Shuster, M. (2007). Assessment of the effects of student response systems on student learning and attitudes over a broad range of biology courses. CBE-Life Sciences Education, 6(1), 29-41. http://www.lifescied.org/cgi/content/full/6/1/29 Reay, N. W., Bao, L., Li, P., Warnakulasooriya, R. & Baugh, G. (2005). Toward the effective use of voting machines in physics lectures. American Journal of Physics, 73(6), 554-558. Richards, D., Braiding, K. & Vaughan, A. (2006). Fun and feedback at the press of a button. In Who's learning? Whose technology? Proceedings ascilite Sydney 2006. http://www.ascilite.org.au/conferences/sydney06/proceeding/pdf_papers/p151.pdf Robertson, L. J. (2000). Twelve tips for using a computerised interactive audience response system. Medical Teacher, 22(3), 237-239. Kay and LeSage 249 Sharma, M. D., Khachan, J., Chan, B. & O'Byrne, J. (2005). An investigation of the effectiveness of electronic classroom communication systems in large lectures. Australasian Journal of Educational Technology, 21(2), 137-154. http://www.ascilite.org.au/ajet/ajet21/sharma.html Siau, K., Sheng, H. & Nah, F. (2006). Use of classroom response systems to enhance classroom interactivity. IEEE Transactions on Education, 49(3), 398-403. Simpson, V. & Oliver, M. (2007). Electronic voting systems for lectures then and now: A comparison of research and practice. Australasian Journal of Educational Technology, 23(2), 187- 208. http://www.ascilite.org.au/ajet/ajet23/simpson.html Slain, D., Abate, M., Hidges, B. M., Stamatakis, M. K. & Wolak, S. (2004).An interactive response system to promote active learning in the doctor of pharmacy curriculum. American Journal of Pharmaceutical Education, 68(5), 1-9. Steinhert, Y. & Snell, L. S. (1999). Interactive lecturing: Strategies for increasing participation in large group presentations. Medical Teacher, 21(1), 37-42. Stuart, S. A. J., Brown, M. I. & Draper, S.W. (2004). Using an electronic voting system in logic lectures: one practitioner’s application. Journal of Computer Assisted Learning, 20(2), 95-102. Trees, A. R. & Jackson, M. H. (2007). The learning environment in clicker classrooms: student processes of learning and involvement in large university course using student response systems. Learning, Media, and Technology, 32(1), 21-40. Uhari, M., Renko, M. & Soini, H. (2003). Experiences of using an interactive audience response system in lectures. BMC Medical Education, 3(12), 1-6. http://www.biomedcentral.com/1472- 6920/3/12 Van Dijk, L. A., Van Den Berg, G. C. & Van Keulen, H. (2001). Interactive lectures in engineering education. European Journal of Engineering Education, 26(1), 15-28. Dr Robin H. Kay University of Ontario Institute of Technology, Faculty of Education 2000 Simcoe St. North, Oshawa, Ontario L1H 7L7, Canada Email: Robin.Kay@uoit.ca Web: http://faculty.uoit.ca/kay/home/ Dr Ann LeSage, University of Ontario Institute of Technology, Faculty of Education Email: ann.lesage@uoit.ca Web: http://education.uoit.ca/EN/main/13738/159923/profiles_lesage.php