A Campus-Wide Investigation of Clicker Implementation: The Status of Peer Discussion in STEM Classes
Abstract
At the University of Maine, middle and high school teachers observed more than 250 university science, technology, engineering, and mathematics classes and collected information on the nature of instruction, including how clickers were being used. Comparisons of classes taught with (n = 80) and without (n = 184) clickers show that, while instructional behaviors differ, the use of clickers alone does not significantly impact the time instructors spend lecturing. One possible explanation stems from the observation of three distinct modes of clicker use: peer discussion, in which students had the opportunity to talk with one another during clicker questions; individual thinking, in which no peer discussion was observed; and alternative collaboration, in which students had time for discussion, but it was not paired with clicker questions. Investigation of these modes revealed differences in the range of behaviors, the amount of time instructors lecture, and how challenging the clicker questions were to answer. Because instructors can vary their instructional style from one clicker question to the next, we also explored differences in how individual instructors incorporated peer discussion during clicker questions. These findings provide new insights into the range of clicker implementation at a campus-wide level and how such findings can be used to inform targeted professional development for faculty.
INTRODUCTION
A number of national reports informed by emerging education research have advocated for active-engagement instruction in postsecondary science, technology, engineering, and mathematics (STEM) courses (American Association for the Advancement of Science, 2010; President’s Council of Advisors on Science and Technology [PCAST], 2012; Singer et al., 2012). Moreover, a recent comprehensive meta-analysis of 225 science education research articles indicates that students learn more in and are less likely to drop out of STEM courses that use these active-engagement instructional approaches (Freeman et al., 2014). One such instructional approach involves instructors posing multiple-choice conceptual questions, fostering peer discussion about these questions among the students, and asking students to indicate their answers via personal response systems or clickers.
Clickers are electronic voting devices that allow instructors to obtain real-time student responses to multiple-choice questions in order to assess student thinking and to inform instruction (e.g., Mazur, 1997; Caldwell, 2007; Smith et al., 2011). A recent nationwide survey found that 86% of U.S. college faculty members are familiar with clickers and 12% of faculty members have adopted clickers in their own classrooms (FTI Consulting, 2015). In addition, a study of student engagement in a large-enrollment undergraduate science class found the use of clicker questions and the follow-up to clicker questions to be the most engaging of all in-class activities observed, as measured by the Behavioral Engagement Related to Instruction protocol (Lane and Harris, 2015).
It is often suggested that clicker questions be supported by an instructional strategy known as peer instruction (Mazur, 1997). In a high-fidelity enactment of peer instruction, 1) the instructor poses a multiple-choice conceptual question; 2) students are given time to think; 3) students determine their individual answers and vote; 4) if there is variation in the student answers, neighboring students discuss their answers with one another; 5) students vote again after peer discussion; and 6) the instructor explains the correct answer to the whole class, often displaying a histogram of all student responses and soliciting explanations from students for incorrect and correct answers. Peer instruction encourages student interactions during lecture and breaks the monotony of passive listening while also offering an opportunity for the instructor to walk around the room and interact with students. These interactions allow the instructor to gauge the level of student understanding and thus gain insight into incorrect lines of reasoning.
Multiple studies have shown that the peer discussion portion of peer instruction increases student performance on clicker questions (Smith et al., 2009; Porter et al., 2011; Knight et al., 2015; Barth-Cohen et al., 2016) and that peer discussion produces higher performance outcomes when compared with other tasks such as quiet reflection (Lasry et al., 2009). Furthermore, the largest gains in student performance occur when peer discussion is immediately followed by an instructor explanation (Smith et al., 2011). Notably, students have more positive attitudes about the utility of clickers when faculty encourage peer discussion and are successfully able to create opportunities for students to discuss the multiple-choice questions (Keller et al., 2007).
Although the learning benefits of using clicker questions with peer discussion have been documented through the use of carefully designed protocols, it has also been noted that faculty often change and modify research-based pedagogies and tools, such as clickers, in their classrooms (Henderson and Dancy, 2007; National Research Council [NRC], 2013). For example, survey results from faculty in multiple disciplines showed that 15% of faculty who use clickers reported that they did not allow or did not encourage peer discussion during clicker questions (Keller et al., 2007). In another observation-based study of undergraduate physics classrooms, researchers found that none of the faculty had students record their individual answers before talking with peers (Turpen and Finkelstein, 2009), a component of peer instruction that is advocated by researchers (Mazur, 1997; Smith et al., 2011).
In this study, we used classroom observation data from 21 different STEM departments to explore the spectrum of instructional practices associated with clickers. Specifically, we asked: 1) Are there differences in instructional behaviors in classes that are taught with and without clicker questions? 2) In classes that use clickers, is there variation in how clickers are implemented? 3) How do individual instructors vary their implementation of clicker questions? The answers to all three questions are critical for identifying common-use cases on which to focus future research and for optimizing faculty professional development so that it may better support the effective implementation of clickers.
METHODS
For this investigation, University of Maine STEM instructors were sent emails asking them if they would allow middle and high school teachers to visit their classrooms and collect observation data. The instructors were receptive, with 74% agreeing to allow the teachers to observe their courses. Faculty who declined typically cited reasons such as giving an exam, canceling class, or having a guest lecturer present on the day of the proposed observation.
Observations were conducted in both February and April during the Spring 2014 and Spring 2015 semesters and in November during the Fall 2014 semester. Altogether, 270 class sessions were observed. These observations represented 119 instructors who taught 138 courses in 21 different departments (biology and ecology; chemical and biological engineering; chemistry; civil and environmental engineering; computer sciences; earth sciences; ecology and environmental sciences; economics; electrical and computer engineering; food and agriculture; forest resources; marine science; mathematics and statistics; mechanical engineering; molecular and biomedical science; new media; nursing; physics and astronomy; plant, soil, and environmental science; psychology; and wildlife, fisheries, and conservation biology). On average there were 12.6 (SE ± 2.5) class sessions observed per department. Demographic information about the types of courses and instructors observed is included in Figure 1. Data from 97 observations from Spring 2014 were reported in an earlier study (Smith et al., 2014).
All faculty members who agreed to be observed were given a human subjects consent form. Approval to evaluate teacher observation data of classrooms (exempt status, protocol no. 2013-02-06) was granted by the institutional review board at the University of Maine. Because of the delicate nature of sharing observation data with other faculty members and administrators, the consent form explained that the data would only be presented in aggregate and would not be subdivided according to variables such as department. Faculty members were given access to observation data from their own course(s) upon request after we collected observation and survey data for this study. In total, 68% of the observed faculty members requested their data and met with a professional development coordinator to discuss the results.
Selection and Training of Middle and High School Teachers
Thirty-eight teachers from the state of Maine conducted the classroom observations. To record instructional behaviors in the classroom, the middle and high school teachers were trained to use the Classroom Observation Protocol for Undergraduate STEM (COPUS) according to the training procedure outlined by Smith et al. (2013). COPUS was adapted from the Teaching Dimensions Observation Protocol (Hora et al., 2013; Hora and Ferrare, 2014). Briefly, at the beginning of the training, the 25 COPUS codes and code descriptions (Table 1) were discussed with the teachers (sample COPUS protocol sheets can be found in Smith et al., 2013, and at www.cwsei.ubc.ca/resources/COPUS.htm). The teachers then practiced coding videos of classrooms and discussed codes that were not unanimously selected. In total, the training took approximately 2 h. Teachers conducted observations in pairs, and we calculated Cohen’s kappa scores to measure interrater reliability (Landis and Koch, 1977) for every in-class observation; more details are included in the next section.
Analyzing COPUS Data
Pairs of teachers observed classes and were instructed to record their COPUS results independently. Cohen’s kappa interrater scores were calculated for each observation pair to establish coder reliability. The mean Cohen’s kappa was 0.89 (SE ± 0.006) for all observations. Because some of the paired observations had low interrater reliability scores, we removed the six lowest paired observations, which all had Cohen’s kappa interrater scores below 0.650. The average Cohen’s kappa interrater score for the remaining 264 observations was 0.91 (SE ± 0.005), indicating strong agreement among paired observers (Landis and Koch, 1977). With strong reliability between coders, only codes that both observers marked during each 2-min interval were included in the data set analyzed for this study.
To determine the relative abundance of each COPUS code, we added the total number of times each code was marked and divided by the total number of codes, resulting in a percent of code. For example, if both observers marked instructor lecture (Lec) during the same 24 time intervals in a 50-min class period and marked 29 total instructor codes for the duration of the class, then 24/29 or 82.8% of the instructor codes correspond to lecture. Because it was difficult to visually represent and compare 25 COPUS codes in 264 different class sessions, we also collapsed the codes into four categories that describe what the instructor is doing and four categories that describe what the students are doing, as reported in Smith et al. (2014) and shown in Table 1. The percentage of each collapsed code was determined by adding the percent code of each individual code within the collapsed category. This collapsed-code representation is advantageous because it allows for a holistic view of multiple COPUS codes at the same time and facilitates comparisons across broad instructional approaches.
However, when trying to determine and compare the frequency of a single code, such as instructor lecturing (Lec) or student listening (L), percent-code calculations can be difficult to interpret, because multiple COPUS codes can be marked at the same time, which in turn can impact the denominator of the calculation (Lund et al., 2015). In particular, some codes are often marked together, such as instructor real-time writing (RtW) and instructor lecturing (Lec). Therefore, we also compared class sessions by calculating the percentage of 2-min time intervals in which specific codes, such as instructor lecturing (Lec) or student listening (L), were observed. The percentage of 2-min time intervals was determined by counting the number of 2-min time intervals in which each code was marked and then dividing that by the total number of time intervals that were coded. For example, if instructor lecture (Lec) was marked during 24 two-minute time intervals out of the possible 25 two-minute time intervals then 24/25 or 96.0% of the possible 2-min time intervals contained lecture.
Analyzing Instances of Clicker Use
For this study, we were particularly interested in class sessions that used clicker questions. At the University of Maine, instructors largely started implementing clickers in their classrooms in 2005 (Strukov, 2008), and the Faculty Development Center estimates that currently more than 5000 students are enrolled in courses that use clickers each academic year. To determine which class sessions used clickers, we looked for instances in which the instructor clicker question (CQ) code was marked by both observers during a single 2-min interval. The CQ code was identified in 80 of the 264 class sessions observed.
To find out more about how the clicker questions were used in each of the 80 class sessions, we looked for blocks of 2-min time intervals with individual or consecutive instructor clicker question (CQ) codes, and we called these “clicker episodes” (Figure 2). In total, 181 clicker episodes were observed, and the duration of each episode was determined by counting the number of consecutive 2-min time intervals marked with clicker question (CQ) codes. Overall, clicker episodes had a mean duration of 2.4 (SE ± 0.11) 2-min time intervals; therefore, the average clicker episode was less than 5 min (2 × 2.4 = 4.8) in duration. To determine how clicker questions were used during these episodes, we examined student behaviors during the same time intervals. Only two student behaviors were selected along with instructor CQ: individual thinking (Ind) and clicker group discussion (CG). Thus, there were three possible combinations of student behaviors during clicker episodes: Individual Thinking Only, Peer Discussion Only, and Individual Thinking and Peer Discussion Combined. For example, Figure 2 shows a class session with three clicker episodes, indicated by three segments of time with instructor clicker question (CQ) codes that are separated by one or more 2-min time intervals. The first clicker episode is solely characterized by individual thinking (Ind) student codes and is classified as Individual Thinking Only. The second clicker episode has clicker group work (CG) only and is classified as Peer Discussion Only. The third clicker episode has individual thinking (Ind) student codes followed by clicker group discussion (CG) codes, and is classified as Individual Thinking and Peer Discussion Combined.
Clicker Use in Class Sessions
Upon analysis of all clicker episodes, two broad class session modes were identified, those with peer discussion during clicker questions and those without peer discussion during clicker questions (Figure 3). Peer Discussion class sessions had at least one clicker episode with peer discussion, indicated by a clicker group discussion (CG) code with a corresponding instructor clicker question (CQ) code (Figure 4A). These class sessions may have also included student individual (Ind) or group work (WG, OG) codes, but the presence of at least one student CG code defined this mode.
The class sessions without peer discussion during clicker questions were further classified into two modes, Individual Thinking and Alternative Collaboration (Figure 3). Individual Thinking class sessions had no clicker episodes with peer discussion, and were thus characterized by the presence of instructor clicker question (CQ) codes paired with student individual thinking (Ind) codes (Figure 4B). In the Individual Thinking class sessions, students never discussed class material in groups. Alternative Collaboration class sessions had no clicker episodes with peer discussion, and therefore individual thinking (Ind) was the only student code that coincided with instructor clicker question (CQ) codes. However, Alternative Collaboration class sessions included worksheet-based group work (WG) or other group work (OG) at another point in the class period (Figure 4C). While students in Alternative Collaboration class sessions voted on clicker questions as individuals, just like students in Individual Thinking class sessions, observer notes suggested that the clicker questions were often tied to the group activities. Thus, students in some Alternative Collaboration class sessions had the opportunity to discuss relevant question material with peers, just not in the context of the clicker questions themselves.
To determine whether there is variation in how individual instructors use clickers at different times, we looked at clicker data for instructors who were observed teaching with at least two clicker episodes (n = 25 instructors). Specifically, we examined the percentage of clicker questions that had Individual Thinking Only, Peer Discussion Only, and Individual Thinking and Peer Discussion Combined. The percent was calculated by dividing the number of episodes in each category by the sum of all episodes for that instructor.
Observer Feedback
To collect additional information about class sessions and to give the middle and high school teachers opportunities to reflect on the instruction, we developed a feedback survey for the teachers to complete after each observation. The survey was developed during Summer 2014, piloted in Fall 2014, revised based on teacher interviews and written feedback, and implemented in Spring 2015. Observers completed the survey for each observation in pairs, discussing their reasoning for each answer. Discussions were audio recorded to monitor the usefulness of the survey and to ascertain whether or not additional clarification was needed on any items.
A section of this survey included a question about how challenging the clicker questions were for students. In general, the clicker questions were:
Challenging for students—the class vote was often split.
Easy for students—the majority of the students answered correctly.
Cannot determine—the instructor did not talk about the class voting results.
To help teachers answer this question, during training sessions, we watched training videos of instructors using clicker questions and showed teachers how the clicker system worked in detail by asking them a few clicker questions. We also demonstrated how the clicker results could be displayed to students and how the teachers could learn about how challenging the questions are for students based on the voting results. Because this survey was not fully implemented until Spring 2015, feedback is provided for 41 of the 80 clicker class sessions. Seventeen middle and high school teachers provided feedback on clicker class sessions.
All statistical analyses were performed using SPSS (IBM, Armonk, NY).
RESULTS
Characterizing Instructional Behaviors in Class Sessions Taught with and without Clicker Questions
To determine whether there are differences in instructional behaviors between classes taught with and without clickers, we first separated out class sessions that included clicker questions. Eighty of the 264 observed class sessions featured at least one clicker question (CQ) instructor code and were thus classified as clicker class sessions (Figure 3). We examined differences in class size and found that the average student enrollment in classes that use clickers (111 students) was significantly higher than that in classes that did not use clickers (72 students) (independent-samples t test, p < 0.05).
To have a holistic view of the COPUS codes at the same time in both clicker and nonclicker classes, we examined the distribution of collapsed-code percentages and found a range of classroom behaviors (Figure 5). Notably, the nonclicker class sessions range from 0 to 100% Instructor Presenting and Student Receiving, whereas the clicker class sessions have a narrower range from 13 to 87% Instructor Presenting and 19 to 89% Student Receiving.
Because percent-code calculations can be difficult to interpret for individual codes (see Methods for further details), we also examined the percentage of 2-min time intervals that included the traditional instructional codes such as instructor lecturing (Lec) and student listening (L). Nonclicker class sessions showed a broader range of percentage of 2-min time intervals that included instructor lecturing (Lec) and student listening (L), and higher median values for these two codes (Figure 6A). However, when comparing means (Figure 6B), there were no statistically significant differences between nonclicker and clicker class sessions (independent-samples t test, p = 0.4 for instructor lecturing and p = 0.2 for student listening).
Taken together, these results indicate that University of Maine STEM classes that use clickers (typically characterized by larger enrollments) displayed a narrower range of Instructor Presenting and Student Receiving collapsed-code behaviors compared with nonclicker class sessions. We also observe that clicker and nonclicker class sessions have a similar mean percentage of 2-min time intervals allocated to instructor lecturing (Lec). As a result, students in classes both with and without clickers spend a similar mean percentage of 2-min time intervals listening (L).
Documenting Variation in How Clickers Are Implemented at a Campus-wide Level
Because faculty may implement clickers in various ways, possibly in alignment with entirely different pedagogical strategies, we examined classes that used clickers in order to identify some common instructional modes. We first separated out class sessions in which the instructors allowed peer discussion during clicker questions from those that did not (Figure 3, further description in Methods), and called this first mode Peer Discussion. The majority of the Peer Discussion class sessions used a combination of both individual and peer discussion during clicker questions. For the class sessions that did not allow peer discussion during clicker questions, we subdivided the class sessions into Individual Thinking (no peer discussion during the entire class period) and Alternative Collaboration (no peer discussion during the clicker questions but peer discussion during other group activities in the class period). Comparisons of the collapsed instructor and student codes betweeen Peer Discussion, Individual Thinking, and Alternative Collaboration class sessions revealed a range of instructional behaviors (Figure 7), with the Alternative Collaboration class sessions showing the lowest abundance of Instructor Presenting and Student Receiving collapsed codes.
To determine whether the three different modes of clicker use impacted the percentage of time allocated to traditional instructional practices such as instructor lecturing (Lec) and student listening (L), we examined the percentage of 2-min time intervals that included these two codes. Peer Discussion, Individual Thinking, and Alternative Collaboration class sessions are all characterized by a range of percent 2-min time intervals that include instructor lecturing (Lec) and student listening (L), with the Individual Thinking class sessions showing the highest median values (Figure 8A). In addition, Individual Thinking class sessions had a significantly greater mean percentage of 2-min time intervals with instructor lecturing (Lec) and student listening (L) compared with the other two types of class sessions that use clickers (Figure 8B, one-way analysis of variance [ANOVA], Tukey’s post hoc test, p < 0.05, in both cases).
By definition, the Individual Thinking and Alternative Collaboration class sessions had no opportunities for clicker-mediated peer discussion, and the Peer Discussion class sessions contained at least one episode that included peer discussion (Figure 3). Because this definition does not account for possible variation in clicker use within the Peer Discussion class sessions, we also examined all 112 clicker episodes that occurred in the 44 Peer Discussion class sessions. Clicker episodes were classified into the following catagories: Individual Thinking Only, Peer Discussion Only, and Individual Thinking and Peer Discussion Combined (Figure 2). In Peer Discussion class sessions, 75% of the clicker episodes included the opportunity for students to talk to one another (Figure 9), with the most common practice including Individual Thinking and Peer Discussion Combined.
In Spring 2015, the teacher observers provided specific feedback for each class session they observed via an online survey, with a portion of this survey specifically focused on clicker use. In particular, the teachers were asked to provide information about how challenging the clicker questions were for the students based on the voting results (Figure 10). The survey responses suggested that Peer Discussion class sessions more commonly included questions that were challenging for students when compared with Individual Thinking and Alternative Collaboration class sessions.
Taken together, these results indicate there are three different predominant modes of clicker use: Peer Discussion, Individual Thinking, and Alternative Collaboration. Among the three different modes of clicker use, the instructors in the Individual Thinking class sessions are spending significantly more time lecturing (Lec) and the students are spending more time listening (L). Alternative Collaboration class sessions tend to include lower abundance of Instructor Presenting and Student Receiving collapsed codes, largely due to the nonclicker group activities. However, the presence of these activities does not result in significant differences in percentage of 2-min time intervals allocated to instructor lecturing (Lec) and student listening (L) between Peer Discussion and Alternative Collaboration class sessions. In addition to providing an opportunity for peer interaction, clicker questions asked during the Peer Discussion class sessions tended to include questions that provided greater levels of challenge to students.
Examining How Individual Instructors Vary Their Implementation of Clicker Questions
Because instructors can vary their instructional style, we also explored the variation in how individual instructors used clickers. For this analysis, we focused on instructors who were observed teaching with at least two clicker episodes, regardless of the type of episode described in Figure 2. This analysis included 25 instructors from nine different departments. Nineteen of these instructors incorporated at least one opportunity for peer discussion (Figure 11). However, nearly all of the instructors used Individual Thinking Only at some point in their instruction, and this strategy accounted for more than 50% of the episodes for 16 of the 25 instructors.
DISCUSSION
Here we discuss the first observation-based, multidisciplinary study of clicker implementation in STEM classes across a single campus. Observations of STEM classrooms revealed that nearly a third of class sessions used clickers (Figure 3). A comparison of class sessions with and without clickers showed that both types of classes had a large range of collapsed-code instructional behaviors (Figure 5) and instructors teaching with clickers allocate a similar percentage of 2-min time intervals to instructor lecturing (Lec) and student listening (L) when compared with the instructors of nonclicker classes (Figure 6). The results from our study appear to confirm the often articulated concern that adding clickers alone does not guarantee that instructors will spend more time overall on active-engagement, student-centered instruction.
We suspect that part of the reason our data do not show dramatic differences between class sessions that use and do not use clickers is because we observed three distinct modes of clicker use: Peer Discussion, in which students had at least one opportunity to talk with one another during clicker questions; Individual Thinking, in which no peer discussion was observed; and Alternative Collaboration, in which students had time for discussion, but it was not paired with clicker questions (Figure 3). Our results indicate that for the Peer Discussion class sessions, the majority included clicker questions that combine both individual thinking and group discussion (Figure 3) and that the questions tended to be challenging for students to answer (Figure 10). Furthermore, instructors in the Individual Thinking class sessions spent a significantly larger percentage of 2-min time intervals lecturing (Lec) and the students spent a larger percentage of 2-min time intervals listening (L; Figure 8B). Finally, the Alternative Collaboration class sessions tended to include fewer Instructor Presenting and Student Receiving collapsed codes (Figure 7), but the presence of these activities did not result in significant differences between percentage of time allocated to instructor lecturing (Lec) and student listening (L) between Peer Discussion and Alternative Collaboration class sessions (Figure 8B).
Although studies have shown that student clicker question performance increases during peer discussion when compared with individual thinking (Lasry et al., 2009), little is known about the potential student learning benefits of the Alternative Collaboration style of clicker use. Future work is needed to explore this mode and other types of clicker-supported group work reported in the literature (e.g., Kryjevskaia et al., 2014), especially in cases in which the clicker questions are used to check class understanding after group activities (e.g., Kolber et al., 2014; Smith and Merrill, 2014).
Inconsistencies in How Peer Discussion Is Used with Clickers
Even though our data indicate that there are a variety of ways clicker questions are being used, Individual Thinking was the only behavior in 28% of the clicker class sessions we observed (Figure 2) and was an instructional strategy used by the majority of instructors (Figure 11). Although these instructors have successfully overcome many obstacles to the implementation of clickers in their classrooms, emphasis on the Individual Thinking strategy may inadvertently limit possible student learning opportunities. For example, if peer discussion is omitted, students may lose the opportunity to build scientific communication skills that are developed by articulating reasoning, evaluating the merits of others’ reasoning, and asking peers questions (Turpen and Finkelstein, 2010). In addition, performance increases attributed to peer discussion are lost (Smith et al., 2009; Lasry et al., 2009; Porter et al., 2011; Knight et al., 2015; Barth-Cohen et al., 2016). Furthermore, when peer discussion is omitted, faculty members do not have the opportunity to circulate around the class and listen to student reasoning (Mazur, 1997).
In addition to the lost student and instructor learning opportunities, previous work has shown that students have more negative attitudes about the utility of clickers when they do not discuss the multiple-choice questions (Keller et al., 2007). Student resistance often impacts instructional decisions, and faculty may abandon clicker-supported instruction and other active-learning pedagogies promoted by discipline-based education researchers (Silverthorn, 2006; Henderson and Dancy, 2007). It has also been documented that faculty prioritize personal experience over empirical evidence when making decisions about teaching strategies (Andrews and Lemons, 2015), and the negative experiences associated with nonoptimal clicker implementation may therefore have a long-lasting impact on future instructional decisions.
How Can We Encourage Faculty to Include Peer Discussion with Clicker Questions?
While research has shown there are benefits to allowing students to talk to one another during clicker questions (Smith et al., 2009, 2011; Porter et al., 2011; Knight et al., 2015; Barth-Cohen et al., 2016), the observed variation in use of the peer discussion portion of clicker implementation is consistent with findings for other research-based pedagogies, which are typically changed and modified during implementation by faculty (Henderson and Dancy, 2007; NRC, 2013). Given that faculty often modify research-based instructional practices, how can we enhance professional development to make sure the peer discussion portion of clicker use is retained during implementation?
One response is to make sure faculty professional development motivates and targets the more nuanced aspects of effective clicker implementation. Considering that 86% of U.S. college faculty recently reported they were familiar with clickers (FTI Consulting, 2015), the majority of professional development audiences likely have some working knowledge of clickers. As such, it is important to move beyond dedicating an entire session to the basics of using a clicker system. Moreover, rather than polling professional development participants about whether or not they have used clickers before, the audience can instead be asked how they use clickers during instruction and can be asked to discuss with their neighbors and report to the group. This approach may reveal innovative ways in which people are using clickers, provides a more detailed picture of participant experience, and can serve as a launching point for motivating the value of peer discussion by drawing upon a combination of findings from research studies (Smith et al., 2009, 2011; Porter et al., 2011; Barth-Cohen et al., 2016) and engaging participants in activities designed to help them identify the features of clicker questions that encourage productive peer discussion.
In addition, recent work has shown that onetime faculty professional development workshops have a limited capacity to create change (Davidovitch and Soen, 2006; Henderson et al., 2011). Instead, faculty need ongoing, in-depth professional development and support (Henderson et al., 2011; PCAST, 2012). For this reason, at the University of Maine, we have started a yearlong faculty professional development program in which faculty meet in rotating pods of three: one individual teaches, one individual observes using the COPUS, and another provides feedback on areas identified in advance by the instructor. Notably, 89% of the faculty members participating in this program have said that encouraging student peer discussion is one of the predominant areas in which they would like assistance, and it will therefore be an ongoing focus of this program.
Finally, there is also a need for clicker question banks that are vetted by the community and include questions that have been shown to encourage productive peer discussion. The work described here indicates that instructors who are using peer discussion in their classes are asking questions that are challenging for students to answer (Figure 10), which are often time-consuming for instructors to write. In addition to presenting clicker questions, a question bank could also include aggregate student voting results, instructor reflections on how to most effectively follow up when student voting results are split among multiple answers, videos of students discussing the clicker questions with one another, and follow-up homework and exam questions that target the concepts from the clicker questions. These supplemental materials, in particular, could 1) foreground the ways in which clicker questions may be used to facilitate student learning, 2) provide some of the scaffolding needed to support effective implementation, and 3) serve as a flexible resource that faculty may adapt based on the needs of their classrooms.
CONCLUSION
Our campus-wide, observation-based study of clicker implementation in STEM classrooms revealed that instructors who used clickers demonstrated variation in implementation, with many instructors eliminating peer discussion during some if not all clicker questions. Omitting peer discussion impacts students’ ability to articulate their reasoning and to work together to solve problems. In addition, instructors who omit peer discussion lose the chance to listen in on student reasoning and may encounter more student resistance to research-supported instructional techniques. To encourage faculty to include peer discussion, we recommend 1) focusing on peer discussion as an essential component of long-term clicker professional development programs that include multiple opportunities for faculty to learn about using peer discussion with clickers, and 2) establishing clicker question banks that include challenging, higher-order questions for faculty to adapt to their instructional needs.
ACKNOWLEDGMENTS
The authors appreciate the contributions of the middle and high school teachers who served as observers, including Elizabeth Baker, Toni Barboza, Tristan Bates, Michele Benoit, Stacy Boyle, Virginia Brackett, Elizabeth Connors, Tracy Deschaine, Lauren Driscoll, Andrew Ford, Bill Freudenberger, Kathryn Priest Glidden, Marshall Haas, Kate Hayes, Beth Haynes, Teri Jergenson, Danielle Johnson, Bob Kumpa, Lori LaCombe-Burby, Melissa Lewis, John Mannette, Nichole Martin, Lori Matthews, Nicole Novak, Jennifer Page, Patti Pelletier, Meredith Shelton, Beth Smyth-Handley, Damian Sorenson, Joanna Stevens, Rhonda Stevens, Nancy Stevick, Thomas White, and Michael Witick; and the University of Maine faculty members who welcomed the teachers into their classes for observations. We also thank Ken Akiha, Karen Pelletreau, and Mindi Summers for helpful feedback on this article. This material is based on work supported by the National Science Foundation under grants 1347577 and 0962805.