Adam G. L.
Schafer
and
Ellen J.
Yezierski
*
Miami University, Department of Chemistry and Biochemistry, Oxford, OH, USA. E-mail: yeziere@miamioh.edu
First published on 17th December 2019
High school chemistry teachers struggle to use assessment results to inform instruction. In the absence of expert assistance, teachers often look to their peers for guidance and support; however, little is known about the assessment beliefs and practices of high school chemistry teachers or the discourse mechanisms used as teachers support one another. Presented in this paper are the results from analyzing a discussion between five high school chemistry teachers as they generated a set of best practices for inquiry assessments. To analyze the discussion, a novel representation called a discourse map was generated to align the analyses conducted on chemistry teacher discourse as they temporally occurred. Results show the utility of the discourse map for evidencing critical friendship and assessment practices evoked by the teachers during the discussion of best practices. Implications for the structural considerations of materials and chemistry teacher professional development are presented as well as potential future investigations of teacher discourse regarding the use of data to inform instruction.
During long-term professional development, teachers have the opportunity to build trust and comradery with critical friends. Trust and experience can lead to new learning opportunities between critical friends as they engage in critical discussions of beliefs and practices (Dunne and Honts, 1998; Curry, 2008; Attard, 2012). Critical friends’ discussions may include conflict and differing perspectives that can stretch beyond their comfort zones, but critical friends are generally able to resolve conflict by collegial means (Snow-Gerono, 2005; Curry, 2008; Attard, 2012). The familiarity of other critical friends’ classroom environments and practices promotes discourse that moves beyond superficial topics to underlying theory that can improve pedagogy. By creating an environment of trust and collective growth, teachers can discuss their chemistry students and chemistry-specific instructional practice in challenging ways. Pairing the real-life experiences of teachers with research and practical inquiry is the most productive integration of educational theory and practice (Richardson, 1996). By engaging critical friends in discussion about chemistry assessment, the nuances of teacher–teacher discourse can be explored to uncover teachers’ chemistry assessment beliefs and practices.
Category | Description (category investigates…) |
---|---|
Discourse moves | The nature of the discursive exchanges |
Logical processes | The logical relationship between contributions to discourse as a social interaction |
Nature of explanation | How the contribution to discourse was constructed |
Cognitive strategies | How the individual approached the contribution they made to discourse |
Discourse moves shed light on the participatory roles of an individual during social interaction by characterizing their conversational turn (Kaartinen and Kumpulainen, 2002). When engaging in social interaction, an individual may take a conversation turn by making a discourse move, such as proposing an idea or continuing conversation based on the idea of a peer. For example, a teacher may initiate a conversation about how best to assess a concept a teacher may ask their peers how they assess the concept. The act of initiating a conversation can be viewed as a single conversational turn.
Logical processes are related to the relationship between conversational turns and how social interaction contributes to collective understanding (Kaartinen and Kumpulainen, 2002). Essentially, logical processes characterize what occurred during a conversational turn. For example, a person who initiates a conversation can do so in a number of ways, such as asking a question or proposing an idea.
The nature of explanation characterizes how the discourse during a conversational turn was constructed (Kaartinen and Kumpulainen, 2002). For example, if a person were to initiate a conversation by asking a question they may do so formally (such as when a student asks their teacher) or informally (such as when a student asks a friend). Although the nature of explanation can change as an individual interacts with other individuals, it can also differ based on the content an individual is discussing or the type of conversational turn they are engaging in. For example, a student may ask a question to their friend in a more formal tone to ensure precision of language but respond to questions from their friend informally.
Cognitive strategies are related to how an individual frames their contribution to social interaction (Kaartinen and Kumpulainen, 2002). Similar to logical processes, cognitive strategies seek to characterize what is occurring during a conversational turn. However, whereas logical processes characterize the relationship between conversational turns, cognitive strategies characterize the relationship of the individual taking the turn to the conversational turn. Cognitive strategies essentially describe how the individual engages in social interaction (Kaartinen and Kumpulainen, 2002). For example, if a student initiated a conversation by asking a question, they could do so by relating to an everyday experience or asking if a concept from class applies in a different situation. Further description of how the Logic of Inquiry framework was used in the study may be found in the Methods section.
(1) To what extent does discourse from a facilitated discussion about assessment best practices reveal characteristics of critical friendships between high school chemistry teachers?
(2a) What best practices about assessment are revealed through a facilitated discussion about best practices for chemistry assessments among high school chemistry teachers?
(2b) How do teachers’ reports of best practices revealed through a facilitated discussion about best practices for chemistry assessments align to those cited in the literature?
(3) What are the chemistry-specific features of the best assessment practices revealed during a facilitated discussion about best practices for chemistry assessments among high school chemistry teachers?
Participant | Years of experience | School size (# students) |
---|---|---|
Anne | 11 | 400 |
Ashton | 25 | 1600 |
Celine | 18 | 650 |
Claude | 16 | 450 |
Emmerson | 10 | 750 |
Both authors facilitated the best practices discussion by asking the teachers to generate assessment practices using the following prompts:
(1) What are the characteristics of a quality assessment?
(2) How do you know the assessment is high quality?
(3) How do you know if the assessment met the goals of the lesson?
While one facilitator mediated the best practices discussion with relevant follow-up questions, another worked as a scribe to record proposed assessment practices by writing them on the board. The scribe consistently conducted member checking with the teachers by asking clarification questions to ensure that written assessment practices accurately reflected the teachers’ ideas. Discussion continued on the first question until the teachers stated they had nothing left to add to their list of practices, at which point the second question was asked. This iterative process continued for about an hour, until the teachers had generated three lists that summarized what they believed to be essential assessment practices for inquiry activities. After the professional development day concluded, the practices were summarized into a set of guidelines for constructing assessments, evaluating assessment quality, and evaluating alignment between the assessment and lesson goals. The guidelines were given back to the teachers during the second professional development day to confirm that the guidelines aligned with the teachers’ original intents. Returning the best practices to the teachers to confirm that the identified practices align to the teachers’ original intents serves as evidence for trustworthiness through member checking (Maxwell, 2013).
The first research question seeks evidence of critical friendship among secondary chemistry teachers. To address this question, several methods of tabulating and organizing frequencies of code applications were considered; however, few offered the opportunity to characterize code occurrences over time. Summaries of overall code applications and cross-tabulations of code applications did not meet the standard for the first research question; however, these were essential benchmarks for addressing the first research question. To simultaneously examine all coding schemes and to represent the temporal relationship among code occurrences, a new representation was developed called a discourse map. Line numbers from the transcript were used to track code location on the discourse map. Code applications were organized using a spreadsheet with each column representing a line of the transcript and each row representing a participant.
The second research question investigates what chemistry assessment best practices were revealed during a facilitated discussion of best practices and how such practices align to high-quality assessment practices cited in the literature. Best practices generated by the teachers were compared to published papers that communicate what teachers should consider when generating assessments, evaluating assessment quality, and determining alignment between the assessment and the lesson goals.
The third research question investigates the chemistry-specific features of best assessment practices revealed during the discussion. To address this question, the discourse contributing to the development of each assessment practice was analyzed for chemistry-specific considerations the teachers discussed when generating their set of best practices.
Code | Ashton | Celine | Claude | Anne | Emmerson | Total |
---|---|---|---|---|---|---|
Initiating | 38 | 31 | 6 | 2 | 0 | 77 |
Continuing | 125 | 118 | 98 | 18 | 1 | 360 |
Agree/disagree | 85 | 56 | 65 | 59 | 40 | 305 |
Replying | 5 | 4 | 1 | 1 | 0 | 11 |
Concluding | 0 | 0 | 5 | 0 | 0 | 5 |
Referring back | 2 | 6 | 6 | 1 | 0 | 15 |
Commenting | 4 | 10 | 6 | 9 | 0 | 29 |
Total | 259 | 225 | 187 | 90 | 41 | 802 |
Emmerson and Anne made fewer contributions compared to their peers (41 and 90 discourse moves). Ashton, Celine, and Claude each more frequently engaged in discourse throughout the best practices discussion (187–259 moves). Since Emmerson and Anne did not contribute as frequently as the others, the assessment practices proposed by these teachers may not align as well to their beliefs. However, Emmerson and Anne's consistent agreement with assessment practices proposed by other teachers is evidence that the practices proposed align to their own. Previous studies about critical friends speak of a dual nature of the relationship (Dunne and Honts, 1998; Schuck and Russell, 2005; Taylor and Storey, 2013). Critical friends support the growth and development of their peers while also often recognizing discussion with critical friends as an opportunity to reflect on their own practice (Dunne and Honts, 1998; Swaffield, 2004; Schuck and Russell, 2005; Taylor and Storey, 2013). While Emmerson and Anne were not vocally contributing to discourse with their peers, they were often writing in their notebooks. Since they were often writing, Emmerson and Anne still seemed engaged.
Nature of explanation codes | ||||||
---|---|---|---|---|---|---|
Action-oriented | Descriptive | Causal | No NoE | Total | ||
Logical processes codes | Provides reasoning | 15 | 41 | 48 | 1 | 105 |
State a goal | 15 | 2 | 3 | 0 | 20 | |
State a result | 3 | 20 | 15 | 0 | 38 | |
Refines | 8 | 23 | 2 | 0 | 33 | |
Presents | 20 | 28 | 14 | 5 | 67 | |
Evaluates | 6 | 8 | 10 | 6 | 30 | |
Contradicts | 0 | 5 | 1 | 1 | 7 | |
Gives example | 44 | 33 | 21 | 2 | 100 | |
Total | 111 | 160 | 114 | 15 | 400 |
Logical processes codes were applied to verbal discourse to identify the relationship between discursive moves. The total counts for the logical processes show that a majority of teacher interactions involved either “providing reasoning” (105 applications) or “giving examples” (100 applications). The “provides reasoning” code was applied to statements of evidence or reasoning for performing a practice or including an assessment characteristic.
The nature of explanation coding scheme helped researchers examine the way statements were constructed. “Action-oriented” codes (111 applications) were applied to statements that encouraged an identifiable action to be performed, such as what to do when receiving confusing results from assessments (Celine: What you have to do is reassess the goal.). “Descriptive” codes (160 applications) were applied to statements that about the qualities of a practice. For example, when clarifying an assessment practice to the group Celine described an overall assessment goal that …it's not about facts, it's about understanding. “Causal” codes (114 applications) were applied to statements that stated a cause-effect implication to the implementation of a practice, such as when Claude was interpreting results from an assessment: …like oh my goodness only 20% of my kids got this right, it must be a bad question. “Descriptive” statements were the most frequently encountered nature of explanation code (160 applications), with only 15 statements receiving “no nature of explanation” code (abbreviated as No NoE in Table 4).
Conducting a statistical test of the frequencies of code applications is inappropriate, since the method of extracting code occurrence results in statements that vary in length. Examining code frequencies qualitatively reveals overall patterns regarding the nature of teacher discourse about assessment development. For example, much of the discourse involved providing context for teacher assessment practices (giving examples) and defending presented practices to critical friends (providing reasoning). Since teachers construct much of their discourse through examples with evidence, they may benefit from materials situated in the classroom. For example, providing a rich description of the environment that pedagogical practices were enacted so that teachers can better reason with how they would apply to their own classrooms. The high frequency of the nature of explanation codes illustrates the variety of support and reflection throughout discussion as teachers consider their own practices and encourage the development of their peers’ practices.
Table 4 shows that teachers more frequently used “action-oriented” (15 co-occurrences) statements than “descriptive” (2 co-occurrences) or “causal” (3 co-occurrences) statements when discussing goals. Alternatively, when discussing the results of implementation teachers were more frequently using “descriptive” (20 co-occurrences) or “causal” (15 co-occurrences) statements than “action-oriented” (3 co-occurrences) statements. The contrast between the nature of explanation code occurrences for “stating goals” and “stating results” can possibly be explained by patterns observed in literature about the use of assessment data to guide instruction. The use of “action-oriented” statements while discussing goals is likely due to the fact that teachers attempt to design assessments to investigate measurable outcomes (Pellegrino, 2012; Sandlin et al., 2015). Alternatively, literature about using assessment results to guide day-to-day instruction is often coarse-grained and does not provide specific guidance on how to use assessment results to guide instruction (Knapp et al., 2005; Irons, 2008; Hamilton et al., 2009; Suskie, 2009; Witte, 2012). The lack of materials about how to use assessment results could negatively influence the teachers’ use of “action-oriented” statements while discussing assessment results.
Although investigating individual (and paired) code occurrences provided insight about the overall nature of critical friendship, a different representation of the data was necessary to reveal features of critical friendship over time. A novel representation of all three coding schemes as applied to all participants, called a discourse map, was used to synchronize data among logic of inquiry coding schemes for each participant, while also representing the code occurrences temporally. A sample selection from the discourse map is shown in Fig. 2. The complete discourse map for the best practices discussion is found in the Appendix 2.
Each cell of the discourse map may contain a symbol, color, and/or shading gradient to indicate code application(s) in that line in the transcript. If a cell is empty, the person in that row was not participating in discourse for that line of the transcript. Discourse moves codes are indicated by symbols, shown in Fig. 1. Logical processes codes are indicated by colors, shown in Fig. 1. Nature of explanation codes are indicated by shading gradients, shown in Fig. 1.
Examination of the discourse map reveals that the codes for “agree/disagree” and “continuing” occur more frequently than others. The “agree/disagree” code was applied when teachers verbally or nonverbally indicated that they agreed or disagreed with another teacher or facilitator. The only non-verbal forms of agreement coded were when a teacher nodded in direct response to a statement made by another teacher. No nonverbal forms of disagreement were detected. In instances when a teacher would nod and provide a verbal remark only one code of “agreement” was applied. The “continuing” code was applied when one teacher extended discourse about the same idea presented by the previous teacher. The high frequency of code applications for the “agree/disagree” and “continuing” codes speak to the relationship among the individuals in the best practices discussion. All teachers have participated in multiple years of professional development together and seemed willing to disagree with their peers or continue the conversation with a differing perspective. For example, some teachers felt they had time to walk question-by-question with students to provide feedback about lab experiments, while others communicated that their school environment necessitated that they find less time-consuming ways to provide feedback to students. Even through disagreement, the teachers were supportive of each other's practices, recognizing the environmental differences that allow for pedagogical differences. Literature on the role of critical friends states that critical friends serve to support and critique each other (Schuck and Russell, 2005; Curry, 2008; Taylor and Storey, 2013; Moore and Carter-Hicks, 2014; Fletcher et al., 2016). Schuck and Russell reflected that the value of critical friendships stems from encouraging the reconsideration of practices and creating space and opportunity to nourish that reconsideration (Schuck and Russell, 2005). The high frequency of “agreement/disagreement” and “continuation” between teachers during the best practices discussion is likely an illustration of the support and ongoing consideration of the beliefs and practices of their peers.
The discourse map shows that a majority of teacher interactions involved either the “provide reasoning” or “giving examples” logical processes. The “provide reasoning” code was applied to statements of evidence or reasoning for performing a practice or including an assessment characteristic. Teachers provided reasoning in various contexts, such as crafting assessments (Celine: The reason it worked is because it forced me to pay attention to what they wrote.), interpreting assessment results (Claude: Like oh my goodness only 20% of my kids got this right…), and the benefits of critical friends (Ashton: …because I gotta be honest with you, for me writing questions is painfully hard.). “Gives example” codes were applied to actual or hypothetical examples teachers stated to illustrate an idea. Teacher examples were often very long and situated discourse in the context of the classroom. For instance, when Claude was talking about goals for his test questions, he stated:
I have a test question where I give data uhm and the data is of four alkaline earth metal salts and what you observe when it's in water. So, one is dissolved and one is a little cloudy and one is thick precipitate. A thick white mix and then I give I think it was three and then I give a fourth and I say with question marks and I say What are you gonna see? Alright, and they have to make a prediction based on periodic trend. And then, I ask a question and I've been fighting with this question for years. I wanna get to the point whether of ‘is this a trend of the atom or is this a trend of the ion formed by that metal?’
Critical friends often require context to adequately discuss each other's aims and perceptions regarding their practice (Swaffield, 2004; Schuck and Russell, 2005). Once critical friends are oriented to their peer's perspective, they encourage each other to think deeper about their practice, giving opportunity for reasoning and examples to arise that emerge through reflection, thus explaining the frequency of lengthy examples.
One of the main benefits of the discourse map is the ability to investigate for discourse patterns over time. For example, examination of the discourse map reveals that teachers often transitioned between several logical processes and/or natures of explanation when elaborating on a single discourse move. In Fig. 3, Ashton is contradicting a statement made by one of the other teachers (indicated by the red shading of cells 146–149). Ashton uses a descriptive explanation for contradiction (shown as vertical shade gradient). Potentially as an alternative, Ashton proposes (brown shading) an assessment characteristic or practice at the conclusion of his remarks. In Fig. 4, Claude is continuing the discourse about an assessment practice proposed by another teacher with an example (blue shading). Claude begins with a descriptive example (vertical gradient) which slowly transitions to a causal explanation (horizontal gradient). Throughout the course of his contribution, Claude's statement even contains a descriptive goal (vertical, purple shading) embedded within his example.
Segments of complex discourse like those shown in Fig. 3 and 4 are able to be identified using the discourse map. Critical friend groups offer space and opportunity to reconsider the aims and practices behind day-to-day instruction that is often a rarity for teachers (Dunne and Honts, 1998; Swaffield, 2004; Schuck and Russell, 2005; Fletcher et al., 2016). Similar to the complexities of their day-to-day instruction, the knowledge and reasoning guiding their practice is likely equally complex. The value of critical friendships is rooted in the context of day-to-day instruction because the beliefs guiding assessment practices are situated with other experiences or knowledge (Dunne and Honts, 1998; Swaffield, 2004; Schuck and Russell, 2005). For instance, the discourse map shows that classroom examples are provided within every segment of discourse and are rarely shorter than three lines of transcript. The frequent use of examples is representative of the importance of context as teachers consider how to change or improve practice.
Overall, the discourse map represents the discourse in a manner that allows for the characterization of discourse trends that are consistent with those of critical friends (Dunne and Honts, 1998; Swaffield, 2004; Curry, 2008; Baskerville and Goldblatt, 2009; Taylor and Storey, 2013). The number of years participating in professional development together serves as additional evidence that these teachers are critical friends. The critical friendship between the teachers participating in the best practices discussion is embodied in the depth and vulnerability of the ideas and practices shared during the discussion (Dunne and Honts, 1998; Schuck and Russell, 2005). The key findings in the study thus far are summarized in Table 5.
Research question | Data sources | Key findings |
---|---|---|
(1) To what extent does discourse from a facilitated discussion about assessment best practices reveal characteristics of critical friendships between high school chemistry teachers? | Tables 3 and 4, discourse map | • Teachers in this investigation have discourse patterns aligned with critical friends as described in the literature |
• Discourse map was more useful than overall code frequencies for revealing discourse characteristics |
Assessment component | Best practice |
---|---|
Creation considerations | a. Assessment items must align to learning goal |
b. State the learning objective in terms that explicitly state what the student should explain/do | |
c. The learning objective should be clear and provided to the students | |
d. The assessment should address different levels of knowledge within one instrument | |
e. Identify the appropriate “level” for demonstrating competency of the learning objective | |
i. The level should consider both Bloom's Taxonomy and Johnstone's Triangle | |
f. Identify the type of data needed from students to give good evidence of understanding | |
g. Focus on conceptual understanding, not just recall of facts | |
h. Assessment items should consider student prior knowledge | |
Evaluation of quality | a. Clarity |
i. Students should understand the point of the item | |
ii. Item should not be “tricky” and should actually get at what you want | |
iii. Not a reading comprehension item, but a chemistry item | |
1. Present no more than 3 ideas per sentence | |
2. At least one idea connects to next sentence | |
3. Consider removing extraneous information | |
iv. Strong links to what has been emphasized during instruction | |
b. Ease of grading | |
i. Needs to be relatively easy to grade | |
c. Consider if the item detects common misconceptions. | |
i. Use both formative and summative to inform teaching and detect misconceptions early | |
Evaluation of meeting goals | a. In thinking about a collection of items you should see a diversity of levels |
i. The level should consider both Bloom's Taxonomy and Johnstone's Triangle | |
b. When comparing student performance on items to goals | |
i. Need to triangulate evidence from multiple items to draw conclusions regarding instruction | |
i. One sketchy item may be a problematic item | |
ii. Verbal student feedback can inform how item is being interpreted by students | |
iii. Consider the distribution of scores on items and who gets the items correct | |
c. Expert validation | |
i. Get insights and evaluation of items from other teachers |
The generated best practices for assessment contain a mix of simple-to-apply rules (e.g., limiting the number of ideas per sentence to a maximum of three and removing extraneous information) as well as broad guidelines without specific actions to follow (e.g., the imperative to align the assessment items to a learning goal). Whether the best practice stated specific actions to follow or not, this set of guidelines aligned to practices cited as high-quality considerations in assessment literature.
To discuss the alignment to assessment literature, the teacher-generated best practices are summarized as three themes:
(1) The importance of assessing for conceptual understanding
(2) Including clear, direct assessment item stems and learning goals
(3) Assessing learning goals multiple times over a variety of conceptual and representational levels
Teachers’ guidelines stating the importance of assessing for conceptual understanding align with several investigations of high-quality assessment practices (Black and Wiliam, 1998; National Research Council, 1999, 2001, 2014; Bell and Cowie, 2001; Stiggins, 2001; Gibbs and Simpson, 2004; Lyon, 2011). The National Research Council states that providing students with assessment opportunities to reveal what they have learned and understood to themselves, their peers, and instructors is more beneficial to achievement than simple recall (National Research Council, 2014). Embedded within the literature is the acknowledgement that teachers are tempted to include assessment items that assess recall (or statements of facts) over items that require application and conceptual knowledge. Recall items can be simple to generate, obtain from assessment sources, and grade, but often do not intellectually engage the student enough to collect adequate data about the students’ true understanding (Stiggins, 2001; Towns, 2014). Advice from the literature says that writing high-quality assessment items can be difficult, and teachers should take advantage of evidence-based resources as well as their peers to generate high-quality, conceptual items (Dunne and Honts, 1998; Towns, 2014).
When discussion about best practices for assessment began, the teachers quickly leapt to the importance of clarity and transparency of the learning goals guiding assessment. Generating clear learning goals, including them in the assessment process, and articulating them to the students are very prominent practices in the literature (Stiggins, 1988, 2001; Black and Wiliam, 1998; Bell and Cowie, 2001; Gibbs and Simpson, 2004; Sato et al., 2008; Hamilton et al., 2009). Learning goals are an expectation for well-crafted learning activities, and a significant body of literature exists depicting the importance of crafting clear, concise, and measurable goals used to link assessment and instruction. The teachers in this study were very aware of the benefits of well-crafted and usable learning goals as evidence by the guidelines they generated.
The third theme from the teacher-generated best practices involved the use of multiple measures of student knowledge to improve the precision of claims made about student understanding. Similar to the other two themes, the idea of assessing learning goals multiple times through a variety of conceptual levels aligns with literature for desirable assessment practices (Black and Wiliam, 1998; Bell and Cowie, 2001; Stiggins, 2001; Gibbs and Simpson, 2004; Hamilton et al., 2009; Lyon, 2011; National Research Council, 2014; Towns, 2014). The use of multiple measures is recommended from several sources, including the National Research Council which calls the use of multiple, related questions with a variety of tasks to assess a single learning goal “multicomponent tasks” (National Research Council, 2014). Studies promoting the use of multiple measures state that the additional data allow the teacher and student to gain insight to student-specific challenges as well as knowledge gained (Black and Wiliam, 1998; Bell and Cowie, 2001; National Research Council, 2001; Stiggins, 2001).
Claude: And so, if I can get away from that question for a second because something's not working, and I've tweaked my instruction.
Celine: Sure. Right, right.
Claude: To try to emphasize that, and it's not working. And so, then I'm like in my brain. If 70% of my kids are missing a question. If 70% of my kids aren't getting the question right. And this may go all the way back to college, right? A 70% a 70 on a chemistry test in college was a low B or a C right?
Anne: That was an A!
Claude: And uhm. If uhm you get the whole bell curve whatever. If 70% of my kids aren't getting it right. Somethings wrong. Somethings wrong with the question something's wrong with the instruction. Something's wrong with something. Uhm. My question is, you know when only 20% are getting it right, like okay, not only is something wrong, something's really wrong. Right? uhm and so what do I do with that? Is that a point where I say, uhm you know, this is a losing battle I can't win it no matter what I do? Is this the point where I say I've gotta change and spend half a week just on the difference between the symbolic?
Although teachers recognize the importance of using multiple measures to investigate student competency, they do not always know what to do with the data they obtain, shown as Claude discusses triangulating evidence.
Claude expresses struggles with understanding how to use results from assessments to guide his instruction. In his remarks, Claude states that he can clearly tell that “something's really wrong” when his students underperform on an assessment but lacks the certainty to identify what that “something” is. Claude poses many questions that illustrate his frustration, leading to the idea that he is “fighting a losing battle.” If 70% of students get an item incorrect, Claude believes that an instructional intervention is needed, but he cannot interpret the 70% value to enact class-level decisions. Previous research showed that when faced with poor student performance, teachers stated they would reteach or re-cover the content, valuing repetition as a means of closing knowledge gaps (Harshman and Yezierski, 2015). Here, Claude is drawing the conclusion that “something” did not work. He believes that a different instructional method is required and that simply repeating the learning activity would possibly result in a similar student performance. Claude's inability to rely on student data to guide instruction was a gap in assessment practice shared by all teachers in the best practices discussion. A recent review found that using assessment results to guide day-to-day instruction is not addressed in current literature (Harshman and Yezierski, 2017). Often sources state that teachers should use assessment results to guide instruction but lack the guidance for how teachers could use data from assessment results to guide instruction (Heady, 2000; Martone and Sireci, 2009; Clinchot et al., 2017). Other sources state that teachers need content-specific professional development opportunities to learn how to use data in their classrooms (Smith, 2013; Herrington and Yezierski, 2014; Banilower et al., 2018). This gap in the literature aligns with the gap in best assessment practices exhibited by the teachers in the best practices discussion. Table 7 summarizes teachers’ best assessment practices uncovered during the teacher discussion.
Research question | Data source | Key findings |
---|---|---|
(2a) What best practices about assessment are revealed through a facilitated discussion about best practices for chemistry assessments among high school chemistry teachers? | Table 6. teacher-generated best practices for assessment | • Teacher-generated practices contain mix of broad guidelines and simple-to-apply rules |
• Teacher-generated practices organize into themes aligned with relevant assessment literature | ||
(2b) How do teachers’ reports of best practices revealed through a facilitated discussion about best practices for chemistry assessments align to those cited in the literature? | • Discourse of assessment practices revealed gaps in teacher practices not addressed by literature |
Claude: So here's the issue, right? I have a test question where I give data and the data is of four alkaline earth metal salts. And what you observe when it's in water. So, one is dissolved and one is a little cloudy and one is thick precipitate. A thick white mix and then I give, I think it was three, and then I give a fourth and I say with question marks. And I say, “What are you gonna see?” Alright, and they have to make a prediction based on periodic trend. And then I ask a question and I've been fighting with this question for years. I wanna get to the point whether of is this a trend of the atom or is this a trend of the ion formed by that metal? Right? and in such in such expression that's what I ask. Okay? because quite honestly for 95% percent of my kids can make the prediction because I've even got it arranged as they are on the periodic table. Uhm, but the last question I'm trying to get at: are these the metals or are these the ions? Uhm and tell me how you know. And
Anne: You're talking the trend is the trend for the metal or for when it's an ion?
Claude: Is the trend is the are we working with magnesium metal, calcium metal, etcetera or are we working with magnesium ion, calcium ion, going down the list. Uhm and inevitably 80% of my kids tell me it's of course it's the metal. Right?
Facilitator: Even though there's no metals in the system.
Claude: Even though there's no metals there! And we've done the experiment. They've seen this. They know they're not working with the metal, but its but this is on my semester exam, whatever. Uhm and I'm like okay. So, I've taught it, we've talked about it, they've worked with it, and they still are missing it. Why? is it because the question in some way is confusing? Is it because I didn't do a good? Is it because, you know why are they missing this question so badly? Uhm and I just don't know, and I don't know how to answer that.
Claude raises uncertainty how to interpret assessment results, as was evidenced before, but Claude shares chemistry-specific details with his critical friends. Claude recognizes that he can share his struggles teaching and assessing periodic trends, whether it stems from his instruction, assessment, or students. Non-chemists would likely need further information about periodic trends to engage in a discussion of pedagogic struggles involving periodic trends, and/or related content. Claude's quotation exemplifies how chemistry teachers could benefit by having discipline-specific critical friends who would understand their chemistry-specific assessment and instructional problems.
Although domain-general discourse was far more common, chemistry-specific conversations occurred during the best practices discussion. For example, when discussing the importance of assessing for conceptual understanding, Claude shared,
Anne: Avoiding over-emphasis of facts but that's, I think that could be added.
Claude: Yeah
Celine: Yeah, it's not about facts it's about understanding.
Anne: Yeah.
Celine: And, sometimes we get caught up in the trap of it's an easy question, everybody should be able to answer it
Claude: Yeah, I remember the first:
Celine: And if it's a bad question, that's meaningless.
Claude: The first test I got, I ever looked at when I started teaching about atomic structure. Uhm it was 90% counting protons, neutrons, electrons, right? Uhm. Why would you do that?
Ashton: Right?
Claude: Why would you do that? You test that about, you can do that in about 4 or 5 questions and then let's get to actually what the, what we know about the atom as opposed to just that because it's easy to do next. I mean that's the thing, its super easy to test.
Anne: 9, 9, 10
Claude: Yes. Yeah, and its super easy to write the question.
Anne: And super boring to grade.
Celine: That's so true
Claude: There's hundreds of examples of question you could write without batting an eye. It's just an easy to write and then it's an easy test to grade, but have you tested that the kids understand anything?
While elaborating on Anne's suggestion to avoid an over-emphasis of assessing factual recall, Claude reflects on how early in his career he fell into the trap of making assessments that were easy to write and grade but did not assess students’ conceptual understanding of chemistry topics. The best practice of assessing for conceptual understanding could be domain-general, however this group of critical friends generated chemistry-specific considerations for the nature and depth of chemistry assessments.
As stated above, the importance of generating clear and direct learning goals is common throughout educational literature. In this discussion of best practices, the topic of learning goals arose several times. Teachers spoke of the importance of being direct, being specific, and potentially giving the learning goals to the students. In one instance from Celine, she discusses how she uses the learning goals to help students recognize knowledge gaps that become evident as students attempt to apply common chemistry symbols.
Celine: When you asked, “How do you make the student aware of the goal?” Okay, so here's an example. We're doing a simple lab. We're getting into reactions again. And uhh reactant sodium bicarbonate. You know one of the things they had to do was write down the equation and translate every symbol from the equation. The arrow, what does it mean? The aq, what does it mean? Then, [the students] had to answer a question after that. This is a pre-lab, and I made them write down the three objectives of the lesson. And one of them was “I can translate the symbols into words for this reaction.” That was. They wrote it. So, you know aq means aqueous right? And the kids were like “means dissolved in water” and they have that answered in just a little chart. Then they have the question, “Looking at the reaction above, what do you predict the reactants will look like with your eyes?” Okay, so we have aqueous baking soda and aqueous hydrochloric acid and they all can tell me the definition of'em in the table, aq means dissolved in water. 75% answered the question in the pre-lab, predict what this will look like with your eyes in the macroscopic, baking soda will be a powder. So, when the kids came up to my desk and I had to stamp it for the pre-lab. I clearly with every kid who got it wrong said, “What's the goal?” and I did this and it was painfully long and awful. This is the goal and you see where you wrote aq, I said, “What does that mean?” and they said it means it's dissolved in water. And I said, “So, this baking soda and this is water. What does it mean?” I said you said powder and I looked at the students and I said, “What I'm telling you and you can tell me what it means, but you can't translate it and apply it to my lab. You haven't met the goal.”
After presenting the importance of generating and articulating clear learning goals, Celine shared a practice she engages in to help her students clarify content by using the learning goal. In the quote, Celine noted that knowing the definition of the symbols used in chemistry are not enough and that students should be assessed on their ability to apply chemistry knowledge in a laboratory environment. Celine's example led teachers to indicate that assessing learning goals using a variety of conceptual and representational levels was important. After Celine's example about how she communicates to students their progress about meeting chemistry-specific expectations within learning goals, she mentioned that the learning goals need to represent specific levels of knowledge expected of the student, as well as the representational levels to be assessed.
Celine: Well, I was also talking about a kid can puke up a definition.
Ashton: Yes!
Celine: They know aq means dissolved in water.
Ashton: But, to get that?
Celine: And they can do that, but to apply.
Ashton: Yes.
Celine: So, I'm talking about the application of the definition.
Ashton: Yes.
Celine: So, when I'm. So, yes it was Johnstone's related, but it was both. It was Bloom's and Johnstone's [levels]
During this dialogue between Celine and Ashton, they shared that although assessing student knowledge of definitions is important, so is assessing the students’ ability to apply the definition. The teachers admitted that assessing student knowledge at a variety of representational levels was difficult, as shown by Ashton's quote,
Ashton: So, just as an aside. I had kids put a puddle of water on a piece of paper like a half page sheet. Conductivity tester, put it in. Nothing. Yeah. Right? They could figure that out. So, I said let's take some sodium carbonate and drop it to the side, push a little bit in. Okay start the conductivity on this side and slowly get closer and they see the light get brighter and brighter and brighter as it gets closer. Okay, now let's try to look at this through Johnstone's triangle. It was like pulling teeth. I mean it was like their brain exploded. Then what happened was after we did that with two items, on one side they put sodium carbonate on the other side they put copper(II) sulfate and they pushed a bit in it and I said just wait. And in the middle, a line appears. We could’ve spent a month on that. I mean the amount of misconceptions, the amount of you know the ions you know about this is what I saw. And then saying okay you saw this. Give me the symbol what does aq mean or what is that line you saw? Is that aqueous? You know is that. It was just like that simple experiment was just like *explosion noise*.
The length and detail of the teacher examples emphasizes the need for critical friends who can interact in content-specific ways. These teachers would not be able to provide such chemistry-rich examples with peers who were not knowledgeable in the content. Although literature about best practices for assessment is discussed in a domain-general manner, the teachers in this study had chemistry-specific conversations about how the practices influenced their ability to generate and use assessments in their classroom. Table 8 provides a summary of the chemistry-specific features of their best-practice discussion.
Research question | Data sources | Key findings |
---|---|---|
(3) What are the chemistry-specific features of the best assessment practices revealed during a facilitated discussion about best practices for chemistry assessments among high school chemistry teachers? | Teacher discourse | • Teachers value assessing for conceptual understanding of chemistry knowledge and identified chemistry-specific pitfalls teachers encounter when designing assessments |
• Teachers shared possible ways to use chemistry experiences to help students better understand learning goals | ||
• Teachers shared the importance of assessing at different levels of chemical representations and chemistry-specific experiences to enact high-quality practices that engage students with multiple levels of representations |
Additionally, teachers often provided reasoning to accompany their examples, evidencing the importance of situating examples within the context of specific classroom environments. Reasoning provided could indicate reflection and discourse about the beliefs driving teachers’ assessment practices. Although the discourse map served as a useful tool to investigate discourse at a bird's eye view, investigation of the best practices generated by the teachers required a thorough examination of the transcript.
The best practices for assessment generated in this study are provided in Table 6. Their generated best practices align to guidelines set by several evidence-based investigations of high-quality assessment practices. Three themes emerged from the teacher best practices: (1) assessing for conceptual understanding, (2) including clear and direct learning goals, and (3) assessing learning goals over a variety of conceptual levels are present throughout the literature. Themes identified represent the beliefs and practices this group of teachers deem important for generating assessments, evaluating assessment quality, and evaluating alignment between instructional and assessment materials. Teachers stated that using assessment data should be used to inform day-to-day instruction, however they reported practice gaps limiting their ability to implement this best practice.
Teachers revealed several chemistry-specific considerations for the generated best practices (as well as gaps in assessment practices) during the best practices discussion. Teachers discussed several seemingly domain-general practices such as developing items that assess conceptual knowledge, generating clear and direct learning goals, and assessing knowledge at a variety of representational levels in a chemistry-specific manner. The chemistry-specific discourse that these teachers engaged in evidences the benefit of having discipline-specific critical friends who are familiar with the content anchoring the discussion of beliefs and practices.
The characteristics of this group align to characteristics of critical friends as described in the literature. The critical friendship shared by this group of high school chemistry teachers was both a limitation and an advantage to this study. Here critical friendship could limit the transferability of study findings, because findings may only apply to teachers who have developed a comfort sharing with one another over several years of interaction. However, critical friendship in this study facilitated greater elicitation and depth of data about teachers’ struggles using assessment results to inform instruction, since teachers shared openly with others who were familiar with their classroom practices.
A limitation of the discourse map is that the line of text as a representational unit is an artifact the data transcription method. Some of the lines of text are sentence fragments while others are complete teacher statements. Additionally, some lines of transcript contain brief (or nonverbal) discourse contributions by from non-speaking teachers. For these reasons, readers are cautioned against interpreting the discourse map quantitatively. Rather, the discourse map should be interpreted qualitatively as a representation of high school chemistry teacher discourse about assessment beliefs and practices.
Results also contribute to a greater understanding of the factors used by high school chemistry teachers while changing and discussing assessment practices. Teachers often rely on the support of their peers while working to improve pedagogy (Black and Wiliam, 1998; Bell and Cowie, 2001; Towndrow et al., 2010). Recognizing the productive discourse patterns used by these critical friends not only helps the design of materials and professional development but also can educate teachers on more productive ways to talk about their beliefs and practices.
The discourse map serves as a tool to characterize discourse patterns over time and can be used to organize complex analyses of content. Toulmin's argumentation pattern has been previously used in chemistry to uncover student ideas embedded in argumentative discourse (Cole et al., 2012). In this study the logic of inquiry framework was used to analyze discourse. Regardless of the discourse framework (such as logic of inquiry or Toulmin), the discourse map can be used to represent the structure of discourse as it occurs over time and/or to identify key discourse fragments.
Additionally, the findings presented here highlight the need for future research investigating gaps in teacher assessment practices regarding the ability to interpret class-level data to inform future instruction. Although the guidelines generated by these critical friends may serve useful for other chemistry educators, the frustration exemplified by Claude and other teachers speaks to the need for quality professional development and materials for high school chemistry teachers regarding the use of data. With more professional development opportunities and materials available to them, teachers would be able to better use assessment results to guide their instruction.
Discourse moves | Codes describe the conversational turn taken by the teachers/facilitators |
---|---|
Initiating | Teacher introduces a new idea (within the topic proposed by the presenter) not previously mentioned |
Continuing | Teacher/facilitator progresses conversation within an initiated idea by elaborating with an example, qualifier, criteria, question |
Referring back | Teacher/facilitator extends to a previously stated idea, links current discourse to previously mentioned idea, or asks a clarifying question about a previously stated idea |
Agree/disagree | Teacher/facilitator doesn't progress conversation, but expresses agreement with the previous statement (can be verbal or nonverbal) |
Replying | Teacher/facilitator responds directly to a question from another teacher or facilitator |
Concluding | Teacher/facilitator makes an ending remark for the current discussion allowing for the initiation of new ideas |
Commenting | Teacher doesn't extend conversation, but adds a personal anecdote to discourse |
Logical processes | Describes the structure of the teachers' statements |
---|---|
Provide reasoning | Teacher gives reasoning or evidence for including an assessment practice or characteristic or why practice or characteristic should be performed a certain way |
State a goal | Teacher states what they are attempting to accomplish when performing an assessment practice or including an assessment characteristic |
State a result | Teacher gives a possible outcome or student results from the implementation of an assessment practice/characteristic (can be hypothetical) |
Refines | Teacher refines assessment technique/assessment characteristic/pedagogical action already stated by proposing a qualifying statement or a restriction OR (connects/defines the boundary between) the assessment technique/assessment characteristic/pedagogical action to a previously stated one |
Presents | Teacher provides a possible assessment technique/assessment characteristic/pedagogical action OR asks the group a question about (the importance of/if) an assessment technique/assessment characteristic/pedagogical action should be included in the best practices |
Evaluates | Teacher/facilitator questions or inquires about an assessment technique/assessment characteristic/pedagogical action made by another to increase depth of explanation, clarify details, or set a boundary from another goal |
Contradicts | Teacher expresses disagreement with a statement from another teacher or expresses that they in some way cannot do a stated assessment practice or include an assessment characteristic |
Gives example | Teacher provides an classroom example to illustrate a practice or characteristic (can be hypothetical) |
Nature of statement | The manner in which teachers discuss certain topics. (Reasoning/characteristic/practice stated as…) |
---|---|
Action oriented | … an action to be performed by the teacher/student/assessment without reference to an identifiable cause or outcome (including practices that they do/try to implement) |
Causal | … either a direct cause or result of an observable phenomena |
Descriptive | … a quality of the assessment/teacher/student |
This journal is © The Royal Society of Chemistry 2020 |