Mihwa
Park
*a,
Xiufeng
Liu
a,
Erica
Smith
b and
Noemi
Waight
a
aDepartment of Learning and Instruction, University at Buffalo, Buffalo, NY, USA. E-mail: mihwapar@buffalo.edu; Tel: +1 7166451007
bTulane University, New Orleans, LA, USA
First published on 25th April 2017
This study reports the effect of computer models as formative assessment on high school students' understanding of the nature of models. Nine high school teachers integrated computer models and associated formative assessments into their yearlong high school chemistry course. A pre-test and post-test of students' understanding of the nature of models using a published measurement instrument on the nature of models were conducted. A four-step hierarchical multiple regression and a two-level (level 1 – student and level 2 – teacher) hierarchical linear modeling were used to test the effect of the intervention on students' understanding of the nature of models. Our analysis revealed a significant effect of frequencies of using computer models for four of the five sub-scales related to the nature of models. The implications of these findings are that, as students have more experience using computer models in their classroom, they develop a better understanding of the nature of models. However, their understanding of models as multiple representations didn't show a significant improvement, possibly due to the lack of support from teachers, who in turn need both content and pedagogical supports within their teaching.
The term model has several different meanings; in general, it refers to objects or people that are worthy of emulation (Chamizo, 2013). In science, Gilbert and Boulter (2000) define a model as a representation of a phenomenon built with a specific purpose. Specific purposes in science are to simplify a phenomenon of scientific inquiries, to show compositions of an object, or to visualize abstractions or ideas as objects (Gilbert and Boulter, 2000). Schwarz et al. (2009) also define a scientific model as “a representation that abstracts and simplifies a system by focusing on key features to explain and predict scientific phenomena” (p. 633).
In this study, we are concerned with a specific type of model, computer models, as a representation of phenomena within macro, submicro, and symbolic domains of chemistry teaching. Many studies reported positive results from using computer models to transform different representations (Dori and Hameiri, 2003) and to reduce cognitive load and visuospatial demands (Wu et al., 2001), which are important to improve conceptual understanding in science (Cook, 2006). In a literature review of 61 empirical studies over the last decade, Smetana and Bell (2012) found that computer models could be as effective, and in many ways more effective, than traditional (e.g., lecture-based, textbook-based, and/or physical hands-on) instructional practices in student knowledge acquisition, developing process skills, and facilitating conceptual change. In chemistry, Barak and Dori (2005) concluded that computer models promote chemistry understanding by illustrating chemical concepts at the macroscopic, submicroscopic, symbolic, and chemical process levels. Ardac and Akaygun (2004)'s study provides further evidence of the positive effect of technology on developing students' understanding of chemistry by enabling students to visualize reactions occurring at a molecular level that would otherwise not be visible to students.
Even though using models has been shown to facilitate student learning in science, studies have also documented students' difficulties in understanding of models or limited use of models within science classrooms (Gilbert and Boulter, 2000; Snir et al., 2003; Coll et al., 2005; Schwarz et al., 2009). Treagust et al. (2002) argue that students' understanding of the role of models should be assessed because their incorrect understanding of models may hinder students from scientific understanding in various science topics. In this respect, Treagust et al. conceptualize students' understanding of the nature of models to consist of five dimensions, and developed a test instrument, students' Understanding of Models in Science (SUMS), to measure student understanding of models in science on the basis of the dimensions:
(1) the models as multiple representations (MR) meaning “students' acceptance of using a variety of representations simultaneously, and their understanding of the need for this variety” (p. 359).
(2) the models as exact replicas (ER) indicating “students' perception of how close a model is to the real thing” (p. 359).
(3) the models as explanatory tools (ET) referring to “what a model does to help the student understand an idea” (p. 359).
(4) the uses of scientific models (USM) referring “students' understanding of how models can be used in science, beyond their descriptive and explanatory purpose” (p. 359).
(5) the changing nature of models (CNM) indicating understanding of the dynamic changes of models related with scientific changes.
Another promising approach to chemistry teaching is formative assessment. Formative assessment involves the process of collecting data from students and using it to improve students' learning (Wiliam, 2011). Black and Wiliam (1998) conclude from their review of more than 250 books and articles that formative assessment with feedback to students could have significant positive effects on student achievement.
While numerous studies have found that computer models and formative assessment can facilitate students' understanding of scientific concepts and enhance classroom instruction, there have been no reported studies of the impact of them on student understanding of the models themselves. This study examined the impact of utilizing computer models as formative assessment tools in classroom instruction on students' understanding of the models in science. This study addressed the following research question: to what extent do computer models and associated formative assessments facilitate student understanding of the nature of models?
We also developed ten sets of computer model-based formative assessments to assess student understanding of matter, energy and models related to the computer models. Each question in the formative assessment addresses either matter, energy or computer models in each chemistry topic. Assessment questions were in an Ordered Multiple-Choice (OMC) format (Briggs et al., 2006), with choices matching different levels of understanding of matter, energy and models. The initial version of the computer models and assessments was pilot-tested with one class of students (n = 15–25) and reviewed by three experts with expertise in chemistry, psychometrics and science assessment during the academic year 2009–2010. The revised models and assessments were then field-tested in three high school chemistry classes during the 2010–2011 academic year. Based on data collected during the year, further revisions to the computer models and formative assessments were made.
During the 2011–2012 school year, the participant teachers incorporated the CCFA into one of their chemistry classes (designated as experimental class), and taught the other chemistry classes in traditional ways (designated as control classes). The recruited teachers received a 1 day training on the formative assessments and computer models prior to the start of the 2011–2012 school year. The training included rationale of formative assessment, nature of the computer models, and how computer models may be incorporated into lessons. A doctoral student in science education was also assigned to each teacher's classroom to support teacher integration of the models in teaching. The doctoral student visited an experimental class 2–3 times during the year; the purpose of the visit was to provide technical support and to check for fidelity of implementation. Note that the doctoral student was not involved in actual teaching. Table 1 presents years of experience teaching chemistry for participant teachers and the number of students who participated in this study.
Teacher | Years in chemistry teaching | N of students in experimental class | N of students in control class | Total N |
---|---|---|---|---|
Teacher 1 | 6 | 22 | 22 | 44 |
Teacher 2 | 18 | 23 | 50 | 73 |
Teacher 3 | 13 | 25 | 55 | 80 |
Teacher 4 | 8 | 11 | 35 | 46 |
Teacher 5 | 20 | 19 | 15 | 34 |
Teacher 6 | 18 | 14 | 23 | 37 |
Teacher 7 | 5 | 10 | 20 | 30 |
Teacher 8 | 8 | 22 | 78 | 100 |
Teacher 9 | 9 | 28 | 52 | 80 |
Total | 174 | 350 | 524 |
Depending on the teacher, these worksheets were either done independently by students, in small groups, or as an entire class led by the teacher. After students had experienced working with the computer models for each chemistry topic, teachers administered the computer model-based formative assessment for each topic to experimental class students. The assessment was administered individually to students online. While taking the assessments, students were asked to open a second window to view the computer models simultaneously and could also refer to their worksheets for assistance if needed. The purpose of the assessment was to determine students' understanding on matter, energy and computer models. The assessments took between 25 and 40 minutes to complete.
Once students completed the online assessments, the researchers generated students' scores remotely, and provided computer printouts of these scores to teachers. Utilizing the students' score results of understanding of matter, energy and models, the participating teachers were then able to plan and implement differentiated instruction during the next unit.
In order to document teachers' implementation of the CCFA, they also recorded their use of the computer models in a log sheet (e.g., the frequencies of using each of the computer models). The range of frequency of using Flash and NetLogo models was four to 20 during the school year.
Teachers in the comparison classes used their traditional instructional approaches. We did not limit teachers to use computer models in the control classes because some teachers might already use computer models for various purposes such as introducing a topic or demonstrating a concept etc. The difference between experimental classes and comparison classes was that experimental classes used CCFA systematically while comparison classes used computer models on an ad hoc base.
Second, predictor variables were submitted to a two-level (level 1 – individual and level 2 – teacher) hierarchical linear modelling (HLM) to simultaneously test the effect of student and teacher predictor variables. HLM7 (Raudenbush et al., 2010) was used to conduct the hierarchical linear modelling analysis. The student-level variable was control/experimental class (0: control class, 1: experimental class) and frequencies of using Flash and NetLogo models. In addition, student pre-test scores on the PUM were included as a control variable in level 1. The teacher-level variable was years of chemistry teaching experience. Other teacher level variables such as teacher certification in chemistry, were not included because there was little to no variation among the nine teachers on these variables.
Outcome variables for both analyses were gain scores between pre and post-test of the five sub-scales in SUMs. The gain scores for each SUMS sub-scale between the pre and post-test were calculated by subtracting pre-test scores from post-test scores. The above outcome variables were tested separately. All variables used in the analysis were summarized in Table 2.
Variable name | Description |
---|---|
Predictor variables | |
PRESURVE | Student pre-test scores on Progression of Understanding Matter (PUM) test |
CONTROLE | Control/Experimental class; coded 0 = control group, 1 = experimental group |
MODELFRE | Frequencies of using Flash and NetLogo models |
CHEMT | Years of chemistry teaching experience of teachers |
Outcome variables | |
Gain MR | Gain scores of models as multiple representations scale |
Gain ER | Gain scores of models as exact replicas scale |
Gain ET | Gain scores of models as explanatory tools scale |
Gain USM | Gain scores of the uses of scientific models scale |
Gain CNM | Gain scores of the changing nature of models scale |
Outcome | Predictora | R 2 | Test of model | Test of predictor |
---|---|---|---|---|
a Predictor: PRESURVE = student pretest scores on PUM, CONTROLE = control/experimental class, MODELFRE = frequencies of using Flash model and NetLogo model, and CHEMT = years of chemistry teaching experience of teachers. | ||||
Gain MR | Step 1 | 0.002 | F(1,363) = 0.617 | |
PRESURVE | β = −0.089 | |||
Step 2 | 0.002 | F(2,362) = 0.323 | ||
PRESURVE | β = −0.088 | |||
CONTROLE | β = 0.153 | |||
Step 3 | 0.002 | F(3,361) = 0.242 | ||
PRESURVE | β = −0.079 | |||
CONTROLE | β = 0.606 | |||
MODELFRE | β = −0.037 | |||
Step 4 | 0.004 | F(4,360) = 0.347 | ||
PRESURVE | β = −0.063 | |||
CONTROLE | β = 0.381 | |||
MODELFRE | β = −0.011 | |||
CHEMT | β = −0.074 |
Outcome | Predictor | R 2 | Test of model | Test of predictor |
---|---|---|---|---|
*p < 0.05. | ||||
Gain ER | Step 1 | 0.001 | F(1,363) = 0.383 | |
PRESURVE | β = −0.075 | |||
Step 2 | 0.017 | F(2,362) = 3.070* | ||
PRESURVE | β = −0.067 | |||
CONTROLE | β = 2.192* | |||
Step 3 | 0.032 | F(3,361) = 4.012* | ||
PRESURVE | β = −0.145 | |||
CONTROLE | β = −1.804 | |||
MODELFRE | β = 0.326* | |||
Step 4 | 0.033 | F(4,360) = 3.063* | ||
PRESURVE | β = −0.135 | |||
CONTROLE | β = −1.947 | |||
MODELFRE | β = 0.343* | |||
CHEMT | β = −0.047 | |||
Gain USM | Step 1 | 0.000 | F(1,363) = 0.150 | |
PRESURVE | β = −0.022 | |||
Step 2 | 0.019 | F(2,362) = 3.531* | ||
PRESURVE | β = −0.018 | |||
CONTROLE | β = 1.133* | |||
Step 3 | 0.029 | F(3,361) = 3.656* | ||
PRESURVE | β = −0.048 | |||
CONTROLE | β = −0.405 | |||
MODELFRE | β = 0.126* | |||
Step 4 | 0.036 | F(4,360) = 3.319* | ||
PRESURVE | β = −0.034 | |||
CONTROLE | β = −0.612 | |||
MODELFRE | β = 0.149* | |||
CHEMT | β = −0.068 |
Outcome | Predictor | R 2 | Test of model | Test of predictor |
---|---|---|---|---|
*p < 0.05. | ||||
Gain ET | Step 1 | 0.000 | F(1,363) = 0.096 | |
PRESURVE | β = −0.27 | |||
Step 2 | 0.011 | F(2,362) = 2.003 | ||
PRESURVE | β = −0.022 | |||
CONTROLE | β = 1.295* | |||
Step 3 | 0.021 | F(3,361) = 2.579 | ||
PRESURVE | β = −0.067 | |||
CONTROLE | β = −0.997 | |||
MODELFRE | β = 0.187 | |||
Step 4 | 0.023 | F(4,360) = 2.147 | ||
PRESURVE | β = −0.053 | |||
CONTROLE | β = −1.190 | |||
MODELFRE | β = 0.209* | |||
CHEMT | β = −0.064 | |||
Gain CNM | Step 1 | 0.000 | F(1,363) = 0.050 | |
PRESURVE | β = 0.013 | |||
Step 2 | 0.000 | F(2,362) = 0.055 | ||
PRESURVE | β = 0.013 | |||
CONTROLE | β = 0.104 | |||
Step 3 | 0.021 | F(3,361) = 2.537 | ||
PRESURVE | β = −0.029 | |||
CONTROLE | β = −2.019* | |||
MODELFRE | β = 0.173* | |||
Step 4 | 0.021 | F(4,360) = 1.944 | ||
PRESURVE | β = −0.033 | |||
CONTROLE | β = −1.961* | |||
MODELFRE | β = 0.167* | |||
CHEMT | β = 0.019 |
We note that the initial models (Step 1s) entering only student pre-test scores on PUM were not significant for all outcome variables.
Table 3 presents results of hierarchical multiple regression model for Gain MR (gain multiple representations scale scores) as an outcome variable.
When including Gain MR as an outcome variable, all models were not significant and none of predictor variables significantly predicted the outcome variable (Gain MR).
When using Gain ER (Gain scores of models as exact replicas scale) and Gain USM (gain the uses of scientific models scale scores) as outcome variables, results from hierarchical multiple regression models were similar, so we present the results together in Table 4.
In the case of Gain ER and Gain USM, all model effects were significant except the first model. In the second models for Gain ER and Gain USM, control and experimental classes (CONTROLE) emerged as a significant predictor, and in the third and fourth models, frequencies of using Flash and NetLogo models (MODELFRE) emerged as a significant predictor. These findings indicated that when the frequency of using Flash and NetLogo models increases, student scores on ER and USM tended to significantly increase.
Lastly, we present results from using Gain ET (gain explanatory tools scale scores) and Gain CNM (gain the changing nature of models scale scores) as outcome variables in Table 5.
In cases of Gain ET and Gain CNM as outcome variables, all models were not significant fits of the data. In the second model of Gain ET, control and experimental classes (CONTROLE) emerged as a significant predictor, and in the fourth model, frequencies of using Flash and NetLogo models (MODELFRE) emerged as a significant predictor. In the case of Gain CNM, frequencies of using Flash and NetLogo models (MODELFRE) were a significant predictor for the third and the fourth models, and control and experimental classes (CONTROLE) were also a significant predictor.
In sum, except Gain MR, frequencies of using Flash and NetLogo models (MODELFRE) emerged as a positively significant predictor for the outcome variables when other predictors were controlled. Years of chemistry teaching experience of teachers (CHEMT) contributed no variance in any outcome variables when student pre-test scores on PUM (PRESURVE), control and experimental classes (CONTROLE), and frequencies of using Flash and NetLogo models (MODELFRE) were taken into account.
Although the hierarchical multiple regression analysis revealed that as the frequency of using Flash and NetLogo models increased, students' scores on the models as exact replicas (ER), the models as explanatory tools (ET), the uses of scientific models (USM), and the changing nature of models (CNM) scales increased, we found that overall experimental classes' gain CNM scores were significantly lower than control classes' gain CNM scores when other predictors were controlled. We assumed that it might be due to the possibility that we didn't consider nested data structure in evaluating teacher effects. So we applied hierarchical linear modelling (HLM) to test relationships within and between grouped data, which tests simultaneously the effect of teacher level predictors and the effect of student level predictors on outcome variables.
First, using an outcome measure as the dependent variable (specifically: Gain MR scores, Gain ER scores, Gain ET scores, Gain USM scores, and Gain CNM scores, see Table 2), the model is as follows:
Level-1 model
Dependent variableij = β0j + rij
Level-2 model
β 0j = γ00 + u0j
Table 6 presents the results of the model for each of the dependent variables. Only when using Gain CNM scores (gain the changing nature of models scale scores) as an outcome variable, there was a statistically significant difference between teachers (χ2(8) = 21.91, p = 0.005). Between-teacher variance was 0.55 and the within-teacher variance was 14.24, thus the intra-class correlation was 3.7%. The intra-class correlation represents the percent of variance in the outcome variable that is between groups (Woltman et al., 2012). So in the study, 3.7% of the variance in the gain CNM scores was at the group level. The statistically significant between-class variance indicates that average gain CNM score varied significantly across teachers. This supports the use of HLM for gain CNM scores.
Type of variability | Variance | χ 2 |
---|---|---|
*p < 0.05. | ||
Gain MR scores | ||
Between-teacher | 0.51 | 10.23 |
Within-teacher | 60.09 | |
Gain ER scores | ||
Between-teacher | 0.03 | 6.05 |
Within-teacher | 68.95 | |
Gain ET scores | ||
Between-teacher | 0.59 | 13.14 |
Within-teacher | 34.51 | |
Gain USM score | ||
Between-teacher | 0.21 | 11.91 |
Within-teacher | 15.25 | |
Gain CNM scores | ||
Between-teacher | 0.55 | 21.91* |
Within-teacher | 14.24 |
When using gain CNM scores as an outcome variable, the model was as follows:
Level-1 model
GAINCNMij = β0j + β1j(CONTROLEij) + β2j(PRESURVEij) + β3j(MODELFREij) + rij
Level-2 model
β 0j = γ00 + γ01*(CHEMTj) + u0j
β 1j = γ10 + γ11*(CHEMTj)
β 2j = γ20
β 3j = γ30 + γ31*(CHEMTj) + u3j
Table 7 presents the results of the above model. From the fixed effect section in Table 7, we can see that there was no statistically significant predictor for the differences in teacher level means (CHEMT) on gain CNM scores. HLM analysis results revealed that there was no cross-level interaction between frequencies of using Flash and NetLogo models and years of chemistry teaching experience (γ31 = 0.01, p > 0.05), meaning that years of chemistry teaching experience had no influence on the strength of the relationship between frequencies of using Flash and NetLogo models and student gain CNM scores. In addition, control and experimental classes didn't have a significant effect on student gain CNM scores (γ10 = −0.44, p > 0.05) when other predictors were controlled. The frequencies of using Flash and NetLogo models showed an insignificant effect on student gain CNM scores (γ30 = 0.003, p > 0.05) either. In sum, the fixed effect section in Table 7 indicates that the degree of teacher experience (CHEMT) had no influence on the strength of the relationship between predictor variables and gain CNM scores; all regression coefficients, γ, were not statistically significant, p > 0.05.
Fixed effect | |||
---|---|---|---|
Independent variables | Gamma coefficient | Standard error | |
a Years of chemistry teaching experience. b Control/experimental class. c Pre-test scores on PUM. d Frequencies of using Flash and NetLogo models.*p < 0.05. Note: gamma coefficients of CHEMT variable indicate as following: student gain CNM scores were higher in classes with teachers who had more experience in teaching chemistry but the result was not statistically significant (γ01 = 0.12, p > 0.05), the cross-level interaction between control/experimental group (CONTROLE) and teacher experiences in teaching chemistry (CHEMT) was not statistically significant (γ11 = −0.17, p > 0.05), and the cross-level interaction between frequencies of using Flash and NetLogo models (MODELFRE) and teacher experiences in teaching chemistry (CHEMT) was not statistically significant (γ31 = 0.01, p > 0.05). | |||
Mean achievement | β 0 | ||
Intercept | γ 00 | 1.56* | 0.51 |
CHEMTa | γ 01 | 0.12 | 0.12 |
CONTROLEb | β 1 | ||
Intercept | γ 10 | −0.44 | 1.05 |
CHEMT | γ 11 | −0.17 | 0.24 |
PRESURVEc | β 2 | ||
Intercept | γ 20 | −0.04 | 0.06 |
MODELFREd | β 3 | ||
Intercept | γ 30 | −0.00 | 0.09 |
CHEMT | γ 31 | 0.01 | 0.02 |
Random effect | Variance | χ 2 | |
---|---|---|---|
Intercept | u 0 | 0.65 | 19.82* |
MODELFRE slope | u 3 | 0.01 | 26.01* |
In the random effect section in Table 7, average gain CNM scores among teachers and among frequencies of using Flash and NetLogo models (MODELFRE) varied significantly (χ2(6) = 19.82, p < 0.05, and χ2(6) = 26.01, p < 0.05, respectively). This result implies that there still remained additional variability between teachers left unexplained.
HLM analysis results showed that overall students' gain scores were not significantly different among teachers except their gain the changing nature of models (CNM) scale scores. Although we used the HLM analysis method to see the effect of teacher level and the effect of student level variables on student gain CNM scores, we didn't find any significant predictor to explain student gain CNM scores. Furthermore, HLM analysis on gain CNM scores revealed that a significant variation remained after controlling for the predictor variables, indicating that other variables not accounted may be important factors. This finding should be considered in light of the limitations of this study, which implies that there were other variables in teachers to predict students' gain CNM scores. Future work should include more teacher level variables such as types or amounts of their feedback to students.
This is one of the first studies to examine both computer models and formative assessments in chemistry teaching and learning in relation to students' understanding of nature of models. The findings could be the result of several factors, both at the student level and teacher level. At the student level, a factor that may have impacted the effectiveness of the CCFA intervention was the frequency of using computer models. Simply being exposed to computer models might not have a significant effect on students' understanding of the nature of models. Students who had greater exposure in the experimental classes showed more improved understanding of the nature of models in four of the five sub-scales in SUMS. Therefore, the amount of time they spent with the models did significantly impact their understanding of the nature of models. Previous studies found positive results from using computer models to promote chemistry learning (e.g., Ardac and Akaygun, 2004) and to develop their understanding of chemical representations (Wu et al., 2001). The current study contributes to the current body of literature in that engaging students in using computer models also have a positive effect to students' understanding of the nature of models. As Snir et al. (2003) pointed out, many students have difficulty in relating scientific models to abstract scientific concepts due in part to a lack of support to engage students in developing correct understanding of models (Duschl et al., 2007). With this regard, we expect that students' better understanding of scientific models will facilitate their understanding of scientific concept as well.
This study also found that students' understanding of models as multiple representations (MR) was not significantly improved after the yearlong intervention. This implies that providing students with different models in chemistry was not enough to improve their MR understanding; they need to be given additional support in understanding the complexity of those models in relation to the concepts they are being taught in class. The use of explicit instruction to promote student understanding of models has been cited by several studies on the effective use of models within classroom instruction (Snir et al., 2003; Coll et al., 2005; Gobert et al., 2011). In particular, a study done by Gobert et al. (2011) on the relationship between high school students' understanding of nature of models and the use of model-based software, they found that explicit instruction that was provided for students within the Connected Chemistry curriculum was effective at enhancing student understanding of models. Waight and Gillmeister (2014) also found that student's lack of chemistry content knowledge and knowledge of scientific models made it difficult for students to make meaning of the different representations illustrated in the models in CCFA. This implies that the teacher's role may be more important in enhancing students' understanding of model as multiple representations. In other words, the ability of the participating teachers to understand and carry out the CCFA intervention could have led to the overall non-significant effect on developing student understanding of models as multiple representations. This ability may have been impacted by not only the teacher's knowledge of chemistry but also their knowledge of models beyond the traditional ball-and-stick models, most often used within high school chemistry curriculum.
Further, the insignificant result of frequencies of using Flash and NetLogo models on the Gain CNM score from HLM analysis may have been due in part to the participating teachers lack of knowledge regarding the CCFA intervention and as a result it was not fully implemented into their classroom practices. Specifically, in our project, although teachers were provided with formative assessment test results regarding students' understanding of matter, energy and models as well as suggested differentiated labs and activities during the chemistry course, not all teachers consistently followed the suggestions. This indicates that feedback from teachers to students may vary considerably in quality and quantity. This difference in providing feedback might be an important factor in explaining the significant variation remaining after controlling for the predictor variables. Wiliam (2011) asserted the importance of providing feedback enabling to move learnings forward, suggesting that focused, specific, and scaffolded feedback is necessary to improve students' learning outcome. Our classroom observation revealed that there was a wide variation in the extent to which teachers used the data to inform their instruction from no impact on subsequent instruction to some forms of differentiated laboratory experiences for students. The difference in using information from formative assessments results into modifying their instructions to meet students' needs suggests a possible lack of teacher knowledge and skills in implementing formative assessments with feedback. The effectiveness of formative assessment depends on the content of learning activities and quality and quantity of feedback (Black and Wiliam, 1998). Additionally, in order for teachers to effectively implement the intervention into practice, they must have a good understanding of the scientific background on matter and energy underlying the computer models, and to relate the computer models to student learning difficulties on specific chemical concepts (e.g., chemical equilibrium) during the unit of instruction. During the field testing, we found that most teachers did not possess this type of knowledge, and thus providing teachers with more extended professional development will be critical to the implementation of the intervention (Waight and Gillmeister, 2014). In this study, we did not collect teacher level variables such as understanding of models and chemistry concepts, and the degree of modification of their instruction or provided feedback to students except for their years of chemistry teaching and teaching certificates. In light of this limitation, future studies should investigate more variables at the teacher level and their impacts on students' understanding of models.
We acknowledge a few potential uncontrolled threats to internal validity of findings of this study. Potential factors jeopardizing the validity of an experimental study (Campbell and Stanley, 1966) are: (1) history – extraneous event's effect on the dependent variable between pre and posttest; (2) maturation – subject's natural growth; (3) testing – effect of taking a pretest on posttest results; (4) instrumentation – changes in the instrument or scores which may affect on the dependent variable; (5) selection bias – differences between groups which are not comparable; (6) statistical regression – due to the selection of subjects on the basis of extremely low or high scores; and (7) experimental mortality – the loss of participants. In the current study, we used a pretest–posttest control–experimental group design in order to control several potential threats. More specifically, (1) history and (2) maturation were addressed by using a control group, (3) testing was addressed by using a long time period between pretest and posttest (1 year), and (4) instrumentation was addressed by using a same published test with comparable reliability coefficients for scales (Campbell and Stanley, 1966; Parker, 1990; Tayler and Asmundson, 2008). However, we note that (5) selection bias, (6) statistical regression, and (7) experimental mortality might be potential threats to internal validity. Those three potential threats could be addressed by randomization of group membership, and providing rewards to groups to prevent attrition (Parker, 1990; Tayler and Asmundson, 2008), which the current study did not apply into our study design. The changes in students' understanding of models might be due to the possibility that the students in the experimental groups altered their behaviour because they were aware of being studied (Adair, 1984). Another possibility for the change might be because they were highly enthusiastic about the new computer simulations, however a novelty effect doesn't last long, rather it dissipates significantly in longer duration studies (Kulik et al., 1983), which implies that this study might be less likely to be affected by the novelty effect.
In conclusion, the findings of this study reveal that the frequency of using different types of computer models had a positive influence on students' understanding of the nature of models. The findings also suggest additional efforts are required in regards to helping teachers develop their understanding of how to effectively use models and formative assessments in their instruction. This finding contributes to the literature by identifying several key factors for computer models as formative assessment to improve student understanding of models, including explicit instruction of models to students and extended teacher professional development with a focus on developing their understanding of models and how to use models to modify their instruction. It is not enough for teachers to be given access to new scientific models, further assistance to teachers in the implementation process is necessary in order for students to develop a better understanding of the nature of models in science and to be able to utilize models effectively to explain and predict phenomena. Moreover, as the importance of formative assessment has been emphasized to improve students' learning, it is necessary to enhance teachers' understanding and skills in incorporating formative assessment involving feedback to students into their instructions.
This journal is © The Royal Society of Chemistry 2017 |