Dimitrios Stamovlasis*a,
George Papageorgioub and
Georgios Tsitsipisb
aAristotle University of Thessaloniki Faculty of Philosophy, Greece. E-mail: stadi@auth.gr
bDemocritus University of Thrace, Department of Primary Education, Greece
First published on 13th June 2013
Research in cognitive psychology and science education has acknowledged two antagonistic theoretical perspectives concerning students' knowledge. One perspective considers students' knowledge as coherent or theory-like, explaining the learning process in terms of framework theories and stable mental models, whereas the other considers it fragmented and based on the existence of knowledge elements that can be activated, interconnected and organized accordingly. Although both perspectives continue to demonstrate empirical support, a third point of view focuses on the role of the particular topic and suggests that the stability of the in situ formed conceptions varies under different circumstances being also affected by the researchers' interpretation of data. Since methodological issues are by far fundamental and crucial for the above theoretical premises, in the present research an attempt is made to test the above hypotheses by applying a robust statistical tool, the Latent Class Analysis (LCA). Data were taken from students' responses (N = 329, aged 14–15) using an instrument designed to access students' understanding of the structure of matter and its changes of states that had previously been used to identify students' progression through a series of mental models reported in the literature. However, the results did not support the hypothesis that participants would hold these as coherent mental models. The present research adds to the ongoing dialogue on such a crucial issue for its pedagogical implications by contributing to the methodological pathways for an advancement of science education research through theory building.
The acknowledgement of the differences between students' and experts' ideas and on the other hand the emphasis on their stability and coherence raised a number of questions. Smith et al. (1993) wondered for instance how the main principles of constructivism could be served by the existence of such stable misconceptions, when these should be replaced by the conceptions of experts during the learning processes. They claimed that many of the assertions of the above perspective are inconsistent with constructivism and thus, they supposed that some other productive resources may exist. Within this context, another perspective for the learning process and so for the knowledge itself was developed, typically known as knowledge in pieces or fragmented knowledge (e.g., diSessa, 1988; Smith et al., 1993; Harrison et al., 1999; diSessa et al., 2004). According to this perspective, during learning or cognitive processes some preexisting productive “implicit knowledge elements” are activated and any understanding or explanatory scheme is produced in situ when a particular phenomenon is being studied. diSessa (1993) established the term phenomenological primitives (p-prims) in order to specify such knowledge elements. These elements, within this perspective, are presented as flexible pieces of knowledge that can be evolved, modified and enlarged over time, however they could also be activated and coordinated with other pieces of knowledge into a bigger structure formulating a conception for a particular phenomenon (diSessa, 1988, 2006; Taber and García Franco, 2010).
Launching a different viewpoint, other researchers (e.g. Hammer, 1996, 2004; Taber, 2000, 2008, 2009; Taber and García Franco, 2010) argued that the two apparently antagonistic perspectives might not be as controversial as they seem to be. Taber (2008) for instance pointed that the researcher's interpretation of data might be what leads to specific evidence supporting one or the other perspective. According to Taber, there are relevant research reports on specific topics supporting the view that students have developed stable conceptual frameworks for some ideas, whereas in other cases, ideas appear to be formed in situ and varied. Thus, the question is shifted from ‘whether students' conceptions are coherent or not’ to ‘which conceptions within the context of a specific topic and under particular circumstances appear to be stable’.
It is pertinent to recall here an epistemological note on Popper's (1979) model of cognition concerning the accessibility of students' knowledge (Taber, 2008). The ‘cognitive structures’, which correspond to representation of knowledge in the ‘materiale substatum’ (brain) can only be accessed indirectly by assessing ‘conceptual structures’ portrayed in the products of human mind (talk, writing, drawing etc.). Conceptual structures are products of thoughts, or outcomes of (thinking) processes. In science education, even though the phenomena under investigation are referred to as processes, research is not designed to examine processes or mechanisms controlling changes over time. Normally, cross sectional studies are carried out and merely the outcomes of thinking processes are evaluated at one point in time and on the basis of response patterns to a specially designed instrument. The above indicate the first methodological deficiency that is expected to limit our research. Since students' responses are created by activating and processing some kind of knowledge elements or conceptual resources and given the dynamical nature of a cognitive process and its inherent nonlinearity (Stamovlasis, 2006, 2011) the notion of in situ formation seems to reasonably justify the variable consistency and stability of conceptual frameworks observed under different circumstances. Human knowledge, represented in contemporary psychological theories as networks of interrelated elements, grows, develops and increases in size and complexity by incorporating more general and/or context-specific components. Thus, a diversity of grain size of knowledge structures might exist (Smith et al., 1993), the coherence and stability of which is a core issue for conceptual change theories and it is directly interwoven with research methodology and data analysis. The notion of coherency deserves a specific attention when referring to a relatively large ‘grain size’ pieces of knowledge, e.g. the mental models, because their hypothesized stability is the origin of students' learning difficulties due to their resistance to change.
Conclusively, although a plethora of research and theoretical endeavors have added to our understanding in this area, further investigation is still needed and what becomes obvious yet is that methodological issues are involved. The majority of the studies supporting the coherent conceptual frameworks' perspective relied methodologically on interviews, small samples or weak statistical analyses. A reexamination of the issue in question suggested that suboptimal analysis and deficiency in interpretation of data can be avoided by implementing more robust statistics methods suitable to provide a more lucid picture of the nature of students' knowledge.
The methodology implemented for the classification of tasks in mental model research is usually based on the following: for each mental model, an expected pattern of responses in a specific questionnaire is formulated, while an observed pattern is recorded. Then, the degree of correspondence between the expected and the observed response pattern is determined and the subjects are classified into groups that hypothetically possess the corresponding mental model. In such a classification, a deviation between expected and observed patterns is usually tolerated and the classification of the response is achieved with a predetermined but also arbitrary degree of agreement (e.g. 85% fit). This is known as Rule Assessment Methodology (RAM) (Siegler, 1976, 1981) and it has been followed in many research studies on mental models (e.g., Johnson, 1998a; Vosniadou and Brewer, 1992; Vosniadou, 1994). There are some weak points and disadvantages in this methodology, which are associated with the allocation of the response patterns to mental models and the statistical treatment usually implemented. First, only the predetermined mental models, i.e. the ones that the researcher proposes, can be detected; possible alternative models might be missing. Second, the assignment of response patterns to mental models is made on the basis of an arbitrary criterion of a minimum percentage of correspondence between observed and expected responses. Since the robustness of the criterion appears to be affected by the number of items used in the questionnaire, a number of alternative classifications might be obtained. A third disadvantage is that no statistical tests are implemented to provide a measure of how effectively the proposed classification fits and accounts for the data (Jansen and van der Maas, 1997).
An appropriate methodology to overcome the above disadvantages is the implementation of Latent Class Analysis (LCA), which belongs to the family of the Latent Structure Models (LSM) (Clogg, 1995; Dayton, 1998; Magidson and Vermunt, 2001; McCutcheon, 1987). In LSM, which can be used in exploratory or confirmatory analyses, a distinction between latent and manifest or observed variables is made, where the assumed underlying latent variables could be considered at different measurement levels. Latent Class Analysis is implemented when both manifest and latent variables are assumed to be categorical. LCA achieves to divide a set a response patterns into groups or clusters named latent classes (LC) where the subjects hypothetically belong to, being qualitatively different from each other. In contrast to the traditional clustering approaches, which use distances between responders, LCA implements similarity between responses and specifically probabilities of response patterns. In LCA, the assignment of cases to a certain class is based on the log of odds of being in a particular class versus another class (Magidson and Vermunt, 2001).
An advantage of LCA is that it offers quantitative statistical measures for goodness-of-fit, indicating how well the LC clusters account for the data. Ergo, LCA tests whether the mental models proposed by the researcher are statistically supported by the data and it could falsify any proposed theoretical account. However, the most important advantage of LCA is that it can detect clusters of unexpected response patterns. That is, the mental models do not have to be known in advance and therefore, unanticipated alternative classes (LCs) might be emerged from the analysis. The fit of a latent class analysis can be assessed by a number of indicators such as the number of parameters (Npar), likelihood ratio statistic (L2), Bayesian information criterion (BIC), degrees of freedom (df) and bootstrapped p-value. Non-significant p-values greater than 0.05 are desired indicating a good fit between the theoretical model and data (Vermunt and Magidson, 2002).
If the mental models of students on a particular topic are coherent, distinct and qualitative different entities responsible for certain response patterns, then the number of subjects possessing the same model, i.e. those who manifest the same response pattern, could be revealed as a distinct cluster (latent Class, LC) by LCA based on the subjects' answers to a selected questionnaire.
In LCA the emerged LCs are characterized by a number of conditional probabilities. A conditional probability is the probability of providing a certain response to an item, given that the subject belongs to a specific LC. In LCA, the classification – a division into clusters (LCs) – is realized by means of statistical criteria applied to conditional probabilities, while in usual procedures concerning mental model research, such as RAM, the classification is based on the number of correct responses. If the LCA procedure converges to a single class, then students cannot be classified into distinct clusters and their knowledge based on their response patterns cannot be classified into qualitatively different mental models. Then, the hypothesized students' mental models are not supported by the data. In contrast, when the LCA procedure converges to a limited number of distinct classes and the hypothesized (and being tested) mental-model account is correct then, the resulted classes should possess conditional probabilities consistent with these hypothesized mental-models. Note that the emerged clusters (if any) do not have to be in line with hypothesized mental-models; however, if they do, this would support a coherent mental model perspective.
Students often consider matter as continuous. Even those who have developed a particulate model of matter frequently show understanding difficulties in important aspects of the particle theory. These are mostly related to the space between particles, the intrinsic motion of the particles, the relative spacing between the particles in the three states, the attractions between particles and the nature of the particles themselves (Johnson, 1998a). In the main, students have problems in conceiving the notion of ‘empty space’ among particles. Thus, many students think that the space among particles is filled with various kinds of ‘stuff’, such as air or dust, as well as with other particles or particles of the same substance (Novick and Nussbaum, 1978; Lee et al., 1993; Johnson, 1998a). Frequently, molecules are thought to be in substances like ‘blueberries in a muffin’ rather than substances to be composed of molecules (Lee et al., 1993). According to this view, ‘particles are additional to the substances’ (Johnson, 1998a, p. 399). The intrinsic motion of particles (structural units – usually molecules) was also found to be a difficult concept (Novick and Nussbaum, 1978; Lee et al., 1993; Johnson, 1998a; Papageorgiou and Johnson, 2005). In the case of the gas state, this was attributed to various causes, e.g. to low specific gravity or to the action of air (Novick and Nussbaum, 1978; Lee et al., 1993). On the other hand, in the solid state, where no motion of the substance is visible, molecules are often thought to be still (Lee et al., 1993). In accordance with the above, Dow et al. (1978) found out that, although the majority of the students in their investigation indicated particle motion in the liquid and gas states, about a third of them advocated that there was no particle movement in the solid state.
Significant conceptual difficulties are reflected in the way that the students make a distinction between states, which are not always clear or complete. Secondary students' misconceptions about the liquid state are due to the fact that they consider liquid to be merely an intermediate state between solid and gas states. In this context, students overestimate the molecular spacing in the liquid state (Dow et al., 1978). In another study, molecular spacing in the gas state has been found to be underestimated (Pereira and Pestana, 1991).
In relation to the nature of the particles themselves, students have shown a great deal of difficulty in understanding that the properties of the states of substances are due to the collective behavior of particles. They often regard a particle, e.g. a molecule, as a small quantity of a substance having all the macroscopic properties of the substance. That is, ice molecules are regarded as frozen or ‘solid molecules’, water molecules as ‘liquid molecules’, and so forth (Lee et al., 1993; Johnson, 1998a). Furthermore, molecules are described to undergo “the same changes as the visible changes in the substances. Thus, molecules start to move when ice melts into water, molecules of water are heated up and make water boil, or molecules expand, contract, melt, evaporate, condense, and so forth” (Lee et al., 1993, p. 268).
In an attempt to further investigate the various students' conceptions related to the nature of matter and organize them into conceptual categories, a longitudinal study with secondary English pupils (ages 11–14) was carried out by implementing qualitative approaches and interview methods (Johnson, 1998a). In this study, students' misconceptions were investigated thoroughly and a number of distinct models for students' classification based on the strength of their particle thinking were proposed. These distinct models are defined below:
(1) Model X: continuous substance. Particle ideas have no meaning. (2) Model A: particles in the continuous substance. The particles are additional to the substance. (3) Model B: particles are the substance, but with macroscopic character. There is nothing between the particles. Individual particles are seen as being of the same quality as the macroscopic sample – literally small bits of it. (4) Model C: particles are the substance. The properties of a state are seen as collective properties of the particles. In addition to the above, there are intermediate models. It was also suggested that these models might represent stages through which students' ideas evolve towards the science model (Johnson, 1998a).
It is worth noting that the above models are described by Johnson within the context of the mental models perspective, which corresponds to a coherent structure that a novice might develop and sustain during the learning process and before he/she attains the scientific view. The present paper reexamines the above mental model account with a confirmatory latent class analysis.
Research in science education, as in all social science, implements a plethora of methodological approaches justified by various epistemological positions. The argument that the variables in social science are mostly categorical, and thus they are not subjected to measurement in a positivistic sense, has prevented many research studies from fostering quantitative methods. However, generalization and theory building are favored only by robust statistical methods. The latest advances in categorical data analysis have overcome many former statistical limitations and they are capable of testing sophisticated hypotheses and elucidating significant theoretical issues. Latent class analysis is one of the advanced methods that could be beneficial for the dialogue on the issue in question, and as such it was selected for the present confirmatory study.
Latent Class Analysis is the Latent Variable approach, where both manifest (observable) and latent variables are considered to be categorical. The way of accessing a latent variable, such as knowledge, is realized by the evaluation of the item responses, which are the manifest variables. If coherent mental models exist, then they are distinct entities/categories and ontologically they could be considered as categorical latent variables. These latent variables are responsible for the item responses (manifest variable), which in turn, if defined and measured at the nominal level of measurement, could be analyzed by LCA. The ‘coherence’ or ‘fragmentation’ of a latent variable itself is impossible to be directly observed by any means and thus, a researcher seeks for a representation at the observable level that could reflect the nature of knowledge as far as its coherency is concerned. Since the item responses are driven by a hypothesized coherent knowledge, the consistency of the responses in a variety of items reflects the coherency of the latent variable. Coherence is expressed by the capability of the latent variable not to contradict with itself when producing responses for a series of questions. By mapping the conditional probabilities of answering a series of question, consistently with a hypothesized mental model, one obtains a portrait reflecting the degree of coherency of the latent variable.
The main hypothesis being tested in this research is about the fragmented nature of students' knowledge on the structure of matter. Since LCA is a model based clustering analysis, it tests a hypothesized LC model and its null hypothesis (Ho) states that the model does not fit to the data; if Ho is rejected (based on the p-value), one could claim a statistically significant model. Thus, from a statistical point of view the fragmented nature of students' knowledge serves as the Ho, while the alterative hypothesis (Ha) supports the coherent mental models perspective.
Before the data collection, a small pilot study (N = 25) was carried out using the particular test followed by interviews with students discussing all items in order to detect possible communication difficulties of the test. A Cronbach's alpha reliability coefficient of 0.78 was obtained for the above instrument and the present sample. The instrument is a section of the full instrument used by Tsitsipis et al. (2010) and Tsitsipis et al. (2012). It is worth noticing that pupils' responses to this test were asked without any notification almost one year after pupils had been taught the relevant subject matter. Thus, the test measures the pupils' residual knowledge. The data collection and analysis was undertaken in Greek.
The test consists of two parts covering the following topics: the particulate nature of matter (first part), a number of properties of each of the three states as a result of the collective behavior of particles (second part). A description of the instrument is presented in the Appendix.
In a second stage, students' answers were assessed and were allocated into different levels of understanding the particulate nature of matter, which correspond to different mental models under examination. That is, the responses in each item were linked directly to one of the mental models and they were coded as X, A, B, C or S by examining each response alone and/or in combination with other responses. For example, co-evaluating students' answers to the questions 1A, 2A and 3A (see Appendix), a student could be linked to model X or not. If not, then a co-evaluation of the answers to the questions 1B, 2B, 3B, 1C, 2C and 3C could link the student to model A or not. If not, then a co-evaluation of the answers to the questions 4 (A–C) and 5 (A, B) could link the student to model B, C or S. In the open-ended question the language used had an influential role in the evaluation. For example, a student who answered that, there are ‘molecules’ (which excludes X), ‘nothing between molecules’ (which advocates B, C, S rather than A) and ‘water molecules are in liquid state’ (which excludes C, S) was linked to model B. In contrast, a student who answered that, there are ‘molecules’ (which excludes X) and ‘water between water molecules’ (which excludes B, C, S) was linked to model A.
The four models X, A, B and C are the ones which had been introduced by Johnson (1998a, 1998b, 1998c), whereas the fifth model (S) was added in order to have a link to the complete scientific view; as also suggested by Johnson, model C does not represent a full scientific view. Subjects with model S would use the proper scientific language in their descriptions and explanations and they would be those who have undoubtedly acquired the scientific view. Finally, in case of peculiar, confused, irrelevant responses or no responses, the student was linked to another model O. In some cases where a students' response could not be undoubtedly linked to merely one mental model, alternative codes were recorded and they were used alternatively as inputs in the LCA procedure.
Some relevant examples are given below showing that students' knowledge, a latent variable that is being accessed by their responses, is not coherent and in line with a specific mental mode:
A student's responses to the questions 1A, 2A and 3A, for instance, might be different by choosing a continuous material for one state (e.g. gas state – answer which was not rare) and the particulate model for others (e.g. solid and/or liquid states); when he referred to motion of particles, he/she accepted the moving particles for the gas but not for the solid state, while talking about the space between particles he refuted the particle nature for example for the solid state, but he accepted it for the liquid state. Note that by examining similar items across physical states the greatest discrepancy in the responses became evident. Furthermore, although a student advocated model X, for instance in 4A, he discussed whether an isolated molecule of water has come from ice, liquid water or water in the gas state, respectively. Similarly, a student stated in 4B that all the three cases of water molecules (in the solid, liquid and gas state) are the same, whereas in 5B he described the molecules of sugar, water and oxygen in a context of macroscopic characteristics (e.g., a solid molecule for sugar, a liquid one for water and so on). The overall degree of inconsistency of a student's responses increases in cases where confused responses or no responses were provided.
Even though the accuracy of linking each response to a mental model might be restricted by the instrument and the students' limited expressions and language use, given that the evaluation scheme was followed strictly and consistently throughout the cases, the final portrait of LC's conditional probabilities (see Fig. 1–3) is still anticipated to reflect the actual inconsistency or consistency in students responses.
![]() | ||
Fig. 1 Conditional probabilities for LC1 (9.2%, dominance of S-model). Items are for the solid state. |
![]() | ||
Fig. 2 Conditional probabilities for LC2 (28.6%). Items are for the solid state. |
![]() | ||
Fig. 3 Conditional probabilities for LC3 (62.2%). Items are for the solid state. |
Attempts to separately analyze the questions concerning the solid, liquid and gas states, respectively, lead to a successful classification for some groups of items, which are presented next. For these groups of items three clusters, i.e. three LCs, emerged, which fitted satisfactorily to the data for the solid and the gas states. Thus, in the case of these two states, a 3-cluster model was selected, while in the case of the liquid state merely one cluster emerged. The fit of such a latent-class model was assessed by the indicators already reported earlier, i.e. the Npar, BIC and bootstrapped p-values, which are presented in Tables 1 and 2 (Vermunt and Magidson, 2000).
L2 | BIC | Npar | df | p-value | Class. err. | |
---|---|---|---|---|---|---|
a The most parsimonious and best fitting model: Npar, number of parameters; L2, likelihood ratio statistic; BIC, Bayesian Information Criterion; df, degrees of freedom; bootstrapped p-value; classification error. | ||||||
1-Cluster | 526.84 | 2776.73 | 17 | 239 | 0.00 | 0.000 |
2-Cluster | 400.47 | 2750.16 | 35 | 221 | 0.00 | 0.047 |
3-Clustera | 331.17 | 2780.68 | 53 | 203 | 0.07a | 0.011 |
4-Cluster | 289.18 | 2838.51 | 71 | 185 | 0.12 | 0.014 |
5-Cluster | 260.26 | 2917.43 | 89 | 167 | 0.07 | 0.011 |
6-Cluster | 240.98 | 2986.64 | 107 | 149 | 0.14 | 0.014 |
L2 | BIC | Npar | df | p-value | Class. err. | |
---|---|---|---|---|---|---|
a The most parsimonious and best fitting model: Npar, number of parameters; L2, likelihood ratio statistic; BIC, Bayesian Information Criterion; df, degrees of freedom; bootstrapped p-value; classification error. | ||||||
1-Cluster | 737.95 | 3032.73 | 17 | 239 | 0.00 | 0.0000 |
2-Cluster | 494.51 | 2982.07 | 35 | 221 | 0.03 | 0.0520 |
3-Clustera | 439.29 | 3011.00 | 53 | 203 | 0.15a | 0.0010 |
4-Cluster | 401.46 | 3074.56 | 71 | 185 | 0.17 | 0.0154 |
5-Cluster | 372.51 | 3144.19 | 89 | 167 | 0.17 | 0.0143 |
6-Cluster | 346.02 | 3221.71 | 107 | 149 | 0.16 | 0.0154 |
In particular, Table 1 shows the results of the LCA for the solid state. Among all the possible cluster models (LC models), the 3-cluster model appears to be statistically significant (p = 0.07) for the solid state, the most parsimonious and the best fitting one. This LCA was based on 6 items: 1A, 1B, 1C, 4A, 4B and 4C. The corresponding three LCs account for 9.2%, 28.6% and 62.2% of the sample respectively. In each of the latent classes, a set of conditional probabilities is assigned for each item, that is, the probability for an item to be answered in line with one of the hypothesized mental models, given that the subject is a member of this latent class. Accordingly, these conditional probabilities for the three emerged latent classes for the solid state are shown in Fig. 1–3. If groups of students respond similarly and consistently with any of the hypothesized mental models then they would be allocated to the same cluster (latent class). Interestingly, the emerged clusters/latent classes, namely the LC1, LC2 and LC3 include subjects that do not respond similarly and consistently with any of the hypothesized mental models; that is, the set of conditional probabilities in each LC is not homogeneous in terms of responses in line to a certain mental model. For example, in Fig. 2 the item 4C has a probability of 0.37 to be answered in line with the A-model, a probability of 0.32 to be answered in line with the B-model, a probability of 0.21 to be answered in line with the C-model and a probability of 0.10 to be answered in line with the S-model. Thus, the emerged LCs do not coincide with any of the hypothesized mental models. Students' knowledge appears to be fragmented, not showing any coherence as the mental models perspective hypothesized. Note that LC1 (Fig. 1) tends to be the most homogeneous compared to LC2 and LC3 (Fig. 2 and 3, respectively), since students in this LC have potentially attained the scientific view (S-model).
Analogous to the solid state are the LC results for the gas state. By using the corresponding indicators for the items 3A, 3B, 3C, 4A, 4B and 4C, the 3-cluster model (see Table 2) is proved to be the most parsimonious and the best fitted (p = 0.15). The three LCs account for 24.6%, 30.3% and 45.1% of the sample respectively. In Fig. 4–6, the conditional probabilities of these three latent classes, namely the LC1, LC2 and LC3, are presented in relation to the items. All three latent classes are characterized by mixed response patterns switched among all the models (O, X, A, B, C, S), demonstrating fragmented students' knowledge.
![]() | ||
Fig. 4 Conditional probabilities for LC1 (24.6%). Items are for the gas state. |
![]() | ||
Fig. 5 Conditional probabilities for LC2 (30.3%). Items are for the gas state. |
![]() | ||
Fig. 6 Conditional probabilities for LC3 (45.1%). Items are for the gas state. |
For the liquid state, LCA did not converge to a significant classification based on the corresponding items (2A, 2B, 2C, 4A, 4B and 4C) and resulted one-cluster model, that is, for the liquid state the students' knowledge appears to possess an even higher degree of fragmentation. Moreover, although the LC analysis was also performed with various item combinations including further items such as 5A and 5B, the analysis did not reach different results.
Conclusively, LCA based on conditional probabilities did not provide evidence that any of the revealed latent classes is directly associated with the hypothesized mental models (Johnson, 1998a). The results of incoherent students' knowledge of course concern the specific mental models, since the coding was based on them. Different hypothesized mental models could be tested analogously by proper coding and the corresponding questionnaire.
In order to explicate how the LCA can serve and facilitate testing related hypotheses on coherence versus fragmented knowledge, a didactic demonstration is provided below with a semi-simulation experiment analogous to the use of a “standard sample” in chemical analysis:
In the present empirical data set an artificial segment of data was added with the following properties: all subjects' responses are in line with the mental model B for all items except item 4B, where 45% of the cases respond according to A and 55% of the cases according to X. From the ensuing analysis of the modified data (original data plus the artificial segment) four latent classes emerged. LCA has captured the artificial segment as the forth LC preserving its properties; Fig. 7 shows its conditional probability pattern, which is almost homogeneous. In this LC the conditional probabilities tend to be or are equal to the unity for mental model B and tend to be or are equal to zero for the other models in all items (except item 4B). The picture shown in Fig. 7 is the one expected when an emerged LC coincides with a coherent mental model; this is substantially different from the ones (Fig. 1 to 6) resulting from the empirical data.
![]() | ||
Fig. 7 The artificially introduced segment (16.4%) in the sample with responses in line with model B for all items except item 4B (45% A and 55% X) was captured as LC4 preserving the conditional probabilities for each item response and its coherency. Items are for the solid state. |
Prior to further discussing the findings and drawing specific conclusions for the fragmented nature of the students' knowledge about the structure of matter, it is useful at this point to make a reference to some limitations of the study of both general and specific interest. One limitation originates from the utilization of the written instrument and the coding scheme, which is based on the given responses without having the opportunity to control random errors or miscommunications by double-checking students' answers, as it is possible for instance in an interview approach (e.g. Engel and Driver, 1986). The techniques of collecting and coding data have been an issue of debate and their advantages and disadvantages have been discussed repeatedly in the past (Vosniadou et al., 2004; Straatemeier et al., 2008).
An approach widely used is the “generative method”, which implements interviews with ‘open-ended’ questions. In this approach subjects are asked to respond to certain questions by producing drawings or describing verbally whatever they have in mind. Thus, the researcher could access the subjects’ knowledge from ‘on the spot’ formation of mental representations (Vosniadou et al., 2004). However, some disadvantages associated with the interview methods, such as the prolonged processes of repeated questions and the social interference, could be sources of errors; moreover, the limited number of subjects under examination restricts generalized conclusions. This method nevertheless is particularly useful for exploratory studies, where the majority of possible mental representations could be detected, categorized and coded for a large scale research endeavor. Alternatively, the ‘forced-choice’ questionnaires were proposed as a better method for data collection (Straatemeier et al., 2008). It could be applied in large samples and it can access subjects' knowledge satisfactorily, given that the choices include all (or the most of) the possible alternatives. The advantage of a forced-choice questionnaire is that it avoids social interaction; it can be better structured and objectively assessed. In the present research the instrument included both open-ended and forced-choice items.
Moreover, Johnson (1998a) himself proceeded to an analysis of his findings in a two dimensional space, introducing a diagram where models were presented as territories in the ‘outcome space’ without distinct ‘boundaries’. Although each of these identifiable territories was defined with a degree of internal coherence, in that graph, the idea is in fact a move beyond the consideration of a ‘model’ in the way presented within the coherence-knowledge perspective.
Ergo, methodological issues existing behind the controversial results in searching for coherent models arise for further discussion. LCA is able to capture and overcome the above methodological complicatedness. Even though particular clusters were revealed, i.e. the null hypothesis was rejected as far the clustering procedure is concerned, the emerged latent classes do not coincide with any of the hypothesized mental models, which are supposed to be homogeneous in terms of response conditional probabilities. Thus, one may conclude that the null hypothesis (Ho) on fragmented knowledge was not rejected for the present data. The only exception might be the S-model, which expectably was shown to exhibit a degree of coherence, since students in this LC have acquired the scientific view. For the present data, where the coherence at the level of a large ‘grain-size knowledge’ – mental models – was tested, LCA supported the fragmented-knowledge perspective, opposing the mental models theory established so far for the students' knowledge on the structure of matter.
Consequently, the central message of the present work is that research in science education and cognitive science might benefit from advanced statistical modelling, getting direct and clear evidence when testing hypotheses such as the one in question. Furthermore, robust statistical modeling such as LCA can serve as a tool to investigate how knowledge-in-pieces is organised (and under what conditions) either into the scientific view in some cases or possibly into conceptual resources of varying grain-sizes in some other circumstances. To that end, longitudinal studies are needed to show how these changes are realised across ages. This work is a confirmatory one, since the coding scheme is based on hypothesized mental models. In a different methodological approach an exploratory LCA could be implemented, in which a different coding scheme could be applied and the emerged LCs could be examined in an attempt to reveal possible mental models that make sense to the theory building. This is a different analysis compared to the present case, which indicates the prospective usability of LCA. At this point, it is also worth mentioning another interesting application and data analysis with LCA, where the calculated conditional probabilities of receiving correct answers were used to demonstrate students' learning progression in various grades (Harlow et al., 2011).
In the current literature, it is apparent that new rounds of discussions have begun on the issue in question (Taber, 2008; Turcotte, 2012; Kirbulut and Beeth, 2013) and in our opinion the present approach opens a new research area that will contribute to this dialogue and ultimately to the theory building. Advancements in conceptual change theories are closely related to the present issue and the implications for education are by far crucial. Learning strategies and teaching practices are strongly dependent on our deeper understanding of the nature of students' knowledge. Thus, the theoretical frameworks in education and science education in particular have to minimize their reliance on metaphors and unfounded epistemological choices and conversely, they have to promptly adapt new research methodology advancements.
Part 1: (The particulate nature of matter)
The first 3 items (1A, 1B, and 1C) concern the solid state.
1A. Pupils are asked to choose among five alternatives (see Fig. 8 of the appendix) the figure that best represents what they would ‘see’ if they could observe a sugar grain using a hypothetical magnifying glass enabling the view of the grain structure.
![]() | ||
Fig. 8 The five alternatives given to the pupils. In each them a note helped to clarify the corresponding representation. A description of these notes follows: (1) continuous material, (2) molecules (of nearly spherical shape) in array, not touching each other, (3) nothing, (4) molecules (of nearly spherical shape) in a random position relatively close to each other, not touching each other, (5) molecules (of nearly spherical shape) in random position, mostly far away from each other. |
1B. Pupils are asked to explain what they think exists in-between molecules, in case they chose a figure depicting molecules. Otherwise, they do not have to answer this question.
1C. Pupils are asked to answer whether or not they think that the view of the sugar structure through the hypothetical magnifying glass would remain still as the time is passing. They are also asked to explain and justify their answers.
The following 3 items (2A, 2B, and 2C) concern the liquid state.
2A. Pupils are asked to choose among five alternatives (see Fig. 8 of the appendix) the figure that best represents what they would “see” if they could observe a drop of pure (liquid) water using a hypothetical magnifying glass enabling the view of the structure of the drop.
2B. Pupils are asked to explain what they think exists in-between molecules, in case they chose a figure depicting molecules. Otherwise, they do not have to answer this question.
2C. Pupils are asked to answer whether or not they think that the view of the water structure through the hypothetical magnifying glass would remain still as the time is passing. They were also asked to explain and justify their answers.
The following 3 items (3A, 3B, and 3C) concern the gas state.
3A. Pupils are asked to choose among five alternatives (see Fig. 8 of the appendix) the figure that best represents what they would “see” if they could observe a very small quantity of oxygen, found inside a vase containing pure oxygen, a hypothetical magnifying glass enabling the view of the structure of the oxygen.
3B. Pupils are asked to explain what they think exists in-between molecules, in case they chose a figure depicting molecules. Otherwise, they do not have to answer this question.
3C. Pupils are asked to answer whether or not they think that the view of the oxygen structure through the hypothetical magnifying glass would remain still as the time is passing. They are also asked to explain and justify their answers.
At that point, pupils are prompted to circumvent the following items 4 and 5 in case they have not adopted a molecular structure of the substances in the previous items.
Part 2: (The properties of a state as a result of the collective behavior of particles)
The following 3 items (4A, 4B, and 4C) concern the same substance in three different temperatures.
4A. Pupils are prompted to make the assumption that they have isolated one single molecule from one of the following: a block of ice, some pure (liquid) water or some pure water in the gas state. They are asked whether or not they could understand if the isolated molecule has come from ice, liquid water or water in the gas state, respectively. Then, they are also asked to explain and justify their answers.
4B. Pupils are prompted to make the assumption that they have isolated one single molecule from a block of ice, another single molecule from a quantity of pure liquid water and a third single molecule from a quantity of water in the gas state. They are asked whether or not they could determine a physical state for each one of the three molecules and if yes, then, what this state is. They are also asked to justify their answers.
4C. Pupils are prompted to make the assumption that they have isolated one single molecule from a block of ice, another single molecule from a quantity of pure liquid water and a third single molecule from a quantity of water in the gas state. They are asked to compare the shape and the magnitude of the three molecules. Then, they are also asked to justify their answers.
The following 2 items (5A and 5B) concern three different substances under the same normal conditions.
Pupils are prompted to make the assumption that they have isolated one single molecule from each one of the following three substances: sugar (solid), water (liquid) and oxygen (gas).
5A. They are asked whether or not they could determine a physical state for each one of the three molecules and if yes, then, what this state is. They are also asked to justify their answers.
5B. They are asked whether they think that the three molecules are different or not. They are also asked to explain and justify their answers.
This journal is © The Royal Society of Chemistry 2013 |