Skip to main content

Particulate Nature Of Matter And Chemical Bonding Diagnostic Instrument

    OVERVIEW
    Overview
    Listed below is general information about the instrument.
    Summary
    Original author(s)
    • Othman, J., Treagust, D.F., & Chandrasegaran, A.L.

    Original publication
    • Othman, J., Treagust, D.F., & Chandrasegaran, A.L. (2008). An investigation into the relationship between students' conceptions of the particulate nature of matter and their understanding of chemical bonding. International Journal of Science Education, 30

    Year original instrument was published 2008
    Inventory
    Number of items 10
    Number of versions/translations 1
    Cited implementations 2
    Language
    • English
    Country Singapore, United States
    Format
    • Multiple Choice
    Intended population(s)
    • Students
    • High School
    • Undergraduate
    Domain
    • Cognitive
    Topic
    • Bonding
    • Nature of Matter
    Evidence
    The CHIRAL team carefully combs through every reference that cites this instrument and pulls all evidence that relates to the instruments’ validity and reliability. These data are presented in the following table that simply notes the presence or absence of evidence related to that concept, but does not indicate the quality of that evidence. Similarly, if evidence is lacking, that does not necessarily mean the instrument is “less valid,” just that it wasn’t presented in literature. Learn more about this process by viewing the CHIRAL Process and consult the instrument’s Review (next tab), if available, for better insights into the usability of this instrument.

    Information in the table is given in four different categories:
    1. General - information about how each article used the instrument:
      • Original development paper - indicates whether in which paper(s) the instrument was developed initially
      • Uses the instrument in data collection - indicates whether an article administered the instrument and collected responses
      • Modified version of existing instrument - indicates whether an article has modified a prior version of this instrument
      • Evaluation of existing instrument - indicates whether an article explicitly provides evidence that attempt to evaluate the performance of the instrument; lack of a checkmark here implies an article that administered the instrument but did not evaluate the instrument itself
    2. Reliability - information about the evidence presented to establish reliability of data generated by the instrument; please see the Glossary for term definitions
    3. Validity - information about the evidence presented to establish reliability of data generated by the instrument; please see the Glossary for term definitions
    4. Other Information - information that may or may not directly relate to the evidence for validity and reliability, but are commonly reported when evaluating instruments; please see the Glossary for term definitions
    Publications: 1 2

    General

    Original development paper
    Uses the instrument in data collection
    Modified version of existing instrument
    Evaluation of existing instrument

    Reliability

    Test-retest reliability
    Internal consistency
    Coefficient (Cronbach's) alpha
    McDonald's Omega
    Inter-rater reliability
    Person separation
    Generalizability coefficients
    Other reliability evidence

    Validity

    Expert judgment
    Response process
    Factor analysis, IRT, Rasch analysis
    Differential item function
    Evidence based on relationships to other variables
    Evidence based on consequences of testing
    Other validity evidence

    Other information

    Difficulty
    Discrimination
    Evidence based on fairness
    Other general evidence
    Review
    DISCLAIMER: The evidence supporting the validity and reliability of the data summarized below is for use of this assessment instrument within the reported settings and populations. The continued collection and evaluation of validity and reliability evidence, in both similar and dissimilar contexts, is encouraged and will support the chemistry education community’s ongoing understanding of this instrument and its limitations.
    This review was generated by a CHIRAL review panel. Each CHIRAL review panel consists of multiple experts who first individually review the citations of the assessment instrument listed on this page for evidence in support of the validity and reliability of the data generated by the instrument. Panels then meet to discuss the evidence and summarize their opinions in the review posted in this tab. These reviews summarize only the evidence that was discussed during the panel which may not represent all evidence available in the published literature or that which appears on the Evidence tab.
    If you feel that evidence is missing from this review, or that something was documented in error, please use the CHIRAL Feedback page.

    Panel Review: Particulate Nature Of Matter And Chemical Bonding Diagnostic Instrument

    (Post last updated 28 December 2024)

    Review panel summary   
    The Particulate Nature of Matter and Chemical Bonding Diagnostic Instrument (PNM-CBDI) is a 10 item instrument with five items assessing students’ conceptions of the particulate nature of matter and five items assessing chemical bonding [1]. It has been used with high school students in Singapore [1] and general chemistry students in a U.S. college [3]. The items used are two-tiered, where the student selects an answer to the question and then selects a reason for their answer. Most items for the PNM-CBDI were taken from existing instruments, with a few developed for this work. The items are intended to identify misconceptions held by students. Single administration reliability evidence is based on coefficient alpha values of 0.66 [1] and 0.62 [3] when the 20 items (10 answers plus 10 reasons) are analyzed separately. When the answers and reason are scored together, resulting in a 10-item set, alpha decreases to 0.42 [3]. Weak evidence of internal structure validity, for either a one- or two-factor solution, was provided using confirmatory factor analysis [3]. Some validity evidence in the form of relation to other variables has been found. In one study, students in higher grades had greater overall scores than students in lower grades [1]. Scores on the assessment, when SAT Math was accounted for, were higher for students that had taken a preparatory chemistry course than for students that had not [3].

    Recommendations for use   
    According to the validity evidence found to date, there is no support for the proposed two-factor structure (i.e., students’ conceptions of particulate nature of matter and students’ understanding of chemical bonding). The lack of validity evidence supporting the proposed factors, and the alpha values suggests that this instrument is more akin to a content assessment. Without further psychometric exploration, the instrument is best used as an item-by-item diagnostic tool for the presence of common misconceptions, and not as a tool to measure the proposed cognitive constructs of conceptions of particulate nature of matter and students' understanding of chemical bonding.

    Details from panel review   
    The PNM-CBDI was developed to measure two proposed cognitive constructs: students’ conceptions of particulate nature of matter and students’ understanding of chemical bonding. The development process followed a four-step process: (1) Identification of propositional knowledge statements, (2) Development of concept map(s), (3) Relating of propositional knowledge to the concept map(s), and (4) Validation of the content (by two academics not in the research team) [2]. One of the concept maps was developed for this instrument (particulate nature of matter) and the other (concepts of chemical bonding) was modified from a concept map from previous study. The items developed were not developed specifically for this instrument, rather they were taken from other instruments. The development paper does not describe in detail the relationship between the selected items and the concept maps [1]. Fit statistics from confirmatory factor analysis models of the proposed two-factor model were acceptable, the item loadings were problematic with items only weakly loading on their proposed factor [3]. Additionally, there was high correlation, > 0.83, between the factors suggesting redundancy in the two factor solution. Fit statistics for the one factor model were slightly worse, again the items only weakly loaded on the one factor [3].

    References

    [1] Jazilah Othman , David F. Treagust & A. L. Chandrasegaran (2008). An Investigation into the Relationship between Students’ Conceptions of the Particulate Nature of Matter and their Understanding of Chemical Bonding. Int. J. Sci. Educ., 30(11), 1531-1550.

    [2] Heredia, K., Xu., X., & Lewis, J. E. (2012). The application and evaluation of a two-concept diagnostic instrument with students entering college general chemistry. Chem. Educ. Res. & Pract., 13, 30-38.

    Versions
    Listed below are all versions and modifications that were based on this instrument or this instrument were based on.
    Name Authors
    • Ozalp, D., & Kahveci, A.

    Citations
    Listed below are all literature that develop, implement, modify, or reference the instrument.
    1. Othman, J., Treagust, D.F., & Chandrasegaran, A.L. (2008). An investigation into the relationship between students' conceptions of the particulate nature of matter and their understanding of chemical bonding. International Journal of Science Education, 30

    2. Heredia, K., Xu, X., & Lewis, J.E. (2012). The application and evaluation of a two-concept diagnostic instrument with students entering college general chemistry. Chemistry Education Research and Practice, 13(1), 30-38.