Skip to main content

Reaction Coordinate Diagram Inventory

RCDI

    OVERVIEW
    Overview
    Listed below is general information about the instrument.
    Summary
    Original author(s)
    • Atkinson, M.B., Popova, M., Croisant, M., Reed, D.J., & Bretz, S.L.

    Original publication
    • Atkinson, M.B., Popova, M., Croisant, M., Reed, D.J., & Bretz, S.L. (2020). Development of the Reaction Coordinate Diagram Inventory: Measuring Student Thinking and Confidence. Journal of Chemical Education, 97(7), 1841-1851.

    Year original instrument was published 2020
    Inventory
    Number of items 19
    Number of versions/translations 1
    Cited implementations 2
    Language
    • English
    Country United States
    Format
    • Multiple Choice
    • Response Scale
    Intended population(s)
    • Students
    • Undergraduate
    Domain
    • Cognitive
    Topic
    • Energy
    • Representations
    Evidence
    The CHIRAL team carefully combs through every reference that cites this instrument and pulls all evidence that relates to the instruments’ validity and reliability. These data are presented in the following table that simply notes the presence or absence of evidence related to that concept, but does not indicate the quality of that evidence. Similarly, if evidence is lacking, that does not necessarily mean the instrument is “less valid,” just that it wasn’t presented in literature. Learn more about this process by viewing the CHIRAL Process and consult the instrument’s Review (next tab), if available, for better insights into the usability of this instrument.

    Information in the table is given in four different categories:
    1. General - information about how each article used the instrument:
      • Original development paper - indicates whether in which paper(s) the instrument was developed initially
      • Uses the instrument in data collection - indicates whether an article administered the instrument and collected responses
      • Modified version of existing instrument - indicates whether an article has modified a prior version of this instrument
      • Evaluation of existing instrument - indicates whether an article explicitly provides evidence that attempt to evaluate the performance of the instrument; lack of a checkmark here implies an article that administered the instrument but did not evaluate the instrument itself
    2. Reliability - information about the evidence presented to establish reliability of data generated by the instrument; please see the Glossary for term definitions
    3. Validity - information about the evidence presented to establish reliability of data generated by the instrument; please see the Glossary for term definitions
    4. Other Information - information that may or may not directly relate to the evidence for validity and reliability, but are commonly reported when evaluating instruments; please see the Glossary for term definitions
    Publications: 1 2

    General

    Original development paper
    Uses the instrument in data collection
    Modified version of existing instrument
    Evaluation of existing instrument

    Reliability

    Test-retest reliability
    Internal consistency
    Coefficient (Cronbach's) alpha
    McDonald's Omega
    Inter-rater reliability
    Person separation
    Generalizability coefficients
    Other reliability evidence

    Vailidity

    Expert judgment
    Response process
    Factor analysis, IRT, Rasch analysis
    Differential item function
    Evidence based on relationships to other variables
    Evidence based on consequences of testing
    Other validity evidence

    Other information

    Difficulty
    Discrimination
    Evidence based on fairness
    Other general evidence
    Review
    DISCLAIMER: The evidence supporting the validity and reliability of the data summarized below is for use of this assessment instrument within the reported settings and populations. The continued collection and evaluation of validity and reliability evidence, in both similar and dissimilar contexts, is encouraged and will support the chemistry education community’s ongoing understanding of this instrument and its limitations.
    This review was generated by a CHIRAL review panel. Each CHIRAL review panel consists of multiple experts who first individually review the citations of the assessment instrument listed on this page for evidence in support of the validity and reliability of the data generated by the instrument. Panels then meet to discuss the evidence and summarize their opinions in the review posted in this tab. These reviews summarize only the evidence that was discussed during the panel which may not represent all evidence available in the published literature or that which appears on the Evidence tab.
    If you feel that evidence is missing from this review, or that something was documented in error, please use the CHIRAL Feedback page.

    Panel Review: Reaction Coordinate Diagram Inventory (RCDI)

    (Post last updated June 24, 2022)

    Review panel summary

    The Reaction Coordinate Diagram Inventory (RCDI) consists of 21 multiple-choice items with 2 of the items being two-tiered answer-reason items, all 21-items include a 0-100% confidence scale (not confident to confident) [1]. The instrument was designed to assess student’s thinking and confidence in interpreting reaction coordinate diagrams and their correspondence to reaction mechanisms. The instrument has been evaluated with second semester general, organic, and physical/biophysical chemistry students. Several aspects of validity and reliability have been assessed for the data generated by the RCDI. Evidence for test content validity was provided by experts (general and organic chemistry instructors) and students [1] and was based on a theoretical framework presented by the authors [2]. During pilot testing, undergraduate students provided feedback on clarity of the RCDI items and this was used to provide evidence for response process validity [1]. Relation with other variables was supported by comparing the different means scored by three groups of students from different classes. Higher mean scores for more advanced classes with significant differences were identified.

    Reliability, item difficulty, and item discrimination of the data generated by the RCDI were examined by using Classical test theory (CTT). In terms of reliability, coefficient alpha has been used to estimate single administration reliability for the whole instrument [1, 2]. Even though the coefficient alpha value reported for the RCDI was in an acceptable range, the designers argued that coefficient alpha is not the most appropriate way to assess the reliability in the context of a concept inventory such as the RCDI. The authors justify this statement by mentioning that concept inventories are designed based on students’ misconceptions and fragmented ideas that are also used when students respond to items. Additionally, coefficient alpha implies an underlying unidimensional construct, which is not the case for concept inventories such as the RCDI. Fergusons’ delta was used to demonstrate that the students could achieve a wide distribution of scores across the range of possible scores on the RCDI. The item difficulty and item discrimination indices showed that two of the items (7 and 8) had values out of the acceptable range [1, 2]. The authors argued that these items should still be retained since they could suggest the prevalence of the misconceptions expressed in these items that are held by the students [1, 2]. Item difficulty and discrimination estimates decreased from general chemistry to organic in a longitudinal study which was an indication that the top-performing and low-performing students answered these items correctly in their later chemistry courses [2].

    Recommendations for use

    The RCDI was designed to assess students’ understanding and confidence with the kinetic and thermodynamic concepts encoded within RCDs [1, 2]. Literature to date provides support for many aspects of validity and reliability to support the use of the RCDI in assessing students’ conceptual understanding pertaining to RCDs. However, while aspects at the item level are well described, evidence to support the internal structure validity of the instrument for use with a total score [1, 2] would further strengthen the interpretation of RCDI data.

    Details from panel review

    The designers of the RCDI gathered and reported several aspects of validity and reliability evidence for the instrument [1, 2]. For example the test content was decided by the authors and then the designed items were checked for clarity and correctness with experts and students [1]. Undergraduate student interviews were used to support item modifications and interpretations, providing evidence in support of the response process validity [1]. The authors used Classical Test Theory to analyze the psychometric properties of individual items within the instrument. Evidence for item difficulty, item discrimination, and reliability of the scores were provided [2]. Two of the items (items 7 and 8) had indices out of the acceptable range. The designers used the Sankey diagrams to show changes in student’s responses to individual items over time [2]. The results of a longitudinal study showed no significant change in the RCDI scores or confidence from the beginning to the end of second-semester general chemistry but the instrument captured changes in students’ performance by the end of first-semester organic [2].

    References

    [1] Atkinson, M.B., Popova, M., Croisant, M., Reed, D.J., & Bretz, S.L. (2020). Development of the reaction coordinate diagram inventory: Measuring student thinking and confidence. Journal of Chemical Education, 97(7), 1841-1851. https://doi.org/10.1021/acs.jchemed.9b01186

    [2] Atkinson, M.B, & Bretz, S.L. (2021). Measuring changes in undergraduate chemistry students’ reasoning with reaction coordinate diagrams: A longitudinal, multi-institution study. Journal of Chemical Education, 98(4), 1064-1076. https://doi.org/10.1021/acs.jchemed.0c01419

    Versions
    This instrument has not been modified nor was it created based on an existing instrument.
    Citations
    Listed below are all literature that develop, implement, modify, or reference the instrument.
    1. Atkinson, M.B., Popova, M., Croisant, M., Reed, D.J., & Bretz, S.L. (2020). Development of the Reaction Coordinate Diagram Inventory: Measuring Student Thinking and Confidence. Journal of Chemical Education, 97(7), 1841-1851.

    2. Atkinson, M.B., & Bretz, S.L. (2021). Measuring Changes in Undergraduate Chemistry Students' Reasoning with Reaction Coordinate Diagrams: A Longitudinal, Multi-institution Study. Journal of Chemical Education, 98(4), 1064-1076.