Skip to main content

Implicit Information From Lewis Structures Instrument

IILSI

    OVERVIEW
    Overview
    Listed below is general information about the instrument.
    Summary
    Original author(s)
    • Cooper, M.M., Underwood, S.M., & Hilley, C.Z.

    Original publication
    • Cooper, M.M., Underwood, S.M., & Hilley, C.Z. (2012). Development and validation of the implicit information from Lewis structures instrument (IILSI): Do students connect structures with properties?. Chemistry Education Research and Practice, 13(3), 195-2

    Year original instrument was published 2012
    Inventory
    Number of items 17
    Number of versions/translations 2
    Cited implementations 4
    Language
    • English
    Country United States
    Format
    • True/False
    • Open Ended
    Intended population(s)
    • Students
    • Undergraduate
    • Graduate
    Domain
    • Cognitive
    Topic
    • General Chemistry
    • Chemical Structure
    Evidence
    The CHIRAL team carefully combs through every reference that cites this instrument and pulls all evidence that relates to the instruments’ validity and reliability. These data are presented in the following table that simply notes the presence or absence of evidence related to that concept, but does not indicate the quality of that evidence. Similarly, if evidence is lacking, that does not necessarily mean the instrument is “less valid,” just that it wasn’t presented in literature. Learn more about this process by viewing the CHIRAL Process and consult the instrument’s Review (next tab), if available, for better insights into the usability of this instrument.

    Information in the table is given in four different categories:
    1. General - information about how each article used the instrument:
      • Original development paper - indicates whether in which paper(s) the instrument was developed initially
      • Uses the instrument in data collection - indicates whether an article administered the instrument and collected responses
      • Modified version of existing instrument - indicates whether an article has modified a prior version of this instrument
      • Evaluation of existing instrument - indicates whether an article explicitly provides evidence that attempt to evaluate the performance of the instrument; lack of a checkmark here implies an article that administered the instrument but did not evaluate the instrument itself
    2. Reliability - information about the evidence presented to establish reliability of data generated by the instrument; please see the Glossary for term definitions
    3. Validity - information about the evidence presented to establish reliability of data generated by the instrument; please see the Glossary for term definitions
    4. Other Information - information that may or may not directly relate to the evidence for validity and reliability, but are commonly reported when evaluating instruments; please see the Glossary for term definitions
    Publications: 1 2 3 4

    General

    Original development paper
    Uses the instrument in data collection
    Modified version of existing instrument
    Evaluation of existing instrument

    Reliability

    Test-retest reliability
    Internal consistency
    Coefficient (Cronbach's) alpha
    McDonald's Omega
    Inter-rater reliability
    Person separation
    Generalizability coefficients
    Other reliability evidence

    Validity

    Expert judgment
    Response process
    Factor analysis, IRT, Rasch analysis
    Differential item function
    Evidence based on relationships to other variables
    Evidence based on consequences of testing
    Other validity evidence

    Other information

    Difficulty
    Discrimination
    Evidence based on fairness
    Other general evidence
    Review
    DISCLAIMER: The evidence supporting the validity and reliability of the data summarized below is for use of this assessment instrument within the reported settings and populations. The continued collection and evaluation of validity and reliability evidence, in both similar and dissimilar contexts, is encouraged and will support the chemistry education community’s ongoing understanding of this instrument and its limitations.
    This review was generated by a CHIRAL review panel. Each CHIRAL review panel consists of multiple experts who first individually review the citations of the assessment instrument listed on this page for evidence in support of the validity and reliability of the data generated by the instrument. Panels then meet to discuss the evidence and summarize their opinions in the review posted in this tab. These reviews summarize only the evidence that was discussed during the panel which may not represent all evidence available in the published literature or that which appears on the Evidence tab.
    If you feel that evidence is missing from this review, or that something was documented in error, please use the CHIRAL Feedback page.

    Panel Review: Implicit Information from Lewis Structures Instrument

    (Post last updated July 21, 2021)

    Review panel summary

    The Implicit Information from Lewis Structures Instrument (IILSI) is a one question instrument that provides students with seventeen options to select as to whether they think the option can be determined from Lewis structures. It has been evaluated with students in the general chemistry sequence [1, 2], the organic chemistry sequence [1, 2], and with upper level undergraduate and graduate students [1] at two universities in the United States. Several aspects of validity and reliability have been investigated for the data generated by the IILSI. The options that are included in the IILSI are based on open-ended question responses in interviews and surveys which provide a solid foundation for the instrument. Evidence in support of test content validity was established by multiple graduate students and faculty agreeing that it measured its intended purpose and that the options to be chosen represent the information that can be determined from Lewis structures. During development of the IILSI, two additional open-ended questions were added to determine if any answer choices were unclear to the students and to allow students to add any additional options they felt were absent from the list [1]. Additionally, the results of students not being able to connect structures with properties, which were found in the data generated by IILSI, were consistent with interview data from previous studies [1]. These provide evidence for response process validity. In looking at evidence to show relation to other variables, students who have not had instruction on the specific option at the time of data collection, such as resonance, do not choose the option and those that have had instruction do choose the option [1]. Along these lines, results from IILSI also show a relation between the options students choose and the type of curriculum used in the course [2]. As for reliability, the authors note that test-retest reliability, as normally defined, would not be the best way to establish reliability [1]. However, they did show consistency in initial IILSI scores for two first-semester general chemistry samples from different fall semesters exposed to the same instructor and instruction [1].  

    Recommendations for use

    The IILSI was developed to determine students’ perceptions of the information that can be determined from Lewis structures [1]. There was a large amount of validity evidence to support that the IILSI does meet its intended purpose. However, due to the inherent nature of the instrument, reliability evidence is lacking, although the instrument does seem to provide consistent results across semesters [1]. Currently, the IILSI data generated has only been used to look at each option for an entire group of students as opposed to looking at the number of options students correctly choose [1, 2]. Overall, it is recommended that additional evidence be investigated to ensure the IISLI functions in samples beyond that in which it was developed.

    Details from panel review

    The developers of the IILSI provided an extensive amount of validity evidence to support that the instrument is measuring students’ perceptions of the information that can be determined from Lewis structures [1]. Additionally, the authors followed best practices in developing the instrument based soundly on results from previous qualitative studies [1]. Evidence for test content validity was established through two separate means: 1) four graduate students and a faculty member agreed that it measured its intended purpose and 2) five graduate students and two faculty agreed the options to be chosen represent the information that can be determined from Lewis structures [1]. The authors established evidence for response process validity during the development of the IILSI by including additional questions at the end to ensure that there were no options that the students felt were unclear and to allow students to add additional information they felt could be determined from Lewis structures that were not already included [1]. Additional evidence for response process validity included that the data generated from the IILSI resulted in the same findings as the previous qualitative studies, which showed that students were not able to connect structures with properties [1]. The IILSI was also able to provide evidence of relation to other variables including exposure to instruction on the topic and the type of curriculum used to teach the topic [1, 2].

    While the authors provide many means of validity evidence for the IILSI, the evidence provided to establish reliability is limited. The authors note, and the panel agrees, that test-retest reliability, which is the typical means of establishing reliability with an instrument such as this, would not be appropriate due to the impact it might have on students’ future responses. Instead, the authors report the consistency of responses from administration of the IILSI to two similar groups pre-instruction at the start of general chemistry. In comparing the groups, the authors found no significant difference [1]. Additionally, while the majority of the evidence comes from students at a research university in the southeastern United States, the authors did compare the responses to students at the end of first-semester general chemistry at a comprehensive university in the southeastern United States and found no significant difference [1]. While these are not traditional means of establishing reliability evidence for an instrument, the panel does agree that the IILSI provides consistent results for similar samples. The question remains as to the reliability of the instrument outside of these very specific groups.

    References

    [1] Cooper, M. M., Underwood, S. M., & Hilley, C. Z. (2012). Development and validation of the implicit information from Lewis structures instrument (IILSI): Do students connect structures with properties? Chemistry Education Research and Practice, 13(3), 195–200. https://doi.org/10.1039/C2RP00010E

    [2] Underwood, S. M., Reyes-Gastelum, D., & Cooper, M. M. (2016). When do students recognize relationships between molecular structure and properties? A longitudinal comparison of the impact of traditional and transformed curricula. Chemistry Education Research and Practice, 17(2), 365–380. https://doi.org/10.1039/C5RP00217F

    Versions
    Listed below are all versions and modifications that were based on this instrument or this instrument were based on.
    Instrument has been modified in:
    Name Authors
    • Stowe, R.L., Herrington, D.G., McKay, R.L., & Cooper, M.M.

    • Kararo, A.T., Colvin, R.A., Cooper, M.M., & Underwood, S.M.

    • Underwood, S.M., Kararo, A.T., & Gadia, G.

    Citations
    Listed below are all literature that develop, implement, modify, or reference the instrument.
    1. Cooper, M.M., Underwood, S.M., & Hilley, C.Z. (2012). Development and validation of the implicit information from Lewis structures instrument (IILSI): Do students connect structures with properties?. Chemistry Education Research and Practice, 13(3), 195-2

    2. Underwood, S.M., Reyes-Gastelum, D., & Cooper, M.M. (2015). Answering the Questions of Whether and When Learning Occurs: Using Discrete-Time Survival Analysis to Investigate the Ways in Which College Chemistry Students' Ideas About Structure-Property Rela

    3. Burrows, N.L., & Mooring, S.R. (2015). Using concept mapping to uncover students' knowledge structures of chemical bonding concepts. Chemistry Education Research and Practice, 16(1), 53-66.

    4. Underwood, S.M., Reyes-Gastelum, D., & Cooper, M.M. (2016). When do students recognize relationships between molecular structure and properties? A longitudinal comparison of the impact of traditional and transformed curricula. Chemistry Education Research