Education Sciences (Nov 2024)
Generative AI for Culturally Responsive Science Assessment: A Conceptual Framework
Abstract
In diverse classrooms, one of the challenges educators face is creating assessments that reflect the different cultural backgrounds of every student. This study presents a novel approach to the automatic generation of cultural and context-specific science assessments items for K-12 education using generative AI (GenAI). We first developed a GenAI Culturally Responsive Science Assessment (GenAI-CRSciA) framework that connects CRSciA, specifically key cultural tenets such as indigenous language, Indigenous knowledge, ethnicity/race, and religion, with the capabilities of GenAI. Using the CRSciA framework, along with interactive guided dynamic prompt strategies, we developed the CRSciA-Generator tool within the OpenAI platform. The CRSciA-Generator allows users to automatically generate assessment items that are customized to align with their students’ cultural and contextual needs. We further conducted a pilot demonstration of item generation between the CRSciA-Generator and the base GPT-4o using standard prompts. Both tools were tasked with generating CRSciAs that aligned with the Next Generation Science Standard on predator and prey relationship for use with students from Ghana, the USA, and China. The results showed that the CRSciA-Generator output assessment items incorporated more tailored cultural and context assessment items for each specific group with examples, such as traditional stories of lions and antelopes in Ghana, Native American views on wolves in the USA, and Taoist or Buddhist teachings on the Amur tiger in China compared to the standard prompt assessment items within the base GPT-4o. However, due to the focus on nationality in the pilot demonstration, the CRSciA-Generator assessment items treated the countries as culturally homogeneous, overlooking subcultural diversity in these countries. Therefore, we recommend that educators provide detailed background information about their students when using the CRSciA-Generator. We further recommend future studies involving expert reviews to assess the cultural and contextual validity of the assessment items generated by the CRSciA-Generator.
Keywords