ASCB logo LSE Logo

General Essays and ArticlesFree Access

Mind the Gap: Investigating Disaggregated Student Performance in an Upper-Division Biochemistry Course Following Team-Based Learning

    Published Online:https://doi.org/10.1187/cbe.24-02-0090

    Abstract

    Team-based learning (TBL) is a highly intense active learning pedagogy that uses a cycle of preclass preparation, formative assessment for readiness, mini-lectures, and complex team exercises. Consistent with the literature on active learning, prior research on TBL consistently shows its benefits for student outcomes as compared with previous lecture. However, little work has examined student outcomes disaggregated by demographic variable. We analyzed assessment and student demographic data (i.e., binary gender, racial/ethnic group, generational status) in three semesters of TBL in an upper-division biochemistry course to understand the degree to which performance could be predicted by student demographics. We use theoretical and empirical research from social psychology to hypothesize that the intense interpersonal interactions of TBL could activate psychological threats, the effects of which would be measurable as differences in student performance that correlate with demographic variables. Our regression analysis did not support this hypothesis. This null result invites a deeper discussion on how we measure the potential effects of active learning on student outcomes, particularly given how important it is to account for intersectional and invisible identities.

    INTRODUCTION

    A hallmark of active learning is student interaction (Driessen et al., 2020). In many active learning classrooms, students are invited to interact with nearby neighbors in transient pairs or small groups (e.g., think-pair-share, peer instruction) which are likely to differ in composition and intensity of interaction from day to day. Team-based learning (TBL) is one type of active learning pedagogy that is distinct in the respect that students are assigned to permanent groups that work together daily. The efficacy of TBL has been demonstrated in a variety of undergraduate contexts, including in majors and nonmajors courses, multiple Science, Technology, Engineering, and Math (STEM) disciplines, and in both introductory and upper-division courses (e.g., St. Clair and Chihara, 2012; Lamm et al., 2014; Hettler, 2015; Yuretich and Kanner, 2015; Behling et al., 2017; Donovan et al., 2018; Parappilly et al., 2019; Lekhi and Varao-Sousa, 2024). However, most studies have generally looked only at course-level aggregate effects. Few have examined outcomes at the student level, limiting our understanding of who benefits from TBL in these contexts.

    In TBL, student groups are intentionally formed by leveraging student strengths, backgrounds, and experiences to construct groups with diverse composition (Michaelsen and Sweet, 2008). Diverse group composition is rationalized from a resource perspective; the greater the range of individual perspectives, the greater the quantity of resources available to the group to solve meaningful problems (Watson et al., 1993; Sweet and Michaelsen, 2007). Yet assignment with an eye toward maximum diversity overlooks the potential implications of intense social interactions between students of diverse social identities on group dynamics and performance. More recent research on active learning suggests that students from historically marginalized racial and ethnic groups experience the group work of active learning classrooms as unwelcoming (Nardo et al., 2022).

    Reinholz and colleagues (2022) observed gendered performance differences between students of binary gender identities in active learning environments, which their evidence suggested was attributable to gendered differences in engagement. Other studies have noted differences in in-class and group work participation and subsequent performance correlating with historically marginalized identification (Eddy et al., 2014; Eddy and Hogan, 2014). Social psychologists postulate that these differences in performance maybe be explained, at least in part, to interactions with social identity. Several studies have demonstrated that the intersection between a student’s social identity and their social context can produce psychological barriers (e.g., self-doubt, fear of failure) which in some cases can reduce sense of belonging, motivation, engagement, that can affect subsequent academic performance (Walton and Cohen, 2011; Cohen and Sherman, 2014; Easterbrook and Hadden, 2021; Nardo et al., 2022). The implications of these prior studies prompted us to hypothesize that membership in a diverse TBL group might increase the likelihood that students with historically marginalized gender, racial/ethnic, or generational identities would perceive psychological barriers, the effects of which would be measurable as reduced academic performance. Therefore, we sought to answer the following research question:

    To what extent are differences in performance in a TBL course explained by students’ historically marginalized identities?

    WHAT IS TEAM-BASED LEARNING AND FOR WHOM IS IT EFFECTIVE?

    TBL is a high-intensity active learning approach that involves small teams of 5–7 students spending most of class time working collaboratively on complex disciplinary problems. Intense teamwork of this nature requires a certain level of team readiness which is achieved through preclass assignments intended to prepare students for in-class teamwork, individual and team readiness assurance tests (iRATs and tRATs) that diagnose student readiness for teamwork, and brief lectures (as needed) to fill in knowledge gaps (Michaelsen and Sweet, 2008). As the name implies, student-student interaction is the cornerstone of TBL course design. Permanent student teams work on two types of activities, the tRAT and a team activity. Both activities involve discussing and negotiating a consensus response to questions but differ in that the readiness assurance process (RAP) tends to focus on comprehension or simple application problems whereas the team activities require higher order cognitive skills.

    TBL has been shown to improve student performance, confidence, motivation, attitudes, teamwork skills, and engagement as compared with other instructional methods.(e.g., Parmelee et al., 2009; Persky and Pollack, 2011; Huitt et al., 2015; Nyindo et al., 2014; Yuretich and Kanner, 2015; Behling et al., 2017; Cho and Kweon, 2017; Jeno et al., 2017; Remington et al., 2017; Parappilly et al., 2019; Sakamoto et al., 2020). However, most of the studies are limited to business, nursing, medical, pharmacy, or other graduate/professional school contexts which are distinct from undergraduate STEM (Haidet et al., 2014). While some studies of TBL report the demographics of their sample (e.g. Parrish et al., 2021), most studies fail to examine disaggregated student data. The disaggregation of student data is an important practice in evaluating educational interventions, including active learning (Bensimon, 2005; Eddy and Hogan, 2014; Theobald et al., 2020). For active learning in general, higher-intensity practices reduce, but do not seem to fully eliminate, gaps in performance between students from advantaged and marginalized backgrounds (Theobald et al., 2020).

    SOCIAL IDENTITY THREAT AND STUDENT PERFORMANCE

    Social identity can be defined as the thoughts, feelings, and beliefs an individual has about themselves based on their membership in a group (Turner et al., 1979; Hogg and Turner, 1987). In educational contexts, social identity shapes how daily experiences in the classroom are perceived. For example, imagine a female-identifying student in a calculus course that also happens to have mostly male-identifying students enrolled. A poor score on a quiz might be experienced by this student as social identity threat, manifesting in thoughts such as, “Students like me don’t belong here” because of gendered societal messages about math ability and the lack of representation of other female-identifying students. Repeated threats to social identity are theorized to lead to a series of negative recursive events (e.g., reduced motivation, withdrawal) that over time may have negative educational outcomes (e.g., reduced performance, leaving the major; Cohen and Garcia, 2008; Murphy and Zirkel, 2015; Easterbrook and Hadden, 2021). While this example is oversimplified—gender is not binary, identities are intersectional, and negative experiences do not always lead to negative educational outcomes like reduced performance—it does reveal that the ways in which daily experiences are perceived are highly context specific.

    The designers of TBL prescribed forming diverse teams based on the principle of resource wealth distribution—the assertion that diverse team composition maximizes the resources available to the group for problem solving (Michaelsen and Sweet, 2008). This approach may also presume a greater likelihood of divergent viewpoints and therefore more careful consideration and deliberation during group tasks. But when viewed through a social identity lens, it is possible the opposite will happen. Some students, particularly those from socially marginalized groups, may experience threats to their social identity that over a semester of intense group work during TBL, produces long-term performance inequalities (Eddy et al., 2015; Hood et al., 2020; Easterbrook and Hadden, 2021).

    MATERIALS AND METHODS

    We sought to understand the degree to which performance in a TBL course could be explained statistically by students’ historically marginalized identity. To this end, we gathered a variety of student performance measures (i.e., exam scores, course grades) and tested whether performance could be explained by demographic variables.

    Study Context

    Data were gathered in the first semester of an upper-division biochemistry two-course series. Typical course enrollment ranges between 45 and 65 students, about two-thirds of whom are junior and senior undergraduates, and one-third of whom are graduate students. Graduate students are generally first-year graduate students studying disciplines outside biochemistry (e.g., animal science, chemical engineering) who are required to take the course to improve their biochemistry background knowledge.

    In the TBL classroom, students were assigned to groups of four to six. In Fall 2019, groups were composed of both undergraduate and graduate students. In the following two semesters, undergraduate and graduate student were separated into their own groups. Students’ prior academic performance in biochemistry prerequisites, major/graduate discipline, gender, and race/ethnicity were factors used to create diverse groups (Watson et al., 1993), with special attention to ensuring that students with historically marginalized identities were not the only individual of that identity in their group. In Fall 2019 and 2020, the course had the same two instructors. In Fall 2021, one of the instructors was replaced with a new instructor. Additionally, the students worked synchronously online during the Fall 2020 semester in their TBL groups because of the COVID-19 global pandemic (Arneson et al., 2022).

    The course structure (Figure 1) was informed by recommendations in the TBL literature (Michaelsen and Sweet, 2008). On the first day of the course, students were introduced to their permanent groups and provided with the rationale for forming permanent groups; students develop as teammates over the semester by completing in-class activities, participating in constructive peer review, and build working relationships to succeed in learning biochemistry (Walker et al., 2020). Students also engaged in a team-building activity focused on behaviors—positive and negative—that are common during TBL group activities. Students identified their own constructive and destructive group behaviors and shared them with their groups with the help of a handout (adapted from Brunt, 1993). As part of these introductory activities, they created paper placards with their name, pronouns, constructive and destructive behaviors they identified, and a personal interest or hobby. Students used these placards to introduce themselves to their TBL team and the instructors, who referred to the placards as a name-, pronoun-, and behavior-learning tool at the beginning of the semester. As a team and with the whole class, students created and shared plans to leverage the team’s constructive behaviors and minimize destructive behaviors.

    FIGURE 1.

    FIGURE 1. The format of Team-Based Learning in the Biochemistry course.

    After the first day, the class consisted of 50-minute periods that met 3 days per week (M, W, F); team activities took place each day. iRATs and tRATs were only completed on Mondays and Fridays. Before each day of class, students reviewed instructor-provided materials to prepare for the RAP and team exercises. These materials typically included a short, assigned reading (Nelson and Cox, 2017), slides, or videos. Before class, students started the RAP by completing the iRAT. The iRAT was administered online via the course learning management site (LMS) and was comprised of questions that predominantly targeted knowledge, comprehension, and low-level application cognitive skills (Anderson and Krathwohl, 2001). When they arrived in class, students completed a tRAT in groups on the LMS. In 2019 and 2020, the tRAT was identical to the iRAT. In 2021, a minimum of 50% of tRAT questions were isomorphic to the iRAT questions. After the tRAT, the instructors reviewed student responses to the tRAT to determine any residual gaps in baseline knowledge and, when necessary, provided a brief lecture (∼15 min) to prepare the students for the team activities. TBL groups completed team exercises in class which commonly consisted of more complex activities such as interpreting data, developing hypotheses or explanations for an example phenomenon, or completing a case study through the LMS. As an addition to the conventional TBL format, once per week, students took an individual quiz. All TBL activities contributed to a students’ grade in the course as prescribed to invite both individual and team accountability throughout the course (Michaelsen and Sweet, 2008). The course grading scheme can be found in the Supplemental Materials.

    Twice during the course, students completed anonymous peer evaluations for their TBL group members. The peer evaluation included two questions asked about each teammate, “Provide a numerical rating for Student X on a scale from 1 to 10 (with 10 being the most favorable). In your score, consider attendance, preclass preparation and participation in team activities.” as well as “Provide constructive feedback for Student X as a teammate. Consider strengths and provide suggestions for areas of improvement.” The instructors checked peer evaluations to monitor group dynamics and ensure peer feedback was constructive and returned feedback to TBL groups anonymously. Completing the peer evaluations with effortful and constructive feedback composed 5% of a student’s final grade.

    Data Sources

    Three quantitative measures of student outcomes in the course were analyzed. We analyzed final course grades, learning gains on the Biochemistry Concept Inventory (BCI) assessment (Villafañe et al., 2011), and students’ performance on unit exams. Final grades were selected because potential performance gaps could relate to subsequent course choice and success. The BCI was used as an external measure of student knowledge of biochemistry content. Unit exam scores were selected because of their potential to reveal performance gaps at discrete timepoints throughout the semester that might otherwise not be revealed when looking at aggregate exam performance. Performance on unit exams represents students’ experience at progressive timepoints during the semester. We examined unit exam performance to explore two possible predictions. One prediction was that no gaps or small gaps would exist in Exam 1 scores, but differences in students’ exam scores connected to demographic identities would increase over time as social interactions in TBL groups led to threats and exacerbated psychological barriers to performance. Another, coming from TBL literature that posits that groups improve as they learn to coordinate over time, would be that gaps would exist early in the semester, but would decrease over the semester due to groups creating positive norms and developing ways of working together (Watson et al., 1991; Letassy et al., 2008; Akl et al., 2012; Persky, 2012).

    The course included two midterm exams and one comprehensive final exam. Exams consisted of upper-level Blooms questions and were mostly multiple-choice questions with some short-answer questions. In 2020, exams were administered online. That semester, time for the final exam was extended by one hour in accordance with university policies as an accommodation for COVID-19 conditions. The three exams together comprised 40% of a student’s grade in the course.

    On the first and last days of class, the BCI was administered as a pretest and posttest. The students received points for completing the BCI with an effortful attempt, but their work was otherwise ungraded. Normalized change scores were calculated for each student using their BCI pretest and posttest scores and range from −1 to 1 (Marx and Cummings, 2007). The BCI normalized change score was used for analysis as it accounts for students who perform lower on the posttest than on the pretest, limiting negative outcomes from skewing further statistical analysis. Student exam scores and course grades had scores between 0 and 100.

    Students’ self-identified demographic data were collected from requests to the registrar, and in some cases, a survey (IRB #19076). After these collections, the demographic information for some students was still not available. Those students’ data were excluded from the final models. The demographic makeups of the three semesters of the course, as well as the University, are listed in the categories used for modeling below (University: 2019 Fall).

    The base, car, effsize, lme4, sjplot, and sjstats packages in R were used to test, model, and visualize the data.

    Modeling Student Performance

    We used the three performance measures to generate linear mixed models which included demographic, assessment, and grouping factors, as well as interactions between factors, as explanatory variables. We chose to use linear mixed models as they allowed us to examine student performance as explained by one factor or interaction between factors while controlling for the other factors of interest. Our analysis was informed by Theobald (2018). Model outcomes were generated using lme4, sjplot, and sjstats in R. Models incorporated the semester each student took the course (three total semesters), three factors describing student demographic information (gender, historically marginalized [HM] status, and generational status), and iRAT performance, as well as relevant interactions between gender and the other two demographic factors. Gender is not a binary construct; however, the demographic information was reported in a binary M-F format. Although transgender and nonbinary students were present in the course, and shared their identities with the instructors, they are not identified or separated in the data because of privacy concerns. HM or historically marginalized status is defined as a group of backgrounds including Black, Latino, Native Hawaiian, Pacific Islander, and Native American/Alaska Native racial/ethnic backgrounds. Students from these backgrounds disproportionately are affected by historical and current marginalization, resulting in underrepresentation in STEM. First-generation students are defined as those for whom none of their parents or guardians earned a bachelor’s degree. This definition of first-generation matches that used by the University where the Study took place.

    We selected three demographic factors based on prior investigations of students from marginalized gender, racial/ethnic, and generational backgrounds in STEM courses with active learning practices (Hood et al., 2020; Theobald et al., 2020; Odom et al., 2021). Scholarly work explaining the interaction between students’ social identities and performance also contributed to our choice of factors, as all three have been correlated with threats to social identity in academic contexts (Easterbrook and Hadden, 2021). The interactions between the factors were selected due to their known joint effects on undergraduate experiences (Charleston et al., 2014; Laajala-Lozano and Jenkins, 2022). In initial models, we included interactions between gender and our other demographic factors of interest to account for intersectionality. Throughout the modeling process, these interaction factors were unable to explain student performance, were non-significant, and because of this were removed from the final models in the Exam 2, BCI normalized change, and course grades models.

    Preparation has been a predictor of student performance in some active learning classrooms, so iRAT performance was selected as a factor as a proxy for students’ preparation before class (Gross et al., 2015). iRAT performance was also more strongly correlated with student exam scores than other potential factors (e.g., student BCI prescores, tRAT scores, team exercise scores; Supplemental Materials), so it was a better control for student work during the unit than those other factors. iRAT scores were removed from the final course grade calculation to prevent unintentional correlations. The factors we selected to represent students’ demographic backgrounds might logically be related as components of identity, so variance inflation factors were calculated between the selected independent variables to assess the degree to which they were correlated. This method suggested that the factors were not at risk of producing error (Supplemental Materials).

    A backward selection method was used to build and test models (Theobald, 2018). First, the random factor (Semester) was tested by itself. Semester is a random factor in many models of student performance as students are nested within semesters, meaning they are not truly independent of each other (Theobald, 2018). In all models except for the comprehensive (third) exam model, the intraclass correlation coefficient (ICC) of Semester was 0 or near 0, meaning that Semester was not a significant factor in explaining student scores. In those models, modeling proceeded without the random factor, using only the fixed factors. In the Exam 3 model, the ICC of the Semester factor was far from 0, and modeling proceeded including Semester as a random factor. Then, models were tested with all fixed factors, and if factors were found not to be significant, new models were generated one by one with the factor furthest from significance removed, and the Akaike information criteria (AIC) of the new models compared to determine which model had the best fit to the data. If the AICs of the two models were significantly different (as denoted by a difference of 2 or more in the AICs of the two models), the model with the lower AIC (and thus the more efficient model) was used in further tests. This process continued with each non-significant factor until a model with only factors that were significant or significantly contributed to the AIC remained. Models were interpreted in order to address the research question concerning differences in performance in the classroom (Table 1).

    TABLE 1. Student demographics in the university and the TBL course

    Demographics2019 – University2019 – Course2020 – University2020 – Course2021 – University2021 – Course
    Gender (Women)53.2%53.1% (17/32)54.4%50% (26/52)54.3%62.5% (30/48)
    HM status30.5%18.8% (6/32)31.5%30.8% (16/52)32.1%20.8% (10/48)
    First-generation32.9%25% (8/32)32.5%25% (13/52)31.9%16.7% (8/48)
    First-generation*WomenNot provided18.8% (6/32)Not provided9.6% (5/52)Not provided10.4% (5/48)
    Women*HMNot provided9.4% (3/32)Not provided11.5% (6/52)Not provided10.4% (5/48)

    RESULTS

    Exam Performance

    The mixed models we generated to explain performance on Exams 1–3 show that the selected demographic characteristics did not explain any differences in performance (Tables 24). The only significant factor that explained exam performance in the initial models (Tables 24) was preparation (i.e., student average iRAT performance). The R2 values of the final exam models indicate that they had weak to moderate performance for explaining scores (Hair et al., 2011). Model outcomes were similar across exams, with iRAT performance explaining exam performance in each unit, and demographic factors unable to explain exam performance in any unit. There was no evidence of any cumulative effects or effects over time that affected student performance during the period of the course.

    TABLE 2. Final regression model explaining student outcomes on Exam 1

    FactorBSE
    Intercept36.786.06
    iRAT average0.570.07
    Gender−1.923.43
    HM status−0.843.09
    Generation1.113.43
    Gender*HM status−1.564.52
    Gender*Generation−0.744.61

    N = 134; Adjusted R2 = 0.35.

    p < 0.001.

    TABLE 3. Final regression model explaining student outcomes on Exam 2

    Fixed factorsBSE
    Intercept48.326.28
    iRAT average0.490.08
    Gender−3.232.21
    HM status−2.722.57
    Generation−3.412.59

    N = 134, Adjusted R2 = 0.2506.

    p < 0.001.

    TABLE 4: Final regression model explaining student outcomes on Exam 3

    Random factorVarianceStd. Dev
    Semester12.763.57
    Fixed factorsBSE
    Intercept49.075.86
    iRAT average0.380.07
    Gender−2.432.22
    HM status−0.632.95
    Generation1.613.01
    Gender*HM status0.514.31
    Gender*Generation−6.454.15

    N = 134; Marginal R2 = 0.23, Conditional R2 = 0.32.

    p < 0.001.

    Normalized Change and Course Grades

    The models explaining normalized change on the BCI and course grades performed poorly compared with those explaining exam scores (Supplemental Materials). Students’ average iRAT scores across the course significantly explained a small amount of the BCI normalized change outcome (β = 0.006, Supplemental Materials). None of the selected factors significantly explained course grades. In further exploratory modeling, students’ BCI prescores were not able to explain student grades. The adjusted R2 values for these models were very low; indicating that the selected factors poorly explained the BCI normalized change scores and course grades.

    DISCUSSION

    TBL assigns students to permanent groups with maximum diversity to work intensely throughout the duration of a course (Watson et al., 1991; Letassy et al., 2008; Michaelsen and Sweet, 2008; Akl et al., 2012; Persky, 2012). We had hypothesized that the intense social interactions characteristic of TBL could produce threats to socially marginalized identities, the effects of which might be observed by differences in performance on internal, instructor-written exams, course grades, and/or external concept inventories. Our measures of marginalized student identity (gender, racial/ethnic group, generational status) could not statistically explain any difference in performance on any of these outcome variables. These results are encouraging because they suggest that, at a minimum, TBL did not introduce performance gaps in this upper-division course. This is an interesting result when juxtaposed to larger, meta-analyses of active learning which often show a reduction—but not elimination—of performance gaps.

    Given the dataset collected, it is difficult to confidently explain this null result. Two potential explanations include survivorship bias and an inability to adequately account for invisible and intersectional identities. The context of this study was a single upper-division biochemistry course typically taken by students who are close to graduation. It is possible that our data suffer from a “survivorship bias” in the population of biochemistry majors. The concept of “survivorship bias” reflects the inability for datasets to include information from subjects who have already left the sample population by the time data can be collected (Brown et al., 1992). Students with socially marginalized identities may have already experienced disproportionate course failures, withdrawals, major changes, and exit from the University before advancing to this upper-division course. Indeed, the biochemistry course had fewer students from HM and first-generation backgrounds than the university as a whole (Demographics Table). Due to their early success in lower-division courses, the remaining students who reached their senior year may have developed greater levels of self-efficacy, providing a protective effect against identity threats. Because our study focused on performance and not other constructs thought to be influenced by the use of TBL like self-efficacy, engagement, or motivation, we were not able to investigate the interaction between the backgrounds we focused on, the TBL context, and these constructs.

    Intersectionality is a crucial concept in understanding equity in STEM teaching and learning because it acknowledges how individuals can hold multiple identities at once, and for individuals with multiple stigmatized identities, different forms of discrimination can combine to create major obstacles to progression in STEM, especially for women of color (Ong et al., 2011). We were limited in our examination of intersectionality due to the course makeup and sample size. Just 16 students were first-generation women students, and just 14 students were women from underrepresented racial/ethnic backgrounds. While our sample was enough to conduct an analysis, it limited the statistical power of our attempts to model interactions among these characteristics or between other combinations of characteristics. Further, our predictor variables were limited to those routinely collected by the institution. We did not have measures for other identities that have been marginalized in educational contexts (e.g., disability, LGBTQ identity) (Cooper and Brownell, 2016; Cooper et al., 2018; Easterbrook and Hadden, 2021; Marques, 2022; Pfeifer et al., 2023). Finally, we were unable to account for invisible identities for which there are no convenient measures (e.g., neurodiversity, religiosity). It is possible that performance could have been explained by a more comprehensive set of variables that better account for a broader diversity of identities and their intersections. Exploring these other identities may reveal differences in performance that could not be detected in our study, and thus lead to additional opportunities to understand how positive group dynamics can be promoted in TBL courses that have diverse groups of students along those identities as well.

    CONCLUSIONS AND FUTURE DIRECTIONS

    This study is the first that has documented comparable performance outcomes among students from diverse identity backgrounds in a TBL context in the undergraduate biosciences. The extent of differences in performance between students with marginalized identities and their advantaged peers during TBL has been underexamined in the literature, and our work investigated just one of many contexts where TBL takes place. To more fully understand the degree that TBL impacts student performance requires studies that look at disaggregated performance effects of TBL in different undergraduate contexts. Additionally, future work should investigate other outcomes (e.g., motivation, engagement, self-efficacy) that also effect student success.

    TBL is just one form of active learning. An extensive body of research demonstrates that active learning reduces, but does not completely close, gaps in performance between students of marginalized and advantaged backgrounds (Theobald et al., 2020). In our analysis of this one TBL context, we were not able to detect any gaps in performance. The multiple caveats of our research design, as well as the many facets of students’ identities we were not able to explore, preclude us from claiming such gaps do not exist in TBL contexts, but they encourage us to call for more examination of TBL as a practice. Further work with stronger controls, deeper disaggregation, and more robust replication is necessary to truly determine the extent to which TBL may reduce or eliminate performance gaps among students from different backgrounds. More evidence indicating TBL reduces or even eliminates gaps in performance would be a major advancement for STEM educators desiring equitable practices for their increasingly diverse courses.

    ACKNOWLEDGMENTS

    The authors would like to acknowledge Dr. Jessie Arneson and the WSU Center for Interdisciplinary Statistical Education and Research for their assistance with the conception and methods of analysis for this research. Funding for this project was provided by a National Science Foundation Graduate Research Fellowship (Award ID: 1842493)

    REFERENCES

  • Akl, I. B., Ghaddar, F., Sabra, R., Parmelee, D., Simaan, J. A., Kanafani, Z. A., & Zgheib, N. K. (2012). Teaching clinical pharmacology using team-based learning: A comparison between third- and fourth-year medical students. The Journal of Clinical Pharmacology, 52, 1806–1814. https://doi.org/10.1177/0091270011428986 MedlineGoogle Scholar
  • Anderson, L. W., & Krathwohl, D. R. (2001). A Taxonomy for Learning, Teaching, and Assessing: A Revision of Bloom’s Taxonomy of Educational Objectives. New York: Longman. Google Scholar
  • Arneson, J. B., Woodbury, J., & Offerdahl, E. G. (2022). Fidelity of implementation as a guiding framework for transitioning research-based instructional practices from on site to online. Journal of Microbiology & Biology Education, 23, e00337–21. https://doi.org/10.1128/jmbe.00337-21 MedlineGoogle Scholar
  • Behling, K. C., Kim, R., Gentile, M., & Lopez, O. (2017). Does team-based learning improve performance in an infectious diseases course in a preclinical curriculum? International Journal of Medical Education, 8, 39–44. https://doi.org/10.5116/ijme.5895.0eea MedlineGoogle Scholar
  • Bensimon, E. M. (2005). Closing the achievement gap in higher education: An organizational learning perspective. New Directions for Higher Education, 2005(131), 99–111. Google Scholar
  • Brown, S. J., Goetzmann, W., Ibbotson, R. G., & Ross, S. A. (1992). Survivorship bias in performance studies. The Review of Financial Studies, 5, 553–580. https://doi.org/10.1093/rfs/5.4.553 Google Scholar
  • Brunt, J. (1993). Facilitation skills for quality improvement. Madison, WI: Quality Enhancement Strategies, University of Wisconsin-Madison. Google Scholar
  • Charleston, L. J., Adserias, R. P., Lang, N. M., & Jackson, J. F. L. (2014). Informing higher education policy and practice through intersectionality. Journal of Progressive Policy and Practice, 2(3), 273–293. Google Scholar
  • Cho, Y. H., & Kweon, Y.-R. (2017). Effects of team-based learning on communication competence for undergraduate nursing students. Journal of Korean Academy of Psychiatric and Mental Health Nursing, 26(1), 101–110. https://doi.org/10.12934/jkpmhn.2017.26.1.101 Google Scholar
  • Cohen, G. L., & Garcia, J. (2008). Identity, belonging, and achievement: A model, interventions, implications. Current Directions in Psychological Science, 17(6), 365–369. https://doi.org/10.1111/j.1467-8721.2008.00607.x Google Scholar
  • Cohen, G. L., & Sherman, D. K. (2014). The psychology of change: Self-affirmation and social psychological intervention. Annual Review of Psychology, 65(1), 333–371. https://doi.org/10.1146/annurev-psych-010213-115137 MedlineGoogle Scholar
  • Cooper, K. M., & Brownell, S. E. (2016). Coming out in class: Challenges and benefits of active learning in a biology classroom for LGBTQIA students. CBE–Life Sciences Education, 15(3), ar37. https://doi.org/10.1187/cbe.16-01-0074 LinkGoogle Scholar
  • Cooper, K. M., Downing, V. R., & Brownell, S. E. (2018). The influence of active learning practices on student anxiety in large-enrollment college science classrooms. International Journal of STEM Education, 5(1), 23. https://doi.org/10.1186/s40594-018-0123-6 MedlineGoogle Scholar
  • Donovan, D. A., Connell, G. L., & Grunspan, D. Z. (2018). Student learning outcomes and attitudes using three methods of group formation in a nonmajors biology class. CBE–Life Sciences Education, 17(4), ar60. https://doi.org/10.1187/cbe.17-12-0283 LinkGoogle Scholar
  • Downing, V. R., Cooper, K. M., Cala, J. M., Gin, L. E., & Brownell, S. E. (2020). Fear of negative evaluation and student anxiety in community college active-learning science courses. CBE–Life Sciences Education, 19(2), ar20. https://doi.org/10.1187/cbe.19-09-0186 LinkGoogle Scholar
  • Driessen, E. P., Knight, J. K., Smith, M. K., & Ballen, C. J. (2020). Demystifying the meaning of active learning in postsecondary biology education. CBE–Life Sciences Education, 19(4), ar52. https://doi.org/10.1187/cbe.20-04-0068 LinkGoogle Scholar
  • Easterbrook, M. J., & Hadden, I. R. (2021). Tackling educational inequalities with social psychology: Identities, contexts, and interventions. Social Issues and Policy Review, 15, 180–236. https://doi.org/10.1111/sipr.12070 Google Scholar
  • Eddy, S. L., Brownell, S. E., Thummaphan, P., Lan, M.-C., & Wenderoth, M. P. (2015). Caution, student experience may vary: Social identities impact a student’s experience in peer discussions. CBE–Life Sciences Education, 14(4), ar45. https://doi.org/10.1187/cbe.15-05-0108 LinkGoogle Scholar
  • Eddy, S. L., Brownell, S. E., & Wenderoth, M. P. (2014). Gender gaps in achievement and participation in multiple introductory biology classrooms. CBE–Life Sciences Education, 13, 478–492. https://doi.org/10.1187/cbe.13-10-0204 LinkGoogle Scholar
  • Eddy, S. L., & Hogan, K. A. (2014). Getting under the hood: How and for whom does increasing course structure work? CBE–Life Sciences Education, 13, 453–468. https://doi.org/10.1187/cbe.14-03-0050 LinkGoogle Scholar
  • Espey, M. (2018). Diversity, effort, and cooperation in team-based learning. Journal of Economic Education, 49, 8–21. https://doi.org/10.1080/00220485.2017.1397571 Google Scholar
  • Gross, D., Pietri, E. S., Anderson, G., Moyano-Camihort, K., & Graham, M. J. (2015). Increased preclass preparation underlies student outcome improvement in the flipped classroom. CBE–Life Sciences Education, 14(4), ar36. https://doi.org/10.1187/cbe.15-02-0040 LinkGoogle Scholar
  • Haidet, P., Kubitz, K., & McCormack, W. T. (2014). Analysis of the team-based learning literature: TBL comes of age. Journal on Excellence in College Teaching, 25(3–4), 303. MedlineGoogle Scholar
  • Hair, J. F., Ringle, C. M., & Sarstedt, M. (2011). Pls-Sem: Indeed a silver bullet. Journal of Marketing Theory and Practice, 19, 139–151. https://doi.org/10.2753/MTP1069-6679190202 Google Scholar
  • Hettler, P. (2015). Student demographics and the impact of team-based learning. International Advances in Economic Research, 21, 413–422. https://doi.org/10.1007/s11294-015-9539-7 Google Scholar
  • Hogg, M. A., & Turner, J. C. (1987). Intergroup behaviour, self-stereotyping and the salience of social categories. British Journal of Social Psychology, 26(4), 325–340. https://doi.org/10.1111/j.2044-8309.1987.tb00795.x Google Scholar
  • Hood, S., Barrickman, N., Djerdjian, N., Farr, M., Gerrits, R. J., Lawford, H., ... & Hull, K. (2020). Some believe, not all achieve: The role of active learning practices in anxiety and academic self-efficacy in first-generation college students. Journal of Microbiology & Biology Education, 21(1). https://doi.org/10.1128/jmbe.v21i1.2075 MedlineGoogle Scholar
  • Huitt, T. W., Killins, A., & Brooks, W. S. (2015). Team-based learning in the gross anatomy laboratory improves academic performance and students’ attitudes toward teamwork. Anatomical Sciences Education, 8(2), 95–103. MedlineGoogle Scholar
  • Jeno, L. M., Raaheim, A., Kristensen, S. M., Kristensen, K. D., Hole, T. N., Haugland, M. J., & Mæland, S. (2017). The relative effect of team-based learning on motivation and learning: A self-determination theory perspective. CBE—Life Sciences Education, 16(4), ar59. https://doi.org/10.1177/0091270011428986 LinkGoogle Scholar
  • Laajala-Lozano, A., & Jenkins, S. R. (2022). Undergraduates’ academic acculturative stress at the intersection of first-generation status and gender in a public university setting. Journal of College Student Retention: Research, Theory & Practice, 1–18. https://doi.org/10.1177/15210251221120396 Google Scholar
  • Lamm, M. H., Dorneich, M., & Rover, D. T. (2014). Team-based learning in engineering classrooms: Feedback form and content adds value to the learning experience. 2014 ASEE North Midwest Section Conference. 2014(1), 1–22. https://doi.org/10.17077/aseenmw2014.1026 Google Scholar
  • Lekhi, P., & Varao-Sousa, T. (2024). Implementing team-based learning in a large environmental chemistry course and its impact on student learning and perceptions. Chemistry Education Research and Practice, 25, 193–211 https://doi.org/10.1039/D3RP00158J Google Scholar
  • Letassy, N. A., Fugate, S. E., Medina, M. S., Stroup, J. S., & Britton, M. L. (2008). Using team-based learning in an endocrine module taught across two campuses. American Journal of Pharmaceutical Education, 72(5), 103. https://doi.org/10.5688/aj7205103 MedlineGoogle Scholar
  • Marques, A. A. G. (2022). Team-Based Learning Impact on College Students’ Anxiety. Universidade Nova de Lisboa, Information Management School. Google Scholar
  • Marx, J. D., & Cummings, K. (2007). Normalized change. American Journal of Physics, 75(1), 87–91. https://doi.org/10.1119/1.2372468 Google Scholar
  • Michaelsen, L. K., & Sweet, M. (2008). The essential elements of team-based learning. In Michaelsen L. K.Sweet M.Parmelee X. (Eds.), Team-Based Learning: Small Group Learning’s Next Big Step. New Directions for Teaching and Learning (pp. 7–27). San Francisco: Wiley. https://doi.org/10.1002/tl.330 Google Scholar
  • Murphy, M. C., & Zirkel, S. (2015). Race and belonging in school: How anticipated and experienced belonging affect choice, persistence, and performance. Teachers College Record, 117(12), 1–40. https://doi.org/10.1177/016146811511701204 Google Scholar
  • Nardo, J. E., Chapman, N. C., Shi, E. Y., Wieman, C., & Salehi, S. (2022). Perspectives on active learning: Challenges for equitable active learning implementation. Journal of Chemical Education, 99(4), 1691–1699. https://doi.org/10.1021/acs.jchemed.1c01233 Google Scholar
  • Nelson, D. L., & Cox, M. M. (2017). Lehninger Principles of Biochemistry, 7th ed. New York, NY: W.H. Freeman. Google Scholar
  • Nyindo, M., Kitau, J., Lisasi, E., Kapanda, G., Matowo, J., Francis, P., & Bartlett, J. (2014). Introduction of team-based learning (TBL) at Kilimanjaro Christian Medical University College: Experience with the ectoparasites module. Medical Teacher, 36, 308–313. https://doi.org/10.3109/0142159X.2013.876490 MedlineGoogle Scholar
  • Odom, S., Boso, H., Bowling, S., Brownell, S. E., Cotner, S., Creech, C., … & Ballen, C. J. (2021). Meta-analysis of gender performance gaps in undergraduate natural science courses. CBE-Life Sciences Education, 20(3), ar40. https://doi.org/10.1187/cbe.20-11-0260 LinkGoogle Scholar
  • Ong, M., Wright, C., Espinosa, L., & Orfield, G. (2011). Inside the double bind: A synthesis of empirical research on undergraduate and graduate women of color in Science, Technology, Engineering, and Mathematics. Harvard Educational Review, 81, 172–209. Google Scholar
  • Parappilly, M., Woodman, R. J., & Randhawa, S. (2019). Feasibility and effectiveness of different models of team-based learning approaches in STEMM-based disciplines. Research in Science Education, 51, 391–405. https://doi.org/10.1007/s11165-019-09888-8 Google Scholar
  • Parmelee, D. X., DeStephen, D., & Borges, N. J. (2009). Medical students’ attitudes about team-based learning in a pre-clinical curriculum. Medical Education Online, 14(1), 4503. Google Scholar
  • Parrish, C. W., Guffey, S. K., Williams, D. S., Estis, J. M., & Lewis, D. (2021). Fostering cognitive presence, social presence and teaching presence with integrated online—Team-based learning. TechTrends, 65, 473–484. MedlineGoogle Scholar
  • Persky, A. M. (2012). The impact of team-based learning on a foundational pharmacokinetics course. American Journal of Pharmaceutical Education, 76(2), 31. https://doi.org/10.5688/ajpe76231 MedlineGoogle Scholar
  • Persky, A. M., & Pollack, G. M. (2011). A modified team-based learning physiology course. American Journal of Pharmaceutical Education, 75(10), 204. MedlineGoogle Scholar
  • Pfeifer, M. A., Cordero, J. J., & Stanton, J. D. (2023). What I wish my instructor knew: How active learning influences the classroom experiences and self-advocacy of STEM majors with ADHD and specific learning disabilities. CBE-Life Sciences Education, 22(1), ar2. https://doi.org/10.17763/haer.81.2.t022245n7x4752v2 MedlineGoogle Scholar
  • Reinholz, D., Johnson, E., Andrews-Larson, C., Stone-Johnstone, A., Smith, J., Mullins, B., … & Shah, N. (2022). When active learning is inequitable: Women’s participation predicts gender inequities in mathematical performance. Journal for Research in Mathematics Education, 53, 204–226. https://doi.org/10.5951/jresematheduc-2020-0143 Google Scholar
  • Remington, T. L., Bleske, B. E., Bartholomew, T., Dorsch, M. P., Guthrie, S. K., Klein, K. C., ... & Wells, T. D. (2017). Qualitative analysis of student perceptions comparing team-based learning and traditional lecture in a pharmacotherapeutics course. American Journal of Pharmaceutical Education, 81(3), 55. MedlineGoogle Scholar
  • Sakamoto, S. R., Dell’Acqua, M. C. Q., Abbade, L. P. F., Caldeira, S. M., Fusco, S. F. B., & Avila, M. A. G. (2020). Team-based learning: A randomized clinical trial in undergraduate nursing. Revista Brasileira de Enfermagem, 73, e20180621. https://doi.org/10.1590/0034-7167-2018-0621 MedlineGoogle Scholar
  • St. Clair, K., & Chihara, L. (2012). Team-based learning in a statistical literacy class. Journal of Statistics Education, 20(1), 1–20. https://doi.org/10.1080/10691898.2012.11889633 Google Scholar
  • Sweet, M., & Michaelsen, L. K. (2007). How group dynamics research can inform the theory and practice of postsecondary small group learning. Educational Psychology Review, 19, 31–47. https://doi.org/10.1007/s10648-006-9035-y Google Scholar
  • Theobald, E. (2018). Students are rarely independent: When, why, and how to use random effects in discipline-based education research. CBE–Life Sciences Education, 17(3):rm2. https://doi.org/10.1187/cbe.17-12-0280 LinkGoogle Scholar
  • Theobald, E. J., Hill, M. J., Tran, E., Agrawal, S., Arroyo, E. N., Behling, S., … & Freeman, S. (2020). Active learning narrows achievement gaps for underrepresented students in undergraduate science, technology, engineering, and math. Proceedings of the National Academies of Sciences, 117, 6476–6483. https://doi.org/10.1073/pnas.1916903117 MedlineGoogle Scholar
  • Turner, J. C., Brown, R. J., & Tajfel, H. (1979). Social comparison and group interest in ingroup favouritism. European Journal of Social Psychology, 9, 187–204. https://doi.org/10.1002/ejsp.2420090207 Google Scholar
  • Villafañe, S. M., Bailey, C. P., Loertscher, J., Minderhout, V., & Lewis, J. E. (2011). Development and analysis of an instrument to assess student understanding of foundational concepts before biochemistry coursework. Biochemistry and Molecular Biology Education, 39, 102–109. https://doi.org/10.1002/bmb.20464 MedlineGoogle Scholar
  • Walker, E. R., Lang, D. L., Caruso, B. A., & Salas-Hernández, L. (2020). Role of team dynamics in the learning process: A mixed-methods evaluation of a modified team-based learning approach in a behavioral research methods course. Advances in Health Science Education, 25, 383–399. https://doi.org/10.1007/s10459-019-09931-3 MedlineGoogle Scholar
  • Walton, G. M., & Cohen, G. L. (2011). A brief social-belonging intervention improves academic and health outcomes of minority students. Science, 331(6023), 1447–1451. https://doi.org/10.1126/science.1198364 MedlineGoogle Scholar
  • Watson, W., Michaelsen, L. K., & Sharp, W. (1991). Member competence, group interaction, and group decision making: A longitudinal study. Journal of Applied Psychology, 76(6), 803–809. Google Scholar
  • Watson, W. E., Kumar, K., & Michaelsen, L. K. (1993). Cultural diversity’s impact on interaction process and performance: Comparing homogeneous and diverse task groups. Academy of Management Journal, 36, 590–602. https://doi.org/10.5465/256593 Google Scholar
  • Yuretich, R. F., & Kanner, L. C. (2015). Examining the effectiveness of team-based learning (TBL) in different classroom settings. Journal of Geoscience Education, 63, 147–156. https://doi.org/10.5408/13-109.1 Google Scholar