Building Better Bridges into STEM: A Synthesis of 25 Years of Literature on STEM Summer Bridge Programs
Abstract
Summer bridge programs are designed to help transition students into the college learning environment. Increasingly, bridge programs are being developed in science, technology, engineering, and mathematics (STEM) disciplines because of the rigorous content and lower student persistence in college STEM compared with other disciplines. However, to our knowledge, a comprehensive review of STEM summer bridge programs does not exist. To provide a resource for bridge program developers, we conducted a systematic review of the literature on STEM summer bridge programs. We identified 46 published reports on 30 unique STEM bridge programs that have been published over the past 25 years. In this review, we report the goals of each bridge program and whether the program was successful in meeting these goals. We identify 14 distinct bridge program goals that can be organized into three categories: academic success goals, psychosocial goals, and department-level goals. Building on the findings of published bridge reports, we present a set of recommendations for STEM bridge programs in hopes of developing better bridges into college.
INTRODUCTION
Summer bridge programs are typically developed to facilitate students’ transition into college and improve student academic success. These programs are often multiweek intensive experiences that occur in the weeks before a student’s first year of college or a transfer student’s entry into a 4-year institution. Such programs are commonly referred to as summer bridge programs, boot camps, summer programs, or college prep programs. Over the past decade, there have been many national calls to increase the retention of students in science, technology, engineering, and mathematics (STEM) and to enhance the diversity of STEM professionals (National Academy of Sciences, National Academy of Engineering, and Institute of Medicine, 2010; American Association for the Advancement of Science [AAAS], 2011; President’s Council of Advisors on Science and Technology, 2012; Chen, 2013). As such, there has been increasing interest in developing bridge programs for STEM disciplines to prepare students to meet the challenging academic demands of these degree programs and to address factors that may contribute to the attrition of students in STEM (Seymour and Hewitt, 1997).
Bridge programs take substantial time, effort, and resources due to their immersive nature, so a question that often arises is, What is the impact of bridge programs on students? Surprisingly, there are few review articles of bridge programs, and the ones that exist only summarize a subset of bridge programs (e.g., Kezar, 2000; Sablan, 2014). To our knowledge, there has been no comprehensive review of the literature on STEM bridge programs. To meet this need, we summarize the literature on bridge programs that target students who are entering or are interested in entering STEM majors. We describe characteristics of existing bridge programs, identify the goals of these programs, and highlight to what extent these programs report success in achieving these goals. We also identify gaps in the current literature on bridge programs and provide a set of recommendations for the STEM community to build better bridges into college for incoming students. We hope that this summary of the current landscape of STEM bridge programs can serve as a useful resource for bridge program developers. Further, we hope this review will stimulate thoughtful consideration of how to improve existing bridge programs and better evaluate the impact of bridge programs on students.
METHODS
We conducted a literature search for published reports that describe STEM bridge programs. In addition to the published literature, we searched the “gray literature” or non–peer reviewed literature, primarily in the form of unpublished dissertations and conference proceedings. Many bridge programs have not led to a peer-reviewed publication in a journal, so we chose to include reports that have not been peer-reviewed in a journal (e.g., conference abstracts). We did this so that we can provide readers with the maximum amount of information about current bridge programs and to avoid only focusing on programs that were most likely to produce favorable or statistically significant results (Rothstein et al., 2005). However, we recognize that there are likely many bridge programs for which no reports exist, and as such, these programs are not included in our review. Specifically, programs developed at master’s colleges and universities, baccalaureate colleges, and community colleges may be underrepresented in this review compared with programs at PhD-granting research institutions due to the greater incentives for faculty at R1 institutions to publish their findings. Between November 2016 and March 2017, we searched three online databases (Google Scholar, ERIC, and Arizona State University [ASU] library One Search) using an array of terms, including: bridge, bridge program, summer bridge, summer bridge program, summer program, boot camp, summer boot camp, summer college program, college preparation program, and summer college preparation program. For consistency, we will refer to all programs as “bridge programs” in this literature review. Further, we used a snowball approach to find additional programs that were referenced in the papers on bridge programs (Biernacki and Waldorf, 1981). We included STEM bridge programs from 1992 (the earliest report we could find) to 2016. The search yielded 46 published reports on 30 unique programs, and while we have tried to be as comprehensive as possible, we acknowledge that it is possible to have missed published reports.
To identify characteristics of each STEM bridge program, two authors (M.A. and K.M.C.) independently reviewed each bridge report and recorded the subject of the program, the type of institution that offered the program, whether the bridge program was designed for first-year students or transfer students, the length of the program, and whether the program targeted a specific population of students as recorded in the report. The reviewers compared their findings and came to consensus about any discrepancies.
To identify the goals of each STEM bridge program, one author (M.A.) reviewed each bridge report; noted any goal of the STEM bridge program as described in the report; and recorded how, if at all, the bridge report indicated that the program goal was measured and whether the program was successful at reaching that goal. A rubric was created to describe each program goal and whether bridge reports indicated that 1) the program succeeded in meeting the program goal, 2) the program did not succeed in meeting the program goal, or 3) the program goal was not measured in the report (please see Supplemental Table 1 for a copy of the rubric). In this review, we include any goal that was reported by at least three bridge programs (10% of all programs). It is important to note that we relied on interpretations of the data made by the authors of the bridge reports in terms of whether the authors perceived their programs were successful at achieving their goals; however, we do highlight some possible issues with methodologies and conclusions. As such, some of these findings should be interpreted with caution as far as the positive impacts of the bridge programs. To establish coding reliability, another author (J.M.C.) reviewed a subset of eight STEM bridge reports and used the rubric to code what goals each program reported and whether the bridge report indicated that the program was successful in meeting the goals. The two authors (M.A. and J.M.C.) compared their coding, discussed any discrepancies, and revised the rubric. This process was repeated on two additional subsets of eight reports until the authors’ consensus estimate was greater than 95% (Stemler, 2004). The two authors then separately reviewed every program and recorded which goals were reported, which goals were measured, and whether the program reported successfully meeting each goal using the revised rubric. The reviewers compared their final codes, achieving a consensus estimate of 99%, and came to consensus about any final discrepancies.
PART 1. CURRENT LANDSCAPE OF STEM BRIDGE PROGRAMS
To our knowledge, there have been 46 published reports on 30 unique STEM bridge programs from 1992, when the earliest report was published, through 2016. Twenty-six of these publications were in peer-reviewed journals (Figure 1A). Over the past 25 years, there has been an increase in the number of published reports on STEM bridge programs, including a rise in the number of peer-reviewed publications in recent years (Figure 1B). STEM bridge programs have been developed for students entering any STEM field (n = 10, 33%), but there are also programs designed for students entering a specific STEM major or interested in a specific STEM major; the highest number of published reports have been in engineering (n = 15, 50%). There are only two chemistry-specific programs (College of Saint Benedict and Saint John’s University FoCuS program and Monmouth College’s SOFIA program), one biology-specific program (Louisiana State University’s [LSU] BIOS program), one geoscience-specific program (Kapiolani Community College’s HaKilo program), and one nursing-specific program (University of Cincinnati’s Leadership 2.0 program) with at least one published report (Figure 2A). STEM bridge programs with published reports have been developed at public doctoral-granting institutions (n = 23, 77%), private doctoral-granting institutions (n = 2, 7%), private master’s degree–granting institutions (n = 2, 7%), and private bachelor’s degree–granting institutions (n = 3, 10%; Figure 2B). Most STEM bridge programs with published reports focus on incoming first-year students (n = 28, 93%); however, two programs focus on students who transfer from 2-year institutions to 4-year institutions (7%; Figure 2C). Bridge programs vary widely in length and last from 3 days to 8 weeks; one online bridge program allows students to participate as long as they want between when they attend new student orientation and when they begin college classes (Figure 2D). These differences among programs highlight that there is no set formula for creating a STEM bridge program, which makes it difficult to draw conclusions about what aspects of bridge programs may lead to positive outcomes.
Of all the bridge programs with published reports, some explicitly welcome any incoming student (n = 3, 10%) or do not specify a specific target population (n = 4, 13%). More commonly, programs target underrepresented minority (URM) students (n = 15, 50%; e.g., Ohio State University’s PREFACE Program), female students (n = 6, 20%; e.g., Bowling Green State University’s AIMS program), academically underprepared students (n = 6, 20%; e.g., Syracuse University’s Summer Bridge Program), or highly academically prepared students (n = 3, 10%; e.g., Monmouth College’s SOFIA program; Figure 2E). URM students and female students are often preferentially recruited into STEM bridge programs because they are historically underrepresented in STEM disciplines (National Science Foundation [NSF], 2017). Students with low academic preparation are often at most risk for attrition from college, particularly during the first year of college (Chen, 2013), so programs may focus on these students to improve overall retention rates. Notably, some programs focus on only high academic–ability URM students (e.g., University of Maryland Baltimore County’s [UMBC] Meyerhoff Scholars Program), but no programs with published reports explicitly report focusing only on low academic–ability URM students. A list of the 30 STEM bridge programs with characteristics of each program is shown in Table 1.
Program | Subject of program | Institution type | First-year or transfer students | Length | Target population of students | Number of published reports | Reports in peer-reviewed journals | Reports in other formats |
---|---|---|---|---|---|---|---|---|
Bowling Green State University: Academic Investment in Math and Science (AIMS) Summer Bridge Program | Across STEM | Public, doctoral-granting university | F | 5 weeks | URM, females | 1 | Gilmer, 2007 | — |
Claremont Colleges: Summer Science Immersion Program (SScIP) | Across STEM | Private, bachelor’s degree–granting institution | F | 1 week | URM, females | 1 | Bhattacharya and Hansen, 2015 | — |
College of Saint Benedict and Saint John’s University: Future Chemists Scholarships and Support (FoCuS) Program | Chemistry | Private, bachelor’s degree–granting institution | F | 7 weeks | URM, females, highly academically prepared | 2 | Graham et al., 2017b | Graham et al., 2013 |
Kapiolani Community College (KCC): School of Ocean and Earth Science and Technology (SOEST) Summer Bridge: HaKilo | Geoscience | Public, doctoral-granting university | T | 1 week | URM | 1 | Bruno et al., 2016 | — |
Louisiana State University (LSU): Biology Intensive Orientation for Students (BIOS) Program | Biology | Public, doctoral university | F | 5 days | Open to anyone | 3 | Wischusen and Wischusen, 2007; Wischusen et al., 2011; Wheeler and Wischusen, 2014 | — |
Massachusetts Institute of Technology: Discover Engineering | Engineering | Private, doctoral-granting university | F | 4–5 days | Did not specify | 1 | Thompson and Consi, 2007 | — |
Middle Tennessee State University: Precollege summer bridge program | Across STEM | Public, doctoral-granting university | F | 10 days | Academically underprepared | 1 | Raines, 2012 | — |
Monmouth College: Summer Opportunities for Intellectual Activities (SOFIA) program | Chemistry | Private, bachelor’s degree–granting institution | F | 3 weeks | Highly academically prepared | 2 | Moore et al., 2016 | Moore, 2013 |
Ohio State University: National Science Foundation Science and Engineering Talent Expansion Program (OSTEP) and Pre-Freshman and Co-operative Education (PREFACE) program | Across STEM | Public, doctoral-granting university | F | 6 weeks | URM | 2 | Tomasko et al., 2016 | Tomasko et al., 2013 |
Syracuse University: Summer Bridge Program | Engineering | Private, doctoral-granting university | F | 6 weeks | Academically underprepared | 1 | Doerr et al., 2014 | — |
University of Alabama: Engineering Math Advancement Program (E-MAP) | Engineering | Public, doctoral-granting university | F | 5 weeks | Did not specify | 1 | Gleason et al., 2010 | Boykin et al., 2010 |
University of California–Santa Barbara: Expanding Pathways in Science, Engineering, and Mathematics (EPSEM) | Across STEM | Public, doctoral-granting university | T | 2 weeks | URM | 1 | Lenaburg et al., 2012 | — |
University of Cincinnati: Leadership 2.0 program | Nursing | Public, doctoral-granting university | F | 6 weeks | URM | 1 | Pritchard et al., 2016 | — |
University of Maryland Baltimore County (UMBC): Meyerhoff Scholars Program | Across STEM | Public, doctoral-granting university | F | 6 weeks | URM, highly academically prepared | 7 | Hrabowski and Maton, 1995; Maton et al., 2000, 2009, 2012, 2016; Stolle-McAllister, 2011; Stolle-McAllister et al., 2011; Summers and Hrabowski, 2006 | — |
University of Memphis: STEM Talent Expansion Program (STEP) Summer Mathematics Bridge Bootcamp | Across STEM | Public, doctoral-granting university | F | 2 weeks | Did not specify | 1 | Russomanno et al., 2010; Windsor et al., 2015 | — |
University of Wisconsin Milwaukee: College of Engineering and Science (CEAS) Summer Bridge Program | Engineering | Public, doctoral-granting university | F | 4 weeks | Academically underprepared | 1 | Reisel et al., 2012 | — |
Arizona State University (ASU): Women in Applied Sciences and Engineering (WISE) | Engineering | Public, doctoral-granting university | F | 5 days | Females | 2 | — | Fletcher et al., 2001a,b |
ASU: Minority Engineering Program (MEP) Summer Bridge Program | Engineering | Public, doctoral-granting university | F | 10 days | URM | 2 | — | Reyes et al., 1998, 1999 |
Morgan State University: Alliance for Minority Participation Math Bridge Program and National Aeronautical Space Administration Morgan Engineering Enrichment Program | Across STEM | Public, doctoral-granting university | F | 4 weeks | URM | 1 | — | Wheatland, 2000 |
Pennsylvania State University: Pre-First Year Engineering and Science (PREF) program | Engineering | Public, doctoral-granting university | F | 6 weeks | URM | 1 | — | Persaud and Freeman, 2005 |
Purdue University: Mathematics Summer Bridge Program | Engineering | Public, doctoral-granting university | F | 1 week | URM | 1 | — | Budny, 1992 |
Saint Edward’s University: Community for Achievement in Science, Academics and Research (CASAR) | Across STEM | Private, master’s degree–granting institution | F | 1 week | URM | 1 | — | Kopec and Blair, 2014 |
Southern Illinois University–Carbondale: Success week | Engineering | Public, doctoral-granting university | F | 3 days | Did not specify | 1 | — | Chevalier et al., 2001 |
Texas A&M: Personalized Precalculus Program (PPP) | Engineering | Public, doctoral-granting university | F | 3 weeks | Academically underprepared | 1 | — | Nite et al., 2015 |
University of Florida: Engineering Freshman Transition Program (EFTP) | Engineering | Public, doctoral-granting university | F | 6 weeks | Open to anyone | 1 | — | Citty and Lindner, 2012 |
University of Missouri–Saint Louis and Washington University: McDonnell Douglas Access to Engineering | Engineering | Public, doctoral-granting university | F | 8 weeks | URM, females | 1 | — | Shields et al., 1996 |
University of New Mexico: Summer Bridge Program | Across STEM | Public, doctoral-granting university | F | 4 weeks | URM, females | 1 | — | Ami, 2001 |
University of North Carolina at Charlotte: Engineering Boot Camp | Engineering | Public, doctoral-granting university | F | No set length | Open to anyone | 1 | — | Harkins, 2016 |
University of Portland: Summer Bridge | Engineering | Private, master’s degree–granting institution | F | 6 weeks | Academically underprepared | 1 | — | Cairncross et al., 2015 |
Wayne State University: Summer Bridge Program | Engineering | Public, doctoral-granting university | F | 8 weeks | Academically underprepared | 1 | — | Grimm, 2005 |
PART 2. GOALS, OUTCOMES, AND SUCCESSES OF STEM BRIDGE PROGRAMS
What are the goals of these STEM bridge programs, and how, if at all, do programs measure how successful they are at achieving such goals? We have identified program goals articulated by at least three bridge programs and categorized them into three main areas: academic success goals, psychosocial goals, and department-level goals. Many of the goals of bridge programs are well aligned with educational theories centered on undergraduate student success and retention. For example, Tinto’s theory of college student departure suggests that students must experience social integration (e.g., form relationships with peers and faculty) and academic integration (e.g., achieve good grades) to maximize their chances of graduating from college (Tinto, 1975, 1987).
We have also identified research designs and assessment strategies that evaluators use to assess bridge student outcomes. While evaluators commonly rely on quantitative methodologies such as surveys or tests to collect data, qualitative measures such as interviews or focus groups can also be used to provide insight into students’ thought processes and experiences. Bridge program evaluators tend use a pre–post research design or a comparison group design to assess student gains. In a pre–post design, bridge evaluators measure students’ academic success or affect at the beginning and end of the program. However, the specific measures used in pre–post designs (e.g., one item from a survey, a previously validated survey, student interviews) and the extent to which evaluators use statistical analyses to determine the effect size and significance of bridge student gains vary widely. Another way to assess the outcomes of bridge programs is to compare the gains of bridge students with the gains of a comparison group of students who did not participate in the program. However, because students usually self-select into bridge programs and often need to meet specific criteria to be eligible for the program, evaluators are usually unable to conduct randomized experiments. Thus, evaluators use quasi-experimental designs and can reduce bias by controlling for the variation between bridge students and the comparison group of students. However, the extent to which evaluators control for potentially confounding variables and the analyses they use vary extensively. For example, some evaluators compare bridge students with a larger group of students with similar demographics, while other evaluators choose to compare bridge students with a matched-pair group, with individual non–bridge students matched with individual bridge students using characteristics such as race/ethnicity, gender, and high school grade point average (GPA). The number of covariates used in the match, the criteria for determining whether the bridge and non–bridge students are appropriately matched, and the technique that evaluators use to create matched-pair groups (e.g., propensity score matching, using statistical tests to compare group averages) also vary extensively.
In the following sections, we outline the goals for STEM bridge programs included in this review, how programs measured these goals, and whether programs reported that they are successful at meeting those goals. We highlight differences in how goals were measured and possible limitations of the study designs where applicable. These findings are summarized in Table 2. We acknowledge that these goals may not be representative of all bridge programs, because not all bridge programs have published reports, so our conclusions are limited to what we can say about the published reports.
Academic success goals | Psychosocial goals | Department-level goals | ||||||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Remediation | Improve content knowledge | Maximize GPA | Increase research participation | Increase retention | Increase graduation rates | Increase interest in the major | Sense of belonging | Sense of preparedness | Self-efficacy | Network with students | Network with faculty | Recruit students to the major | Enhance diversity of the major | |||||||||||||||
Programs with reports in peer-reviewed journals | ||||||||||||||||||||||||||||
Bowling Green State University: Academic Investment in Math and Science (AIMS) Summer Bridge Program | – | * | – | * | – | S | ✓ | S | ✓ | S | ✓ | S | – | * | – | * | – | * | – | * | – | * | – | * | – | * | ✓ | * |
Claremont Colleges: Summer Science Immersion Program (SScIP) | – | * | – | * | – | * | – | * | ✓ | * | – | * | – | * | ✓ | * | – | * | – | * | – | * | – | * | – | * | – | * |
College of Saint Benedict and Saint John’s University: Future Chemists Scholarships and Support (FoCuS) Program | – | * | – | * | – | S | ✓ | S | ✓ | * | – | * | – | * | – | * | – | * | – | * | ✓ | * | ✓ | * | ✓ | * | ✓ | * |
Kapiolani Community College (KCC): School of Ocean and Earth Science and Technology (SOEST) Summer Bridge: HaKilo | – | * | – | * | – | * | – | * | – | * | – | * | ✓ | S | – | * | – | * | ✓ | S | – | * | ✓ | * | ✓ | S | ✓ | S |
Louisiana State University (LSU): Biology Intensive Orientation for Students (BIOS) Program | – | * | ✓ | * | ✓ | S | – | * | ✓ | S | – | * | – | * | – | * | – | * | – | S | ✓ | S | – | * | – | * | – | * |
Massachusetts Institute of Technology: Discover Engineering | – | * | – | * | – | * | – | * | – | * | – | * | ✓ | S | – | * | – | * | – | * | ✓ | S | ✓ | S | ✓ | S | – | S |
Middle Tennessee State University: Precollege summer bridge program | – | * | ✓ | S | – | S | – | * | ✓ | S | – | * | – | * | ✓ | * | – | S | – | U | – | * | – | * | – | * | – | * |
Monmouth College: Summer Opportunities for Intellectual Activities (SOFIA) program | – | * | – | * | – | * | ✓ | * | – | * | – | * | ✓ | * | – | * | – | * | – | * | – | * | – | * | ✓ | * | – | * |
Ohio State University: National Science Foundation Science and Engineering Talent Expansion Program (OSTEP) and Pre-Freshman and Co-operative Education (PREFACE) program | – | * | – | * | – | U | – | * | – | S | – | * | – | * | ✓ | S | – | S | – | * | – | * | – | * | – | * | ✓ | * |
Syracuse University: Summer Bridge Program | – | * | ✓ | S | ✓ | S | – | * | – | * | – | * | – | * | – | * | – | * | – | * | – | * | – | * | – | * | – | * |
University of Alabama: Engineering Math Advancement Program (E-MAP) | ✓ | S | ✓ | * | – | U | – | * | ✓ | S | ✓ | * | ✓ | * | – | * | – | * | – | * | – | * | – | * | – | * | – | * |
University of California Santa Barbara: Expanding Pathways in Science, Engineering, and Mathematics (EPSEM) | – | * | – | * | – | * | – | * | – | * | ✓ | * | – | S | – | * | – | * | ✓ | * | ✓ | S | ✓ | S | ✓ | * | ✓ | * |
University of Cincinnati: Leadership 2.0 program | – | * | ✓ | * | – | S | – | * | ✓ | S | – | * | – | S | ✓ | S | – | S | – | * | ✓ | S | ✓ | S | – | * | ✓ | * |
University of Maryland Baltimore County (UMBC): Meyerhoff Scholars Program | – | * | ✓ | * | ✓ | S | ✓ | S | ✓ | S | ✓ | S | – | * | ✓ | S | – | * | ✓ | S | ✓ | S | ✓ | S | – | * | ✓ | S |
University of Memphis: STEM Talent Expansion Program (STEP) Summer Mathematics Bridge Bootcamp | – | * | ✓ | S | – | S | ✓ | S | ✓ | S | ✓ | * | – | S | – | * | – | S | – | S | ✓ | S | ✓ | * | ✓ | * | – | * |
University of Wisconsin Milwaukee: College of Engineering and Science (CEAS) Summer Bridge Program | ✓ | S | – | * | – | * | – | * | – | * | ✓ | * | ✓ | * | – | * | – | * | – | * | – | * | – | * | – | * | – | * |
Programs with no reports in peer-reviewed journals | ||||||||||||||||||||||||||||
Arizona State University (ASU): Women in Applied Sciences and Engineering (WISE) | – | * | ✓ | * | ✓ | * | – | * | ✓ | S | – | * | – | * | – | * | – | * | – | * | – | * | – | * | – | * | – | S |
ASU: Minority Engineering Program (MEP) Summer Bridge Program | – | * | – | * | – | S | – | * | – | S | ✓ | * | – | * | ✓ | * | – | * | ✓ | * | – | * | – | * | – | * | ✓ | S |
Morgan State University: Alliance for Minority Participation Math Bridge Program and National Aeronautical Space Administration Morgan Engineering Enrichment Program | – | * | – | * | ✓ | S | – | * | ✓ | S | – | * | – | * | – | * | – | * | – | * | – | * | – | * | – | * | – | * |
Pennsylvania State University: Pre-First Year Engineering & Science (PREF) program | – | * | – | * | – | S | – | * | ✓ | * | ✓ | S | – | * | – | * | – | * | – | * | – | * | – | * | – | * | ✓ | * |
Purdue University: Mathematics Summer Bridge Program | – | S | ✓ | * | ✓ | S | – | * | ✓ | * | – | * | – | * | – | * | – | * | – | * | – | * | – | * | – | * | – | * |
Saint Edward’s University: Community for Achievement in Science, Academics and Research (CASAR) | – | * | – | * | ✓ | * | – | * | ✓ | S | ✓ | S | – | * | – | * | – | * | – | * | – | * | – | * | – | * | – | * |
Southern Illinois University Carbondale: Success week | – | * | – | * | – | * | – | * | – | S | ✓ | * | – | * | – | * | – | * | – | * | ✓ | * | ✓ | * | – | * | – | * |
Texas A&M: Personalized Precalculus Program (PPP) | ✓ | S | – | * | – | U | – | * | ✓ | * | – | * | – | * | – | * | – | * | – | * | – | * | – | * | – | * | – | * |
University of Florida: Engineering Freshman Transition Program (EFTP) | – | * | – | * | – | S | – | * | ✓ | S | – | * | – | * | – | * | – | * | – | * | ✓ | * | – | * | ✓ | * | – | * |
University of Missouri–Saint Louis and Washington University: McDonnell Douglas Access to Engineering Program | – | * | ✓ | * | – | * | – | * | ✓ | * | – | * | – | * | – | * | – | * | – | * | – | * | – | * | – | * | ✓ | * |
University of New Mexico: Summer Bridge Program | ✓ | S | – | * | – | * | – | * | ✓ | * | ✓ | * | – | * | – | * | – | * | – | * | – | * | – | * | – | * | ✓ | * |
University of North Carolina at Charlotte: Engineering Boot Camp | – | * | ✓ | * | – | S | – | * | ✓ | S | ✓ | * | – | * | – | * | – | * | – | * | – | * | – | * | – | * | – | * |
University of Portland: Summer bridge | ✓ | S | – | * | – | S | – | * | ✓ | * | – | * | ✓ | * | ✓ | * | – | * | – | S | – | S | – | * | – | * | – | * |
Wayne State University: Summer Bridge Program | – | * | – | * | – | * | – | * | ✓ | * | – | * | – | * | – | * | – | * | – | * | ✓ | * | – | * | – | * | – | * |
Academic Success Program Goals
Remediation: Providing Students with Foundational Knowledge in a STEM Domain.
Incoming students may not be academically prepared for the level of difficulty of college STEM course work, so some bridge programs try to remediate students to meet the requirements for entry into the major (Chen, 2013). Most STEM bridge programs with this goal target academically underprepared students and often use a placement test or incoming Scholastic Aptitude Test (SAT) or ACT score to determine whether students are eligible for the program. Many programs use a pre–post research design to measure remediation by having students take a placement exam before or at the beginning of the program and at the end of the program. Some programs determined they were successful if students placed into a higher-level course based on their placement exam score after participating in the program. This remediation often focused on basic math skills, so students would take a math placement test at the end of the program to measure whether they were better prepared to enter their STEM majors.
Five STEM bridge programs had remediation as a goal of their program; all five programs assessed student remediation and concluded that they were successful in remediating students (Ami, 2001; Boykin et al., 2010; Gleason et al., 2010; Reisel et al., 2012; Cairncross et al., 2015; Nite et al., 2015). One additional program reported successfully remediating students, but was explicit that remediation was not an original goal of the program (Budny, 1992).
Improving Student Content Knowledge in a Discipline.
The first year of college STEM courses is notoriously difficult, and students are prone to struggle with the sheer amount of information presented in introductory courses (AAAS, 2011; Brownell et al., 2014). To give students a head start on the content they will encounter during their first semester in college, some bridge programs strive to increase student content knowledge in a particular STEM discipline. Eleven bridge programs reported improving student content knowledge as a goal of the program, and three programs reported measuring student content knowledge. All three programs that reported measuring content knowledge gains used a pre–post assessment of math ability. One of these programs gave students a previously validated pre–post concept inventory on rate of change and found that students achieved significantly higher scores at the end of the bridge program (Doerr et al., 2014). Another program gave students a pre- and posttest with questions taken from the math section of the ACT and showed that bridge participants increased test scores by more than 40% (Raines, 2012). A third bridge program gave students a math pre- and posttest that the program developers created themselves; they found that students’ algebra and trigonometry scores increased by 15% and 14%, respectively (Russomanno et al., 2010).
Maximizing Student GPA.
One way to measure student academic success is to examine students’ GPAs. Students’ GPAs are typically obtained by accessing institutional data. Seven bridge programs reported increasing the GPAs of students who participated in the program compared with students who did not participate in the program as an explicit goal of the program; five of the seven programs measured student GPAs. An additional 13 programs did not explicitly state maximizing bridge students’ GPAs as a goal of the program, but measured GPA as a way to determine student academic success.
Fifteen of the 18 programs that measured students’ GPAs (83%) showed that their programs were successful in maximizing bridge student GPAs. It is important to note that programs used different ways of measuring gains in GPA. One program used propensity score matching to compare bridge students with students with similar characteristics (including gender, race/ethnicity, and prior academic ability) who did not participate in the bridge program (Windsor et al., 2015). Another program matched bridge participants to current and historical students who did not participate in the bridge program yet had similar characteristics (including gender, race/ethnicity, and prior academic ability; Maton et al., 2000). One program controlled for both gender and a measure of prior academic ability between bridge students and comparison students who did not participate in the program (Wischusen and Wischusen, 2007). Four additional programs did not use matched pairs, but did control for a measure of prior academic ability of students who did and did not participate in the bridge program (Wheatland, 2000; Gilmer, 2007; Gleason et al., 2010; Doerr et al., 2014).
However, seven programs compared students who completed the bridge program with students in other courses without any control for prior academic ability (Budny, 1992; Citty and Lindner, 2012; Cairncross et al., 2015; Nite et al., 2015; Pritchard et al., 2016; Tomasko et al., 2016; Graham et al., 2017), and two programs did not use a comparison group at all (Raines, 2012; Harkins, 2016). Because prior academic ability is often a strong predictor of future GPA, it is important to control for a measure of prior academic ability to examine the impact of a program. It is also important to note that students volunteered to participate in all of these programs, so there may be a volunteer effect that is leading to some of these gains (Brownell et al., 2013), because most programs compared students who volunteered to participate in the program with students who did not volunteer to participate. However, one program did compare bridge participants with “declined” students (students who planned to participate, but ended up deciding to attend another institution) to reduce volunteer-bias effects (Maton et al., 2000).
Increase Research Participation.
Five STEM bridge programs reported increasing students’ participation in research as a goal. Four of these programs measured students’ participation in research, and all reported successfully meeting this goal. Three of these programs reported the number of students who had participated in their bridge programs who went on to participate in undergraduate research (Gilmer, 2007; Russomanno et al., 2010; Windsor et al., 2015; Graham et al., 2017), and another program reported the number of former bridge students who went on to conduct PhD-level research (Summers and Hrabowski, 2006; Maton et al., 2009, 2012, 2016; Stolle-McAllister, 2011; Stolle-McAllister et al., 2011). Of note, most of these programs did not report a change in the percentage of students who were conducting research after implementing the bridge program and did not use a comparison group of students. Although one program examined the rates of success of applying to undergraduate research programs between bridge students and students who did not participate in the program, they did not control for a measure of academic ability between the two groups (Russomanno et al., 2010). Controlling for academic ability would strengthen their results, because academic ability has been shown to predict who is selected into undergraduate research programs (Russell, 2006a,b; Hurtado et al., 2014). The program that reported whether students went on to conduct PhD-level research did use a comparison group of students who had declined attending the program.
Increase Student Retention.
Student retention in a major can be defined in a number of different ways: retention through one semester, one year, multiple years, or up until graduation. Retention is usually determined by accessing institutional data on student degree progression. Twenty-one STEM bridge programs had a goal of improving student retention in college, and 12 of these programs measured student retention. An additional three programs reported measuring student retention, though they did not report retention as a goal of their program. Out of the 15 programs that measured retention, all reported having positive impacts on student retention.
There was variation in the measures for how long students were retained. Three of these programs measured retention at every year after students participated in the program until they graduated (Chevalier et al., 2001; Boykin et al., 2010; Gleason et al., 2010; Tomasko et al., 2013, 2016), and one program measured retention percentages of different cohorts of students at a single time point (Maton et al., 2000). Nine programs measured retention 1 year after the bridge program (Reyes et al., 1998, 1999; Wheatland, 2000; Fletcher et al., 2001a,b; Citty and Lindner, 2012; Raines, 2012; Windsor et al., 2015; Kopec and Blair, 2014; Harkins, 2016; Pritchard et al., 2016), one program measured retention each semester after the bridge program through students’ fourth semester in college (Wischusen and Wischusen, 2007; Wischusen et al., 2011), and one program measured retention after the first year, the third semester, the fifth semester, and the seventh semester (Gilmer, 2007).
There was also variation in the comparison groups used to evaluate bridge student retention. One program used matched pairs, individually matching students in the bridge program to students of similar characteristics who did not participate in the bridge program (Windsor et al., 2015). One program compared bridge participants with current and historical nonparticipants with similar characteristics (including gender, race/ethnicity, and prior academic ability; Maton et al., 2000). Two programs only controlled for average academic ability between the group of students who participated in the bridge program and the group of students who did not participate in the program (Gilmer, 2007; Boykin et al., 2010; Gleason et al., 2010). However, three programs used students as a comparison group to bridge students without controlling for other variables such as gender, race/ethnicity, or prior academic ability that could predict retention differences between the two groups (Chevalier et al., 2001; Citty and Lindner, 2012; Pritchard et al., 2016). One program did not use a comparison group but compared bridge participant retention with first-year student retention at the university as a whole (Raines, 2012).
Increase Student Graduation Rate from College.
Bridge program developers often stated increased graduation rates as a distinct goal from retention, so we considered it a separate category. Graduation rates are typically obtained from accessing institutional data. Twelve bridge programs had a goal of increasing graduation rates, and four of the programs measured it (Maton et al., 2000, 2009, 2012; Persaud and Freeman, 2005; Summers and Hrabowski, 2006; Gilmer, 2007; Kopec and Blair, 2014). All four of these programs reported successfully increasing graduation rates of bridge students. Notably, one of these programs focused on higher academic–ability students from underrepresented backgrounds, while the other three focused on underrepresented students of no specific academic ability.
Psychosocial Program Goals
Increase Interest in the Major.
One reason that college students choose to leave STEM majors is because of a lack of interest in the discipline (Seymour and Hewitt, 1997). Thus, bridge programs often aim to increase student interest in the major in hopes that increased interest will lead to increased recruitment and persistence. Six STEM bridge programs stated that their goal was to increase student interest in the major. Only two of these programs reported measuring student interest, and both demonstrated success in increasing student interest (Thompson and Consi, 2007; Bruno et al., 2016). Three additional bridge programs measured student interest and reported increasing student interest, though this was not an explicit goal of those programs (Russomanno et al., 2010; Lenaburg et al., 2012; Pritchard et al., 2016).
Two of the programs that measured bridge student interest in the major used a single-item pre–post measure (Lenaburg et al., 2012; Bruno et al., 2016); two other programs used postprogram surveys to measure student interest. One of these programs used a single-item measure of student interest in the major on the postprogram survey (Russomanno et al., 2010); for the other program, it is not clear from the published paper how many items were used to measure interest (Thompson and Consi, 2007). Another program measured student interest by interviewing nursing students after the program and coding the interviews for student interest; 92% of students were more interested in nursing after the program (Pritchard et al., 2016).
Improving Student Sense of Belonging.
Student sense of belonging to a community has been shown to influence student academic motivation, student well-being, and student academic achievement (reviewed in Anderman and Freeman, 2004; Trujillo and Tanner, 2014). Seven bridge programs cited sense of belonging as a goal, and three of those programs assessed student sense of belonging after the program (Maton et al., 2000, 2012, 2016; Stolle-McAllister, 2011; Stolle-McAllister et al., 2011; Tomasko et al., 2013, 2016; Pritchard et al., 2016). Two of these programs conducted one-on-one or focus group interviews with bridge students and analyzed interviews for themes of belonging and used survey items for identifying a sense of community (Maton et al., 2000, 2012; Stolle-McAllister, 2011; Stolle-McAllister et al., 2011; Pritchard et al., 2016). The other study assessed students’ written responses to open-ended questions about the impact of the program and coded for evidence of student sense of belonging (Tomasko et al., 2013, 2016). This program also gave students a pre–post survey that measured constructs of sense of belonging such as student social support on campus, feelings of connectedness, and being part of a group. All of these programs showed that students developed a sense of belonging by participating in the program.
Increasing Student Sense of Preparedness.
A student’s sense of preparedness relates to how well a student perceives that he or she is prepared for a degree program. Bridge programs could help prepare students for college, especially if students enter college from high schools that may not have adequately prepared them for the rigors of college. One program examined preparedness for precalculus (Raines, 2012), another program assessed students’ sense of preparedness regarding future STEM courses (Pritchard et al., 2016), and two programs focused on preparedness for college in general (Russomanno et al., 2010; Tomasko et al., 2013, 2016).
While no programs specifically reported sense of preparedness as being a goal, these four programs did report measuring students’ sense of preparedness, and all reported having a positive impact. Three programs measured this through a single post survey question (Russomanno et al., 2010; Raines, 2012; Pritchard et al., 2016), while the fourth program used a pre–post survey design, although they only provided the post survey score in the published report (Tomasko et al., 2013, 2016).
Increasing Student Self-Efficacy.
Self-efficacy is a student’s perception of his or her ability to complete a task (Bong and Skaalvik, 2003) and has been suggested to be important for retention in science (Seymour and Hewitt, 1997). It is distinct from a sense of preparedness, because it is the actual confidence that a student has in being able to complete a task, not the extent to which he or she feels prepared for an academic experience. Four programs reported increasing student self-efficacy as a goal, and two of those programs measured student self-efficacy (Bruno et al., 2016; Maton et al., 2016). Additionally, four programs measured self-efficacy, even though they did not explicitly state increasing student self-efficacy as a goal (Wischusen and Wischusen, 2007; Russomanno et al., 2010; Raines, 2012; Wheeler and Wischusen, 2014). One program measured student research self-efficacy using six items from a scientific self-efficacy scale, which probes student confidence in performing common research practices such as technical skills or generating a research question (Maton et al., 2016). However, this program did not use a pre–post survey design; they reported students’ research self-efficacy scores but did not measure how the program influenced the level of student research self-efficacy. Another study used the self-efficacy subscales of the MSLQ (Motivated Strategies for Learning Questionnaire), which measures students’ perceptions of their abilities to be successful in their introductory biology class, in a pre–post program design (Pintrich and De Groot, 1990). Surprisingly, students’ self-efficacy decreased over the duration of the program (Wheeler and Wischusen, 2014). While a decrease in self-efficacy could be interpreted as a failure of the program, some literature suggests that overestimation of self-efficacy can result in lower probability of completing a task and lower student motivation. Therefore, a student may benefit from calibrating his or her self-efficacy (Schunk and Pajares, 2009), which is what the authors described as possibly happening in this program. Thus, the authors interpreted this to be an unintended benefit of the program.
Networking with Students.
Building relationships with their peers can help students feel as though they are part of the college community. Conceptualized as social integration in Tinto’s theory of college student departure (Tinto, 1975), student relationships can lead to enhanced student sense of belonging, sense of community, and retention in a degree program (Tinto, 1987; Cabrera et al., 1993; Gerdes and Mallinckrodt, 1994). Of the 10 STEM bridge programs to report a goal of increasing student relationships, five of these programs measured it using qualitative methods (Maton et al., 2000; Thompson and Consi, 2007; Stolle-McAllister, 2011; Stolle-McAllister et al., 2011; Wischusen et al., 2011; Lenaburg et al., 2012; Pritchard et al., 2016), and one measured it by providing students with a postprogram survey in which students rated their agreement to a question regarding networking with other students (Russomanno et al., 2010). One additional program that did not have networking with students as a goal conducted post–bridge focus group interviews, and they found that students reported networking with their peers during the bridge program (Cairncross et al., 2015).
Networking with Faculty.
Building relationships with faculty can be particularly helpful to students in navigating STEM disciplines, because students sometimes perceive STEM faculty as “chilly” and “unapproachable” (Seymour and Hewitt, 1997). There is evidence to suggest that few college students develop relationships with instructors (Snow, 1973; Pascarella, 1980; Lamport, 1993; Kuh and Hu, 2001). As an example, only 15% of students at a large research institution reported that an instructor knew their names in a typical high-enrollment biology course (Cooper et al., 2017b). Eight STEM bridge programs had a goal of students getting to know faculty. Of the four programs that measured this goal, two programs measured this by coding for themes of students networking with faculty from one-on-one or focus group interview data (Maton et al., 2000; Pritchard et al., 2016), while the other two programs issued a postprogram survey with questions about the extent to which students met and networked with faculty (Thompson and Consi, 2007; Lenaburg et al., 2012). All four programs showed positive gains in student reports of networking with faculty.
Department-Level Goals
Recruit Students to the Major.
Some STEM departments used the bridge program as a way to attract more students to a specific major and advertised the bridge program as a way for students to learn more about the major. Seven STEM bridge programs had a goal of using the program to recruit students, and the two programs that measured this goal were both successful in recruiting more students to the major. One program measured recruitment by counting the total number of students the bridge program recruited (Bruno et al., 2016), and the other program reported an increase in the number of students in the major after the bridge program was implemented (Thompson and Consi, 2007).
Enhance Diversity in the Major.
There is a mismatch between the diversity of the current scientific workforce and the general public, so there have been efforts aimed at increasing the diversity of STEM majors (NSF, 2017). Diversity in STEM has been proposed to be important for ameliorating socioeconomic disparities, increasing the total number of talented scientists, and minimizing bias in scientific findings by having individuals from diverse backgrounds counteract each other’s biases (Intemann, 2009). Although more than half of the STEM bridge programs targeted an underrepresented group in STEM (e.g., URMs or women), only 11 bridge programs (37%) explicitly stated that increased diversity in a STEM major was a goal of the program. Three of the 11 bridge programs reported the results of measuring increased diversity and stated that they were successful in enhancing diversity (Maton et al., 2000, 2012; Summers and Hrabowski, 2006; Stolle-McAllister et al., 2011; Bruno et al., 2016). Two additional programs also measured and reported success in enhancing diversity, though this was not an explicit goal of those programs (Fletcher et al., 2001a,b; Thompson and Consi, 2007). Success was determined by increased numbers of underrepresented groups in the major after implementation of the bridge program.
PART 3. RECOMMENDATIONS FOR FUTURE DIRECTIONS
While the STEM bridge literature is growing and the number of peer-reviewed journal publications has increased in recent years, we have identified some areas for improvement in the current literature on bridge programs. In the following sections, we outline a set of four recommendations for enhancing the quality of the literature on STEM bridge programs, so bridge program developers can make more informed and evidence-based decisions in creating and refining bridge programs.
Recommendation 1: A Call to Document and Publish Bridge Program Descriptions, Goals, and Outcomes
We encourage bridge developers and evaluators to publish their findings in peer-reviewed journals. The process of peer review will help strengthen the literature base for bridge programs, so others can build upon what other bridge programs have learned. We chose to include both peer-reviewed and non–peer reviewed reports in our analysis of bridge reports, because, out of the 30 STEM bridge programs we identified, only 16 generated peer-reviewed publications. We also acknowledge that there are bridge programs that exist that do not have any published reports written about them, so they are missing from our analysis. This highlights the need for bridge developers and evaluators to publish on successful bridge programs currently absent from the peer-reviewed literature.
It is possible that bridge developers and evaluators are not publishing their findings because most of their evaluation is focused internally on improving their programs. Yet, in order to establish evidence for successful bridge programs nationally, it is important for these findings to be documented in peer-reviewed journals that others can easily access. A possible solution to this problem is to move away from only internal evaluation of bridge programs and instead use design-based research (Anderson and Shattuck, 2012). Design-based research explores the impact of interventions that are implemented in educational settings, such as summer bridge programs, to both inform the process of designing and revising such programs and to contribute to our broader understanding of education theory (Anderson and Shattuck, 2012).
Recommendation 2: Development of a Program Is Iterative, So Report Lessons Learned from Prior (Unsuccessful) Iterations to Guide the Development of More Successful Future Programs
Design-based research recognizes the complexity of interventions such as bridge programs and acknowledges that design and evaluation are iterative processes that help to inform each other (Anderson and Shattuck, 2012). There are many moving parts in the development of a successful bridge program, and it is quite possible that first iterations of programs will not be successful in meeting those programs goals. Much can be learned from formative assessment of bridge programs, particularly pilot versions of programs, so that changes can be made to help improve the experience for a first cohort of students. It is important for bridge developers to realize that with a limited literature base on bridge programs that is not comprehensive of all types of students or institutions, it is likely that the first iteration of a program will not go smoothly. For example, we could not find one published example of a summer bridge program that is focused on underrepresented students who are academically underprepared, even though we have met multiple people at conferences who are interested in starting bridge programs for this specific population of students. Without any prior evidence to build on, bridge program developers may find out that what works well for academically prepared underrepresented students may not work as well for academically underprepared underrepresented students. However, many insights can be gained from “failed” programs with small pilot groups of students, especially if bridge developers can identify specific aspects of the program that did or did not work well. We encourage bridge developers to publish lessons learned from early pilots of bridge programs with enough detail so that others may be able to avoid the same issues.
Unlike some teaching innovations that are relatively low cost and easy to implement, bridge programs tend to require significant time, funding, and resources to develop and implement. The lack of current information available, combined with the zeal for developing these bridge programs, means that we are likely developing new programs without building on the experiences and advice of others. Given the current funding pressures and limited resources on campuses, particularly at public institutions, perhaps we should prioritize funding to programs that are using evidence to design and revise their programs.
Generally positive gains were reported from bridge programs, and while we do not doubt that positive gains can result from bridge programs, we worry that some of these reported positive gains were not significant and/or that negative results may not have been reported. To maximize the effectiveness of STEM bridge programs, we need to learn from both the successes and failures of these programs. The bridge community would greatly benefit from examples of programs that were not successful and ideas about why they may not have been successful. Further, there are two programs (LSU’s BIOS and UMBC’s Meyerhoff Scholars Program) that make up almost 50% of the peer-reviewed journal publications on STEM bridge programs. Although BIOS and the Meyerhoff Scholars Program are very different in their target populations (open to anyone compared with targeting high academic–ability URM students, respectively) and length of time (5 days compared with 6 weeks), there may be particular elements of these programs that do not generalize to other STEM bridge programs. Thus, we need additional examples in peer-reviewed journals of bridge programs that take place in different contexts with different students. We can learn a lot from the elements of programs that do not lead to positive results as well as from elements that are successful, especially if we know enough about the details of the programs. All of this information may help other bridge developers design their own programs.
Recommendation 3: Report More Information about the Details of Implementing Bridge Programs
When we surveyed the literature on bridge programs, we were struck by how little information was presented about how to run the day-to-day activities of a bridge program. How much do these programs cost? How are students recruited and selected into the program? What is the average size of the program, who staffs the program, and who develops the curriculum? A team developing a bridge program would benefit from more information about program logistics. While we recognize that there are constraints (e.g., word limits) for papers and concerns about relevance to the main points of a paper, we would encourage bridge developers to include this type of information in the supplemental material of a paper, on a linked website, or in a stand-alone curriculum description of the program. For more information about some of the details of the bridge programs in this review, please see individual program websites listed in Supplemental Table 2.
Costs and Resources.
Costs involved in running bridge programs can include funds to support curriculum development, stipends/salaries for staff, housing/food for students, materials for the program, and transportation. We had difficulty finding information about the costs of programs, including how much programs cost overall, program costs per individual student, and where the funds to run the program come from, although some reports do provide specific information (e.g., Wischusen and Wischusen, 2007). Some programs may be funded by an institutions, and if an institution is interested in increasing retention rates, then the focus of the program may be on academically underprepared students who are predicted to not be retained through their first year of college. Alternatively, programs may be underwritten by national funding agencies. In such cases, the priorities of national funding agencies may influence the demographics of students recruited to a bridge program. For example, funding from the National Science Foundation is awarded in part based on broader impacts and often focuses the recruitment into the program on underrepresented or underserved groups in science. Some programs may depend on participants to fund the bridge program, although requiring students to pay to participate in bridge programs may limit the participation of students from low socioeconomic backgrounds who are less likely to persist in college than their economically advantaged peers (Thayer, 2000).
Bridge program developers would also benefit from learning about programs that use a combination of funding sources, especially if they receive initial funding from a national funding source and succeed in becoming self-sustaining. To our knowledge, there is little information about which programs have been able to become self-sustaining after national funding runs out. Some programs seem to extend their programming by combining national grant funding with private donor funding, while others have their institutions take on the financial responsibility for their programs. It is worth noting that different institutions have varying levels of institutional support for applying for grants or developing relationships with potential donors, so conclusions about how best to fund and sustain programs are likely specific to institution type. Nevertheless, bridge program developers would likely benefit from learning how current programs are sustained after initial national funding is depleted.
Recruitment and Selection.
Recruiting students to bridge programs can be difficult, because by definition, students who are recruited for bridge programs have not yet attended the 4-year institution, so communication can be complicated. More information about how programs recruit students would be useful to future bridge program developers (e.g., do they coordinate with admissions departments or do they hire staff who are in charge of recruitment?). Additionally, if programs cannot accommodate all students who are eligible and willing to participate in the program, then the bridge report should include how participants are selected. This information can be especially important when interpreting results, because selecting for students with certain characteristics (e.g., high motivation or commitment to the major) may influence outcome measures (e.g., retention), and programs would need to control for these characteristics to determine whether the intervention had an impact on a particular outcome.
Size.
From the published literature, it was often difficult to ascertain the size of programs, because sizes were not reported or programs often scale up over time and reports did not break down program enrollment by year. However, bridge program developers would benefit from other programs that support a similar number of students, because scaling up comes with its own set of challenges (Elias et al., 2003; Klingner et al., 2003).
Curriculum Development.
Bridge programs that have the same goals as a previously published bridge program may want to adapt previously developed curricular materials. If programs develop unique activities or exercises for students, especially ones that have been shown to influence student cognitive or affective gains, it would be helpful to share them by publishing in an outlet such as CourseSource, a website devoted to evidence-based teaching resources for biology education (CourseSource, 2017). The bridge community would benefit from creating and building upon resources that can be easily shared.
Follow-up and Concurrent Interventions.
All of these bridge programs take place during the summer before students enter college. However, 40% of the programs have a follow-up component that stretches into students’ first semesters or first years, or even their entire college careers (see Supplemental Table 2 for information about the follow-up for the 30 bridge programs). Thus, the length of summer bridge programs outlined in Figure 2 may be a bit misleading if a program continues to support students for an additional 4 years. While some bridge programs reported this follow-up, the level of detail about the follow-up varies. Sometimes it consists of a few social events spread out over the first year of college, other times it is a weekly class that has a set curriculum. Thus, the impact of these bridge programs may be the result of more than the summer experience. Further, institutions may implement multiple strategies to improve first-year student retention, including first-year mentoring programs (Rodger and Tremblay, 2003) or first-year learning communities (Hotchkiss et al., 2006), which can make it difficult to disaggregate how bridge programs influence student success. Information about bridge follow-up and other concurrent interventions would be useful to include in bridge reports or in supplemental material to give bridge developers a greater level of information about what aspects of the program may have contributed to specific gains and to help inform decisions about how long to support bridge students.
Recommendation 4: Align Bridge Goals and Measured Outcomes
From our analysis of the peer-reviewed published reports on STEM bridge programs, we have identified a misalignment between stated goals of the programs and the outcomes that are measured. In some cases of peer-reviewed publications of bridge programs, programs have goals that they do not report measuring, while in other cases, programs report outcomes, but do not explicitly state that those outcomes were goals of the program. It is possible that the program evaluators measured the goal and did not report an unsuccessful outcome, that the data that they needed to measure a goal were not available at that time, or that the authors felt that the data were not relevant to that particular publication. However, it is also possible that some programs may not have aligned their assessments of their programs with their intended goals.
This disconnect between stated goals and outcomes may indicate that some STEM bridge programs are not being backward designed. Backward design is a model for course or program development that recommends first outlining desired goals, then determining acceptable evidence, and finally planning experiences and instruction (Wiggins and McTighe, 1998). Backward design has primarily been recommended for lecture courses. Recently, we have made specific recommendations to use backward design when designing course-based undergraduate research experiences and independent research experiences because of disconnects between goals and measured outcomes in the literature on undergraduate research (Cooper et al., 2017c). It appears as though similar disconnects may be happening for bridge programs, and a backward design approach could help program developers ensure that the activities in their program are being designed to meet their larger program goals.
Specifically, bridge developers may benefit from first selecting the goals for their programs. Theoretical frameworks, such as Tinto’s theory of college student departure, can guide developers in selecting short-term goals (e.g., social integration) that are theorized to lead to accomplishing long-term goals (e.g., increased retention in the major). Different institutions (e.g., community colleges compared with research universities) may have different challenges for incoming students and thus may want to consult the literature to identify what short-term goals may be most likely to lead to accomplishment of long-term goals given the unique characteristics of their institution.
Once goals are selected, then bridge developers can decide what constitutes acceptable evidence that a goal has been met. This may take the form of a pre–post assessment design wherein students’ progress toward a particular goal is measured at the beginning and end of the program, or a matched-pair design wherein bridge student outcomes are measured by comparing them with the outcomes of students who did not enroll in the bridge program or students in a previous year when the bridge program did not exist. Bridge programs are often voluntary, so any comparison of the participants with nonvolunteers should be viewed with caution, as there are likely motivation differences among these students (Brownell et al., 2013). Once bridge developers have selected goals and determined acceptable evidence, they can use STEM education literature to plan student experiences and instruction that are most likely to lead to program goals. If bridge developers observe an unexpected result from their bridge program (e.g., Cooper et al., 2017a), then they could build these desired outcomes into future versions of bridge programs and then use backward design to assess what specific elements of the program led to that particular result. As such, this backward design process becomes a highly iterative one in which novel outcomes could lead to a revision of the goals for the next year’s bridge program.
Backward design can also help bridge developers be more critical about how they are assessing the impact of their programs. A major theme from our literature review on STEM bridge programs is that outcomes are being measured in many different ways, some of which are not aligned with best practices and may not be the most direct or rigorous way to measure a specific outcome. While it is unsurprising that researchers in an emerging area are using different approaches to assess bridge programs, discussions about what to consider when measuring certain outcomes and what are best practices for measuring those outcomes would benefit those interested in improving bridge programs.
CONCLUSIONS
In this paper, we synthesized the literature on STEM bridge programs over the past 25 years. We found 46 total published reports on 30 unique bridge programs. While these programs report success in achieving certain goals and provide a foundation for bridge developers, there is much that still needs to be established about the impact of STEM bridge programs. We hope that this review will spark more research and conversations about the potential impact of bridge programs.
ACKNOWLEDGMENTS
We thank the developers of the bridge programs discussed in this article for providing an important foundation that future programs can build upon. We also thank the members of the ASU Biology Education Research Lab, especially Sai Tummala and Erin Shortlidge, for their contribution and feedback on earlier drafts of this paper.