Existing Program Evaluation (EPE) Application Reviews

General Review Information

Each specific review criterion (see below) is rated on a 5-point scale; higher scores indicate higher criteria strength. Reviewers consider strengths and weaknesses within each criterion when scoring (1 = Poor, 2 = Fair, 3 = Good, 4 = Excellent, and 5 = Outstanding). A reviewer’s final recommendation concerning whether to invite or fund a proposal (depending on the Stage level) is not necessarily based on the average of the criterion scores. For example, a reviewer may give only moderate scores to some of the review criteria but still recommend inviting/funding because one criterion critically important to the project is rated highly; or a reviewer could give mostly high criterion ratings but still not recommend inviting/funding because one criterion critically important to the project being proposed is not highly rated.

Review Criteria Categories (see below for detailed information concerning each criterion)

1. Overall fit and potential impact

  • Specific aims / scope of work and fit with BEF mission
  • Collaboration
  • Potential impact and dissemination plans
  • Rationale (need for the program and need for this particular evaluation)

2. Program to be evaluated

  • Program implementation and goals
  • Program leadership
  • Target population
  • Feasibility of implementation from practitioner / service provider perspective
  • Accessibility from perspective of potential participants / target population
  • Affordability and sustainability
  • Strength-based orientation

3. Evaluation research criteria

  • Evaluation design and methodology
    • Design
    • Intended study sample (size, recruitment, maintenance, demographics)
    • Procedures and measures
    • Analytic plan
    • Feasibility of proposed evaluation
  • Research environment and team
    • Research environment
    • Expertise of the PI / research team

4. Budget and budget justification

Detailed Information Concerning Review Criteria Specific to Existing Program Evaluation (EPE) Applications

Below, the factors considered when scoring each of the criteria are described below. The questions in the application most pertinent to each criterion are provided; however, responses throughout the application may be considered when scoring each criterion as well

1. Overall fit and potential impact

The mission of the Brady Education Foundation is to improve educational outcomes for children (ages birth through 18), including those living in marginalized communities. One way the Foundation pursues its mission is by promoting collaboration among researchers, educators, and other stakeholders via the funding of program evaluations that have the potential of informing private funders and public policy. When determining the overall fit and potential impact of an EPE proposal, the following criteria are considered:

  • Specific aims / scope of work and fit with BEF mission
    • Factors considered when scoring these criteria include what is/are the specific aim(s) of the proposed project (including the primary aim of evaluating effectiveness as well as any secondary aims concerning possible mediators, moderators and/or cost-benefit analyses), clarity of the aim(s), and the goodness of fit with the aim(s) with the mission of the Foundation.
    • Proposals that seek to test moderators, mediators and/or conduct cost-benefit analyses, in addition to evaluating the impact of the program on children’s outcomes, tend to score higher on these criteria.
    • Section 2, questions 1 and 2; Section 1, questions 8 is also considered if other current or potential funding partners exist and/or if the proposed evaluation is part of a larger project.
  • Collaboration
    • Scoring of this criterion is based on the extent to which the proposed project reflects strong collaborations between the evaluation team and practitioners / service providers (and other community stakeholders as appropriate such as parents / families, economists, policy makers, other community members, etc.) throughout the proposed project (e.g., developing questions, recruitment, data collection, analyses and interpretation, dissemination).
    • Proposals that demonstrate authentic collaboration among the researchers, practitioners and other stakeholders score higher on this criterion.
    • Section 2, question 3.
  • Potential Impact and dissemination plans
    • Factors considered when scoring these criteria include potential dissemination products (e.g., solid contribution to lists such as the “What Works Clearing House” at the US Department of Education’s Institute of Education Sciences, likelihood of publications in high quality journals), and the potential impact for the results from the project to inform practice, programmatic funding decisions by private foundations, and/or public policy.
    • Proposals that have clear and specific plans for how results might be disseminated to researchers, practitioners, and policy makers, and those that are able to identify funders (e.g., specific foundations) and/or policy makers (e.g., public school districts) with whom they are in contact and whose work will be informed by the findings tend to score higher on these criteria.
    • Proposals that have the potential to generalize beyond the specific program to be evaluated and inform how it might be scaled up to serve other communities tend to score higher as well.
    • Section 2, questions 4 and 5.
  • Rationale (need for the program and need for this particular evaluation)
    • Factors considered when scoring this criterion include:
      • the extent to which compelling rationale is provided to justify the aim(s) of the proposed project,
      • clarity of the overall Theory of Change guiding the proposed project and how this specific project fits within it,
      • and the strength of the empirical literature supporting the aim(s) of the proposed project.
    • Proposals that do the following score higher on this criterion:
      • demonstrate a clear need for this program to address disparities in educational opportunities
      • provide evidence supporting the components of the program to be evaluated (e.g., specific program activities, specific program services),
      • and demonstrate the clear need for the specific evaluation being proposed (particularly if any prior evaluations have been conducted or other current evaluations are being conducted).
    • Also considered are what “next steps” results from the proposed project might inform (e.g., providing pilot data that might lead to further program development and/or larger evaluation study).
    • Section 3, questions 1-3.

2. Program to be evaluated

Several factors concerning the program itself are considered when assessing whether to invite an ECE application to Stage 2.

  • Program implementation and goals
    • Scoring of this criterion is based on the extent to which the program to be evaluated is likely to have a meaningful impact on children’s educational outcomes. 
    • Factors considered include the intended goals of the program and the activities participants engage in to achieve these goals, the duration / intensity of the program, and the methods the program uses to reach and enroll potential program participants.
    • Proposals that provide compelling evidence that via the program outreach methods, the duration and intensity of the program, and the activities program participants engage in will likely have meaningful impacts on ultimately promoting academic, achievement, and cognitive outcomes score higher on these criteria. 
    • Section 3, questions 1 and 4a-h.
  • Program leadership
    • BEF favors the evaluation of programs that include representation of the communities served in the leadership (e.g., executive director, administrative team, board of directors, advisory board). 
    • Programs that demonstrate higher representation of the communities served in their leadership score higher on this criterion. 
    • Section 3, question 4i.
  • Target population
    • Factors considered when scoring this criterion include the extent to which the program to be evaluated serves children and families in a variety of communities.
    • Programs that serve high proportions of children and families from these communities score higher on this criterion. 
    • Section 3, question 9.
  • Feasibility of implementation from practitioner / service provider perspective:
    • Scoring of this criterion is based on the strength of the evidence provided concerning how well the program to be evaluated has been / can be implemented in the “real world,” considering the opportunities and challenges faced by practitioners and service providers implementing the program, and the likelihood it could be scaled up across a variety of settings / communities. 
    • This criterion specifically concerns factors that facilitate the ability of those providing the program to implement it (factors that facilitate program participation are considered in the next criterion concerning accessibility). 
    • Proposals that identify specific challenges those providing the program potentially face implementating the program and how those are / might be addressed score higher on this criterion. 
    • Section 3, question 5.
  • Accessibility from perspective of potential participants / target population
    • Scoring of this criterion is based on the strength of the evidence provided concerning how well the program participants can access the program, considering both barriers to and incentives for program participation (such as the extent to which potential participants view the program as useful). 
    • Proposals that provide data concerning the percentage of the target population that actually participate (both starting the program as well as continuing through to its completion) and identify specific potential barriers to access and how those are / might be addressed score higher on this criterion. 
    • Section 3, question 6.
  • Affordability and sustainability
    • Affordability refers to the both the start-up costs required to initiate a program as well as the on-going operational costs of a program, weighing costs to potential benefits. 
    • Sustainability in this regard refers to the extent to which sources of funding sufficient to meet those costs can be identified (i.e., “financial stakeholders”). 
    • Scoring of this criterion is based on the strength of the evidence provided concerning the affordability and sustainability of the program to be evaluated. 
    • Proposals that provide estimates for both start-up and on-going implementation costs and specify how those costs might be met by a community implementing the program (e.g., philanthropic donations; local, state, or federal funding) tend to score higher on this criterion.  
    • Section 3, question 7.
  • Strength-based orientation
    • Scoring of this criterion is based on the extent to which the program to be evaluated is grounded in a strength-based approach (rather than a deficit-based model), identifying specific strengths individuals and families bring that the program builds upon. 
    • That is, while the program may intend to build strengths by participating in the program, the program also recognizes strengths that participants bring that can be built upon to promote outcomes. 
    • Factors considered when scoring this criterion include the extent to which the program considers the specific and unique needs, challenges, and strengths of children and families in various communities, (recognizing that one-size does not necessarily fit all). 
    • Section 3, question 8.

3. Evaluation research criteria

Two general categories of evaluation research criteria are considered when assessing whether to invite an application to Stage 2: research design / methodology and the research environment / team.

  • Evaluation design and methodology:
    •  Scoring of these criteria are based on the extent to which the proposed design and methodology are likely to provide high quality data that will inform the specific aim(s) of the proposal.
      • Section 4, questions 1-9; Section 7, timeline with benchmarks.
    • Design
      • Scoring for this criterion is based both on the clarity and the rigor of the evaluation design. 
      • In general, RCTs are considered more rigorous than comparison group designs, and comparison group designs are considered more rigorous that pre-post designs that only include program participants. 
      • Thus, RCTs tend to score higher than QED studies on this criterion, with both scoring higher than pre-post designs. Further:
        • For RCTs:
          • Scoring is also based on: 
            • the clarity of the descriptions of the treatment and control groups (e.g., activities), and
            • the clarity of the description of how and when randomization will take place. 
          • Note: The Foundation does not favor withholding services in order to conduct an RCT but does recognize that RCTs can be ethically done in a number of circumstances such as when programs are oversubscribed and thus cannot serve all individuals who would want to participate, or when wait-list RCTs are conducted such that the control group is able to participate in the program after post data are collected.
        • For comparison group designs:
          • Scoring is also based on
            • the strength of the rationale for not conducting an RCT, and 
            • the extent to which possible confounding variables (e.g., due to selection bias) are identified and can be controlled for and baseline equivalence among the groups can be determined. 
          • When methods such as propensity score matching are proposed, projects that demonstrate the ability to match on not only demographic factors score higher on this criterion (e.g., also match on performance measures prior to program participation).
        • For pre-post designs that only include program participants:
          • Scoring is also based on:
            • the strength of the rationale for conducting neither an RCT nor a comparison group design, and 
            • the rigor of the methods proposed to determine program effectiveness. 
          • Note: The Foundation very rarely funds proposals with pre-post designs.
  • Secondary aims:
    • Proposals that include secondary aims of investigating possible mediators and/or moderators of effects and/or conduct cost/benefit analyses tend to score higher than those only investigating main effects of the program.
    • Scoring is based on design / methodology as well as the extent to which compelling rationale for including these secondary aims is provided in Section 3, question 1.
    • Further:
      • Mediators:
        • Proposals that articulate clear hypotheses concerning what factors may account for the program effects tend to score higher.
        • Such factors may be operationalized such that the data collected can be entered into quantitative data analyses to test these hypotheses (e.g., path modeling) or investigated by collecting and analyzing qualitative data
        • Section 4, question 5
      • Moderators:
        • Such proposals may articulate clear hypotheses concerning how the strength or direction of a program’s effects may vary for different subgroups or may aim to test whether the program is equally effective across different subgroups. 
        • Proposals that provide evidence of sufficient power to test these hypotheses and provide specific analytic plans for testing them (e.g., interaction terms entered into regression models) score higher.
        • Section 4, question 6.
      • Cost / benefits:
        • Proposals that are able to identify and quantify all costs of the program (both start-up and ongoing) as well as monetary benefits (could be at the individual or societal levels) and articulate a clear analytic plan score higher.
        • Section 4, question 7.
  • Intended study sample (size, recruitment, maintenance, demographics)
    • Scoring for this criterion is based on:
      • the extent to which the sample size is judged to be appropriate for all aims of the proposed project including both main effect and moderation research questions (if applicable), and 
      • the extent to which the intended demographic characteristics are consistent with the aims of the study. 
    • Proposals that do the following score higher:
      • provide power analyses where appropriate (stating assumptions made) that demonstrate that the target sample size results in the power needed to detect main effects (and test moderation hypotheses if applicable), and 
      • articulate specific strategies for recruiting (and procedures for retaining over time if applicable) a sample representative of the intended population in sufficient numbers to investigate the aim(s) of the study. 
      • Further:
        • For RCTs, also considered when scoring this criterion:
          • the sample sizes for each treatment and control group at the level of randomization, and
          • the strength of the evidence provided that the sample size will be large enough to detect meaningful effects (e.g., if randomizing at the classroom level, evidence that the number of classrooms included in each group is sufficient to detect effects).
        • For comparison group designs, also considered when scoring this criterion:
          • the strength of the evidence provided that the sample size will be large enough to detect meaningful effects, and 
          • the strength of the evidence provided that the sample size is large enough to control for potential confounding variables.
  • Procedures and measures
    • Scoring for these criteria are based on:
      • whether the proposed procedures and data to be collected are consistent with the aims of the proposed project. 
      • the extent to which outcome measures are consistent with the goals of the program (e.g., including procedures to assess student achievement if the ultimate goal of the program is to increase student achievement),
      • the purposes of the data to be collected (e.g., to assess child outcomes),
      • the methods used to collect data (e.g., direct assessment, parent-report, administrative data), and
      • the strength of the evidence provided concerning the psychometric properties of the methods proposed (including evidence of the validity of the proposed measures for use with the specific populations represented in the sample).
    • Proposals that do the following score higher:
      • intend to obtain data from multiple sources,
      • obtain child outcome data by directly assessing children, and
      • demonstrate that the measures proposed are reliable and valid for the populations represented in the sample score higher on these criteria.  
    • Note: if sources for data include administrative data sets or other existing data sets currently managed by other parties, be sure to attach the data sharing agreements to the application. (Section 7, required attachment) 
  • Analytic plan
    • Scoring for this criterion is based on the level to which the analyses for each specific aim of the proposed project are clearly articulated and appropriate (e.g., accounting for clusters in modeling). 
    • Proposals that do the following score higher:
      • specify the statistical procedures that will be used to test the hypotheses (e.g., regression analyses, growth curve modeling, path analyses),
      • explain the proposed statistical procedures using the variables that will be included and in terms that educated individuals not familiar with the analytic methods can understand, and
      • articulate the rationale for selecting the analytic methods (providing references as appropriate). 
  • Feasibility of proposed evaluation
    • Scoring of this criterion is based on the extent to which compelling evidence is provided to support that the proposed project can be completed as planned. 
    • Proposals that do the following score higher:
      • identify factors that increase the likelihood that the project will be completed as promised,
      • identify specific potential challenges that may be faced in attempting to conduct the project at each stage of the work, and 
      • articulate action plans for addressing potential challenges should they arise. 
    • Note: This is a different question than the one concerning feasibility of the program.
      • Feasibility of the program (see above) concerns the feasibility of implementing the program itself (regardless of any evaluation activities that would be conducted as part of this project, what is the evidence that the program can be implemented in the “real world”?).  
      • Feasibility of the proposed evaluation concerns the likelihood that the proposed evaluation project can be completed and provide meaningful data and findings that can then be disseminated to inform practice, funders, and/or policy.
  • Research environment and team: Scoring of these criteria are based on the extent to which the PI’s research environment can support the proposed project and the evidence provided that the evaluation team has a demonstrated record that provides confidence in their ability to do the project.
    • Research environment
      • Scoring for this criterion is based on the strength of the evidence provided concerning the level of support the PI’s institution is able to provide for the proposed project.
      • Considered is the research ranking of the institution in general (similar to the Carnegie Classification of Institutions of Higher Education in which universities are classified by research activity as measured by research expenditures, number of research doctorates awarded, number of research-focused faculty).
      • Resources that would be provided by the institution to support the evaluation team while carrying out the proposed project are also considered (e.g., administrative support such as grant support services, office and lab space, personal computers and equipment, software, and/or technological support).
      • Proposals that provide evidence of a high-quality research environment and identify specific resources provided by the institution to support proposed work score higher on this criterion.
      • Section 5, question 1.
    • PI / research team
      • Scoring for this criterion is based on the strength of the evidence provided (e.g., record of publications, prior grants) that provides confidence in the ability of the PI / Research Team to do the proposed project.
      • Considered when scoring this criterion are:
        • evidence of expertise concerning the topic area of the proposed project (e.g., the target population, the intended goals of the program),
        • evidence of expertise concerning the research methods that would be used in the proposed project, and
        • the level to which the intended procedures have been previously pilot tested by the evaluation team with the intended target population.
      • Proposals that provide evidence that at least one member of the team (not necessarily the PI) has experience successfully leading projects of similar or greater scope score higher on this criterion.
      • Also considered is whether at least one member of the leadership team (PI / co-PI level) is representative of the populations included in the study sample.
      • Section 5, questions 2 and 3; Section 7, CVs / Resumes of key personnel.

4. Budget criteria

Scoring for this criterion is based on the extent to which the proposed budget is in line with specific aim(s) / scope of work proposed and is reasonable and justified.

  • Considered are
    • the total budget amount requested relative to the funding capacity of BEF,
    • whether other support for the project has been secured and/or is pending,
    • whether all activities proposed are represented in the budget,
    • whether unjustified costs are included,
    • the extent to which the FTE percent requested for each key personnel is reasonable given his/her role and responsibilities (i.e., is neither too high nor too low), and
    • the extent to which estimates for supplies, equipment and other costs (e.g., incentives for participants, costs for assessment materials) are reasonable given the scope of work proposed.
  • Also considered is whether operational funding for the program to be evaluated has been secured for the period during which the evaluation would occur.
    • Proposals that demonstrate that operational funding for the program has been secured from other sources (or very likely will be secured given past funding history) throughout the program evaluation period score higher on this criterion. 
    • Further, all funds requested from BEF must be for evaluation costs (i.e., no program operation costs should be included in the budget). 
    • If the proposal includes a subcontract to the organization that operates the program, very clear articulation of the evaluation activities that would be supported by the subcontract should be provided. 
  • Proposals that are able to identify clear links between each cost and the specific tasks needed to complete the proposed work score higher.
  • Section 1, questions 5-9; Section 6; Section 7, budget justification.