JobDescription.org

Public Sector

Program Evaluator

Last updated

Program Evaluators design and conduct systematic assessments of government programs, policies, and interventions to determine whether they are achieving intended outcomes and how efficiently they use public resources. Working in federal agencies, state governments, nonprofits, and consulting firms, they translate evaluation findings into actionable recommendations that inform budget decisions, program redesigns, and legislative oversight. The role sits at the intersection of social science methods, public policy, and performance accountability.

Role at a glance

Typical education
Master's degree in public policy, public health, or related field; PhD preferred for senior roles
Typical experience
5-7 years for mid-career or Bachelor's + experience
Key certifications
None typically required
Top employer types
Federal agencies, research and evaluation centers, management consulting firms, state government units, philanthropic organizations
Growth outlook
Steady growth driven by the Evidence Act of 2018 and increased state/philanthropic demand
AI impact (through 2030)
Augmentation — AI accelerates routine tasks like literature searches and qualitative coding, increasing productivity expectations while requiring evaluators to maintain high-level methodological rigor.

Duties and responsibilities

  • Design evaluation frameworks — logic models, theory of change, and evaluation questions — aligned to program goals and stakeholder needs
  • Select appropriate evaluation methods: randomized controlled trials, quasi-experimental designs, or qualitative case studies based on feasibility and purpose
  • Develop data collection instruments including surveys, interview protocols, and administrative data extraction specifications
  • Manage data collection from program records, federal administrative databases, grantee reports, and primary fieldwork
  • Conduct quantitative analysis using regression, difference-in-differences, or propensity score matching to isolate program effects
  • Perform qualitative data analysis through coding, thematic synthesis, and key informant interview summaries
  • Write evaluation reports synthesizing findings, conclusions, and recommendations for program managers and policymakers
  • Present findings to agency leadership, legislative staff, and OMB reviewers in briefings and formal hearing testimony
  • Coordinate with program staff, grantees, and contracted research partners to ensure data quality and timely deliverable completion
  • Ensure evaluation activities comply with OMB Statistical Policy Directives, IRB requirements, and agency data privacy protocols

Overview

Program Evaluators answer one fundamental question on behalf of taxpayers and policymakers: is this program actually working? That sounds simple until you try to do it rigorously — defining what "working" means for a workforce development program serving 40,000 participants across 15 states, or determining whether a juvenile justice intervention reduces recidivism when you can't randomly assign kids to treatment and control conditions.

The job starts well before any data is collected. A good evaluation begins with a clear articulation of the program's theory of change — what inputs are supposed to produce what activities, what activities are supposed to produce what outcomes, and over what time horizon. Getting that logic model right requires working closely with program managers who often haven't had to make their assumptions explicit before. That negotiation, done well, is some of the most valuable work evaluators do.

Once the evaluation design is set, the work becomes methodological. Evaluators specify data needs, navigate the bureaucratic process of getting access to administrative records (which can take months at federal agencies), design surveys that will actually get responses from grantees who are already drowning in reporting requirements, and make defensible decisions about comparison groups when random assignment isn't feasible.

Analysis follows data collection, and the outputs range from descriptive statistics and simple cross-tabulations in formative work to difference-in-differences models and instrumental variable approaches in impact evaluations where causal inference is at stake. The analysis only matters if it gets communicated effectively — and translating regression output into a finding a congressional staffer will act on is a distinct skill that takes years to develop.

The institutional context matters too. A GAO evaluator operates under different constraints than an evaluator at a state health department or a management consulting firm doing OMB PART work. Federal evaluators navigate agency politics, data access bureaucracy, and formal review processes. State government evaluators often work with smaller teams and less data infrastructure but closer proximity to programs. Consulting evaluators work on faster timelines with more varied methodological demands.

Qualifications

Education:

  • Master's degree in public policy (MPA/MPP), public health (MPH), social work research, applied economics, or a related field — the standard credential for mid-career federal and state positions
  • PhD preferred for research director and senior fellow roles at federal agencies, think tanks, and evaluation research centers
  • Bachelor's degree plus 5–7 years of directly relevant evaluation experience accepted at some state agencies and nonprofits

Methods competency:

  • Quantitative: regression analysis, difference-in-differences, propensity score matching, interrupted time series — at minimum one causal inference method at publication-ready proficiency
  • Qualitative: semi-structured interview design, thematic coding, key informant analysis, document review
  • Survey design: questionnaire development, sampling, nonresponse bias assessment
  • Statistical software: Stata, R, or Python for quantitative work; NVivo or Atlas.ti for qualitative coding

Institutional knowledge:

  • OMB Circular A-11 requirements for agency performance management
  • Evidence Act (Foundations for Evidence-Based Policymaking Act of 2018) — learning agendas, evidence tiers, and evaluation officer roles
  • IRB protocols and informed consent requirements for human subjects research
  • Federal data governance: data use agreements, FISMA compliance for research data systems

Preferred experience backgrounds:

  • GAO, agency Inspector General, or Congressional Research Service
  • Federally-funded research and evaluation centers (Mathematica, Abt Associates, MDRC, Urban Institute, RAND)
  • State legislative fiscal offices or state evaluation units
  • University research institutes with policy evaluation portfolios

Soft skills:

  • Ability to write clearly for non-technical audiences without distorting methodological nuance
  • Diplomatic persistence in obtaining data access from reluctant program offices
  • Comfort delivering findings that contradict what program sponsors hoped to hear

Career outlook

Demand for rigorous program evaluation has grown steadily since the passage of the Foundations for Evidence-Based Policymaking Act of 2018. The Evidence Act required federal agencies to designate Evaluation Officers, publish learning agendas, and build evaluation capacity — creating a structural hiring driver that didn't exist before. Most CFO Act agencies now have dedicated evaluation units that are still staffing up.

Budget pressure is a double-edged force. In tight fiscal environments, oversight bodies and appropriators want evidence that programs are delivering value — which supports evaluation demand. At the same time, evaluation budgets are discretionary and can be cut when agencies need to trim. The net effect is that evaluation work funded by statute (as in many HHS and Education programs) is more durable than evaluation funded through agency administrative accounts.

The evidence ecosystem beyond the federal government is also expanding. States have significantly increased evaluation capacity under federal matching incentives in Medicaid, SNAP, and child welfare. Philanthropic organizations — particularly those focused on economic mobility, criminal justice reform, and K–12 education — are major funders of independent evaluation. The result is a labor market with meaningful demand from government, nonprofits, and consulting firms simultaneously.

Automation is reshaping the work without eliminating it. Tasks that once consumed evaluation staff time — literature searches, qualitative coding passes, boilerplate report sections — are being accelerated by AI tools. This is compressing timelines and allowing evaluators to take on more projects, but it is also raising expectations for productivity. Evaluators who embrace these tools while maintaining methodological rigor are well-positioned; those who treat methods as the only value they offer will find the role harder to justify at its current cost.

For evaluators with strong causal inference skills and familiarity with administrative data systems, the job market in 2026 is favorable. The career ladder runs from analyst to senior evaluator to evaluation director or Chief Evaluation Officer — a role that now exists by statute at every major federal agency. Consulting paths lead to principal and director-level positions at firms that hold large federal evaluation contracts, where total compensation can substantially exceed the GS pay scale.

Sample cover letter

Dear Hiring Manager,

I'm applying for the Program Evaluator position at [Agency/Organization]. I have six years of experience designing and conducting evaluations of federally-funded social programs, most recently as a senior analyst at [Firm/Agency] where I led a multi-site impact evaluation of a DOL workforce development demonstration.

That evaluation used a difference-in-differences design with propensity score matching to construct a comparison group from state unemployment insurance wage records — a methodological choice we made after a randomized design was ruled out mid-project due to program expansion. Managing that redesign while keeping the evaluation credible for OMB review was the hardest problem-solving I've done, and it gave me strong working knowledge of administrative data linkage and the data use agreement process across three states.

On the qualitative side, I managed a portfolio of 60 site visit interviews with program operators and participants, supervised coding in NVivo, and wrote the implementation findings chapter that ended up driving the program's mid-course corrections more than the impact findings did. That experience shaped how I think about what different methods are actually useful for — impact estimates tell you whether something worked, implementation analysis tells you why and for whom.

I've followed [Agency]'s learning agenda closely and I think the evaluation gap around [specific program area] you identified in the FY2025 agenda is one I could contribute to directly. I'd welcome the opportunity to discuss how my background fits what your evaluation team is building.

[Your Name]

Frequently asked questions

What is the difference between program evaluation and performance monitoring?
Performance monitoring tracks ongoing metrics against targets — things like number of participants served or cost per output — and is typically managed by program staff. Program evaluation is a structured, episodic inquiry into whether a program is producing its intended outcomes, for whom, and why. Evaluation answers causal or explanatory questions that routine metrics cannot.
What credentials or certifications matter for Program Evaluators?
The American Evaluation Association (AEA) does not administer a single required license, but the Certified Government Evaluator (CGE) credential from the Defense Acquisition University matters in federal defense contexts. Graduate training in public policy, public health, social work research, or applied statistics is the standard academic credential. Demonstrated methods experience — especially causal inference or mixed-methods design — matters more in hiring than any specific certification.
How is AI changing the work of Program Evaluators?
Large language models are accelerating literature reviews, qualitative coding, and draft report writing, which reduces time on tasks evaluators previously spent days on. The more significant shift is in administrative data linkage: agencies are increasingly connecting program records across systems, creating richer evaluation datasets than were previously feasible. Evaluators who can specify and validate AI-assisted data pipelines while maintaining rigor in causal claims are becoming more productive than those relying solely on traditional methods.
Do Program Evaluators need a security clearance?
Many federal agency positions — particularly at DOD, DHS, and intelligence community offices — require at minimum a Secret clearance, and some positions require Top Secret/SCI. GAO and IG offices often require baseline suitability investigations. Evaluators at civilian domestic agencies like HHS or Education typically do not need clearances, though access to restricted administrative data may require data use agreements and background checks.
What is the difference between an internal evaluator and an independent evaluator?
Internal evaluators are agency employees who assess programs their own organization runs, which provides deep institutional knowledge but raises independence questions. Independent evaluators — typically from GAO, Inspectors General offices, or external contractors — have more credibility with oversight bodies but less access to program staff and internal documents. Most large federal programs use both: internal staff for formative work and external evaluators for high-stakes summative assessments.
See all Public Sector jobs →