SOCW 6311 Week 10 Assignment: How to Write a Complete Outcome Evaluation Plan
A section-by-section guide to building the 3–4 page outcome evaluation plan — how to write each required component, choose and justify your group research design, identify stakeholders, select valid instruments, and describe data collection and analysis with peer-reviewed APA support.
The SOCW 6311 Week 10 outcome evaluation plan is the capstone assignment of the course — a 3–4 page document that draws on every prior week’s work, including the program you proposed in Week 4 and the process evaluation plan from Week 9. Students consistently lose marks on this assignment because they describe their research design without justifying it, list instruments without explaining why they are appropriate, treat the stakeholder section as a formality rather than an analytical task, or write a data analysis plan so vague it cannot be evaluated. This guide walks through every required component in sequence, explains what each one actually demands, and identifies the rubric-level distinctions between a submission that meets expectations and one that exceeds them.
This guide explains how to approach each section of the plan. It does not write the plan for you. The program, outcomes, and evaluation decisions must be grounded in your actual Week 4 proposal and your specific program context. Use this guide to understand what each required component requires structurally, how to justify your design and instrument choices with peer-reviewed support, and where most submissions fall short of the rubric’s “Meets Expectation” threshold.
What This Guide Covers
What the Assignment Requires
The assignment produces a single 3–4 page plan — not a reflection, not a literature review, and not a repetition of the Week 4 proposal. It is a prospective planning document: you are designing a complete evaluation that could, in principle, be implemented. Every section must be specific enough to be usable, and every major decision (design choice, instrument selection, data analysis approach) must be justified through reference to the course Learning Resources or peer-reviewed scholarly sources.
The Rubric: Where Points Come From
The rubric allocates points unevenly. Understanding the weight of each criterion before you write tells you where to invest the most depth and specificity.
The group research design (14 pts) and data collection/analysis (14 pts) sections together account for 40% of the assignment grade. Both rubric rows distinguish between meeting expectations and exceeding them based on whether your choices are “fully justified through reference to the Learning Resources or other peer-reviewed research.” A design choice stated without citation — even if it is the correct choice — does not meet the rubric’s expectations. Every design and method decision needs a scholarly source attached to it.
Before You Write: Connect to Prior Work
The assignment instructions are explicit: recall the program you proposed in Week 4 and review your process evaluation plan from Week 9. The Week 10 plan is not built from scratch — it is built on top of what you have already developed. Before writing a single sentence of the outcome evaluation plan, pull both documents and note the following from each.
From Your Week 4 Program Proposal
- The program name, target population, and service setting
- The client need the program addresses, with supporting data
- The program’s stated goals — these become your outcomes to be evaluated
- The theoretical or evidence base you used to justify the program
- The activities and services the program delivers
- Any eligibility criteria or intake procedures that define your sample
From Your Week 9 Process Evaluation Plan
- What process indicators you identified — these complement but differ from outcome indicators
- The stakeholders you identified for the process evaluation — your Week 10 list may expand on these
- Data collection procedures you already planned — avoid duplication, build on existing infrastructure
- Any design constraints your process evaluation identified (access issues, sample size concerns, attrition risk)
The Critical Connection: Goals Become Outcomes
The goals stated in your Week 4 proposal are the direct source of your outcomes to be evaluated in Week 10. If your program proposed to reduce housing instability among recently released individuals, the outcome to be evaluated is a measurable reduction in housing instability — defined precisely enough that an instrument can measure it. If your program proposed to increase parenting self-efficacy among first-time parents in a home visiting program, that is the outcome, and you need a validated instrument that measures parenting self-efficacy specifically. Before you write the Week 10 plan, convert each Week 4 goal into a measurable, specific outcome statement.
Program Outline Section
The program outline is brief — the assignment says so explicitly. This section is not a re-submission of the Week 4 proposal. It is a condensed orientation for a reader who has not read your previous work. Aim for one tightly written paragraph that covers the essential facts a reader needs to understand everything that follows.
Target Population
Who the program serves — age group, circumstances, needs profile, eligibility criteria. Be specific: “low-income adults experiencing food insecurity” is more useful than “community members.”
Services Provided
What the program actually does — the specific interventions, activities, and services delivered. One to three sentences covering the core program model or approach.
Intended Goals
What the program is designed to achieve for clients — stated as goals here, then converted to specific measurable outcomes in the next section. Should connect directly to the client need.
The program outline should take up no more than a quarter of a page. Including multiple paragraphs of program background, literature review content, or needs assessment detail from the Week 4 proposal uses page space that belongs to the evaluation planning sections — and those sections carry the rubric points. Brief means brief: summarize the program in enough detail that the evaluation plan is intelligible, then move to the evaluation components.
Evaluation Purpose and Outcomes to Be Evaluated
The purpose of the evaluation answers the question: why conduct this evaluation, and what will the results be used for? In social work program evaluation, outcome evaluations typically serve one of several functions — accountability to funders, evidence for program continuation or expansion, evidence for program modification, or contribution to the knowledge base on effective interventions with a specific population. Your purpose statement should be specific to your program and its evaluation context.
[Vague — Does Not Meet Expectations] The purpose of this evaluation is to determine whether the program is effective. The results will be used to improve the program and help clients.
[Specific — Direction to Aim For] The purpose of this outcome evaluation is to determine whether participation in the 12-week cognitive-behavioral parenting skills program produces measurable improvements in parenting self-efficacy and reductions in harsh discipline behaviors among first-time parents referred through child protective services. The evaluation results will be used to report outcomes to the county Department of Children and Family Services, which funds the program, and to determine whether the program model warrants replication at additional sites serving the same population.
The specific version names the program, the timeframe, the outcomes, the population, and the two uses of the results. A reader knows exactly what question this evaluation answers and what happens with the answer.
The outcomes to be evaluated are the specific, measurable changes the program is designed to produce in clients. Each outcome must meet three criteria: it must be measurable (an instrument exists to capture it), it must be attributable to program participation, and it must connect directly to the program’s stated goals from Week 4. Write each outcome as a clear statement — not a goal (“improve mental health”) but a measurable endpoint (“a statistically significant reduction in PHQ-9 depression scores from pre- to post-program among adult clients completing at least 8 of 12 group sessions”).
Group Research Design: Choosing and Justifying
The group research design section carries the most rubric points of any content criterion (14 points) and is the section where the distinction between meeting and exceeding expectations most directly depends on the quality of your justification. You must name a specific design, explain its structure, and provide a peer-reviewed rationale for why it is appropriate for your evaluation context — not just describe what it is.
Group research designs in program evaluation fall into three broad categories — experimental (randomized controlled), quasi-experimental (non-randomized comparison), and pre-experimental (one group, before-and-after or post-only). The choice among them is driven by what is feasible in your program context: your sample size, whether random assignment is ethically and practically possible, whether a comparison group is accessible, and the strength of causal claim your evaluation needs to support.
Comparing the Major Design Options
| Design | Structure | When to Choose It | Key Limitation |
|---|---|---|---|
| One-group pretest–posttest (pre-experimental) | Measure the same group before and after the program. No control or comparison group. | When a comparison group is not feasible, the sample is small, or the evaluation priority is demonstrating change rather than attributing it. Most common in community-based social service programs. | Cannot rule out alternative explanations for change — maturation, historical events, or regression to the mean. Weak causal inference. |
| Non-equivalent comparison group (quasi-experimental) | Measure both a program group and a similar group receiving no intervention (or a different intervention), pre and post. Groups are not randomly assigned. | When a naturally occurring comparison group exists — a waitlist, a site that does not yet offer the program, or clients who declined participation. Stronger causal inference than pre-experimental. | Groups may differ on unmeasured variables that explain outcome differences. Selection bias remains a threat to validity. |
| Randomized controlled trial (experimental) | Random assignment of participants to program or control condition. Measures both groups before and after. | When random assignment is ethically feasible, the sample is large enough, and the evaluation must support a strong causal claim. Typically used in efficacy trials of manualized interventions. | Ethically complex when withholding effective services. Logistically demanding. Often not feasible in standard social service program contexts. |
| Time-series design (quasi-experimental) | Multiple measurements before and after the program to track trends. Can include a comparison series from a similar group. | When program delivery is ongoing and repeated measurement is feasible. Strong for detecting sustained change over time. | Requires consistent measurement over an extended period. Sensitive to instrumentation changes and attrition. |
How to Write the Justification
After naming your design, the justification must do three things: explain why this design fits your specific program (sample size, population characteristics, feasibility of comparison group), acknowledge what alternative designs you considered and why you did not choose them, and cite at least one peer-reviewed source that discusses this design’s appropriateness in social work program evaluation contexts. A justification that says only “I chose a one-group pretest–posttest design because it is the most common design in social work” does not meet the rubric. One that says “A one-group pretest–posttest design was selected because random assignment would be ethically problematic given the vulnerability of the population and the lack of a clear waitlist from which to draw a comparison group; this design, while limited in causal inference, is appropriate for early-stage program evaluation aimed at demonstrating feasibility and magnitude of change (Royse et al., 2016)” is moving toward meeting expectations.
Key Stakeholders and Their Potential Concerns
The stakeholder section requires you to identify who has a stake in the evaluation — not just in the program — and to describe at least two specific concerns each stakeholder group is likely to bring. The rubric’s “Exceeds Expectation” level requires more than two concerns and in-depth analysis of how to address them. This section rewards specificity and critical thinking, not a generic list.
Common Stakeholder Groups in Social Work Program Evaluation
- Program funders (government agencies, foundations, private donors) — Primary concern: Are program resources producing measurable client outcomes? Secondary concern: Is the program reaching the intended population at the projected scale? They want efficiency evidence, not just effectiveness evidence.
- Program clients and client communities — Concern about confidentiality of their data, cultural appropriateness of the evaluation instruments, and whether results will be used in ways that benefit them or affect their services. Often underrepresented in evaluation planning despite being the program’s reason for existing.
- Program staff and administrators — Concerned that evaluation findings will be used punitively — to reduce funding, eliminate positions, or challenge their professional judgments. Also concerned about the data collection burden placed on them and on clients during service delivery.
- Oversight bodies and regulatory agencies — Concerned with compliance: does the program meet legal, ethical, and accreditation standards? Is data collection conducted in accordance with human subjects protections?
- Partner organizations and referral sources — Concerned about whether the evaluation findings will reflect on their referral decisions and organizational reputations. May be concerned about data sharing and attribution of outcomes across program components.
Potential concerns are not complaints or objections to the evaluation. They are the legitimate interests, questions, and risks that each stakeholder group brings to the evaluation context — concerns that a well-designed evaluation should anticipate and address. A funder’s concern about whether the sample size is sufficient to detect meaningful effects is a legitimate concern that affects evaluation design decisions. A client community’s concern about whether survey instruments have been validated with culturally similar populations is a legitimate concern that affects instrument selection. Your stakeholder analysis should name these concerns and indicate, at least briefly, how your evaluation plan addresses them.
Indicators and Instruments to Measure Outcomes
The instrument selection section requires you to identify what you will use to measure each outcome and to justify why each instrument is appropriate for your population and purpose. The rubric’s “Meets Expectation” level requires choices to be “fully justified through reference to the Learning Resources or other peer-reviewed research.” That means citing scholarly sources that document the instrument’s validity, reliability, and appropriateness with your target population — not simply naming the instrument and moving on.
Types of Instruments Used in Social Work Outcome Evaluation
- Standardized scales and assessments — Psychometrically validated tools that measure constructs like depression (PHQ-9), anxiety (GAD-7), self-efficacy (General Self-Efficacy Scale), functioning (Global Assessment of Functioning), or family strengths. These are the most defensible instrument choices because published literature documents their validity and reliability.
- Structured interviews and surveys — Researcher-developed or program-developed tools used when no validated scale exists for a specific outcome. Require documentation of how content validity was established.
- Administrative and service records — Case notes, attendance records, housing placements, employment records, and child welfare case outcomes. Useful for behavioral outcomes that can be objectively tracked without self-report.
- Observational protocols — Structured observation of client behavior or functioning, useful for parenting programs, early childhood programs, and skill-based interventions.
How to Justify Each Instrument Choice
- Name the instrument and what it measures — be specific about the construct, the number of items, and the response scale
- State whether the instrument has been validated with a population similar to yours — cite the validation study
- Note whether the instrument has published reliability data (Cronbach’s alpha, test-retest reliability)
- Address whether the instrument is available in languages spoken by your client population if applicable
- State when you will administer it — pre-program, post-program, and/or at follow-up — and connect that timing to your research design
- Note any limitations of the instrument for your specific context
Every outcome you listed in the earlier section must have at least one corresponding instrument. If you identified two outcomes — for example, reduced depressive symptoms and improved social support — you need one instrument for each. A plan that identifies three outcomes but provides only one instrument has not completed the assignment. Map the section explicitly: state the outcome, then name the instrument you will use to measure it, then justify that choice. This makes the correspondence between outcomes and measurement clear to the evaluator.
Data Collection, Organization, and Analysis
The data methods section carries the same rubric weight as the research design section (14 points each). It requires you to specify who collects the data, how, when, and from whom — then explain how data will be organized and stored, and what analysis approach will be used to determine whether outcomes were achieved. The rubric distinguishes between “Meets Expectation” (specific details and examples, fully justified through peer-reviewed sources) and “Fair” (general details, vague justification). Every element needs specificity.
Data Collection: Who, How, When, and From Whom
Name the data collector — program staff, an external evaluator, or trained research assistants — and explain why that choice minimizes bias. Name the administration method — self-administered paper survey, digital platform, interviewer-administered — and connect it to your population’s characteristics (literacy level, access to technology, language). State specifically when instruments will be administered: at intake (pre-program), at program completion (post-program), and/or at a follow-up point. For each instrument, indicate the administration window — for example, “the PHQ-9 will be administered within the first two sessions and again within the final two sessions of the 12-week program.” Identify who is in your sample — the full program enrollment, a subset, or clients who meet specific completion criteria — and what your expected sample size is.
Data Organization and Management
Explain how collected data will be stored — paper forms, a secured electronic database, a program management system. Address who has access to the data, how client confidentiality will be protected (de-identification procedures, password protection, data sharing restrictions), and how long data will be retained. If you are using a comparison group, explain how you will track which participants are in which condition. This section is often treated as a formality — but the rubric’s “Meets Expectation” language requires “specific details and examples,” meaning a sentence like “data will be stored securely” does not meet the threshold. Specify the platform, the de-identification procedure, and the access control.
Data Analysis: Matching the Method to the Design
The analysis method must match your research design. A one-group pretest–posttest design analyzed with a paired-samples t-test or Wilcoxon signed-rank test (for non-normally distributed data) is a defensible and common approach in social work outcome evaluation. A quasi-experimental design with a comparison group requires an independent samples t-test, ANCOVA (to control for pre-test differences), or multilevel modeling if participants are nested within sites. If you have categorical outcomes (housed vs. not housed, employed vs. not employed), chi-square or logistic regression is appropriate. Name the specific statistic you will use, explain why it matches your design and outcome type, state what significance threshold you will apply (typically p < .05), and cite a source that supports this analysis approach. Also note who will conduct the analysis — the program evaluator, a contracted statistician, or the program director.
Interpreting and Reporting Results
Address how results will be interpreted: what constitutes a meaningful outcome (statistical significance and/or practical significance measured by effect size), and what will be reported to each stakeholder group. A results section that reports only p-values without effect sizes is increasingly seen as insufficient in social work research — note whether you will calculate and report Cohen’s d or eta-squared alongside significance tests. Indicate whether results will be shared through a formal evaluation report, a presentation to funders, or a feedback session with program staff — and how findings will be used to inform program decisions.
[Vague — Does Not Meet Rubric Expectations] Data will be analyzed using statistical methods to determine whether the program achieved its outcomes. Results will be reviewed by program staff and shared with stakeholders.
[Specific — Direction to Aim For] Because this evaluation uses a one-group pretest–posttest design and the PHQ-9 produces continuous interval-level data, a paired-samples t-test will be used to compare mean depression scores from intake to program completion. If the assumption of normality is violated, a Wilcoxon signed-rank test will be substituted. Effect size will be calculated using Cohen’s d to assess practical significance beyond statistical significance, with d ≥ 0.50 defined as a meaningful clinical effect consistent with benchmarks used in depression intervention research (Cuijpers et al., 2014). Analysis will be conducted by the program evaluator using SPSS. Results will be reported to the program director and the county funding agency in a structured evaluation report delivered 60 days after the final cohort completes the program.
The specific version names the statistic, explains why it matches the design and data type, identifies the effect size measure, states who conducts the analysis and with what software, sets a specific reporting timeline, and cites a peer-reviewed source. Each of those elements corresponds to a rubric distinction between meeting and not meeting expectations.
Using Learning Resources and APA Citations
Every rubric criterion for this assignment — except the program outline — includes language about justifying choices “through reference to the Learning Resources or other peer-reviewed research.” That requirement is not satisfied by a general reference list at the end of the paper. It requires in-text citations at the point where each major decision is explained: when you name your research design, when you identify each instrument, when you explain your analysis approach.
Course Learning Resources First
Your course text and weekly readings are the primary expected sources. The rubric specifically mentions Learning Resources before peer-reviewed research. Cite chapter pages for design descriptions, evaluation frameworks, and measurement guidance. Know what your course readings cover and cite them directly — do not replace them with outside sources.
Peer-Reviewed Articles for Instruments
For each instrument you select, find the original validation study or a major study using it with your target population. Databases: PsycINFO, Social Work Abstracts, PubMed. Search the instrument name plus your population. This citation is the evidence that the instrument is valid and appropriate — not just that you have heard of it.
APA 7th Edition Throughout
Title page, running head (check whether your program requires it), double spacing, 12-point Times New Roman or 11-point Calibri, one-inch margins, in-text citations for every paraphrased source, and a full reference list. The Writing rubric criterion covers all of this — and penalizes both missing citations and over-reliance on direct quotation.
Where Most Submissions Lose Marks
Research Design Named but Not Justified
“I will use a pre-experimental one-group pretest–posttest design.” Full stop. No explanation of why this design fits the program, no acknowledgment of alternatives, no citation. This satisfies the “Needs Improvement” row but does not reach “Meets Expectation,” which requires the choice to be “fully justified through reference to the Learning Resources or other peer-reviewed research.”
Instead
Name the design, explain why it fits your specific context (sample size, feasibility of control group, ethical constraints on random assignment), acknowledge what stronger design would have been used if feasible, and cite a peer-reviewed source supporting this design’s use in similar program evaluation contexts. That structure meets the rubric and positions you for the “Exceeds Expectation” tier.
Instruments Listed Without Validation Evidence
“I will use the PHQ-9 to measure depression and the GAD-7 to measure anxiety.” Naming widely used instruments without citing validation studies does not justify the choice — it assumes the reader already knows these instruments are valid for your population. The rubric requires justification, not recognition.
Instead
For each instrument, cite the validation study or a key peer-reviewed article documenting its psychometric properties with a population similar to yours. Identify the reliability coefficient (Cronbach’s alpha), the construct it measures, and any known limitations with your specific population (e.g., a scale normed on clinical populations being used with a community sample).
Stakeholder Concerns Listed at the Surface Level
“Funders will be concerned about whether the program is effective. Clients will be concerned about their privacy.” These are the minimum possible content — they could apply to any program evaluation anywhere. The rubric’s “Meets Expectation” requires full description of at least two concerns; “Exceeds Expectation” requires in-depth analysis and discussion of how to address them.
Instead
For each stakeholder group, name a concern that is specific to your program’s context and population. Explain what drives that concern (the funder’s reporting requirements, the population’s history of research exploitation, the staff’s prior experience with evaluations used punitively) and indicate how your evaluation plan addresses it. That depth is what separates adequate from strong performance on this criterion.
Data Analysis Section Uses Vague Language
“The data will be analyzed to determine whether clients improved.” No statistic named. No match to the research design. No significance threshold. No effect size measure. This does not meet the rubric — and the data methods section is worth 14 points.
Instead
Name the specific statistical test, explain why it matches your design and outcome type, state the significance threshold, name the effect size statistic, identify who conducts the analysis and with what software, and cite a source. Every one of those elements corresponds to language in the rubric that distinguishes adequate from strong performance.
Paper Falls Below Three Pages or Exceeds Four
A two-and-a-half-page submission almost certainly means one or more sections are underdeveloped — the research design section alone, done well, requires at least half a page of justified content. A five-page submission likely includes too much program background from Week 4 and not enough analytical depth in the evaluation components.
Instead
Before writing, allocate approximate page space to each section based on rubric weight. The program outline gets a quarter to a third of a page. The research design, stakeholder analysis, instruments, and data methods sections — the four high-point criteria — should together account for the bulk of the three to four pages. Write to the rubric’s depth requirements, not to fill or trim page count.
Direct Quotation Instead of Paraphrasing
The Writing rubric criterion explicitly addresses this: the paper should “appropriately paraphrase sources, using one or fewer quotes” to meet expectations, and penalizes reliance on multiple short or long direct quotations. A paper that strings together quoted passages from the textbook or course readings is not demonstrating comprehension — it is demonstrating avoidance of the intellectual work of synthesis.
Instead
Every source you use should be paraphrased — restated in your own words and cited. Direct quotation should be reserved for cases where the original wording is essential and cannot be restated without losing meaning. In a planning document of this type, there are almost no such cases. Read the source, close it, write what it says in your own words, then add the citation.
Submission Checklist
- File named correctly: WK10Assgn + last name + first initial (e.g., WK10AssgnjohnsonA)
- Paper is 3–4 pages of body content, double-spaced, APA 7th edition formatting
- Title page included with running head (if required by your program)
- Program outline present — brief, one paragraph, covering population, services, and goals
- Evaluation purpose stated specifically — names the question, the population, and the use of results
- Outcomes are specific and measurable — not goals, but endpoint statements an instrument can capture
- Research design named, structured, and justified with at least one peer-reviewed citation
- Alternative designs acknowledged and reasons for not choosing them stated
- At least two stakeholder groups identified, each with at least two specific concerns
- Each outcome has at least one named instrument with a validation citation
- Data collection procedures specify who collects, when, how, and from whom
- Data storage and confidentiality procedures described with specific detail
- Analysis method named, matched to design and outcome type, and justified with citation
- Effect size measure identified alongside significance test
- All sources paraphrased — one or fewer direct quotations
- In-text citations present for every claim requiring attribution
- Reference list complete, in APA 7th edition format, matching all in-text citations
- Draft submitted through Turnitin Drafts before final submission
Frequently Asked Questions
The Research Foundation: Why Outcome Evaluation Design Decisions Require Scholarly Justification
The assignment’s insistence on peer-reviewed justification for design and instrument choices is not a citation formality — it reflects how evidence-based program evaluation actually works in the social work field. The question of which research design produces credible evidence of program effectiveness is an empirical question with a documented literature, not a matter of preference. Designs vary in their ability to control for threats to internal validity — selection bias, history, maturation, testing effects — and the choice among them has direct implications for how confidently a funder, policymaker, or practitioner can attribute client outcomes to the program rather than to other factors.
The Campbell Collaboration, an international research network that produces systematic reviews of evidence in social policy and social welfare, evaluates intervention research according to the strength of the research design used to generate outcome evidence. Its evidence standards — widely cited in social work research — distinguish between randomized controlled trials (the strongest design for causal inference), quasi-experimental designs (moderate strength), and pre-experimental designs (weakest causal claim). Understanding where your chosen design sits within that hierarchy helps you write a justification that acknowledges the design’s limitations honestly while explaining why it is the appropriate choice for your evaluation context. That intellectual honesty is precisely what the rubric’s “Exceeds Expectation” level rewards.
Continue with: social work assignment help · research paper writing services · professional coursework writers · proofreading and editing services · proofread my paper.