Support Planning and Reviews in Person-Centred Care: Turning Conversations into Measurable Change
Share
🗂️ Support Planning and Reviews in Person-Centred Care: Turning Conversations into Measurable Change
A support plan isn’t a document — it’s a shared map of what matters, what’s working, and what’s next. When planning and reviews are genuinely person-centred, they turn everyday conversations into measurable change. This guide shows how to make planning and reviews consistent, meaningful and evidence-based across all service types — from learning disability and mental health to reablement, older people’s and complex care.
If you’re strengthening your support planning frameworks, we can help you translate values into inspection-ready routines through Proofreading & Compliance Checks. You can also download editable, person-centred templates via our Editable Method Statements and Editable Strategies. For sector-specific builds, see Learning Disability, Home Care and Complex Care.
🎯 Why Support Planning Still Defines Quality
Every CQC framework — from Key Question 1 to the new single-assessment model — hinges on how well providers plan and review support. Planning is the practical test of dignity, autonomy and control. Reviews are the test of learning and improvement.
- Planning = turning what matters to the person into clear, deliverable actions.
- Review = checking whether those actions worked — and proving change through data and narrative.
When both work together, care feels consistent and people experience progress, not maintenance.
🧭 Core Principles of Person-Centred Planning
- Co-production: The plan belongs to the person — it uses their words, priorities and language.
- Outcomes not tasks: Describe what success looks like, not what staff do.
- Strengths-based: Build on what the person can do, not on deficits.
- Accessible: Plans must be understandable to the person — easy-read, pictorial or audio where needed.
- Dynamic: Treat the plan as live — updated after every review, incident, or new goal.
🧩 What a Strong Support Plan Looks Like
High-scoring or inspection-ready plans show clear logic from the person’s voice to measurable actions:
- 💬 “What matters to me” — the person’s priorities in their words.
- 🎯 “My goals” — short-term, achievable, phrased positively.
- 🧠 “How you can support me” — specific prompts, frequency, and responsible staff.
- 📅 “Review date” — always current and linked to outcomes.
- 📈 “How we’ll know it’s working” — qualitative and quantitative evidence (feedback, observation, data).
Example: “I want to cook my own evening meal twice a week. Staff will provide step-by-step support using my visual recipe cards. We’ll review in four weeks. Success = I cook independently using prompts less than once per meal.”
📋 Turning Conversations into Evidence
Person-centred planning meetings should feel informal but yield formal evidence. Each conversation becomes data when you log what was discussed, what changed and what worked.
- 🗣️ Record people’s own phrases — direct quotes add authenticity.
- 🧾 Note decisions, dates and responsibilities in plain English.
- 🔁 Track follow-ups in the same record — continuous story, not scattered notes.
- 🕵️♀️ Sample monthly: managers audit two plans per staff member to ensure follow-through.
💬 Co-Production That Feels Real
Families, advocates and staff all contribute — but the person leads. Practical structures help balance voices:
- Invite chosen participants; clarify who will speak when.
- Use easy-read agendas and short meetings.
- Include photos or videos if that helps express progress.
- Summarise actions verbally and in writing within five days.
Example assurance line: “All support plans co-produced and signed by the person (or advocate). Summaries issued within five days; review compliance 97 %.”
🧠 Embedding Reviews into Daily Rhythm
Planning is only meaningful if reviews close the loop. Treat reviews as part of your service rhythm, not as paperwork deadlines:
- Monthly mini-reviews (15–30 mins): check what’s working; record one measurable change.
- Quarterly full reviews (60 mins): reset outcomes; update risk and health data; capture feedback.
- Ad-hoc updates after incidents or achievements: confirm learning and next step.
For regulated services, align this rhythm with supervision and governance calendars so data and reflection flow both ways.
📊 Measuring Progress — The “So What?” Test
Every review should answer: so what changed?
- 🧩 Independence: prompts reduced ? new skills gained ?
- 💬 Confidence: person’s feedback score or qualitative quote.
- 📈 Safety: fewer incidents, improved attendance or engagement.
- 🤝 Relationships: family/advocate feedback or involvement level.
Example: “Prompt frequency halved over eight weeks; person now initiates activity unprompted twice weekly; confidence self-score 2 → 4.”
📘 Before / After — Make Tender Lines Score
Before: “We review care plans regularly.”
After: “All support plans reviewed monthly and co-signed by the person or advocate; actions updated within five working days; 92 % of plans show measurable outcome change in last quarter.”
Before: “Support is person-centred.”
After: “Each plan begins with ‘What matters to me’ in the person’s words, linked to two measurable outcomes; audits verify voice present in 100 % of current plans.”
🛠️ Audit and Assurance Framework
- Quarterly audit = 10 % of plans per service or min 10 records.
- Check: person’s voice, outcome clarity, review timeliness, follow-through.
- Rate 0–2 and plot improvement trend by service.
- Governance minutes to capture learning actions and NI sign-off.
🏗️ Linking Planning to Risk and Enablement
Person-centred plans balance choice and safety through positive risk-taking. Reviews should always ask whether restrictions still fit the least-restrictive principle.
- Record how risks are managed collaboratively (“we agreed I’d travel alone with a phone check-in”).
- Track reductions in restrictions as progress metrics.
- Include advocates for complex decisions or capacity assessments.
🧱 Digital Planning & Reviews
Digital tools can enhance visibility and consistency — provided they stay person-centred:
- Use secure, cloud-based systems compliant with DSPT “Standards Met”.
- Give the person controlled access (view or comment) where feasible.
- Auto-track deadlines, reviews and outcome metrics.
Metric: “Digital planning platform introduced Q2; 94 % of reviews completed on time; feedback positive for transparency.”
💡 How to Evidence “Lived Planning” During Inspection
- Show one person’s plan, their quotes, the data trend, and how it changed their life.
- Cross-reference with supervision notes and governance dashboards.
- Explain your review rhythm and escalation process confidently — it proves control and learning culture.
🧩 Real-World Examples (Cross-Service)
- Reablement: “Goal – walk to the shop unaided. Outcome: achieved within 3 weeks; therapy discharged early; independence sustained 12 weeks.”
- Learning Disability: “Goal – prepare breakfast independently. Outcome: visual schedule used; prompts reduced 60 %; family reports pride and routine.”
- Mental Health Support: “Goal – manage anxiety before appointments. Outcome: grounding script recorded; avoided cancellations 3 months running.”
- Older People’s Care: “Goal – reconnect with church group. Outcome: volunteer escort arranged; attendance weekly; improved wellbeing score.”
- Complex Care: “Goal – safely self-administer medication. Outcome: graded exposure; double-sign checks → self-med stage 3 approved.”
📈 Governance & Board Visibility
- Monthly service dashboard: review timeliness, outcome achievement rate, overdue actions.
- NI samples two plans per service/quarter; verifies co-production and measurable progress.
- Learning shared via a “what changed” report; drives training refresh if themes recur.
🧮 Self-Score Grid (0–2; target ≥ 17/20)
| Dimension | 0 | 1 | 2 |
|---|---|---|---|
| Person’s voice | Absent | Quoted once | Present + linked to goals |
| Goals measurable | None | Partially | SMART + dated |
| Review cadence | Ad-hoc | Quarterly | Monthly + quarterly deep-dive |
| Outcome evidence | Generic | Qualitative | Qual + quant verified |
| Advocacy/family input | None | Occasional | Documented + timely |
| Risk review | Static | Annual | Dynamic least-restrictive |
| Audit | Absent | Annually | Quarterly + trend |
| Digital access | None | Staff only | Shared + secure |
| Governance reporting | Verbal | Spreadsheet | Dashboard + NI sample |
| Training/reflection | Once | Annual | Annual + supervision link |
🧰 30-Minute Uplift Checklist
- Add “What matters to me” and “How we’ll know it’s working” boxes to every plan.
- Introduce monthly mini-reviews with one measurable change logged.
- Audit five random plans this week for outcome clarity – give feedback in supervision.
- Set a five-day SLA for updating plans after reviews.
- Publish a service-level dashboard showing % reviews on time and % outcomes achieved.
🚀 Key Takeaways
- 🗂️ Support plans are living documents — update them whenever life moves.
- 🧭 Co-production, measurable outcomes and fast follow-up prove quality.
- 📈 Reviews are where learning happens; treat them as part of your governance rhythm.
- 📋 Evidence with small, dated metrics and direct quotes — they score and inspire.
- 👥 Link planning to enablement, safety and family involvement for full assurance.
Ready to make your support plans inspection-ready? We’ll help you embed measurable, person-centred planning through Proofreading & Compliance, or supply ready-to-edit Method Statements and Strategies. For full bid builds or service mobilisation, see Learning Disability, Home Care and Complex Care.