19 th EPSO Conference – 16-17 April 2015 - Oslo The evaluation of the Care Quality Commission’s acute hospital regulatory model Rachael Addicott Alan Boyd.

Slides:



Advertisements
Similar presentations
Intelligence Step 5 - Capacity Analysis Capacity Analysis Without capacity, the most innovative and brilliant interventions will not be implemented, wont.
Advertisements

Improving outcomes for older people: Monitoring and regulating standards Ann Close 8 th June 2011.
Vanessa Pinfold and Terry Hammond Developing a carer strategy for the UK Mental Health Research Network.
CQC into the future Malcom Bower-Brown
Service User Survey 2011 Service User Survey Results 2011 Toni Martin – Senior Consultant Quality Health.
RNHA What CQC expect 16 October 2014 Nick Kerswell 1 1.
Service Users subject to s. 41 of the Mental Health Act Their views of risk and risk assessments Jeremy Dixon.
Formative Evaluation of the first 12 months of the PfPS Project in England & Wales Anna Allford, Project Manager, AvMA Formative Evaluation of the first.
About CQC Sarah Seaholme Ram Sooriah 1 1.
Telehealthcare: Integrating service development and evaluation Tracy Finch, Carl May, Frances Mair, Maggie Mort, Linda Gask Universities of Newcastle,
The Francis Report and its impact for care providers Professor Ian Peate © e-GNCS Limited All rights reserved. No part of this publication may be.
CQC’s new approach to inspecting and regulating GP and OOH providers
Perceptions of the Role of Feedback in Supporting 1 st Yr Learning Jon Scott, Ruth Bevan, Jo Badge & Alan Cann School of Biological Sciences.
Patient Experience: Why does it matter?
Regulating the dental sector Tracy Norton Compliance Manager (Central Region) 4 October 2012.
The New Inspection Framework The Multi agency arrangements for protecting children The multi-agency arrangements for the protection of children The multi-agency.
Measuring the Quality of Long-Term Care in England Juliette Malley Personal Social Services Research Unit LSE Health and Social Care London School of Economics.
Achieving Quality in a New Era Paul Dunnery Operations Director Alzheimer’s Society.
“Healthy Conversation Skills” training Dr Wendy Lawrence PhD CPsychol Senior Research Fellow MRC Lifecourse Epidemiology Unit The intervention Delivery.
Planning for challenge. Key issues addressed by the study This study explored:This study explored: –the extent to which pupils are challenged in their.
Effective Strategies for Reducing the Symptoms of ADHD in Children in Mainstream Education: Lessons from children, parents and teachers 1 Dr Lesley Hughes.
Teachers mentoring teachers: A process of reflection and rejuvenation
Theory & Practice – the new Common Inspection Framework and what it means to governors UCU-LSIS-UNISON FE Staff Governors’ Conference 3 December 2012 Lorna.
1 CQC – the next phase Alan Rosenbach Special Policy Lead.
Professional Standards Authority conference – 28 March 2014 Regulatory design: the role of programme theory in evaluating healthcare regulation Kieran.
Medical Audit.
Your Ambulance Service Foundation Trust Consultation.
Irene Khan – Secretary General Building effective and responsive INGOs, the strategic role of HR: The IS Job Value Review 8 February 2008.
4 th National Patient Safety Conference Friday 7 th November, 2014 Inspecting for improvement? Lessons from evaluating the Care Quality Commission's new.
1 CQC: The journey to excellence and The new approach to inspection of ambulance services Professor Sir Mike Richards Chief Inspector of Hospitals April.
CQC Thematic Activity Emma Steel Bernadette Hanney.
Helping to make care better Cynthia Bower, CEO National Care Association Conference 11 November 2009.
SECHA AGM 29 September Louise Bushell and Janet Ortega Compliance Managers CQC Regulation under the Health and Social Care Act 2008.
Force Results – August 2012 Sussex Police Employee Survey 2012.
AN INTRODUCTION Managing Change in Healthcare IT Implementations Sherrilynne Fuller, Center for Public Health Informatics School of Public Health, University.
Transforming Patient Experience: The essential guide
The New Ofsted Framework Pupil Achievement Quality of Leadership and Management Quality of Teaching Behaviour and Safety.
WATERFALL DEVELOPMENT MODEL. Waterfall model is LINEAR development lifecycle. This means each phase must be completed before moving onto the next!!! WHAT.
Kathy Corbiere Service Delivery and Performance Commission
CQC activity in Coventry Coventry Cares Learning Network 1 March 2013.
Induction toolkit 2. WHAT DOES MY TRUST LOOK LIKE? © GovernWell
Making it Count! Program Evaluation For Youth-Led Initiatives.
Improving Lives in Our Communities Leading through the CQC Inspection Process.
Registration Speaker Susan Robinson Job Title Area Manager
How to avoid a warning notice 4 December 2012 Jennifer Pattinson Compliance Manager.
1 CQC’s new approach to inspecting and regulating GP and OOH providers 29 April 2015.
1 1 Care Quality Commission Vicki Wells Head of General Practice The Patients' Voice Conference.
1 To help each team member feel more confident when the CQC Inspector comes to visit your practice © The Dental Business Academy
Inspection of General Practice Andy Brand Inspection Manager 1.
Assuring quality in health services for people with learning disabilities Dr Theresa Joyce CQC National Professional Advisor – Learning Disabilities.
Week 2: Interviews. Definition and Types  What is an interview? Conversation with a purpose  Types of interviews 1. Unstructured 2. Structured 3. Focus.
HEALTH AND CARE STANDARDS APRIL Background Ministerial commitment 2013 – Safe Care Compassionate Care Review “Doing Well Doing Better” Standards.
Registration and monitoring compliance Michele Golden Compliance Manager 2 November 2010.
……………………………………………………………………………. Chief Inspector of Hospitals visit Quality Summit 11 June 2015.
Preventing avoidable inpatient admissions: a qualitative study of mental health liaison nurse practice using the Think Aloud technique. Iain Hepworth Linda.
Raising standards improving lives The revised Learning and Skills Common Inspection Framework: AELP 2011.
CQC’s New Inspection Approach. Learning objectives: A reminder of who the CQC are and what they do To build up an understanding of the CQC’s new approach,
First Wave Findings: Jobseekers in Scotland. Background View shared by successive governments of need to tackle ‘welfare dependency’. Increased requirements.
CQC matters: Regulating the safe and effective use of medicines
Raising standards, putting people first
Vicky Blomfield, Msc Health Service
Our new quality framework and methodology:
CQC: The new approach to inspection
Public engagement strategy
So you’ve been inspected…. communicators driving improvement
Rosemary Smyth Interim Chief Executive Mental Health Commission
CORE 3: Unit 3 - Part D Change depends on…
Regulatory relationships: regulators, organisations and health professionals Kieran Walshe Professor of Health Policy and Management University.
Clerks’ Update 22 April 2013 The current HMCI and ministers are focussing on governance and its effectiveness in an unprecedented way. While legal responsibilities.
Presentation transcript:

19 th EPSO Conference – April Oslo The evaluation of the Care Quality Commission’s acute hospital regulatory model Rachael Addicott Alan Boyd Ruth Robertson Shilpa Ross Kieran Walshe

The Care Quality Commission Established in 2009 to regulate health and social care in England Legislative duty “to protect and promote” the health, safety and welfare of service users and the “general purpose of encouraging the improvement of health and social care services” Much criticised in for its performance and effectiveness – impact of mid-Staffs/Francis report etc Change of leadership in 2012, and new strategy and regulatory model in 2013, first in acute healthcare and now being rolled out to other sectors

CQC’s new approach to hospital inspection

Important features of the new model Greater scale and depth of prior data collection (large, complex, heterogeneous, risky, organisations) Greater expertise and resources of inspection teams (complexity, content knowledge, credibility) Eight key clinical service areas covered (heterogeneity, risk, clinical focus) Defined “key lines of enquiry” as framework for fieldwork/data collection/judgement making (consistency vs discretion) Quantitative ratings in five domains and narrative report (accountability, leverage/impetus for change, improvement)

An example of inspection ratings

Methods About 16 interviews with people in CQC and outside about the new acute hospital regulatory model Six observed hospital inspections in late 2013 – about 48 days of non-participant observation, review of documents, attending QA group meetings, quality summits etc – and 4 followup observations in 2014 About 65 1:1 telephone interviews with CQC inspection team members and NHS trust staff following inspections in 2013/14 Surveys of CQC inspection team members and trust staff following inspections in 2014(approx 350 and 130 respondents respectively at 21/5/2014) Interim report in Feb 2014, final report end May 2014

Some key findings – more detail in report Evidence and the use of information Inspection teams Inspection processes Collection and use of evidence for judgements, ratings and reports Reporting, impact and follow-up Full report at

Evidence, KLOEs and information Data pack impressive collation of performance data – some issues with timeliness, level of granularity, not well understood and used KLOEs – mostly general, not specific to the organisation/clinical area, quite high level Wider sources of data – up to 500+ documents from trust and others, some not read, hard to sift/sort and synthesise Reliance on direct observation, narrative and commentary

Evidence, KLOEs and information “...when you’ve done a lot of inspections and you know this probably as much as I do, if you look hard enough you can find something with everything, everywhere requires improvement…” [CQC inspector] I think that probably still needs some work on it, because when you’ve got a lot of information and then you have to decide, so do we mean that’s good or outstanding or not so good? It’s hard then, it’s trying to not to move into subjectivity I think. …But I think it’s challenging, because particularly when you’ve got conflicting information trying to be proportionate, because you might have a, you know, a certain amount of information that you think, well, that’s not very good that they’re doing, but hen you’ve got this other information that’s saying, but that’s really good! [CQC inspection lead] If there’s going to be a process where people can challenge, which I can’t see that there can’t be, then I think they need to be tight. There needs to be tight criteria, which is very difficult when you’ve got very complex organisations. “ [CQC inspector]

Inspection teams Large teams maximise content expertise and “feet on the ground”, at some cost to selectivity, coherence and manageability Selection/training/matching, skills gaps evident, prior experience largely shapes approach Team roles and management unclear and variable – best teams have clearly defined complementary roles Team members’ competence, confidence and credibility crucial – learning from experience Roles of analysts and CQC team leads/chair challenging

Inspection processes Initial day of familiarisation/team building Actual fieldwork process – highly inductive, experiential, intuitive, dependent on team skills, little structured or directed inquiry, variable across teams but scope for specialisation Focus groups, interviews, listening events all used to gather information from staff, patients, public Corroboration sessions to triangulate/challenge - but hard in large teams under time pressure Relentless activity – no time to think/reflect/discuss

Inspection process “I think really there’s no hiding place, so if the inspection is carried out thoroughly, there’s not a lot you can hide from it, it was far broader than anything I’ve experienced..” [Trust chief executive] “It felt like a full on week, obviously it was a week out of our working lives that was pretty much dominated by it, so it was a big week in that respect. But actually it was very positive, the staff really enjoyed it by and large, you know, the feedback was there seemed to be a buzz around the place,... The staff felt they were being inspected by peers, people who understood, they enjoyed showing what was good, what they were proud of, they felt they were talking to people who understood them. I’d say during the week it was a pretty positive feeling actually.” [Trust chief executive]

Evidence, judgements, ratings, reports Inspector observation/experience tends to predominate Core service areas – good clinical focus but don’t cover all clinical services, some corporate issues Domains have some ambiguities and hard to know what fits where; well-led an organisational cover-all Rating levels not explicitly defined and so interpreted variably within teams, across teams, across inspections Process for integrating and weighting disparate evidence to form rating (and back it up) implicit A lot left to CQC team leads to write up afterwards

Evidence, judgements, ratings, reports “I think as a trust, we’re disappointed with our ratings. They are clearly not linear scales. They don’t add up, so we’ve got some areas where we’ve got several greens and then a yellow, and then the total scores are yellow. And then there’s somewhere else where we’ve got two yellows and two greens, and the total scores are green. And it’s not very clear…” [Trust staff] “I think there does need to be a bit of calibration of these teams, I think there has to be some guidance given about the extent to which they are at liberty or not to make qualitative judgements, the extent to which they actually need to triangulate and base on evidence, and there’s probably something about how you kind of calibrate across the reports as well.” [Trust staff]

Evidence, judgements, ratings, reports “Everybody was really keen to know how we've done really. And the feedback was very limited and I understand why, but it was extremely limited. So people walked away almost a little bit frustrated. And particularly then frustrated when you get the report and there's a lot more detail in there that you then think oh, it doesn't feel as good as the feedback was on the day. So managing that internally has been a challenge and I don't think we saw that coming if I'm honest with you. I thought we'd get more detailed feedback on the day. …people really want to know how have we done. People have got pass/fail mentality, did we pass? Did we do okay? And so when they're left up in the air a little bit and then by the time you get the report in for factual accuracy it feels for most staff, I mean not for senior managers, for most staff it feels a little bit too distant and oh right, what does that mean then? Did we pass? Did we fail? [Trust staff]

Reporting, impact and follow-up Trusts respond in preparation, by seeking to manage the process and by responding to reports Reports and ratings seen as mostly right, though some findings contested especially those of poor performance Narrative reports scrutinised in detail by trusts for consistency and accuracy – qualitative rigour challenged Quality summits – used to launch report and initiate follow-up and action planning – more in cases of poor performance

Reporting, impact and follow-up Well, I think for us, for the majority of their findings I think it was holding up a mirror to us. We wouldn't disagree, we haven't disagreed with anything they've said in the report, we've just accepted it. Part of that is because we believe what they've said and part of it is political, that it's pointless fighting over minor stuff actually; you miss the point really. So everything in there is either already in an action plan, a priority or a risk register, and therefore associated actions. What I think this does is helps put a focus on it and gets stuff happening quicker. It focuses staff attention; people don't want to be deemed to be not good. So I think it's helpful in that respect. (Trust staff) But it is expensive and I think that it's going to be one of those things, unless we can make it…if we can make it more constructive and less punitive then people will see it as a value. If we don't I think it's going to be subjected to criticism from the host organisations and from the wider public I think. People will sit there and say bloody hell, there's 50 people with clipboards trawling round a hospital! It's costing bucketfuls of money, to tell us what we already know. So I think it's trying to really use it to make a difference. (CQC inspector)

Conclusions Rapid development of complex new regulatory model at scale and pace – ambitious undertaking Conscious intent to learn from piloting – concurrent internal and external evaluation Perceived rigour, comprehensiveness and credibility of new model high – especially in comparison with prior approach Issues about costs and value for money – further research now examining new model in four sectors (acute, mental health, social care, primary care)

Conclusions: possible improvements Maximise the quality of inspection team members through selection, training, deployment, experience and review/feedback Maximise validity, reliability, efficiency and utility of inspections through greater structure/process without constraining flexibility or losing scope for professional judgement Maximise validity/reliability of ratings through simplification, definition, training and testing Start to measure impact after the inspection to drive change and improvement