NDIS independent assessments are off the table for now. That’s a good thing – the evidence wasn’t there
12 July 2021 at 3:50 pm
The proposed NDIS independent assessments have been controversial. And data from a pilot doesn’t really tell us whether they’d deliver positive outcomes for Australians with disability, write Helen Dickinson and Anne Kavanagh.
Federal, state and territory disability ministers met on Friday to debate a proposal to introduce independent assessments into the National Disability Insurance Scheme (NDIS).
Big news: NSW families minister Alister Henskens confirms the government’s NDIS independent assessments plan “will not proceed”. pic.twitter.com/gJByatuma2
— Luke Henriques-Gomes (@lukehgomes) July 9, 2021
Remind me, what are independent assessments?
Independent assessments were proposed as a new way of determining the functional level of an individual with disability, which would then be then used to inform what level of funding that person would receive.
Under the current system, people demonstrate this by gathering evidence from their own specialists.
The government has argued this is not fair, because those with greater means can bypass waiting lists in the public system and see private specialists. They also believe professionals known to an individual may be affected by “empathy bias”, resulting in their clients ultimately being awarded larger funding packages.
Independent assessments would instead have seen people assessed by a government-contracted allied health professional unknown to them in a three-hour interview. This short assessment would have been used to determine what level of funding that person was entitled to.
Discussions about the introduction of independent assessments started under the previous NDIS minister, Stuart Robert.
The proposal attracted strong backlash from the disability community who argued the assessments were dehumanising and would lead to inappropriate plans, with a significant risk of traumatising participants.
Read more: ‘We know it’s bullsh*t’: Advocates say independent assessments will be a disaster for NDIS participants
The NDIS Joint Standing Committee received more than 320 submissions on the issue from a range of individuals, advocacy organisations, academics and more – the vast majority of which were highly critical of the proposed assessments.
On taking over the NDIS portfolio in April, Linda Reynolds announced she intended to pause the rollout of independent assessments until further piloting was completed and evaluated, and she had an opportunity to consult with stakeholders across the country.
Trialling independent assessments
The National Disability Insurance Agency has recently been piloting independent assessments (the NDIA is the independent agency responsible for implementing the NDIS).
It asked for volunteers to undertake an assessment, offering to compensate participants with a $150 payment.
The NDIA Research and Evaluation Branch this week released its interim report on the independent assessment pilot.
The evaluation sought to understand the experience of participants through the independent assessment process, including whether the report accurately reflected both what they told their assessor and their functional capacity.
A second aim was to get feedback on the independent assessment tools used in the process, and whether these tools were collecting the right information.
Depending on the age group and disability type there are eight tools which can be used in various combinations. An example of one of these tools is the WHODAS (World Health Organisation Disability Assessment Schedule), which explores how well an individual has been able to do certain activities with or without support (such as self-care and mobility).
While there are just under 450,000 NDIS participants, the pilot included a very small number of these.
Some 3,759 people across Australia took part in the pilot assessments, but of those, only 948 participants and support people provided survey responses which were analysed in the report.
More than 100 pilot participants were interviewed. And the evaluation also included surveys and interviews with assessors.
When we drilled down into the figures from the survey, we found some concerning results. Less than half of respondents reported that:
- their experience of independent assessments was excellent or very good (46 per cent)
- their assessor seemed to know a lot about the participant’s disability (49 per cent)
- the independent assessment report they received was an excellent or very good reflection of their meeting (48 per cent)
- the results of the independent assessment were a very good or excellent reflection of their functional capacity (42 per cent).
Assessors were not overly positive about the experience either. Only 39 per cent rated their training as very good or excellent and some wanted more practical training.
Many rated the assessment tools poorly including concerns about accuracy and relevance of the selected tools, echoing the concerns of Occupational Therapy Australia in its submission to the Joint Standing Committee.
None of this makes a compelling case for such a significant reform. But the most serious problem is this does not constitute a full and rigorous evaluation of independent assessments.
Are independent assessments fit for purpose? We don’t really know
The NDIA has attempted to be more transparent with its evaluation and commissioned a University of Sydney team to provide independent validation of their findings.
On the surface, this looked like a step forward for transparency. But this “validation” largely involved checking the way the NDIA analysed the data – they were not asked to critique the design of the evaluation.
The key question the disability community needed answering was whether independent assessments could be used to deliver funding packages that enabled participants to purchase reasonable and necessary services and supports that enabled them to live “ordinary” lives.
But this evaluation merely explored the experience of taking part in independent assessments; it didn’t answer this question. No participant was given a budget based on their results. No participant outcomes were measured.
If we did want to know whether independent assessments created fairer funding decisions, we would need to compare funding allocations and participant outcomes between a group who received independent assessments and a control group who did not. Without that, we simply don’t know if independent assessments are fit for purpose.
So it’s a good thing they’re off the table.