Article Text

Download PDFPDF

Building trust in artificial intelligence and new technologies in mental health
Free
  1. Bessie O'Dell1,2,
  2. Katherine Stevens1,2,
  3. Anneka Tomlinson1,2,
  4. Ilina Singh1,
  5. Andrea Cipriani1,2,3
  1. 1 Department of Psychiatry, University of Oxford, Oxford, Oxfordshire, UK
  2. 2 Oxford Precision Psychiatry Lab, Oxford Health Biomedical Research Centre, Oxford, UK
  3. 3 Oxford Health NHS Foundation Trust, Warneford Hospital, Oxford, UK
  1. Correspondence to Professor Andrea Cipriani, Department of Psychiatry, University of Oxford, Oxford OX1 2JD, UK; andrea.cipriani{at}psych.ox.ac.uk

Statistics from Altmetric.com

Request Permissions

If you wish to reuse any or all of this article please use the link below which will take you to the Copyright Clearance Center’s RightsLink service. You will be able to get a quick price and instant permission to reuse the content in many different ways.

In 2019, the Topol review was published on behalf of the secretary of state for health and social care in the UK, preparing the healthcare workforce to deliver the digital future.1 A multidisciplinary team of experts, including clinicians, researchers, ethicists, computer scientists, engineers and economists, reviewed the available data and projected into the future (ie, next 20 years) two key questions: what impact technological developments (including genomics, artificial intelligence (AI), digital medicine and robotics) will have on the roles and functions of National Health System clinical staff? How could this innovation (ie, biosensors, electronic patient record, smartphone apps, digital infrastructure and virtual reality) ensure safer, more productive, more effective and more personalised care for patients? It is now widely recognised that data science and information technologies enable understanding of the uniqueness of each individual and the ability to deliver healthcare on a far more timely, efficient and tailored basis.

Mental health is a top priority in the UK national research agenda2 and presents a unique opportunity because it is in the next wave of adoption of digital health and innovation technologies.3 The Topol review noted that innovation can ‘bring a new emphasis on the nurturing of the precious interhuman bond, based on trust, clinical presence, empathy and communication’. Patients must occupy a central role when assessing and implementing any new technologies.4 This is especially true in our field, as people can be more vulnerable due to brain or cognitive disorders, for example, and the patient–clinician interaction has long been at the core of the therapeutic relationship. In this relationship, ‘trust’ has a specific role, as recently highlighted by a European Commission White Paper.5

What is ‘trust’? In its most basic form, ‘trust’ is a willingness to rely on something or someone. However, how exactly we conceptualise ‘trust’ will depend on the theoretical lens that we are using—and as such the concept of ‘trust’ has been applied differently across a variety of disciplines, ranging from sociology and psychology to philosophy and economics. The importance of trust in human interactions has been repeatedly recognised in the last 50 years: ‘the entire fabric of our day-to-day living, of our social world, rests on trust, as almost all of our decisions involve trusting someone else’.6 Trust not only helps to facilitate collaboration among people, but also underpins the formation and maintenance of social relationships.7 More recently, accounts of trust have developed to delineate the more generic idea of ‘trust’ from trust specifically developed in digital contexts and/or involving artificial agents—including a concept identified as e-trust.8 A related, yet distinct notion, is that of ‘trustworthiness’—the extent to which something or someone is deserving of trust or confidence.9 The terms ‘trust’ and ‘trustworthiness’ are often conflated in bodies of the literature discussing these concepts, even though implicitly much of it is primarily about trustworthiness, not trust.10

Why then, are interhuman bonds so important when considering the adoption of technology? Interhuman bonds (eg, between a patient and clinician) and the trusting relationships underpinning them are key to the successful adoption and implementation of digital health and innovation technologies. For example, research has shown that in inherently uncertain territories (such as the use of virtual doctors, as well as other algorithmic decision-makers), people will favour human judgement. Therefore, in medical decision-making, people may be unwilling to use even the best possible algorithm.11 Patients may therefore be more likely to defer to, or rely on (ie, trust) clinicians to mediate their interactions with digital health and innovation technologies. Such deference could be particularly pronounced in mental health, given both the vulnerability of patients, and the importance of patient–clinician interaction in the therapeutic relationship. This highlights the need for both clinicians and technology to be deemed ‘trustworthy’ (deserving of trust) in mental health settings; and, indeed, for clinicians to also consider the technology they bring into the clinical relationship to be trustworthy.

Not paying sufficient attention to the importance of interhuman bonds based on trust is detrimental to the development and adoption of technologies. Taking AI technology as an example, recent years have seen an exponential growth in the number of AI algorithms and projects published in the medical literature. AI systems have consistently demonstrated as being more beneficial than clinical care without a tool,12 13 and are key to the delivery of personalised, evidence-based care. However, this academic interest in AI technologies does not appear to translate well to clinical settings, where the ‘clinical impact in terms of patient outcomes remains to be demonstrated’.14 Keane and Topol have considered how the lack of uptake in AI technologies (despite their potential) could be due to a so-called ‘AI chasm’, in which there is an overemphasis on the technical aspects of the proposed algorithms, with insufficient attention given to the factors that affect the interaction with their human users.15 Others have also called for a better focus on what has been deemed the ‘softer’ or more ‘qualitative’ impacts of AI technologies in clinical care.16 In order to better aid in the translation of novel technologies from research into clinical settings and address this ‘AI chasm’, relevant stakeholders (including researchers and developers) should therefore take into account the more ‘qualitative’ impacts of these technologies. This includes potential impacts on interhuman bonds, trust and trustworthiness.

In conclusion, trust underpins interhuman bonds. These bonds are central to clinical care in mental health, and this is equally the case when digital health and innovation technologies are used.17 Ultimately, it is the role of the clinician to bridge the gap between the technology and their patient(s), and only in doing so can digital health and innovation technologies be better used in mental health.18

Ethics statements

Patient consent for publication

Ethics approval

Not applicable.

References

Footnotes

  • Twitter @And_Cipriani

  • Contributors AC and BO drafted the editorial. All other authors critically revised the text. All authors approved the final version of the article.

  • Funding AC is supported by the National Institute for Health Research (NIHR) Oxford Cognitive Health Clinical Research Facility, by an NIHR Research Professorship (grant RP-2017-08-ST2-006), by the NIHR Oxford and Thames Valley Applied Research Collaboration and by the NIHR Oxford Health Biomedical Research Centre (grant BRC-1215-20005).

  • Disclaimer The views expressed are those of the authors and not necessarily those of the UK National Health Service, the National Institute for Health Research or the UK Department of Health.

  • Competing interests None declared.

  • Provenance and peer review Not commissioned; internally peer reviewed.