ICU Management & Practice, Volume 23 - Issue 5, 2023

img PRINT OPTIMISED
img SCREEN OPTIMISED

This article explores how pervasive and persuasive the internet is in current critical care practice, offers insights into how healthcare professionals, patients and fami­lies can critically appraise where information comes from and its content producers and discusses the opportunities and threats posed by AI on the physicians/team-patient/family relationships.

 

 

         We shape our tools, and thereafter, our tools shape us.
                                                               -Marshall McLuhan

 

Introduction

In ancient times, the Oracle of Delphi was believed to be the greatest source of knowledge, wisdom and prophecy. In modern times, it is the internet, and in the future, it will be artificial intelligence (AI). In ancient times, one had to make an arduous trek. Nowadays, we merely access any connected computer, tablet or phone. The internet, free online medical education (FOAMed), and AI are increasingly the dominant sources for medical data, medical education, medical opinion, and medical predictions. Unfortunately, the same is true regarding medical disinformation. We now all – patient and provider alike-live in a digital age. This means we all need a degree of digital literacy; perhaps even a degree in digital literacy.

 

The use of the internet and AI is a massive and increasingly important topic. The inter­net is both pervasive and persuasive, and yet accessing information and opinion is not the same thing as knowledge or wisdom. At the very least, we healthcare providers (HCPs) need to direct our patients and students (and ourselves) towards higher quality resources and away from half-truths and nonsense. This also should engage reflection on our own online presence: what is expertise vs mere opinion, and how might this influence our current and future relationships with patients and families.

 

Patients, families and HCPs want fast access to reliable information, and the internet has certainly revolutionised our collective ability to do so. AI has the poten­tial to dramatically increase both the speed and the access to information and appears set to revolutionise how we care for patients and families. Yet, above all, humans want to be seen, heard and cared for. There is an ever-present danger that, as technology grows, it becomes master rather than servant. The goals of this article are therefore: 1) to discuss how pervasive and persuasive the internet is in current critical care practice, 2) to offer insights into how HCPs, patients and families can critically appraise where information comes from and its content producers and 3) to explore the opportunities and threats posed by AI on the physicians/ team-patient/family relationships.

 

Consulting Dr Google: Medical Information on the Internet

The enormous increase in online medical information can make it difficult for every­one to distinguish signal from noise. This digital revolution is especially profound and disruptive because there now is a single, albeit vast, portal for education, opinion, information, and disinformation. Unless there are internet firewalls and restricted journal access, we also have a single portal for HCPs, patients and families. When compared to traditional medical forums (i.e. textbooks or journals), online mate­rial can be disseminated faster, wider and cheaper. The downside is that there is often less time for expert review and contemplation.

 

While the old ways of disseminating medical information (textbooks, peer-review, expensive journals) were far from perfect, there are dangers with relying solely upon internet searches. For example, it is unclear who or how information is filtered and who or how they are monetising the process. In short, most search algorithms are proprietary, and many revenue streams are opaque. In the 1960s, the philoso­pher Marshall McLuhan argued that “the medium is the message” (McLuhan 1964). Accordingly, part of being a modern human- a member of “homo digitalis”- is understanding how the digital medium not only spreads messages but morphs them.

 

Health is one of the most searched topics online. There are at least 500 million daily tweets (Twitter Stats 2023) and four billion daily Google searches (Internet Stats 2023). This single search engine constitutes nine out of every ten searches, and the company is valued at over $1.5 trillion. We live in an attentional economy, where more profit is made by keeping our attention rather than ensuring the truth. Moreover, anybody with a laptop, tablet or smartphone can now promote their ideas, regardless of whether these are grounded in science or mere opinion and regardless of whether or not they have any expertise.

 

Nowadays, medical information is less owned by HCPs or medical journals. Producers can game the algorithm using techniques like keyword stuffing and link-building, thereby making content look more relevant and more accurate than is warranted. An example during the pandemic was the Frontline COVID-19 Critical Care Alliance (FLCCC) website. This site promoted unproven, potentially harmful, therapies. The issue is that its professional appearance likely worsened relations because patients and families expected to receive harmful therapies such as ivermectin and hydrochloroquine. In the same vein, Google’s algorithms recently shifted from quality content to content optimised for advertising. The move from organic content to paid advertisements can make it harder to find useful, unbiased information. Search engines can also diminish diversity of opinion, given that over three-quarters of us never scroll past the first page of suggested sites. In short, if a source is not ranked on the first page, then it is unlikely to be read or to influ­ence the debate.

 

The internet not only makes it quicker to find resources but also easier to criticise and dismiss (Wilkinson et al. 2019). Medical information has been democratised, but the cautious, iterative, scientific method is under threat. Traditional crucibles of medical debate have less of a monopoly over what is considered mainstream or trustworthy. Journals also have fewer paid subscribers and, therefore, face unpredict­able futures. Digital publication is now associated with over 30,000 journals on PubMed alone. This makes it harder for most of us to remain current or claim broad competence.

 

Anyone with a significant online pres­ence can have an influence disproportion­ate to their scholarly standing or clinical expertise (Cameron et al. 2017). Moreover, social media is playing an ever-increasing mainstream role. Accordingly, FacebookTM, WhatsAppTM, WikipediaTM, YouTubeTMand XTM(formerly Twitter) are increasingly important sources of medical information for the public. Importantly, however, the same is true for HCPs. Blogs, podcasts and websites are increasingly recommended over textbooks and journals for teaching and learning (Cadogan 2014; Eysenbach 2011; Thoma 2015).

 

Because of the deluge of information, readers, HCPs or patients/families are, nowadays, more likely to scan, rather than fully engage, absorb and reflect. There may be less patience for nuance or interest in ideas that challenge bias. Virtual communication also makes it easier for humans to be anonymous, rude, and dismissive. Accordingly, traditional publishers and academic institutions face an existential dilemma. Namely, to what degree should they embrace digital media to be popular or stay relevant? Websites, blogs and videos can receive thousands more views than journal articles and reach previously untapped audiences. In contrast, an excessive online presence can also tarnish standards, reputation and brand. Universities also face uncertainty regarding how best to recognise and reward non-traditional scholarly output from their faculty and whether they can (or should) censure online heretics. In other words, we must be cautious not to equate worth with ease of access or popularity (Cameron et al. 2017).

 

Returning to Marshall McLuhan, the digi­tal revolution has affected not only how we report medical ideas but also whether they persist. Ideas (i.e. packages of information) can now spread like viruses (i.e. packages of genetic information). Similarly, memes (units of cultural information) compete like genes (units of genetic information). Regardless, the internet has dramatically affected how ideas are birthed, raised, and matured. It has also affected how we interact as humans.

 

Families Searching Doctors; Doctors Searching Patients

Many families of critically ill patients turn to internet searches when some­one they care for is admitted to the ICU. They go online to better understand the disease and to seek out support. However, many also search for their physicians to understand who they are, their expertise (or lack thereof), their publications, and their ratings. Upon being introduced, it is not infrequent -though it can be discon­certing- to have family members inform HCPs that they have ‘read all about them’. Being the subject of internet searches by patients feels, well… awkward. Even if we have posted the information ourselves, it can feel like starting a conversation with a person whose name you have forgotten, or, once met but cannot place.

 

What is less well-known is that many physicians also “Google” their patients (Belisomo 2015; Brown 2019). Sometimes searches are performed at the patient’s request to understand their career (e.g. artist, photographer), to determine if that patient is delirious, or, more controversially, to look up details of their lives (Belisomo 2015; Brown 2019). Regardless, the recipro­cal nature of these online searches shows how internet searches have profoundly affected how modern humans interact.

 

HCPs might go on ward rounds without a stethoscope now, but few of us are ever far from an internet connection. Along with looking up patient records and journals, HCPs also search the internet to better understand behaviours, fads or health crazes. These could range from beliefs in alternate drugs (e.g. ivermectin in COVID), dietary cleanses (e.g. high dose baking soda), or new illicit drugs (i.e. carfentanyl, xylazine) or eccentric prac­tices (e.g. ingesting Tide pods). Of note, these Google searches usually yield faster results than traditional medical searches. They also show exactly what people are reading and exploring.

 

It can be challenging to answer every internet-related question a family has. Regardless, it is a way to show that we are eager to partner, maintain dialogue, and build trust. For patients and families, it is a way to feel empowered, but it can also result in misinformation, anxiety, and, at times, an exaggerated sense of understanding. The challenge for both parties is to navigate this landscape with trust, transparency, integrity, and patient-centredness. It starts with a basic understanding of what is qual­ity online information and what is not.

 

Online Medical Information: It’s Popular, But Is It Any Good?

The term “Free Open Access Medical Education”, aka “FOAM” or “FOAMed", reportedly originated in an Irish pub. Apparently, a doctor was preparing a talk and stared at a half-emptied beer glass. This individual wanted to encapsu­late the proliferation of free online open access medical education, and FOAM was coined (Shaw 2013). These online resources include blog posts, podcasts, online videos, Facebook groups, Twitter feeds, and Google Hangouts. To date, the largest proportion of FOAM comes from emergency medicine, with lesser amounts from critical care medicine and anaesthesiology. North American sources currently predominate.

 

While some have raised concerns with FOAMed and emphasised the need to wait for qualified experts, others argue that online medical material is now unavoid­able. If so, the focus should be on whether digital resources are higher versus lower quality and higher versus lower influence.

 

Ways to gauge the quality of online work include the Medical Education Translational Resources Impact and Quality (METRIQ) study collaboration (https://metriqstudy. org/) and the Critical Care Medical Educa­tion Website Quality Evaluation Tool (CCMEWQET) (Wolbrink 2019). Many scoring systems also exist and include the Social Media Index, the ALiEM AIR score, and the Revised METRIQ Score. Others (Ting et al. 2020) have also identified ten tools, categorised into those that help readers and those that rate producers.

 

In addition to assessing the quality of online products, we can assess producers (i.e. authors). The h-index is a traditional metric used to estimate the productivity of an individual scholar. It is the maximum value where a given author has published h-papers each cited h-times (Hirsch 2005). Albeit slightly tongue-in-cheek, a similar index was developed for the social media age. The Kardashian Index (KI) (Hall 2014) refers to Kim Kardashian, someone with innumerable online followers but no official scientific credentials. The serious point is that in the digital age “influencers” (whether celebrities or academics) can have a greater impact- or lesser impact-than their academic standing warrants (Brindley et al. 2022).

 

AI and Big Data in Critical Care Medicine: Servant, Not Master

AI refers to computer systems that perform tasks that would otherwise require human intelligence. These include, but are not limited to, pattern recognition and decision-making. These are usually powered by big data, namely huge data sets that can be analysed computationally to reveal trends and associations. Critical care medicine is on the cusp of an AI explosion (Hong et al. 2022; Saqib et al. 2023; van de Sande 2021). Potential AI applications are currently limited only by our human imagination and programming power. In time, however, AI holds the prospect of devices updating their own algorithms and generating their own searches.

 

AI already has the ability to predict patient deterioration (Chen et al. 2022; Cho et al. 2020), diagnose/predict the development of sepsis (DeCorte et al. 2022; Pai et al. 2022), predict the development of surgical site infections (Hopkins 2022), prognosticate ICU outcomes from a variety of critical illnesses, predict the effective­ness of triage, determine the best time to initiate intubation (Im et al. 2023; Nopour et al. 2023; Siu et al. 2020), and predict weaning from mechanical ventilation, extubation and safe ICU discharge (Abad et al. 2021; de Vos et al. 2022; Fabregat et al. 2021; Liu et al. 2022). It can predict the impact of ICU surge on patient mortality (Greco et al. 2022), the time to death after withdrawal of life support, and the success of organ transplantation (Yu et al. 2022). It may soon be able to anticipate the factors that create ICU physician/team stress and strain, identify training needs and predict future outcomes. The potential implica­tions for patient care and the HCP-patient/ family relationship could be staggering.

 

Rather than blindly accepting every­thing AI purports to offer, we should embrace our traditional academic scepti­cism. In other words, we should insist on high-quality research and wise reflection before implementation. We should accept that AI’s potential is virtually unlimited and, therefore, both exciting and scary. The anxiety surrounding AI includes what data it accesses/includes/excludes, and what algorithms are used. It matters what AI deems to be important/irrelevant (cost, quality of care, quality of life, etc.) in arriving at its results/predictions; an opacity often called its ‘black box’. As with any decision-making algorithm, whether human or machine, there is concern regard­ing bias and discrimination (Hong et al. 2022; Lorenzini et al. 2023; Mittelstadt 2021). The ability of AI to predict and its quality of output depends on the quality of its input. Without good quality research (input) the effect of internet popularity may impair its usefulness (output).

 

There is also anxiety about how human values, beliefs and lived experiences will be taken into account, if at all (Hong et al. 2022; Lorenzini et al. 2023; Mittelstadt 2021). To date, relatively few research studies have explored the lived experi­ences of critical illness. While some are now including patients and families as co-researchers to better understand their needs and priorities (Douma et al. 2021; Douma et al. 2023), AI risks moving us further from such considerations due to the dearth of such research and the speed of its own development. The concern is that CCM will see an erosion of patient-centred care and a loss of its humanity.

 

Many have argued that HCPs do not need to fully understand how AI works to use it as a tool. We disagree. This is because AI may provide more than just data that we can choose to ignore. It may guide decision-making and be far more determinative, especially if used to prog­nosticate, and especially if/when AI exceeds human intelligence.

 

AI’s Potential Effects on HCP-Patient/Family Relationships

While the potential of AI is not yet realised, discussions of its potential impact on HCP-patient/family relationships (Mittlestadt 2021; Nagy and Sisk 2020; Saqib et al. 2023; Sauerbrei et al. 2023) have begun. These have focused on AI’s ability to expand knowledge and understanding of health and illness. It has also been suggested that AI may enhance HCP-patient/family relationships by freeing up HCPs from more administrative tasks, allowing them to spend more time with patients, though whether this is realistic is not clear.

 

AI could supplant physicians in core knowledge, could be better at generating differential diagnoses, and could be quicker with decision-making. It is anticipated that physicians who use AI as an assistive tool will outperform those who do not. In time, we may even be mandated to use AI. Its enhanced ability to prognosticate may also have implications for how we triage scarce resources. Families currently, and understandably, raise concerns if they believe physicians are making decisions based on imperfect prognosticators. They are not likely to be any happier when algorithms and computers decide, even if the prognostications are more informed. Fears have already been raised about a return to more paternalistic care, this time governed by machines (Lorenzini et al. 2023).

 

Alternatively, AI offers many putative advantages. For example, with better prognostication, we could spare families weeks of organ support if we know the outcome will be bad or increase their resolve if we know there is a good chance. For those wrestling with whether to offer organ donations after cardiac death, they could be spared the distress if we knew that the patient would not die in the necessary timeframe. AI predictions of survival post-transplant could also result in better matching of donors and recipients, thereby improving both the likelihood and quality of survival.

 

Whether its benefits will ultimately outweigh its risks, it seems certain that AI could challenge how trust is earned and kept within HCP-patient/family relationships. To maximise the potential of AI, perhaps the best way forward is to learn from our past. Rather than wait for patients and families to come to us, we should accept they will have searched the internet and are likely to have questions. This means that -like it or not- part of our modern job includes reviewing AI searches/data together, translating what it means, and ensuring it is discussed in human terms.

 

Conclusion

AI’s opportunities and its risks to the doctor-patient relationship were explored In the report, “The Impact of Artificial Intelligence on the Doctor-Patient Rela­tionship”; commissioned by the Steering Committee for Human Rights in the field of Biomedicine and Health. In brief, these included: (1) Unequal access to this technology; (2) Insufficient transparency regarding inconclusive and misguided evidence; (3) The risk of social bias; (4) Diluting the patient’s account of well-being; (5) The risks of automation bias, de-skilling, and displaced liability; and (6) A loss of privacy (Mittelstadt 2021). The report also questions what standards AI will be held to in relationship to profes­sionalism and duty of care.

 

The use of machines and computers is central to critical care medicine, and these devices are becoming increasingly smart in their nature. While exploring technol­ogy’s cutting edge, it is crucial to keep the focus on the relationship between provider and recipient. AI is a remarkable tool we should harness, yet its ultimate benefit or harm rests with how we control it. Patients and HCPs want reliable information, but above all, humans want to be seen, heard and cared for. We cannot forget that healthcare is, and should always be, about human connections, not just online connections. We just need to remember that with the tremendous power of AI comes great responsibility, and we need to apply it wisely.

 

Conflict of Interest

None.

«« Prone Positioning During ECMO in Severe ARDS


Sleep and Circadian Disruption in the ICU »»

References:

Abad ZSH, Maslove DM, Lee J (2021) Predicting Discharge Destination of Critically Ill Patients Using Machine Learning. IEEE J Biomed Health Inform. 25(3):827-837.


Belisomo R (2015) Is it OK for doctors to google their patients? Scientific American. Available at https://www.scientificamerican.com/article/is-it-ok-for-doctors-to-google-patients/


Brindley PG, Byker L, Carley S, Thoma B (2022) Assessing online medical education resources: A primer for acute care medical professionals and others. Journal of the Intensive Care Society. 23(3):340–344.

 

Brown S (2019) Is it ethical for doctors to google their patients? CMAJ. 191(13):

 

Cadogan M, Thoma B, Chan TM, Lin M (2014) Free Open Access Meducation (FOAM): the rise of emergency medicine and critical care blogs and podcasts (2002-2013). Emerg Med J.

 

Cameron P, Carley S, Weingart S, Atkinson P (2017) CJEM Debate Series: #SocialMedia –Social media has created emergency medicine celebrities who now influence practice more than Published evidence. Can J Emerg Med Care. 19(6):471–4.

 

Chen YW, Li YJ, Deng P et al. (2022) Learning to predict in-hospital mortality risk in the intensive care unit with attention-based temporal convolution network. BMC Anesthesiol. 23;22(1):119.


Cho KJ, Kwon O, Kwon JM et al. (2020) Detecting Patient Deterioration Using Artificial Intelligence in a Rapid Response System. Crit Care Med. 48(4):e285-e289.


De Corte T, Van Hoecke S, De Waele J (2022) Artificial Intelligence in Infection Management in the ICU. Crit Care. 26(1):79.


de Vos J, Visser LA, de Beer AA et al. (2022) The Potential Cost-Effectiveness of a Machine Learning Tool That Can Prevent Untimely Intensive Care Unit Discharge. Value Health. 25(3):359-367.


Douma MJ, Graham TAD, Ali S et al. (2021) What are the care needs of families experiencing cardiac arrest? A survivor and family-led scoping review. Resuscitation. 168:119-141.


Douma MJ, Myhre C, Ali S et al. (2023) What Are the Care Needs of Families Experiencing Sudden Cardiac Arrest? A Survivor- and Family-Performed Systematic Review, Qualitative Meta-synthesis, and Clinical Practice Recommendations. J Emerg Nurs. S0099-1767(23)00180-0.


Eysenbach G (2011) Can tweets predict citations? Metrics of social impact based on Twitter and correlation with traditional metrics of scientific impact. J Med Internet Res.


Fabregat A, Magret M, Ferré JA et al. (2020) A Machine Learning decision-making tool for extubation in Intensive Care Unit patients. Comput Methods Programs Biomed. 200:105869


Greco M, Angelotti G, Caruso PF et al. (2022) Outcome prediction during an ICU surge using a purely data-driven approach: A supervised machine learning case-study in critically ill patients from COVID-19 Lombardy outbreak. Int J Med Inform. 164:104807.


Hall N (2014) The Kardashian index: A measure of discrepant social media profile for scientists.

Genome Biol.

 

https://journalofethics.ama-assn.org/article/patient-physician-and-dr-google/2012-05


Hirsch JE (2005) An index to quantify an individual's scientific research output. Proc Natl Acad Sci USA.102(46):16569-16572.


Hong N, Liu C, Gao J et al. (2022) State of the Art of Machine Learning-Enabled Clinical Decision Support in Intensive Care Units: Literature Review. JMIR Med Inform. 10(3):e28781.


Hopkins BS, Mazmudar A, Driscoll C et al. (2020) Using artificial intelligence (AI) to predict postoperative surgical site infection: A retrospective cohort of 4046 posterior spinal fusions. Clin Neurol Neurosurg. 192:105718.


Im JE, Park S, Kim YJ et al. (2023) Predicting the need for intubation within 3 h in the neonatal intensive care unit using a multimodal deep neural network. Sci Rep. 13(1):6213.


Internet Live Stats (n.d.) Available at https://internetlivesstats.com

Liu CF, Hung CM, Ko SC et al. (2022) An artificial intelligence system to predict the optimal timing for mechanical ventilation weaning for intensive care unit patients: A two-stage prediction approach. Front Med (Lausanne). 9:935366.


Lorenzini G, Arbelaez Ossa L, Shaw DM, Elger BS (2023) Artificial intelligence and the doctor-patient relationship expanding the paradigm of shared decision making. Bioethics. 37(5):424-429.


McLuhan M (1964) Understanding Media: The Extensions of Man. McGraw Hill.

Mittlelstadt B (2021) The Impact of artificial intelligence on the doctor-patient relationship, Report commissioned by the Steering Committee for Human Rights in the fields of Biomedcine and health (CDBIO), Council of Europe. Available at https://rm.coe.int/inf-2022-5-report-impact-of-ai-on-doctor-patient-relations-e/1680a68859


Nagy M, Sisk B (2020) How will artificial intelligence affect patient-clinician relationships? AMA Journal of Ethics. Available at https://journalofethics.ama-assn.org/article/how-will-artificial-intelligence-affect-patient-clinician-relationships/2020-05


Nopour R, Shanbezadeh M, Kazemi-Arpanahi H (2023) Predicting intubation risk among COVID-19 hospitalized patients using artificial neural networks. J Educ Health Promot. 12:16.


Pai KC, Wang MS, Chen YF et al. (2021) An Artificial Intelligence Approach to Bloodstream Infections Prediction. J Clin Med. 10(13):2901.


Saqib M, Iftikhar M, Neha F et al. (2023) Artificial intelligence in critical illness and its impact on patient care: a comprehensive review. Front Med (Lausanne). 10:1176192.


Sauerbrei A, Kerasidou A, Lucivero F, Hallowell N (2023) The impact of artificial intelligence on the person-centered, doctor-patient relationship: some problems and solutions, BMC Medical Informatics and Decision-Making. 23:73.


Shaw G (2013) Don’t Call It Social Media: FOAM and the Future of Medical Education. Emerg Med News. 35:29–30.


Siu BMK, Kwak GH, Ling L, Hui P (2020) Predicting the need for intubation in the first 24 h after critical care admission using machine learning approaches. Sci Rep. 10(1):20931.


Thoma B, Chan T, Benitez J, Lin M (2014) Educational Scholarship in the Digital Age: A Scoping Review and Analysis of Scholarly Products (2014). The Winnower.

 

Ting DK, Boreskie P, Luckett-Gatopoulos S et al. (2020) Quality Appraisal and Assurance Techniques for Free Open Access Medical Education (FOAM). Resources: A Rapid Review. Semin Nephrol. 40:309–19.

 

Twitter Statistics (n.d.) Available at https://www.omnicoreagency.com/twitter-statistics/


Van de Sande D, van Genderen ME, Huiskens I et al. (2021) Moving from Bytes tyo Bedside: a systematic review of the use of artificial intelligence in the intensive care unit, Int Care Med. 47(7): 750-760

 

Wilkinson J, Wong AVK, Malbrain M (2019) Social Media in Critical Care: Entering an Exciting Era. Annual Update. Intensive Care Emerg Med. 665–76.

 

Wolbrink TA, Rubin L, Burns JP, Markowitz B (2019) The Top Ten Websites in Critical Care Medicine Education Today. J Intensive Care Med. 34:3–16.

 

Yu YD, Lee KS, Man Kim J et al. (2022) Artificial intelligence for predicting survival following deceased donor liver transplantation: Retrospective multicenter study. Int J Surg. 105:106838.