Artificial intelligence language model ChatGPT has garnered much buzz and attention for its ability to understand and answer open-ended questions about any topic. Choong Zi Chao, Lead Specialist, and Alvin Yuen, Senior Specialist from Synapxe Data aNalytics & AI team, look at how ChatGPT can be used in healthcare and its limitations.
What is ChatGPT?
ChatGPT is a large language artificial intelligence (AI) model developed by OpenAI, a US-based AI research company. A variant of the GPT (Generative Pre-trained Transformer) model, it uses deep learning to generate human-like responses to natural language prompts. Since it was released to the public in November 2022, it has generated much interest for its high coherent and contextually relevant replies in wide-ranging domains.
How can ChatGPT be used in healthcare?
The release of ChatGPT has prompted discussions regarding its capabilities in the healthcare industry. As ChatGPT can be used for different natural language processing (NLP) tasks, such as text summarisation, question answering and classification, it could be useful for potential applications such as chatbots and virtual assistants.
Tools such as ChatGPT could possibly help to automate repetitive tasks such as answering frequently asked questions from patients, generate clinical notes and write medical reports, triage patients based on criticality of their condition, thereby freeing up resources for healthcare professionals to focus on providing primary clinical care. On top of that, it can aid in clinical decision support by extracting and analysing clinical notes to identify trends and patterns based on patients' medical history in electronic medical records.
Patients can receive personalised information and assistance in a timely manner. The complementary clinical expertise with predictive analytics can also lead to better health outcomes, as the treatment plan can be tailored to every patient's unique health condition.
For administrators, it can be a cost-effective solution for healthcare providers to offer virtual assistance and clinical decision support at scale while relieving the strain on our healthcare professionals, especially during manpower crunches such as during the Covid-19 pandemic.
Limitations
ChatGPT is not intended to be used as a tool to retrieve real time information as its data is limited up to 2021. Hence, it is unable to produce accurate answers for recent events such as the latest status of Covid-19 in Singapore or the winner of the 2022 FIFA World Cup.
ChatGPT is not intended to be used as a tool to retrieve real time information as its data is limited up to 2021. Hence, it is unable to produce accurate answers for recent events such as the latest status of Covid-19 in Singapore or the winner of the 2022 FIFA World Cup.
Due to the popularity of ChatGPT, users may occasionally face accessibility issues due to the high volume of traffic. In such instances, users would have to wait for access which can be an issue when prompt responses are needed.
As ChatGPT is trained on a large dataset of text from the Internet, it may unintentionally reflect the bias that is present in the training data. There may be examples of stereotypes and assumptions for gender or ethnicity that can be reflected in the responses based on the training data. The chatbot may not always generate responses which are relevant or accurate as well, and users can overwrite the response given by ChatGPT with an inaccurate one.
Additionally, ChatGPT may also have limited knowledge in expertise areas such as science or medicine. For example, if a user asked about a disease/drug which was discovered recently, ChatGPT may not have information about it in the training data and hence, is unable to provide an accurate answer.