Highlights

In brief

A personalised medical dialogue generation system takes into account the uniquely evolving conditions of patients with diabetes, enabling customised health coaching.

Photo by DC Studio | Freepik

Doctor AI on call

10 Jul 2025

An AI-based framework adds a personal touch to medical dialogues, supporting the health management plans of patients with diabetes.

If you have ever dreamt of having your own robot healthcare assistant like Baymax from Big Hero 6, the idea may not be as farfetched as you think. Amidst advancements in generative artificial intelligence (AI) such as the algorithms that power ChatGPT and Gemini, researchers are also leveraging large language models (LLMs) to develop AI assistants that can converse with patients.

“Personalised medical dialogue generation can improve medical care by tailoring conversations to patients’ specific needs, medical histories and preferences,” said Zhengyuan Liu, Tech Lead of the Multimodal Generative AI Group at the A*STAR Institute for Infocomm Research (A*STAR I2R).

These LLM-based agents can be particularly beneficial in supporting follow-up health communications beyond the clinic, such as diabetes management plans.

According to Liu, personalising these conversations can help encourage patients to adhere to treatment plans, taking into account, and adapting to, changes in their lifestyle or day-to-day health condition.

With the support of the Diabetes Clinic of the Future initiative, Liu and A*STAR I2R colleagues including Senior Principal Scientist Nancy F. Chen developed a generative AI-based medical dialogue system to provide customised coaching for patients with diabetes. To optimise the model, they analysed 856 telehealth conversations between nurses and patients with diabetes to identify essential topics covered in follow-up calls.

“Our approach uses topic-focused summarisation to distil core information from lengthy dialogues, making it better at handling the long conversational contexts that are typical in medical settings,” Liu explained. The framework also incorporates patient profiles from demographic data and health condition changes, all leading up to more personalised dialogue generation.

When compared against real-life healthcare calls, the researchers found that their system effectively extracted key information from noisy dialogue contexts to guide the subsequent conversation flow. This contextualised approach significantly enhanced dialogue generation quality based on standardised metrics for language models, representing a step towards more human-centred AI.

“By adapting to individual patient contexts and preferences, the framework creates more natural, relevant and empathetic healthcare conversations that respect the uniqueness of each patient’s situation,” said Liu. For example, the model might ask more specific follow-up questions on the patient’s diet based on their dietary habits.

The team now aims to expand their personalised dialogue generation system, with one potential development being speech capabilities. “Emotion-aware text-to-speech generation could enhance dynamic and empathetic interactions between the system and patients,” Liu said.

The A*STAR-affiliated researchers contributing to this research are from the A*STAR Institute for Infocomm Research (A*STAR I2R).

Want to stay up to date with breakthroughs from A*STAR? Follow us on Twitter and LinkedIn!

References

Liu, Z., Salleh, S.U.M., Krishnaswamy, P. and Chen, N.F. Context aggregation with topic-focused summarization for personalized medical dialogue generation. Proceedings of the 6th Clinical Natural Language Processing Workshop, 310-321 (2024). | article

Liu, Z., Salleh, S.U.M., Oh, H.C., Krishnaswamy, P. and Chen, N. Joint dialogue topic segmentation and categorization: A case study on clinical spoken conversations. Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing: Industry Track, 185-193 (2023). | article

About the Researchers

View articles

Zhengyuan Liu

Tech Lead, Multimodal Generative AI group

A*STAR Institute for Infocomm Research (A*STAR I2R)
Zhengyuan Liu is currently a Tech Lead in the Multimodal Generative AI group and the Asst. Head of AI for Education programme at the A*STAR Institute for Infocomm Research (A*STAR I2R). He has published over 30 research papers in top-tier AI and Natural Language Processing conferences including ACL, NAACL, EMNLP, COLING, ICASSP and INTERSPEECH. He serves as the reviewer at conferences including NeurIPS, ICLR and ACL; and journals including IEEE TASLP, ACM CSUR and Neurocomputing. He has been elected as an IEEE senior member for his significant professional achievements and won the Best Paper Award at SIGDIAL 2021, C3NLP in ACL 2024, and SUMEval in COLING 2025; and the Outstanding Paper Award at EMNLP 2023 and EMNLP 2024.
View articles

Nancy F. Chen

Senior Principal Scientist and Lead Principal Investigator

A*STAR Institute for Infocomm Research (A*STAR I2R)
Nancy F. Chen is a Senior Principal Scientist and Lead PI at the A*STAR Institute for Infocomm Research (A*STAR I²R). She heads the Multimodal Generative AI group and AI for Education programme. A serial best paper award winner, her AI research spans culture, healthcare, neuroscience, social media, education and forensics. Chen's multilingual tech led to commercial spinoffs and adoption by Singapore’s Ministry of Education. She is Program Chair for NeurIPS 2025, ICLR 2023, APSIPA Governor (2024–2026), IEEE SPS Distinguished Lecturer (2023-2024), ISCA Board Member (2021-2024), and Singapore’s 100 Women in Tech (2021). Previously, she worked at MIT Lincoln Lab during her PhD studies at MIT and Harvard, US.

This article was made for A*STAR Research by Wildtype Media Group