Electronic Health Records (EHR) have become an essential part of clinical workflows but consume a significant portion of physicians’ scheduled clinic time. Nearly half of all clinic time is now spent interacting with EHR systems, according to recent studies, leaving clinicians burdened with extensive clerical duties.
Added to this is the sharp rise in digital patient messages, which surged from 2020 to 2022, as more patients used tools like MyChart to communicate with their care teams.
This influx of messaging has led to increased clinician workload and stress, making it harder to maintain compassionate, human-centered care during busy clinical days.
Artificial intelligence (AI) is emerging as a solution with the potential to assist healthcare teams by automating routine communications and improving response efficiency.
Companies like Simbo AI specialize in front-office phone automation and answering services enhanced by AI, allowing practices to handle calls more efficiently while maintaining essential patient contact.
Yet, integrating AI in healthcare messaging brings important considerations about maintaining empathy, ensuring human oversight, and protecting patient privacy.
This article examines these issues from the perspective of healthcare administrators, owners, and IT managers who must balance innovation with patient-centered care.
The introduction of Electronic Health Records transformed documentation and communication within healthcare but has had unintended consequences for physician workload.
Approximately 50% of scheduled clinic time is consumed by EHR-related tasks.
Studies from Health Affairs and research published in JAMA Network Open show that clinicians spend hours each day managing inbox messages via patient portals such as MyChart.
The volume of these messages rose sharply during the COVID-19 pandemic and remains higher compared to pre-pandemic years.
Many messages received by clinicians contain complex or emotionally charged content. These include requests for medication refills, test results, appointment scheduling, and general questions.
More concerning, some inbox messages include aggressive language and personal attacks.
These contribute to clinician burnout and lower workplace satisfaction.
This volume and emotional load negatively affect well-being and can reduce time available for direct patient care.
Consequently, medical practice managers and healthcare IT specialists search for methods to help lighten this burden without sacrificing quality or patient trust.
AI technology, specifically generative AI (GenAI), is being tested for use in drafting replies to patient messages.
A pilot study conducted by the University of California San Diego involved 52 primary care physicians using GenAI-drafted replies with 70 physicians in a control group.
The study aimed to clarify if AI assistance could ease communication demands while maintaining care standards.
Physicians who used GenAI replies reported spending more time reading patient messages, which showed careful attention to patient concerns.
However, the average response time did not change much compared to those who responded on their own.
This suggests that AI tools might not reduce the total time spent communicating but could improve clinician understanding and message quality.
GenAI was especially helpful in mental health communications, where carefully worded responses are important.
Many physicians liked the AI-provided drafts as starting points for their replies.
However, they also said the AI-generated language sometimes sounded robotic and lacked the empathy needed in patient interactions.
Physicians stressed the need for ongoing human review and editing before sending responses—showing that AI should help, not replace, human communication.
The use of large language models (LLMs) in healthcare also brings important risks:
To address these issues, the study’s researchers recommended adding disclaimers with AI-generated messages to tell recipients about automation.
They also urged healthcare systems to set fair standards so all patients can access and benefit from AI tools.
For medical practice administrators and IT managers, AI offers a chance to make workflows smoother and improve efficiency.
Beyond just drafting patient messages, AI-powered systems can connect with Electronic Health Records and phone systems to automate common front-office tasks.
Simbo AI shows this approach with its front-office phone automation and AI-driven answering service made for healthcare settings.
This kind of AI lowers the workload of receptionists by answering frequently asked questions, scheduling appointments, and handling simple patient triage over the phone.
Automating these tasks can improve response times for patient inquiries and free staff to focus on harder tasks that need human judgment.
Linking this automated system with EHRs makes sure relevant patient info is accurate and updated without manual work.
Also, AI-based workflow automation helps manage the growing number of patient communications.
While doctors spend a lot of time reading and replying to clinical messages, AI tools can filter, prioritize, and draft first responses to less urgent requests like medication refills or paperwork.
This lowers the chance that important messages get missed and helps keep care continuous.
It is important that workflow automation solutions used in the U.S. healthcare system follow HIPAA rules to protect patient privacy.
Data inside AI platforms should be encrypted and access controlled to stop breaches.
Health systems using AI must work with vendors who know healthcare privacy standards and do risk checks before starting.
While AI brings operational benefits, keeping the human part is very important.
Research by Dr. Ming Tai-Seale at UC San Diego shows that doctors liked AI-drafted replies but needed to review them before sending.
This balance is especially needed in sensitive areas like mental health where patients need thoughtful and kind communication.
Healthcare leaders should create rules making sure AI-generated messages are closely checked and edited by clinical staff.
Training office teams on how to use AI tools and check replies helps keep quality high.
Telling patients about AI involvement helps keep trust and shows that a caring person is behind the final message.
At the same time, depending too much on AI can reduce the skills of clinicians.
Doctors must still get chances to talk with patients directly and write personal messages.
AI should support clinicians by doing routine tasks, but not take over their role as caring communicators.
Using AI in healthcare messaging creates important privacy concerns.
AI systems handle lots of patient data that, if leaked, could cause harm.
The study stresses the need for strong security rules when using AI-drafted replies and answering services.
Healthcare providers must make sure all AI vendors follow HIPAA and other rules.
Data encryption, secure cloud storage, strong login methods, and constant monitoring are needed to protect data.
Patients need to know how their info is used and kept and should have the option to say no to automated messages.
Regular checks of AI performance and privacy controls help find problems early.
Since patient trust is key to healthcare, protecting data is very important when adding new technology.
Medical leaders should work with IT security experts to build strong systems that watch data flows and lower risks.
The changing role of AI in healthcare messages needs clear rules and guidance at all levels.
Dr. Tai-Seale and others support creating fair policies to make sure all healthcare groups and patients can use AI safely.
Professional groups, healthcare systems, and regulators in the U.S. must make sure that:
By following these standards, healthcare providers can use AI tools like Simbo AI’s phone automation while keeping empathy, privacy, and care quality.
The use of AI in healthcare patient communication is growing.
These tools can reduce clerical work for clinicians and help front-office teams work better.
But it is important to keep the human parts that make care good.
Healthcare leaders and IT managers in U.S. medical practices must use AI carefully, balancing new technology with empathy, human review, and privacy.
With good planning and teamwork, AI can help build better, patient-focused workflows in today’s digital world.
EHR consumes nearly 50% of scheduled clinic time, significantly affecting physicians’ productivity and well-being.
The volume of patient messages in MyChart rose significantly from 2020 to 2022 and remains higher than pre-pandemic levels, contributing to physician burnout.
A study examined the association between GenAI-drafted replies and the time physicians spent answering messages, focusing on various types of patient communications.
Eligible messages included requests for refills, results, paperwork, and general questions that could be addressed with AI-generated replies.
Physicians using GenAI-drafted replies read patient messages more thoroughly, although their average reply time did not change.
Recent versions of GenAI provided more useful replies for mental health issues, pleasing physicians who valued them as compassionate starting points.
Physicians noted that while GenAI could help, it risked creating a robotic tone and emphasized the necessity for human oversight in communications.
Potential risks include loss of empathy, overreliance on AI, deskilling of clinicians, and privacy and security concerns.
The team plans to leverage LLMs for deeper analysis of patient-clinician communications and enhance understanding of mental health interactions.
A disclaimer was included with AI-generated replies to inform patients that part of the communication was automated, ensuring transparency in the process.