Artificial Intelligence (AI) is being used more in healthcare in the United States. The American Medical Association (AMA) says that in 2024, two out of three doctors (66%) use AI in their work. This is a big jump from 38% in 2023. AI helps doctors spend less time on paperwork, improve patient care through tools that assist with diagnosis, and make billing and records more accurate. The AMA survey found that 57% of doctors think AI’s best use is cutting down administrative work. Common uses of AI include automating billing documentation (used by 21% of doctors), creating discharge instructions and care plans (20%), translation services (14%), and diagnosis help (12%).
More doctors are trusting AI. In 2024, 35% of doctors were more excited than worried about AI, up from 30% in 2023. Those more worried than excited dropped from 29% to 25%. Still, about 40% of doctors have mixed feelings. Their main worries are about keeping patient data private, problems with connecting AI to Electronic Health Records (EHRs), and mistakes made by AI that could cause legal problems.
Data privacy is very important when using AI in healthcare. AI uses large amounts of patient information from sources like EHRs, manual entries, health information exchanges, and cloud storage. Doctors and healthcare groups must make sure AI follows strict privacy laws like HIPAA.
In 2024, about 87% of doctors said strong data privacy is needed for them to trust and use AI more. This worry comes because AI creators and other parties often get access to sensitive patient data to build and run AI. While these teams use data encryption and security, risks like unauthorized access, data leaks, and unclear data ownership still exist.
Healthcare leaders and IT staff should protect patient data by following key steps:
The HITRUST AI Assurance Program helps healthcare groups manage AI risks. It focuses on transparency, responsibility, and privacy. It follows standards from groups like the National Institute of Standards and Technology (NIST), which provides guidance for safe AI use.
EHR systems are a central part of most medical offices in the United States. For AI to be useful, it must connect smoothly to EHRs. The AMA survey says 84% of doctors believe this is necessary for AI to be widely used. When AI does not connect well, doctors may have to enter data twice, which wastes time and cancels out AI’s benefits.
Problems with connecting AI to EHRs include:
Managers and IT leaders should pick AI vendors who focus on working well with popular EHRs. Testing AI for easy data sharing, simple user interfaces, and low disturbances to current workflows helps doctors trust and use AI more.
AI in healthcare brings new legal questions. Doctors worry that wrong AI results might cause incorrect diagnoses or treatments. Liability concerns also include data breaches, misuse of patient information, and mistakes made by AI.
In 2023, 87% of doctors said they want protection from legal consequences caused by AI errors. Also, 86% want good medical liability insurance. The AMA says we need more rules to clarify who is responsible when AI affects patient care.
Medical leaders must work with legal teams and insurers to address:
AI can help automate many tasks in healthcare offices. Simbo AI, for example, uses AI to automate phone answering and scheduling. Automated systems can manage appointment booking, patient reminders, and simple questions without needing staff. This helps cut down on work and lets staff focus on tasks that need human attention.
In clinical work, AI automation supports:
Administrators and IT staff who want to set up AI automation should:
Using AI for routine tasks can lower costs and reduce paperwork, which is a big cause of frustration for doctors.
It is important to address doctor concerns about AI, especially about data privacy, EHR connection, and legal issues. As more doctors use AI, healthcare leaders must choose tools that protect patient data, work well with current systems, and handle legal and ethical matters carefully. This approach can help AI improve healthcare without creating extra risks.
In 2024, 66% of physicians reported using health care AI, a significant increase from 38% in 2023.
Physicians are using AI for various tasks including documentation of billing codes, medical charts, creation of care plans, translation services, and assistive diagnosis.
The sentiment towards AI has become more positive, with 35% of physicians expressing more enthusiasm than concerns, up from 30% in the previous year.
More than half of physicians, 57%, identified reducing administrative burdens through automation as the biggest area of opportunity for AI.
The most commonly cited task is the documentation of billing codes, medical charts, or visit notes, with 21% of physicians using AI for this in 2024.
Physicians are concerned about data privacy, potential flaws in AI-designed tools, integration with EHR systems, and increased liability concerns.
Physicians indicated that data privacy assurances, seamless integration, adequate training, and increased oversight are essential for building trust in AI.
The use of AI for the creation of discharge instructions, care plans, and progress notes increased to 20% in 2024, up from 14% in 2023.
The AMA advocates for making technology an asset to physicians, focusing on oversight, transparency, and defining the regulatory landscape for health AI.
In 2024, only 33% of physicians reported not using AI, a drastic decrease from 62% in 2023.