AI technologies bring benefits to healthcare operations and patient care. A survey by Becker’s Healthcare shows that nearly 68% of healthcare workers use AI, and 19% use it every day. AI helps providers by making tasks like claims processing and prior authorization faster, which are usually slow and can have mistakes. It also helps with diagnostics by analyzing medical images and patient data to find diseases early.
AI could save a lot of money. Private payers might save between $80 billion and $110 billion each year. Physician groups could reduce costs by 3% to 8%, saving between $20 billion and $60 billion. These numbers show how much money AI could save if used well.
Hospitals like Jackson Health System in Miami use custom AI to better use resources and keep their finances steady for years. Joe DiMaggio Children’s Hospital uses AI that records and summarizes doctor-patient talks. This reduces the work doctors have to do on paperwork and gives them more time to care for patients.
These examples show how AI can help with efficiency and costs, but adopting AI is not without problems, especially in different healthcare settings.
One big problem for healthcare leaders is the lack of clear rules about AI technologies. Without clear laws, it is hard to know who is responsible if something goes wrong. As AI starts to affect clinical decisions and administration, providers need to understand these legal worries.
The U.S. Department of Justice (DOJ) is watching AI misuse in healthcare more closely. Deputy Attorney General Lisa O. Monaco said stronger punishments may come for crimes made worse by AI. For example, in 2020, the Practice Fusion case showed how AI was used unethically. The company admitted to taking bribes to program electronic health records to encourage opioid prescribing, focusing on profit instead of patient safety.
The American Medical Association has called for more rules on how AI is used in prior authorization. AI can make this process faster, but some worry AI might wrongly reject valid claims, limiting doctor choices and hurting patient care. Lawsuits involving payors like United Healthcare and Humana over their AI use show that courts are paying close attention to AI decisions in healthcare.
The Department of Health and Human Services (HHS) made a new rule starting January 1, 2025. It requires AI product makers to share clear information about how they create, test, and check their AI for bias. The Office of the National Coordinator for Health Information Technology (ONC) will certify these disclosures. This aims to make AI tools more clear and responsible in clinical use.
Healthcare leaders and IT managers must keep up with these rules and think about them when choosing AI products. Checking vendors carefully and understanding AI’s limits and rules is now very important.
AI in healthcare depends a lot on the data used to teach its algorithms. If training data does not include all groups of people, AI predictions can be biased and wrong. For example, if African American or some female patients are missing from the data, AI might give weaker or unfair advice for them.
Researcher Min Chen says this is an important problem. Biased AI can increase unfairness in healthcare, especially in the U.S. where patients come from many backgrounds. This can lower trust in AI tools and make healthcare less equal.
There are also worries about patient data privacy and security. AI needs a lot of patient data, so questions come up about getting permission, how data is used, and protecting it from being stolen. Health centers must have clear rules to keep patient trust and follow laws like HIPAA.
A big problem with AI is making sure its results are accurate and reliable. Experts warn against relying only on AI in clinical decisions. Alberto Jacir of CANO Health says AI predictions must be checked by humans carefully to avoid mistakes like missing a diagnosis or making wrong ones.
Using AI in diagnostics has potential but risks remain if faulty algorithms affect decisions. For example, AI could mistake tumors in images, which might lead to wrong treatments. Hospitals and doctors should use AI results as help, not as the only answer.
This is important because legal cases for malpractice may rise if bad AI causes harm. The healthcare field needs rules to balance AI help with ongoing human checks.
Although AI promises savings, many healthcare groups face money problems when adopting AI. Rural systems often have less funding and find it hard to pay for AI despite its future benefits.
AI also struggles because electronic medical record (EMR) systems don’t always work well together. Without standard ways to share data, AI can’t use all patient information fully, limiting its usefulness. Miriam Weismann from FIU Business says lack of interoperability is a big problem that makes AI adoption harder and costlier.
Healthcare leaders must balance money spent on AI with other urgent needs. Planning carefully is needed to choose AI uses that improve work or patient care right away.
AI can help automate front-office and admin tasks in medical offices. AI phone systems and answering services reduce work for staff, better patient contact, and smoother scheduling and follow-ups.
Companies like Simbo AI use AI to handle common patient calls, book appointments, refill prescriptions, and check insurance. Using natural language processing, these systems understand what callers want and answer quickly without needing a person for every call.
Using AI workflow systems offers benefits:
Still, adopting AI automation needs careful checks to make sure it fits the office workflow and follows healthcare rules, including privacy laws.
Because of complex rules and technology, healthcare leaders must carefully check AI vendors. Kate Driscoll stresses the need to confirm the openness, accuracy, and safety of AI tools.
Vendors should share clear information about how their algorithms are created, where data comes from, how bias is handled, and test results. Many healthcare groups do not have the tech skills to check these themselves, so working with trusted vendors who meet ONC certifications is helpful.
Also, after AI is introduced, continuous checking of its performance helps find and fix problems fast. This lowers risks like fraud, misuse, or tech failures that could hurt patient care or cause legal issues.
Healthcare leaders must train staff to work well with AI tools. Training should cover what AI can and can’t do, ethical issues, and how to use AI properly in daily work.
Ashwin Kumar Singh from Jackson Health System says custom AI made to fit an organization’s needs works better than generic products. Staff knowledge supports smooth use of AI and keeps patients safe.
Organizations should have clear rules about using AI, ethical standards, privacy protection, and ways to check AI results. Regular sharing of information and clear communication with patients about AI’s role helps build trust.
Even though more healthcare workers use AI, patients still have mixed feelings. A Pew Research Center survey found that 60% of Americans are unsure about letting AI decide their medical care.
Building patient trust means being open about how AI helps doctors and making clear that final decisions are made by humans. Healthcare practices should explain AI’s supportive role and reassure patients about privacy and data protection.
Healthcare administrators and IT managers in the U.S. face challenges when bringing AI technologies into their work. By understanding these risks and planning carefully, they can use AI advantages while lowering problems. Thoughtful use of AI can lead to smarter operations, better clinical help, and improved patient care in healthcare settings across the country.
AI can streamline clinical operations, automate mundane tasks, and assist in diagnosing life-threatening diseases, thus improving efficiency and patient outcomes.
Risks include misuse for fraud, algorithmic bias, and reliance on faulty AI tools which may lead to improper clinical decisions or denial of legitimate insurance claims.
Government enforcers are developing measures to deter AI misuse, including monitoring compliance with existing laws and using guidelines from past prosecutions to inform their actions.
AI can make the prior authorization process more efficient, but it raises concerns about whether legitimate claims may be unfairly denied and if it undermines physician discretion.
AI can analyze medical data and images to identify diseases and recommend treatments, but its effectiveness hinges on the integrity and training of the models used.
The case serves as a cautionary tale showing how AI tools can be exploited for profit by influencing clinical decision-making at the expense of patient care.
While AI can expedite drug development, there is a risk of manipulating data to overstate efficacy, leading to serious consequences and potential violations of federal laws.
Proper vetting is necessary to ensure accuracy, transparency, and compliance with regulatory requirements, as healthcare providers often lack the technical expertise to assess AI tools.
The ONC requires AI vendors to disclose development processes, data training, bias prevention measures, and validation of their products to ensure compliance and accountability.
Companies should maintain strong vetting, monitoring, auditing, and investigation practices to mitigate risks associated with AI technologies and prevent fraud and abuse.