Recent research by the National Institutes of Health (NIH) looked at an advanced AI model called GPT-4V. This AI can understand both text and images. This helps it diagnose medical problems using clinical images and patient information. The study had GPT-4V answer 207 medical questions, including some that involved image diagnosis from the New England Journal of Medicine.
The results showed that the AI often did as well as or better than human doctors when answering questions without extra resources (closed-book setting). But when doctors had access to resources (open-book setting), they did better than the AI, especially in harder cases. This shows where AI diagnostic tools stand now and their possible role in medicine.
Doctors in the study said the AI made many correct diagnoses but had trouble explaining its thinking clearly. Sometimes, the AI misinterpreted images or failed to connect similar lesions from different views. This shows a difference between getting the right answer and fully understanding the medical problem. Because of this, AI cannot replace the experienced judgment of doctors yet.
Stephen Sherry, Ph.D., Acting Director of the National Library of Medicine (NLM) at NIH, said AI could help speed up diagnosis and help doctors start treatment faster. But he pointed out that human skill is still needed for accurate decisions and patient care. Many experts agree: AI is a helpful tool for healthcare teams but cannot replace human providers now.
Using AI in healthcare is not just about how well it works. There are many ethical, legal, and regulatory issues that must be handled. A review by Ciro Mennella and others, published by Elsevier Ltd., points out these challenges. They say these problems must be dealt with so AI can be used safely and well in clinics across the U.S.
One big concern is patient privacy. Healthcare systems must make sure AI tools keep patient information safe and follow laws like HIPAA. AI systems can also have bias. This happens if they are trained on data that does not represent all groups well. For example, if an AI is not trained with images from many types of patients, it might not work as well for some groups.
Another issue is informed consent. Patients and doctors need to know how AI helps with diagnosis or treatment. This keeps things clear and builds trust. If AI decisions are unclear, it could hurt the relationship between patients and healthcare providers.
Regulators are trying to find ways to approve AI medical tools. It is important to have standard rules to check AI safety, effectiveness, and responsibility after they are used. Without clear rules, healthcare places may find it hard to know which AI products are safe and legal.
Mennella and team suggest making a framework that covers ethics, laws, and continuous review. This can help healthcare managers and IT leaders in the U.S. use AI carefully and avoid problems from rushing or using AI the wrong way.
AI can do well on some diagnoses, but it still cannot handle all the complex parts of clinical decisions. Human experience, sound judgment, and deep knowledge are still very important in care.
Experts in the NIH studies say that AI’s problems with explaining its decisions or fully reading medical images show why doctors are essential. Hard cases often need a full review that goes beyond just matching patterns. Doctors use many years of training to think about a patient’s history, physical exam results, social background, and subtle symptoms.
AI and human doctors should work together. AI can quickly process large amounts of data, suggest possible diagnoses, and spot image patterns that might be missed. Doctors check, interpret, and talk to patients while considering both ethics and feelings.
For clinic owners and managers, this means AI should be seen as a helper, not a replacement. Policies should require that humans always review AI results. This protects against depending too much on AI, which can cause trouble if the AI makes mistakes that go unnoticed.
Apart from helping with medical decisions, AI also helps automate front-office and administrative tasks. Simbo AI is a company that makes AI phone systems to reduce work for office staff. Their tools aim to improve patient experiences and make operations run better.
In busy medical offices, front desk phone calls take up a lot of time. Staff handle appointment bookings, patient questions, and insurance checks on the phone. AI phone systems can manage these tasks by understanding the caller and responding right away. This frees staff to handle harder work.
For practice managers and IT staff, AI front-office tools offer clear benefits:
For U.S. medical offices handling many patients and complex schedules, these AI tools can help improve overall management and quality.
However, automation must work securely with current healthcare IT systems and meet all rules about patient privacy and data safety. Using a Human-in-the-Loop (HITL) approach is smart. This means having staff watch the AI and step in if needed to handle unusual cases or keep patients satisfied.
Healthcare leaders thinking about AI must carefully weigh its pros and cons. NIH studies show AI can speed up diagnosis and improve accuracy in some tasks but is not fully reliable yet. Ethics and regulations are still developing, which makes careful use important.
Key points for U.S. healthcare decision-makers include:
By keeping these factors in mind, medical practice owners and managers in the U.S. can make smart choices about AI. This can support better healthcare while keeping patient safety and ethics important.
Adding AI to healthcare is still a work in progress. It needs careful attention to both how well it works and how it affects daily practice. AI models like GPT-4V may help with diagnosis, but human judgment is still needed. At the same time, AI tools like those from Simbo AI can improve front-office work by automating phone tasks. When used with proper oversight, ethical care, and rules, AI can help make healthcare more efficient and improve patient care.
The NIH study found that the AI model GPT-4V performed well in diagnosing medical images but struggled with explaining its reasoning, highlighting both its potential and limitations in clinical settings.
The AI selected correct diagnoses more frequently than physicians in closed-book settings, while physicians using open-book resources performed better, particularly on difficult questions.
The AI often misinterpreted medical images and failed to correlate conditions despite accurate diagnoses, demonstrating gaps in its interpretative capabilities.
It’s crucial to assess AI’s strengths and weaknesses to understand its role in improving clinical decision-making and ensure effective integration into healthcare.
The study was led by researchers from NIH’s National Library of Medicine (NLM) in collaboration with several prestigious medical institutions including Weill Cornell Medicine.
The tested model was GPT-4V, a multimodal AI capable of processing both text and image data, relevant to diagnosing medical conditions.
NLM supports biomedical informatics and data science research, aiming to improve the processing, storage, and communication of health information.
Despite AI’s capabilities, human experience is essential for accurately diagnosing patients, as AI may lack contextual understanding necessary for correct interpretations.
Further research is required to compare AI capabilities with those of human physicians to fully understand its potential in clinical settings.
The findings suggest that while AI can enhance diagnosis speed, its current limitations necessitate careful evaluation before widespread implementation in healthcare.