Artificial intelligence (AI) is slowly becoming a tool used in medicine to help doctors take better care of patients. One important area where AI is helping is in making medical diagnoses. In the United States, hospital managers, clinic owners, and IT staff are interested in how AI can help healthcare workers, reduce their workload, and improve results. This article looks at recent studies from leading U.S. hospitals that compare AI models like Chat GPT Plus with traditional diagnostic methods. The research includes hundreds of doctors and clinical cases.
Recent studies from places like the University of Virginia, Stanford University, and Harvard Medical Center have worked together to see how Chat GPT Plus helps doctors diagnose patients. Chat GPT Plus is a type of large language model (LLM) AI.
One important finding is that AI alone can suggest diagnoses very well. A study at UVA Health included fifty doctors from family medicine, internal medicine, and emergency medicine. Chat GPT Plus by itself reached an accuracy of over 92%. This was better than doctors using traditional methods, who had an accuracy of 73.7%. Doctors who used Chat GPT Plus as a helper improved only a little, to 76.3% accuracy.
This shows AI can sometimes match or do better than doctors in making diagnoses. But when doctors combine AI advice with their own judgment, it doesn’t always improve results. The studies say doctors need more training to learn how to work well with AI and to trust it.
Studies that look closely at how doctors use AI tools show interesting points. A Stanford and UVA study found that doctors who used Chat GPT Plus and other AI tools improved their diagnostic accuracy only a little—from a score of 74 without AI to about 76 with AI.
However, using AI made the doctors faster. They finished their case reviews about 46 seconds quicker (519 seconds with AI vs. 565 seconds without). This time saved is important because doctors are very busy. Even small time savings can help patients get seen faster and reduce doctor burnout.
Ethan Goh, a researcher from Stanford, said that Chat GPT can make doctors’ work faster, and this time saved might be a good reason to use AI tools in the clinic.
But there is a limit. The study also found that doctors don’t always explain how they use AI to make decisions. Jonathan H. Chen, a Stanford professor on the study, said doctors often don’t explain their reasoning clearly. This can make it harder to use AI well. Healthcare groups should give doctors training and make clear ways to use AI suggestions.
The U.S. healthcare system has many specialty areas where correct diagnosis is very important. AI has also been tested outside general medicine. For example, in eye care, a study tested ChatGPT-3.5 and GPT-4.0 on 208 patient cases. GPT-4.0 did as well as, and sometimes better than, eye doctors in training.
GPT-4.0 showed 77.9% accuracy in recommending the right subspecialty based on patient history. The eye doctor trainees got 69.2% correct. GPT-4.0 also did well at catching serious diseases like glaucoma and lens problems, with 76% accuracy. GPT-4.0 made fewer errors than GPT-3.5. For example, completely wrong diagnoses went down from 43% to 17%, and partly correct ones went up from 11% to 50%.
This means advanced AI can help eye doctors and clinical teams with patient registration and diagnosis. But before fully trusting AI, hospitals need to keep checking AI’s accuracy and train their staff carefully.
One main idea from all the research is that doctors and medical staff need good training to use AI tools well. Studies suggest formal training on how to talk to AI and use it properly to get the best diagnostic help.
Dr. Andrew S. Parsons from UVA Health said that although AI helps doctors diagnose faster and better, training is needed for doctors to trust and use AI correctly. Right now, some doctors don’t fully trust AI or ignore its suggestions. This lowers the benefits AI can bring.
Building trust between doctors and AI is key to using AI more. Healthcare managers should add AI training to help workers feel confident and understand AI’s limits and strengths. This focus also keeps patient safety and doctor responsibility in mind, since AI should help doctors, not replace them.
AI can do more than improve diagnoses. It can also help manage clinic tasks. Companies like Simbo AI make AI tools that automate front-office jobs like handling phone calls and answering patient questions. These tools schedule appointments and respond to service questions quickly.
For clinic managers and IT staff, AI automation can reduce call volume and lower the workload on workers. This means staff can focus on harder jobs, feel less stressed, and work better. The time saved in diagnosis with AI is similar to the time saved in administrative work.
Simbo AI’s technology supports human staff by doing routine communication, not replacing people. This kind of AI helps early in the patient’s visit, right when they contact the clinic.
By combining front-office automation and AI diagnostic tools, medical centers can make the whole patient experience smoother—from check-in to diagnosis and care. This helps reduce wait times and holds on the phone.
In the U.S., AI tools like Chat GPT Plus, when used well, give medical workers a way to improve care quality and efficiency. Clinics that use AI might see:
Because healthcare is complex, hospital leaders and IT managers must plan carefully when adding AI. They need to protect patient data and fit AI into work flows. By doing this, AI can become a useful partner in healthcare management, improving care and operations.
The future of AI in healthcare depends on how well doctors and managers accept it. In the U.S., where healthcare workers face many challenges, AI tools like Chat GPT Plus and workflow automation from companies like Simbo AI offer new ways to improve care delivery.
Healthcare leaders should think about the benefits of AI but also focus on good training and careful testing. Investing in education, updating policies, and using automation together can help build stronger care systems. This helps improve diagnosis and patient care with reliable technology, supporting better health for patients across the country.
The study aimed to determine whether using Chat GPT Plus could improve the accuracy of doctors’ diagnoses compared to traditional methods.
Fifty physicians specializing in family medicine, internal medicine, and emergency medicine participated in the study.
One group used Chat GPT Plus for diagnoses, while the other relied on traditional resources like medical reference sites and Google.
The median diagnostic accuracy for the Chat GPT group was 76.3%, while the conventional methods group had 73.7%.
The Chat GPT group reached their diagnoses in a median time of 519 seconds, compared to 565 seconds for the conventional group.
Chat GPT Plus showcased a median diagnostic accuracy of over 92% when used by itself.
The study found that adding a human physician to the mix actually reduced diagnostic accuracy despite improved efficiency.
The researchers suggest that physicians will benefit from formal training on effectively using AI and prompts.
The ARiSE network aims to further evaluate AI outputs in healthcare to optimize their use in clinical environments.
The results were published in the scientific journal JAMA Network Open.