Healthcare data generally falls into two categories: structured and unstructured. Structured data includes clearly defined information such as lab results, coded diagnoses, and billing codes. These are easy to store in databases and analyze using standard software tools. However, according to recent research, structured data represents only 50% to 70% of relevant clinical information. The majority—about 80%—is unstructured, found in free-text clinical notes, physician narratives, pathology reports, imaging descriptions, and other non-standardized formats.
This imbalance has caused problems. Traditional clinical trial recruitment and research methods mostly rely on structured data. This limits patient identification and slows down study progress. For example, in clinical trial recruitment, relying only on structured data has caused delays of four to six months on average. Such delays can cost pharmaceutical sponsors millions per day—estimates range from $600,000 to $8 million per day in some phase III trials.
NLP (Natural Language Processing) and multi-modal AI technologies help fix these issues by pulling useful information from unstructured data. Being able to understand different kinds of data allows a better look at patient groups and supports clinical decisions more effectively.
Multi-modal AI means artificial intelligence systems that handle and combine different types of data at once. In healthcare, these data might include clinical text, medical images, lab results, genetic profiles, and more. Mixing many data types helps make better diagnoses and predictions.
A recent review looking at over 400 studies showed that multi-modal AI models work about 6.2% better in diagnostic accuracy than single-source models. These models give a fuller picture of patients compared to only using coded data.
In the United States, healthcare providers must deal with many kinds of electronic health record (EHR) systems and follow strict rules like HIPAA. Multi-modal AI helps by joining different types of data in real time and connecting datasets that are often separated.
One big problem in using multi-modal data is the lack of common data standards. To help with this, the Observational Medical Outcomes Partnership (OMOP) Common Data Model (CDM) was created. This model sets rules for how healthcare data should be collected and stored, making it easier for different systems to work together.
NLP tools change unstructured data into structured formats that fit OMOP CDM. For example, doctor notes about genetic changes or treatments can be turned into structured data. This lets data systems talk to each other across hospitals, research centers, and clinical trials.
In Europe, within the European Health Data Space (EHDS), systems like IOMED use NLP and automatic text mining to change unstructured clinical data from many hospitals into OMOP format. Similar methods are becoming important in the U.S., where the Office of the National Coordinator for Health Information Technology (ONC) leads efforts to improve data sharing.
Clinical trials test new drugs and treatments. But many trials fail or get delayed because patient recruitment is too slow or not effective. Traditional methods mostly use structured data and miss eligible patients whose important health details are in unstructured records.
Real examples show what advanced data extraction can do:
For U.S. medical administrators and IT managers, using technology that reads both structured and unstructured data means clinical research can move faster and be more accurate. This is very important in areas like cancer and rare diseases, where patients are few and hard to find.
Working with healthcare data means you must protect patient privacy and follow strict rules. Any use of AI in clinical research must follow laws like HIPAA and meet ethical standards.
IQVIA, a healthcare technology company, shows how AI can be used responsibly. Their Healthcare-grade AI™ platform uses privacy tools and complies with legal rules while analyzing large amounts of health data. IQVIA works with NVIDIA to use AI Foundry and multi-modal data extraction to safely access data that was hard to reach before.
Because U.S. laws are complex and health data is sensitive, medical managers should check the privacy features of any multi-modal data extraction tools before using them.
Using AI, including multi-modal data extraction, is changing how healthcare work gets done. Automating routine and hard tasks saves time and lets staff focus on important medical decisions.
AI-driven workflow automation includes:
Simbo AI, a company using AI for phone automation and answering, helps by making appointment scheduling and call handling automatic. This lowers the work load on office staff and frees them to focus on patient care and daily operations.
By combining multi-modal data extraction with AI workflow automation, medical offices and research groups can clear blockages, improve data quality, and get better results in both operations and research.
Outside clinical care, multi-modal data extraction with AI also helps drug discovery and research. IBM Research uses AI, quantum computing, and cloud systems to study patient data on a big scale.
For example, IBM’s AI methods looked at data from over 195,000 Parkinson’s patients and found existing drugs that might slow down dementia. This shows how AI can help find new uses for drugs and shorten research times.
Also, proteochemometric modeling—a machine learning method that studies how molecules and proteins interact—has sped up finding new drugs. Such models might lower costs and time needed to bring new medicines to market.
In the U.S., where drug and healthcare businesses are very active, using these technologies can give an edge. They help research work faster and bring better treatments to patients.
Even though multi-modal AI and data extraction have promise, they face some challenges in healthcare IT:
Still, research and partnerships show steady progress. Many groups are learning how to use multi-modal AI well in real settings.
Medical practice administrators and IT managers in the U.S. can gain from multi-modal data extraction in these ways:
Investing in these tools today fits the bigger goals of healthcare change in the U.S., like making data sharing better, precision medicine, and value-based care.
Medical administrators, owners, and IT managers in U.S. healthcare are at a point where using multi-modal data extraction is key for operations and clinical progress. These systems open up patient information hidden in unstructured data. That leads to better clinical research and patient care possibilities. By carefully adopting and using these technologies, U.S. medical practices can improve workflows, follow rules, and boost research—while helping healthcare improve faster.
The collaboration aims to accelerate the development of AI-powered Healthcare-grade AI solutions, enabling agentic automation of complex healthcare and life sciences workflows to improve efficiency, scalability, and patient outcomes throughout the therapeutic lifecycle.
IQVIA grounds its AI-powered capabilities in privacy, regulatory compliance, and patient safety, ensuring Healthcare-grade AI is trustworthy, reliable, and meets industry-specific standards for data protection and ethical use.
IQVIA offers unparalleled information assets, advanced analytics, domain expertise, and the IQVIA Connected Intelligence™ platform, which supplies high-quality healthcare data and insights critical for building effective AI solutions.
NVIDIA provides its AI Foundry service, NIM microservices, NeMo, DGX Cloud platform, and AI Blueprint for multi-modal data extraction, enabling the creation and optimization of custom AI agents specialized for healthcare and life sciences workflows.
AI agents will serve as digital companions to researchers, doctors, and patients, unlocking productivity, enhancing workflow automation, expanding access to care globally, and facilitating faster, data-driven decision-making.
AI agents are designed to automate and optimize thousands of complex, time-consuming workflows across the healthcare and life sciences therapeutic lifecycle, including research, clinical trials, and commercialization processes.
Healthcare-grade AI™ refers to AI engineered specifically to meet healthcare and life sciences needs, combining superior data quality, domain expertise, and advanced technology to deliver precise, scalable, and trustworthy insights and solutions.
By deploying NVIDIA AI Blueprint for multi-modal data extraction, the collaboration enables AI agents to access and leverage diverse data formats that were previously unreachable by traditional AI models, enriching analysis and insights.
The partnership accelerates innovation by automating workflows, enabling new operational models, improving data-driven decisions, and thereby shortening the time and cost required to bring treatments to market and improve patient outcomes.
IQVIA employs a variety of privacy-enhancing technologies and safeguards to protect individual patient information, ensuring large-scale data analysis is conducted ethically and securely without compromising privacy or regulatory compliance.