Artificial Intelligence (AI) technologies are being used more and more in many areas, like healthcare. These technologies help by automating tasks, making better decisions, and improving efficiency. But AI also comes with risks that need careful handling. To manage these risks, governments and organizations have made AI Risk Management Frameworks (AI RMFs). These frameworks guide the responsible design, use, and deployment of AI systems. They aim to promote transparency, safety, security, and fairness across different fields.
In the United States, public and private groups work together to create voluntary frameworks that build trust in AI systems. This teamwork tries to lower AI risks while still supporting new ideas. For people who manage medical practices, healthcare owners, and IT managers, it is important to know these frameworks and how they affect healthcare and technology services. This article talks about how clear and cooperative methods in AI risk management are shaping partnerships between public and private groups in the U.S. It also looks at how AI-driven workflow automation fits into these efforts, especially in healthcare.
The National Institute of Standards and Technology (NIST) leads the work on creating AI risk management standards for the U.S. On January 26, 2023, NIST released its AI Risk Management Framework (AI RMF) after working openly with people from private companies, government, universities, and community groups. This was done through public comments, workshops, and draft releases starting in 2021. The framework is voluntary, so organizations can decide if they want to use it to make their AI systems more trustworthy. No laws require using it.
NIST’s AI RMF focuses on managing risks AI might create for people, organizations, society, and the environment. It includes several trust principles such as validity, safety, security, accountability, transparency, explainability, privacy, fairness, and reducing harmful bias. These should be considered throughout the AI system’s whole life—from design to deployment and ongoing checks.
The open process NIST used to develop the AI RMF is important. It helps many groups understand risks and find ways to avoid them. Bringing in views from varied sectors adds different kinds of knowledge and addresses many challenges faced by AI creators, users, and evaluators. For healthcare administrators and IT staff, this open work means there is clear and simple guidance when adding AI tools to workflows or patient care.
NIST also offers a Playbook and a Trustworthy and Responsible AI Resource Center, launched on March 30, 2023. These help those using the framework to follow best practices and keep updated with new research or tools for checking AI risks.
Besides NIST, the Department of Homeland Security (DHS) created a “Roles and Responsibilities Framework for Artificial Intelligence in Critical Infrastructure” in November 2024. This was made with help from industry leaders, community groups, and government officials. The framework defines clear, voluntary roles for all involved in the AI supply chain: cloud providers, AI developers, operators of critical infrastructure, community members, and public agencies.
The DHS Framework is important for critical infrastructure like energy, water, transport, and communications. These services are key for all Americans, including healthcare facilities that depend on power, communication, and IT every day. AI systems here must resist attacks, avoid design mistakes, and be clear and accountable.
The framework tells AI developers to follow “Secure by Design” rules. This means building security and safety into AI from the start and regularly checking for weaknesses or unsafe actions. Infrastructure operators must use strong cybersecurity, watch AI systems’ work, and clearly talk to the public about AI. Cloud and computing providers must keep their environments safe and protect important infrastructure from misuse or attack.
The DHS framework supports teamwork by defining roles but does not require adoption. These shared efforts create trust and open communication lines, which are key to managing AI risks well.
Clear and cooperative AI risk management frameworks make partnerships between government and private organizations stronger. These partnerships are very important in healthcare, where medical practice managers and health IT leaders often work with outside AI vendors and tech providers. This teamwork helps make sure AI tools work safely and fairly in strictly controlled healthcare settings.
Leaders like Marc Benioff, CEO of Salesforce, and Dr. Rumman Chowdhury, CEO of Humane Intelligence, support the DHS Framework. They say that clear shared responsibilities and voluntary promises from stakeholders build trust and promote responsible AI use.
In healthcare technology, partnerships help match AI development goals with rules and patient safety needs. Public agencies give guidance, fund research, and sometimes supervise rules. Private companies bring technical knowledge, operational know-how, and market solutions suited to healthcare.
The open model used by NIST also encourages ongoing input. Medical groups can give feedback, share problems, and help shape AI management best practices through public comments or workshops. This ongoing input is valuable because AI changes fast and risks shift with new uses.
One big effect of AI in healthcare is on workflow automation. Tools like AI front-office phone systems, scheduling, and virtual assistants improve how work gets done and help with patient communication. Simbo AI, for example, offers AI-based phone answering that handles common calls. This lets staff focus more on clinical and patient care.
When healthcare places want to use these AI tools, understanding and using AI risk management frameworks is key. Workflow automation affects:
Because these AI tools interact directly with patients, they play a key role in healthcare. Using NIST and DHS framework principles when choosing, deploying, and checking these systems helps lower risks from failures, privacy problems, or communication errors.
Also, teamwork between healthcare providers and AI vendors benefits from these risk models. Both sides can share duties for compliance, data safety, and watching how well the systems work over time.
Healthcare providers in the U.S. work in a complex system of rules that focus on patient safety, privacy, and fair access. Public agencies, including federal ones, play a big part in supporting workable frameworks for AI use. By backing voluntary risk management models, these agencies offer resources without forcing strict rules. This lets healthcare groups of all sizes adopt AI tools carefully and at their own speed.
This method respects the challenges faced by medical practice administrators who must balance patient care, staff duties, and rules. The frameworks provide advice that fits both small clinics and large hospital systems. Small practices might begin AI risk steps by focusing on privacy and security checks. Larger systems could carry out full evaluations covering transparency, explainability, and bias management.
Public sector help also includes promoting new ideas through research funding and field testing AI tools in real situations. This helps make sure AI development matches healthcare needs while keeping safety and risk control in mind.
For medical practice administrators and IT managers in the U.S., AI risk management frameworks give useful ways to check and use AI tools, like automated phone answering and workflow automation.
As AI tools become common in healthcare management and services, frameworks like those from NIST and DHS guide safe use and operation. Clear and cooperative development builds trust between AI makers, healthcare providers, government, and patients. Voluntary frameworks give medical practice owners and IT managers practical help to meet ethical, legal, and operational needs.
Knowing these frameworks supports effective teamwork between healthcare groups and AI tech providers. This helps make sure new tools serve patients well without risking safety or privacy. Continued cooperation between public and private sectors is important to protect critical infrastructure and key healthcare services as AI moves forward.
The AI RMF is designed to help individuals, organizations, and society manage risks related to AI. It promotes trustworthiness in the design, development, use, and evaluation of AI products, services, and systems through a voluntary framework.
It was created through an open, transparent, and collaborative process involving public comments, workshops, and a Request for Information, ensuring a consensus-driven approach with input from both private and public sectors.
The AI RMF was initially released on January 26, 2023.
NIST published a companion AI RMF Playbook, an AI RMF Roadmap, a Crosswalk, and Perspectives to facilitate understanding and implementation of the framework.
Launched on March 30, 2023, this Center aids in implementing the AI RMF and promotes international alignment with the framework, offering use cases and guidance.
On July 26, 2024, NIST released NIST-AI-600-1, a Generative AI Profile that identifies unique risks of generative AI and proposes targeted risk management actions.
No, the AI RMF is intended for voluntary use to improve AI risk management and trustworthiness.
It builds on and supports existing AI risk management efforts by providing an aligned, standardized framework to incorporate trustworthiness considerations.
NIST provides a public commenting process on draft versions and Requests for Information to gather input from various stakeholders during framework development.
The goal is to cultivate trust in AI technologies, promote innovation, and mitigate risks associated with AI deployment to protect individuals, organizations, and society.