Government Responds to Growing Concerns Over Advanced AI, Plans Independent Assessments
Amid increasing apprehensions voiced by industry leaders regarding the perils associated with unrestrained advanced AI systems, the White House has acted.
Following the introduction of ChatGPT, the human-like chatbot developed by OpenAI last autumn, concerns over safety and accountability have been raised by experts and commentators. These concerns have gained momentum in recent times, especially in light of a series of remarkable AI breakthroughs, as leading AI entities such as Microsoft and Google engage in a competitive push to incorporate advanced AI technology across their products and services.
However, a prevailing worry centers on the swift progress of generative AI without the presence of regulatory oversight, potentially resulting in misuse, societal harm, or, as some of the most skeptical critics contend, even the endangerment of humanity. For instance, within the past ten days, current and former executives from OpenAI, Microsoft, and Google have joined a growing chorus of critics expressing concerns about the risks associated with unbridled AI development.
They joined more than 27,000 others -- including prominent industry figures Elon Musk and Steve Wozniak -- who have warned about AI dangers by signing an open letter that calls on all AI labs to immediately pause for at least six months the training of AI systems more powerful than GPT-4, the latest large language model (LLM) from OpenAI.
Now the federal government is addressing the issue on a number of fronts, including independent evaluations of advanced AI systems.
For example, Vice President Kamala Harris is today meeting with the CEOs of Microsoft, Google, OpenAI and Anthropic (another AI leader). Reuters reported the meeting invitation noted the "expectation that companies like yours must make sure their products are safe before making them available to the public."
Prior to that meeting, the White House published a fact sheet on other government initiatives to "promote responsible AI innovation that protects Americans' rights and safety," including:
- New investments to power responsible American AI research and development (R&D). The National Science Foundation announced $140 million in funding to launch seven new National AI Research Institutes. They catalyze collaborative efforts across institutions of higher education, federal agencies, industry, and others to pursue transformative AI advances that are ethical, trustworthy, responsible and serve the public good.
- Public assessments of existing generative AI systems. Anthropic, Google, Hugging Face, Microsoft, NVIDIA, OpenAI and Stability AI will participate in a public evaluation of AI systems at DEFCON 31, a hacking conference to be held in Las Vegas in August. The evaluations will explore how respective AI models align with the principles and practices outlined in the Biden-Harris Administration's Blueprint for an AI Bill of Rights and AI Risk Management Framework. "This independent exercise will provide critical information to researchers and the public about the impacts of these models, and will enable AI companies and developers to take steps to fix issues found in those models," the fact sheet said. "Testing of AI models independent of government or the companies that have developed them is an important component in their effective evaluation."
- Policies to ensure the U.S. government is leading by example on mitigating AI risks and harnessing AI opportunities. The Office of Management and Budget (OMB) will release draft policy guidance on the use of AI systems by the U.S. government for public comment, aiming to help establish specific policies for federal departments and agencies to follow in order to ensure their development, procurement, and use of AI systems centers on safeguarding the American people's rights and safety. "It will also empower agencies to responsibly leverage AI to advance their missions and strengthen their ability to equitably serve Americans -- and serve as a model for state and local governments, businesses and others to follow in their own procurement and use of AI," the fact sheet said. The draft guidance will be released this summer for public comment this summer in order to incorporate input from advocates, civil society, industry, and other stakeholders before finalization.
As far as today's meeting between Harris and other senior administrative officials with the CEOs of AI-leading companies, the White House published a background press call on the new AI announcements. It quotes a senior administration official as saying about that meeting: "We aim to have a frank discussions about the risks we see in current and near-term AI development. We're working with -- in this meeting, we're also aiming to underscore the importance of their role on mitigating risks and advancing responsible innovation, and we'll discuss how we can work together to protect the American people from the potential harms of AI so that they can reap the benefits of these new technologies."
David Ramel is an editor and writer for Converge360.