Biden Announces New Actions to Promote Responsible AI Innovation (04/05/2023)
Home ยป Law Library Updates ยป Sarvarthapedia ยป National ยป Biden Announces New Actions to Promote Responsible AI Innovation (04/05/2023)
Today, the use of automated systems, including those sometimes marketed as
โartificial intelligenceโ or โAI,โ is becoming increasingly common in our daily lives. We
use the term โautomated systemsโ broadly to mean software and algorithmic processes,
including AI, that are used to automate workflows and help people complete tasks or
make decisions. Private and public entities use these systems to make critical decisions
that impact individualsโ rights and opportunities, including fair and equal access to a job,
housing, credit opportunities, and other goods and services. These automated systems
are often advertised as providing insights and breakthroughs, increasing efficiencies
and cost-savings, and modernizing existing practices. Although many of these tools
offer the promise of advancement, their use also has the potential to perpetuate
unlawful bias, automate unlawful discrimination, and produce other harmful outcomes. [JOINT STATEMENT ON ENFORCEMENT EFFORTS AGAINST DISCRIMINATION AND BIAS IN AUTOMATED SYSTEMS-MAY 2023]
FACT SHEET: Biden-โ Harris Administration Announces New Actions to Promote Responsible AI Innovation that Protects Americansโ Rights andย Safety
May 04, 2023
Today, the Biden-Harris Administration is announcing new actions that will further promote responsible American innovation in artificial intelligence (AI) and protect peopleโs rights and safety. These steps build on the Administrationโs strong record of leadership to ensure technology improves the lives of the American people, and break new ground in the federal governmentโs ongoing effort to advance a cohesive and comprehensive approach to AI-related risks and opportunities.
AI is one of the most powerful technologies of our time, but in order to seize the opportunities it presents, we must first mitigate its risks. President Biden has been clear that when it comes to AI, we must place people and communities at the center by supporting responsible innovation that serves the public good, while protecting our society, security, and economy. Importantly, this means that companies have a fundamental responsibility to make sure their products are safe before they are deployed or made public.
Vice President Harris and senior Administration officials will meet today with CEOs of four American companies at the forefront of AI innovationโAlphabet, Anthropic, Microsoft, and OpenAIโto underscore this responsibility and emphasize the importance of driving responsible, trustworthy, and ethical innovation with safeguards that mitigate risks and potential harms to individuals and our society. The meeting is part of a broader, ongoing effort to engage with advocates, companies, researchers, civil rights organizations, not-for-profit organizations, communities, international partners, and others on critical AI issues.
This effort builds on the considerable steps the Administration has taken to date to promote responsible innovation. These include the landmarkย Blueprint for an AI Bill of Rightsย andย related executive actionsย announced last fall, as well as theย AI Risk Management Frameworkย and aย roadmap for standing up a National AI Research Resourceย released earlier this year.
The Administration has also taken important actions to protect Americans in the AI age. In February, President Biden signed anย Executive Orderย that directs federal agencies to root out bias in their design and use of new technologies, including AI, and to protect the public from algorithmic discrimination. Last week, the Federal Trade Commission, Consumer Financial Protection Bureau, Equal Employment Opportunity Commission, and Department of Justiceโs Civil Rights Division issued aย joint statementย underscoring their collective commitment to leverage their existing legal authorities to protect the American people from AI-related harms.
The Administration is also actively working to address the national security concerns raised by AI, especially in critical areas like cybersecurity, biosecurity, and safety. This includes enlisting the support of government cybersecurity experts from across the national security community to ensure leading AI companies have access to best practices, including protection of AI models and networks.
Todayโs announcements include:
- New investments to power responsible American AI research and development (R&D).ย The National Science Foundation is announcing $140 million in funding to launch seven new National AI Research Institutes. This investment will bring the total number of Institutes to 25 across the country, and extend the network of organizations involved into nearly every state. These Institutes catalyze collaborative efforts across institutions of higher education, federal agencies, industry, and others to pursue transformative AI advances that are ethical, trustworthy, responsible, and serve the public good. In addition to promoting responsible innovation, these Institutes bolster Americaโs AI R&D infrastructure and support the development of a diverse AI workforce. The new Institutes announced today will advance AI R&D to drive breakthroughs in critical areas, including climate, agriculture, energy, public health, education, and cybersecurity.
- Public assessments of existing generative AI systems.ย The Administration is announcing an independent commitment from leading AI developers, including Anthropic, Google, Hugging Face, Microsoft, NVIDIA, OpenAI, and Stability AI, to participate in a public evaluation of AI systems, consistent with responsible disclosure principlesโon an evaluation platform developed by Scale AIโat the AI Village at DEFCON 31. This will allow these models to be evaluated thoroughly by thousands of community partners and AI experts to explore how the models align with the principles and practices outlined in the Biden-Harris Administrationโs Blueprint for an AI Bill of Rights and AI Risk Management Framework. This independent exercise will provide critical information to researchers and the public about the impacts of these models, and will enable AI companies and developers to take steps to fix issues found in those models. Testing of AI models independent of government or the companies that have developed them is an important component in their effective evaluation.
- Policies to ensure the U.S. government is leading by example on mitigating AI risks and harnessing AI opportunities.ย The Office of Management and Budget (OMB) is announcing that it will be releasing draft policy guidance on the use of AI systems by the U.S. government for public comment. This guidance will establish specific policies for federal departments and agencies to follow in order to ensure their development, procurement, and use of AI systems centers on safeguarding the American peopleโs rights and safety. It will also empower agencies to responsibly leverage AI to advance their missions and strengthen their ability to equitably serve Americansโand serve as a model for state and local governments, businesses, and others to follow in their own procurement and use of AI. OMB will release this draft guidance for public comment this summer, so that it will benefit from input from advocates, civil society, industry, and other stakeholders before it is finalized.