FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER.

REUTERS. VP Harris tells Microsoft, Google they have legal responsibility to ensure safety of AI products. 04 MAY 2023

“As I shared today with CEOs of companies at the forefront of American AI innovation, the private sector has an ethical, moral, and legal responsibility to ensure the safety and security of their products,” Harris said in a statement.

FOR EDUCATIONAL PURPOSES

WASHINGTON, May 4 (Reuters) – U.S. Vice President Kamala Harris on Thursday told the chief executives of tech companies including Microsoft (MSFT.O) and Google (GOOGL.O) they have a “legal responsibility” to ensure the safety of their artificial intelligence products.

Harris and President Biden are open to advancing new regulations and supporting new legislation on artificial intelligence, she said.

The White House is hosting CEOs of top artificial intelligence companies on Thursday to discuss risks and safeguards as the technology catches the attention of governments and lawmakers globally.

NEWS. VP Harris tells tech CEOs they have ‘moral, legal responsibility’ for AI safety. France 24

FOR EDUCATIONAL PURPOSES

US Vice President Kamala Harris on Thursday told the chief executives of tech companies including Microsoft and Google they have a “legal responsibility” to ensure the safety of their artificial intelligence products.

“As I shared today with CEOs of companies at the forefront of American AI innovation, the private sector has an ethical, moral, and legal responsibility to ensure the safety and security of their products,” Harris said in a statement.

Harris and President Joe Biden are open to advancing new regulations and supporting new legislation on artificial intelligence, she said.

The White House is hosting CEOs of top artificial intelligence companies on Thursday to discuss risks and safeguards as the technology catches the attention of governments and lawmakers globally.

Generative artificial intelligence has become a buzzword this year, with apps such as ChatGPT capturing the public’s fancy, sparking a rush among companies to launch similar products they believe will change the nature of work.

Millions of users have begun testing such tools, which supporters say can make medical diagnoses, write screenplays, create legal briefs and debug software, leading to growing concern about how the technology could lead to privacy violations, skew employment decisions and power scams and misinformation campaigns.

“We aim to have a frank discussion about the risks we see in current and near-term AI development,” said a senior administration official, speaking on the condition of anonymity because of the sensitivity of the matter. “Our North Star here is this idea that if we’re going to seize these benefits, we have to start by managing the risks.”

Thursday’s meeting will include Google’s Sundar Pichai, Microsoft’s Satya Nadella, OpenAI’s Sam Altman and Anthropic’s Dario Amodei along with Vice President Kamala Harris and administration officials including Biden’s Chief of Staff Jeff Zients, National Security Adviser Jake Sullivan, Director of the National Economic Council Lael Brainard and Secretary of Commerce Gina Raimondo.

Ahead of the meeting, the administration announced a $140 million investment from the National Science Foundation to launch seven new AI research institutes and said the White House’s Office of Management and Budget would release policy guidance on the use of AI by the federal government.

Leading AI developers, including Anthropic, Google, Hugging Face, NVIDIA, OpenAI, and Stability AI, will participate in a public evaluation of their AI systems at the AI Village at DEFCON 31 – one of the largest hacker conventions in the world – and run on a platform created by Scale AI and Microsoft.

Shortly after Biden announced his reelection bid, Republican National Committee produced a video featuring a dystopian future during a second Biden term, that was built entirely with AI imagery.

Such political ads are expected to become more common as AI technology proliferates.

United States regulators have fallen short of the tough approach European governments have taken on tech regulation and in crafting strong rules on deep fakes and misinformation that companies must follow or risk hefty fines.

“We don’t see this as a race,” the administration official said, adding that the administration is working closely with the US-EU Trade & Technology Council on the issue.

In February, Biden signed an executive order directing federal agencies to eliminate bias in their use of AI. The Biden administration has also released an AI Bill of Rights and a risk management framework.

Last week, the Federal Trade Commission and Department of Justice’s Civil Rights Division also said they would use their legal authorities to fight AI-related harm.

Tech giants have vowed many times to combat propaganda around elections, fake news about the COVID-19 vaccines, racist and sexist messages, pornography and child exploitation, and hateful messaging targeting ethnic groups.

But they have been unsuccessful, research and news events show. Just about one in five fake news articles in English on six major social media platforms were tagged as misleading or removed, a recent study by activist NGO Avaaz found, and articles in other European languages were not flagged.

FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER.