Mondo News

All of the latest tech and science news from all over the world.

Microsoft Calls for AI Rules to Minimize Risks

Its president, Brad Smith, said companies needed to “step up” and governments needed to “move faster” as artificial intelligence progressed.

David McCabe reports on tech policy from Washington.

“Companies need to step up,” Brad Smith, Microsoft’s president, said in an interview about the push for regulations. “Government needs to move faster.”

Lawmakers have publicly expressed worries that such A.I. products, which can generate text and images on their own, will create a flood of disinformation, be used by criminals and put people out of work. Regulators in Washington have pledged to be vigilant for scammers using A.I. and instances in which the systems perpetuate discrimination or make decisions that violate the law.

The maneuver echoes calls for new privacy or social media laws by internet companies like Google and Meta, Facebook’s parent. In the United States, lawmakers have moved slowly after such calls, with few new federal rules on privacy or social media in recent years.

In the interview, Mr. Smith said Microsoft was not trying to slough off responsibility for managing the new technology, because it was offering specific ideas and pledging to carry out some of them regardless of whether government took action.

He endorsed the idea, supported by Mr. Altman during his congressional testimony, that a government agency should require companies to obtain licenses to deploy “highly capable” A.I. models.

“That means you notify the government when you start testing,” Mr. Smith said. “You’ve got to share results with the government. Even when it’s licensed for deployment, you have a duty to continue to monitor it and report to the government if there are unexpected issues that arise.”

Microsoft, which made more than $22 billion from its cloud computing business in the first quarter, also said those high-risk systems should be allowed to operate only in “licensed A.I. data centers.” Mr. Smith acknowledged that the company would not be “poorly positioned” to offer such services, but said many American competitors could also provide them.

Microsoft added that governments should designate certain A.I. systems used in critical infrastructure as “high risk” and require them to have a “safety brake.” It compared that feature to “the braking systems engineers have long built into other technologies such as elevators, school buses and high-speed trains.”

In some sensitive cases, Microsoft said, companies that provide A.I. systems should have to know certain information about their customers. To protect consumers from deception, content created by A.I. should be required to carry a special label, the company said.

Mr. Smith said companies should bear the legal “responsibility” for harms associated with A.I. In some cases, he said, the liable party could be the developer of an application like Microsoft’s Bing search engine that uses someone else’s underlying A.I. technology. Cloud companies could be responsible for complying with security regulations and other rules, he added.

“We don’t necessarily have the best information or the best answer, or we may not be the most credible speaker,” Mr. Smith said. “But, you know, right now, especially in Washington D.C., people are looking for ideas.”

Category: Technology

Source: NYTimes Technology

Leave a Reply

Your email address will not be published. Required fields are marked *

%d bloggers like this: