acm-header
Sign In

Communications of the ACM

ACM News

Defense Department Issuing AI Ethics Guidelines for Tech Contractors


View as: Print Mobile App Share:
Aerial view of the Pentagon.

The work could change how artificial intelligence is developed by the U.S. government, if the DoDs guidelines are adopted or adapted by other departments.

Credit: Getty Images

In 2018, when Google employees found out about their company's involvement in Project Maven, a controversial US military effort to develop AI to analyze surveillance video, they weren't happy. Thousands protested. "We believe that Google should not be in the business of war," they wrote in a letter to the company's leadership. Around a dozen employees resigned. Google did not renew the contract in 2019.

Project Maven still exists, and other tech companies, including Amazon and Microsoft, have since taken Google's place. Yet the US Department of Defense knows it has a trust problem. That's something it must tackle to maintain access to the latest technology, especially AI—which will require partnering with Big Tech and other nonmilitary organizations.

In a bid to promote transparency, the Defense Innovation Unit, which awards DoD contracts to companies, has released what it calls "responsible artificial intelligence" guidelines that it will require third-party developers to use when building AI for the military, whether that AI is for an HR system or target recognition.

From MIT Technology Review
View Full Article

 


 

No entries found

Sign In for Full Access
» Forgot Password? » Create an ACM Web Account