Google announces new AI code of ethics
Web Log: Principles focus on socially beneficial AI and built-in privacy measures
Google has outlined a code of ethics it will abide by as it develops AI
“How AI is developed and used will have a significant impact on society for many years to come. As a leader in AI, we feel a deep responsibility to get this right,” said Pichai in an official Google blog post.
These seven principles outline that AI developed by Google should be socially beneficial and accountable to people while being built and tested for safety and absence of algorithmic bias.
The principles also focus on built-in privacy measures while adhering to high standards of scientific excellence and finally, making sure that the resulting AI technologies will not be used for purposes that fall outside these principles, ie Google won’t be selling off its AI or APIs to companies that might use them for unsavoury purposes.
“With some caveats, and recognising that the proof will be in their application by Google, we recommend that other tech companies consider adopting similar guidelines for their AI work,” commented Peter Eckersley, writing for the Electronic Frontier Foundation.