The Biden administration is exploring a “bill of rights” to regulate facial recognition and other potentially harmful uses of artificial intelligence, but AI’s problems loom larger than the way a new technology is regulated.
big picture: There is no good way to regulate the role of AI in shaping a fair and equitable society without deciding how power should be balanced between individuals, corporations and government.
running news: White House Office of Science and Technology Policy launched a fact-finding mission OSTP director Eric Lander and his deputy Alondra Nelson wrote in a Op-Ed published by Wired yesterday.
What are they saying: Sanjay Gupta, global head of product and corporate development at Mitek Systems, a pioneer in digital identity verification, says, “Before it is too late, the conversation about what is acceptable – and unacceptable – about AI and our personal data can begin. It’s important to do.” .
- “Companies will still find agreed ways to innovate and integrate with these technologies,” he said.
AI’s biggest booster can be a victim technical solutionism – Expecting technology to efficiently solve structural, social problems.
- yes but: At the same time, however, focusing too narrowly on the applications of AI risks an opposite techno-solutionism – the belief that the fastest way to fix social problems is to change the technologies that affect them. do, and not because of the often complex issues that affect them.
reminder: The Fundamental Rights Bill is almost 230 years old, and we are still debating the meaning of each of its roughly 652 words.
- If the AI Bill of Rights is our ultimate goal, we are still in the bargaining stage on the Articles of Confederation.