Why did the Pope launch the “AI Bible”?

AI For Business


Pope Francis, among other things, participated in a dialogue over the social impact of artificial intelligence (AI). The Pope’s commitment to AI ethics is the result of a collaboration between the Vatican and the Markula Center for Applied Ethics at Santa Clara University. Together they founded the Institute of Technology, Ethics and Culture (ITEC), an institution that mediates ethical concerns in the technology sector. This unexpected development has ripples all over the world. This demonstrates the Vatican’s determination to be an influential voice in the rapidly evolving field of AI. Already, the handbook Ethics in the Age of Disruptive Technology: An Operational Roadmap has been released.

This handbook advocates a synthesis of ethical values ​​rooted in principles that both technology and the organizations that develop it need to be ingrained from the outset. This handbook emphasizes that the “common good of mankind and the environment” must drive all corporate actions. The Grand Principles themselves are divided into seven main guidelines, including respect for human dignity and rights, and promotion of transparency and accountability. The guidelines are further detailed into 46 specific ‘Actionable Steps’, each accompanied by detailed definitions, illustrative examples and ‘Actionable Strategies’. Not bad for a church-led technology document, right?

The intersection of the Vatican, a centuries-old institution representing religion and spirituality, and the latest marvel of technology, AI, may seem a little strange to many. But the Vatican’s effort marks a long-standing interest in technology within the church coming to fruition. The Vatican enjoys a unique authority status (not only moral), and its influence garners appropriate seriousness and support for deliberating important issues concerning the future of technology and its development. I always felt that I could.

The handbook guidelines are well thought out. For example, the handbook emphasizes the importance of privacy and confidentiality under the guidelines of respect for human dignity and rights. It emphasizes the need to “do not collect more data than necessary” and advises that “collected data should be stored in a manner that optimizes protection of privacy and confidentiality.” In addition, it focuses on the responsibilities companies have to their users, as well as the fulfillment of legal requirements, and advocates specific safeguards for sensitive personal, medical and financial data.

The ethical challenges of AI applications are becoming increasingly apparent and subject to scrutiny. The use of various AI technologies can have unintended harmful consequences, such as invasion of privacy. Discrimination based on gender, race/ethnicity, sexual orientation, or gender identity. There are issues such as opaque decision-making. It has never been more important for humanity to address existing ethical challenges and build responsible and equitable AI innovations before deployment.

Going forward, AI must first and foremost learn how to eliminate data bias. Researchers cite an interesting example to illustrate this. The ImageNet database contains far more white faces than non-white faces. When experts train AI algorithms to recognize facial features using databases that don’t contain well-balanced faces, the algorithms don’t perform well on non-white faces, with huge social implications There is a possible built-in bias. impact.

One of the most important issues centers around AI control and morality. Consider drone technology. If you have a drone that can shoot rockets and potentially kill someone, you need a human to intervene in the decision-making process before the missile is deployed. But that is not the case, as drones are allowed to have their own minds. The problem is that AI is increasingly having to make split-second decisions. In high-frequency trading, more than 90 percent of all financial transactions are now driven by “algorithmic intelligence,” so there is no chance of humans controlling decision-making. The same is true for self-driving cars. If a child runs out onto the road, the car needs to react immediately, but AI lacks the human guidance (or reflexes, or instinct) about danger to human life at that critical moment, and it It is a necessary control mechanism for AI.

Ownership is another interesting aspect of AI. The question today is, when AI writes new music, who owns it? Who owns the intellectual property rights to it, and should AI potentially be rewarded for it?

Artificial intelligence bots are getting better at modeling human conversations and relationships. In 2015, a bot named Euse Goostman won the Turing His Challenge. In this challenge, human raters used text input to chat with an unknown entity and guess whether the chatter was a human or a machine. Eugene Goostman tricked more than half of human evaluators into thinking they were talking to humans. Will AI impersonate humans? Or become human. It is becoming an everyday reality.

No wonder the Pope is worried.

The author is MD of Rediffusion



Source link

Leave a Reply

Your email address will not be published. Required fields are marked *