07/22/2024 | Press release | Distributed by Public on 07/22/2024 13:30
AI systems must be accountable, transparent, fair, explainable, and ethical, it is often demanded. But according to professor Pınar Yolum, that is not realistic. Instead, she advocates for trustworthy AI: systems that demonstrate certain properties that allow people to trust their results. For example, by sometimes answering, "I don't know." Yolum delivered her inaugural lecture on July 2.
AI has become an integral part of our daily lives. This brings about wonderful things, such as improving healthcare, facilitating complex calculation or planning tasks, or reducing energy consumption and food waste. But there is also a downside: AI can exhibit biases against certain groups and be used for wrong purposes, such as warfare and the spread of misinformation.
Where does it go wrong? According to Pınar Yolum, the newly appointed professor of Trustworthy AI at the department of Computing & Information Sciences, it mainly lies in the AI techniques we develop. In her inaugural lecture on Tuesday, July 2, she argued that the systems we build today may resemble how humans reason and argue, but they do not replicate how humans attempt to do these tasks. Yolum: "AI systems can perform complex calculations much faster than humans. But at the same time it is tricky, because they can't apply all the checks and balances we do as humans. This means that AI can make mistakes that humans, with their broader understanding of situations, would likely avoid."
In response, many people say that we need to develop AI that is accountable, transparent, fair, explainable, and ethical. According to Yolum, this is not realistic: "For some reason, we accept institutions that are not transparent, people who do not act ethically, and a world that is not fair. But when it comes to AI, the systems must possess all these golden properties. That is an unattainable standard to live up to."
Instead, Yolum advocates for trustworthy AI. "Just as people show they are trustworthy by exhibiting certain behaviors, doing certain things, or refraining from others, an AI system should act in the same way." According to Yolum, this can be achieved by giving AI more response options, such as "I don't know". "A phrase I wish men would say more often," jokes Yolum. "Current AI systems are designed to provide an answer. If there is no clear answer, the system chooses quite arbitrarily, increasing the likelihood of providing an incorrect answer and thus becoming less reliable. If AI systems better understand their expertise or limitations, they can hand certain tasks back to the user."
We accept institutions that are not transparent and a world that is not fair. But AI systems must possess all golden properties?
For instance, Yolum's group developed an AI app that acts as a personal privacy assistant on your phone or computer. "This system learns the user's privacy preferences and can then make recommendations, such as not sharing certain photos or locations. The algorithm can also explain, both in text and visually, why the user might want to keep a particular photo private." The researchers found that when the system could say "I don't know" for certain photos and leave the choice to the user, the other predictions became more accurate and thus more reliable.
Yolum envisions future AI as a trustworthy employee to whom you delegate tasks because they have demonstrated reliability, integrity, and helpfulness. "We are currently in a situation where we have AI perform tasks that we then have humans check because mistakes are made. I believe we should move towards AI that we can confidently assign tasks to, knowing that the system will say 'I don't know' if there are uncertainties. Then you truly collaborate with AI as a team, with everyone doing what they do best."
According to Yolum, it is a political choice to develop AI in this way. "Computer science is a field where most funding comes from industry. And at the end of the day, the industry says, 'It's nice that you're helping people better protect their privacy, but what's in it for me? I want to collect and process data.' I therefore hope that in the future, more public funding will be available for this type of research."
According to Yolum, there are five response options that trustworthy AI must be able to provide: