Navigation
Recherche
|
AI Pioneer Announces Non-Profit To Develop 'Honest' AI
mercredi 4 juin 2025, 01:00 , par Slashdot
![]() Describing the current suite of AI agents as 'actors' seeking to imitate humans and please users, he said the Scientist AI system would be more like a 'psychologist' that can understand and predict bad behavior. 'We want to build AIs that will be honest and not deceptive,' Bengio said. He added: 'It is theoretically possible to imagine machines that have no self, no goal for themselves, that are just pure knowledge machines -- like a scientist who knows a lot of stuff.' However, unlike current generative AI tools, Bengio's system will not give definitive answers and will instead give probabilities for whether an answer is correct. 'It has a sense of humility that it isn't sure about the answer,' he said. Deployed alongside an AI agent, Bengio's model would flag potentially harmful behaviour by an autonomous system -- having gauged the probability of its actions causing harm. Scientist AI will 'predict the probability that an agent's actions will lead to harm' and, if that probability is above a certain threshold, that agent's proposed action will then be blocked. 'The point is to demonstrate the methodology so that then we can convince either donors or governments or AI labs to put the resources that are needed to train this at the same scale as the current frontier AIs. It is really important that the guardrail AI be at least as smart as the AI agent that it is trying to monitor and control,' he said. Read more of this story at Slashdot.
https://slashdot.org/story/25/06/03/2149233/ai-pioneer-announces-non-profit-to-develop-honest-ai?utm...
Voir aussi |
56 sources (32 en français)
Date Actuelle
ven. 6 juin - 23:52 CEST
|