What an intellectually stimulating two days at the AIST’s AI International Symposium 2024 in Japan! ๐ฏ๐ต๐ง The symposium focused on the “Future Direction for Trustworthy and Responsible AI”. Day 2 also covered the latest progress in AI for Science.
Yoshua Bengio ‘s talk on “Towards Quantitative Safety Guarantees and AGI Alignment” was thought-provoking and sobering all at once. He explored how we can build an AI that holds multiple alternative theories/world models that fit with current observations/data, using it to avoid catastrophic outcomes when selecting the best/safest theories to proceed.
I had an interesting discussion with Bengio afterwards about whether we can/should limit the world models/theories that AI learns to a human-understandable level. If not, how can we trust AI when it explains a human-understandable version to us? Bengio had the answer, but I can’t share it here in case a mischievous AI is eavesdropping! ๐๐คซ
On a more optimistic note, I gave my talk on “๐๐ง๐ฌ๐ฎ๐ซ๐ข๐ง๐ ๐๐ ๐๐๐๐๐ญ๐ฒ: ๐ ๐๐ฒ๐ฌ๐ญ๐๐ฆ-๐๐๐ฏ๐๐ฅ ๐๐ฉ๐ฉ๐ซ๐จ๐๐๐ก ๐ญ๐จ ๐๐๐ฌ๐ญ๐ข๐ง๐ , ๐๐ซ๐๐ง๐ฌ๐ฉ๐๐ซ๐๐ง๐๐ฒ, ๐๐ง๐ ๐๐๐๐จ๐ฎ๐ง๐ญ๐๐๐ข๐ฅ๐ข๐ญ๐ฒ.” ๐๏ธ I emphasized CSIRO’s Data61 system-level and engineering approach beyond model training as the key to safe AI guarantees. I also highlighted how National AI Centre is developing Australia’s AI Safety standard that caters to AI deployers, small & medium enterprises, and prioritizes diversity, inclusion, and First Nations perspectives. ๐ฆ๐บ๐ The positive feedback and acknowledgment of Australia’s approach were heartening! โค๏ธ
book:ย https://lnkd.in/g9BCu6nn
research:ย https://lnkd.in/gyzjE4-i
One response to “AI Safety Talk at AIST AI International Symposium 2024”
Stop asking. ๐ note https://three-body-problem.fandom.com/wiki/Wallfacer (้ขๅฃ่ ) program