βοΈβ¨ Itβs been an incredible week in San Francisco discussing AI Safety with experts across AI labs, civil society, government, and academia. A few key learnings:
π ππ ππππππ² ππ§π₯π² ππππππ«π¬ ππ ππ‘π ππ²π¬πππ¦ πππ―ππ₯: Real capabilities emerge only when AI is tested in realistic environments with access to powerful tools, real-world settings, agentic planning, and ample inference time/self-correction. The new US-UK joint testing of Sonet-3.5 highlights this, particularly in tool use ablation analysis showing the capability difference.
βοΈ ππ§πππ«ππ§ππ ππ’π¦π ππππ₯π’π§π > ππ«ππ’π§π’π§π ππ’π¦π ππππ₯π’π§π ?: Scaling at inference time, including temporary model updates, raw compute, and agent iteration numbers, is shaping up to be potentially more powerful than just training scaling laws. This opens up new possibilities for βsovereign AI.β No one entity controls the full βsovereign AIβ stackβfocusing on strategic elements like inference capabilities/stack is more vital than ever for many.
π§ͺ ππ‘π πππ’ππ§ππ π¨π ππ ππππππ²: Robust evaluation is core to AI Safety. No amount of ad hoc red-teaming or probing will suffice. It demands a deep, multidisciplinary scientific approachβmeasurement science, risk quantification, and the study of complex systems.
All of this deeply resonates with CSIRO’s Data61 approach (which I presented at various occasions) to AI safety engineering https://lnkd.in/gPhid9tX:
β‘οΈ End-to-end system evaluation
β‘οΈ Focus on inference-time capabilities as key sovereign AI leverage
β‘οΈ Multidisciplinary scientific rigor for AI safety evaluation
And while in SF, I couldn’t miss full autonomous drivingβevery trip was in a Waymo! πβ¨ The ride? Smooth and confidently decisive, not what I had worried about being overly cautious or jerky. From the outside, itβs still a spectacle at timesβpeople smiling, staring, waving, taking photos.
But one intriguing thingβsome social behaviours around autonomous vehicles are emerging. Jaywalkers or people crossing backstreets seem more comfortable stepping in front of an AV than in front of a human-driven car. Is it that they trust the AV more or just don’t feel the need for politeness towards a machine? π€ Interesting times!
Leave a Reply