๐ฅ๐ถ๐๐ธ-๐ฏ๐ฎ๐๐ฒ๐ฑ ๐๐ ๐ฝ๐ผ๐น๐ถ๐ฐ๐/๐ฟ๐ฒ๐ด๐๐น๐ฎ๐๐ถ๐ผ๐ป ๐๐ผ๐๐ป๐ฑ๐ ๐ผ๐ฏ๐๐ถ๐ผ๐๐. ๐๐๐ ๐ถ๐ ๐ฟ๐ฒ๐๐๐ ๐ฎ ๐ณ๐ฟ๐ฎ๐ด๐ถ๐น๐ฒ ๐ฎ๐๐๐๐บ๐ฝ๐๐ถ๐ผ๐ป: ๐๐ต๐ฎ๐ ๐๐ฒ ๐ฐ๐ฎ๐ป ๐ฎ๐ฐ๐๐๐ฎ๐น๐น๐ ๐บ๐ฒ๐ฎ๐๐๐ฟ๐ฒ ๐ฟ๐ถ๐๐ธ.
Yet risk assessment is often put forward as the first thing to answer, as if itโs the easy part.
Some frameworks push โuse caseโbasedโ solutions with a pre-defined list of high-risk use cases, but then smuggle in the magic word โ๐๐ถ๐ด๐ป๐ถ๐ณ๐ถ๐ฐ๐ฎ๐ป๐โ โ effectively saying: the given high-risk use cases are only high-risk if it causesโฆ significant risk. A tautology that collapses under scrutiny. Or Look closely and โlow-riskโ use cases are never really low:
๐ญ. ๐ ๐๐๐ถ๐ฐ ๐ฟ๐ฒ๐ฐ๐ผ๐บ๐บ๐ฒ๐ป๐ฑ๐ฎ๐๐ถ๐ผ๐ป ๐๐ฒ๐ฒ๐บ๐ ๐ต๐ฎ๐ฟ๐บ๐น๐ฒ๐๐ ๐๐ป๐๐ถ๐น ๐ฏ๐ถ๐ฎ๐ ๐๐๐๐๐ฒ๐บ๐ฎ๐๐ถ๐ฐ๐ฎ๐น๐น๐ ๐๐ถ๐ฑ๐ฒ๐น๐ถ๐ป๐ฒ๐ ๐บ๐ถ๐ป๐ผ๐ฟ๐ถ๐๐ ๐ฎ๐ฟ๐๐ถ๐๐๐.
๐ฎ. ๐๐ฟ๐ฎ๐บ๐บ๐ฎ๐ฟ ๐ฐ๐ต๐ฒ๐ฐ๐ธ๐ถ๐ป๐ด ๐น๐ผ๐ผ๐ธ๐ ๐๐ฎ๐ณ๐ฒ ๐๐ป๐๐ถ๐น ๐ฎ ๐๐ถ๐ป๐ด๐น๐ฒ ๐ฑ๐ถ๐ด๐ถ๐ ๐ผ๐ฟ ๐๐ผ๐ฟ๐ฑ ๐ฐ๐ต๐ฎ๐ป๐ด๐ฒ ๐ถ๐ป ๐ฎ ๐บ๐ฒ๐ฑ๐ถ๐ฐ๐ฎ๐น ๐ป๐ผ๐๐ฒ ๐ฎ๐น๐๐ฒ๐ฟ๐ ๐ฎ ๐ฑ๐ถ๐ฎ๐ด๐ป๐ผ๐๐ถ๐ ๐ผ๐ฟ ๐ฑ๐ผ๐๐ฎ๐ด๐ฒ.
Yesterday at the Department of Health, I used Digital Scribe as a concrete example of navigating through some of the challenges.
At CSIRO’s Data61, weโre moving beyond vague ๐น๐ถ๐ธ๐ฒ๐น๐ถ๐ต๐ผ๐ผ๐ฑโconsequence matrices and building approaches that work in practice:
โข ๐ฃ๐ฟ๐ฒ๐ฐ๐ถ๐๐ฒ ๐ฑ๐ฒ๐ณ๐ถ๐ป๐ถ๐๐ถ๐ผ๐ป๐ of consequence, severity, scale, and impact
โข ๐ฆ๐ฒ๐ฝ๐ฎ๐ฟ๐ฎ๐๐ถ๐ป๐ด ๐ถ๐ป๐๐ฟ๐ถ๐ป๐๐ถ๐ฐ ๐ณ๐ฟ๐ผ๐บ ๐ฑ๐ฒ๐๐ถ๐ด๐ป ๐ฟ๐ถ๐๐ธ๐ โ whatโs intrinsic in any system (human, traditional software, or AI doing it) vs what emerges from AI-specific design choices
โข ๐๐๐ป๐ฎ๐บ๐ถ๐ฐ ๐๐ผ๐ผ๐น๐ that cut through noise and thousands of best practices to surface the few risks that matter and corresponding treatments
โข ๐ข๐๐ฒ๐ฟ๐๐ถ๐ด๐ต๐ ๐บ๐ผ๐ฑ๐ฒ๐น๐ that test effectiveness, not just presence
โข ๐ ๐ฎ๐ฟ๐ด๐ถ๐ป๐ฎ๐น ๐ฟ๐ถ๐๐ธ ๐ฐ๐ผ๐บ๐ฝ๐ฎ๐ฟ๐ถ๐๐ผ๐ป๐ โ evaluating AI against existing systems even without full ground truth
โข Applying all these in public administration, health, and high-stakes decision making, where decisions range from fact-finding to rule-applying to deep deliberation and discretionary judgment
Iโve shared a few redacted slides below. Always happy to discuss with others wrestling with the hardest question: if we want AI to be risk-based, how do we stop โriskโ from being a vague label?
https://www.linkedin.com/feed/update/urn:li:activity:7376738527293685762

