From 2acebe0350261255fe6a0d204ed4c15685bb2e34 Mon Sep 17 00:00:00 2001 From: ravenscroftj Date: Mon, 19 Dec 2022 14:15:05 +0000 Subject: [PATCH] Add 'brainsteam/content/annotations/2022/12/19/1671458951.md' --- .../annotations/2022/12/19/1671458951.md | 73 +++++++++++++++++++ 1 file changed, 73 insertions(+) create mode 100644 brainsteam/content/annotations/2022/12/19/1671458951.md diff --git a/brainsteam/content/annotations/2022/12/19/1671458951.md b/brainsteam/content/annotations/2022/12/19/1671458951.md new file mode 100644 index 0000000..cf9bba9 --- /dev/null +++ b/brainsteam/content/annotations/2022/12/19/1671458951.md @@ -0,0 +1,73 @@ +--- +date: '2022-12-19T14:09:11' +hypothesis-meta: + created: '2022-12-19T14:09:11.863238+00:00' + document: + title: + - My AI Safety Lecture for UT Effective Altruism + flagged: false + group: __world__ + hidden: false + id: tmH8RH-mEe27ArstPwKXEA + links: + html: https://hypothes.is/a/tmH8RH-mEe27ArstPwKXEA + incontext: https://hyp.is/tmH8RH-mEe27ArstPwKXEA/scottaaronson.blog/?p=6823 + json: https://hypothes.is/api/annotations/tmH8RH-mEe27ArstPwKXEA + permissions: + admin: + - acct:ravenscroftj@hypothes.is + delete: + - acct:ravenscroftj@hypothes.is + read: + - group:__world__ + update: + - acct:ravenscroftj@hypothes.is + tags: + - nlproc + target: + - selector: + - endContainer: /div[2]/div[2]/div[2]/div[1]/p[43] + endOffset: 779 + startContainer: /div[2]/div[2]/div[2]/div[1]/p[43] + startOffset: 174 + type: RangeSelector + - end: 16443 + start: 15838 + type: TextPositionSelector + - exact: " And famously, self-driving cars have taken a lot longer than many people\ + \ expected a decade ago. This is partly because of regulatory barriers and\ + \ public relations: even if a self-driving car actually crashes less than\ + \ a human does, that\u2019s still not good enough, because when it does crash\ + \ the circumstances are too weird. So, the AI is actually held to a higher\ + \ standard. But it\u2019s also partly just that there was a long tail of\ + \ really weird events. A deer crosses the road, or you have some crazy lighting\ + \ conditions\u2014such things are really hard to get right, and of course\ + \ 99% isn\u2019t good enough here." + prefix: ' the last jobs to be automated. ' + suffix: ' + + + + + We can maybe fuzzily see ahe' + type: TextQuoteSelector + source: https://scottaaronson.blog/?p=6823 + text: I think the emphasis is wrong here. The regulation is secondary. The long + tail of weird events is the more important thing. + updated: '2022-12-19T14:09:11.863238+00:00' + uri: https://scottaaronson.blog/?p=6823 + user: acct:ravenscroftj@hypothes.is + user_info: + display_name: James Ravenscroft +in-reply-to: https://scottaaronson.blog/?p=6823 +tags: +- nlproc +- hypothesis +type: annotation +url: /annotations/2022/12/19/1671458951 + +--- + + + +
And famously, self-driving cars have taken a lot longer than many people expected a decade ago. This is partly because of regulatory barriers and public relations: even if a self-driving car actually crashes less than a human does, that’s still not good enough, because when it does crash the circumstances are too weird. So, the AI is actually held to a higher standard. But it’s also partly just that there was a long tail of really weird events. A deer crosses the road, or you have some crazy lighting conditions—such things are really hard to get right, and of course 99% isn’t good enough here.
I think the emphasis is wrong here. The regulation is secondary. The long tail of weird events is the more important thing. \ No newline at end of file