{ localUrl: '../page/dwim.html', arbitalUrl: 'https://arbital.com/p/dwim', rawJsonUrl: '../raw/2s1.json', likeableId: '1698', likeableType: 'page', myLikeValue: '0', likeCount: '2', dislikeCount: '0', likeScore: '2', individualLikes: [ 'EricBruylant', 'EricRogstad' ], pageId: 'dwim', edit: '9', editSummary: '', prevEdit: '8', currentEdit: '9', wasPublished: 'true', type: 'wiki', title: 'Do-What-I-Mean hierarchy', clickbait: 'Successive levels of "Do What I Mean" or AGIs that understand their users increasingly well', textLength: '5376', alias: 'dwim', externalUrl: '', sortChildrenBy: 'likes', hasVote: 'false', voteType: '', votesAnonymous: 'false', editCreatorId: 'EliezerYudkowsky', editCreatedAt: '2016-06-06 19:38:45', pageCreatorId: 'EliezerYudkowsky', pageCreatedAt: '2016-03-23 20:21:56', seeDomainId: '0', editDomainId: 'EliezerYudkowsky', submitToDomainId: '0', isAutosave: 'false', isSnapshot: 'false', isLiveEdit: 'true', isMinorEdit: 'false', indirectTeacher: 'false', todoCount: '0', isEditorComment: 'false', isApprovedComment: 'true', isResolved: 'false', snapshotText: '', anchorContext: '', anchorText: '', anchorOffset: '0', mergedInto: '', isDeleted: 'false', viewCount: '271', text: '[summary: "Do What I Mean" or "DWIM" refers to the degree to which an AGI can rapidly [2s3 identify] an intended goal and find a [2l safe] plan to it, based on the AI's understanding of what the user *means* or *wants.*\n\nLevels of DWIM-ness could range over:\n\n- Having a general idea of which parts of the world the user thinks are significant (so that the AI warns about [2pf impacts] on significant things);\n- Having a psychological model of the user's beliefs, and flagging/reporting when the AI thinks the user has a false belief about the consequences of a plan;\n- Having a psychological model of the user's desires, and trying to fulfill what the AI thinks the user *wants* to accomplish by giving the AGI a task;\n- At the extreme end: [313 Extrapolated volition] models of what the user(s) *would* want\\* under idealization conditions.]\n\nDo-What-I-Mean refers to an aligned AGI's ability to produce better-aligned plans, based on an explicit model of what the user wants or believes.\n\nSuccessive levels of DWIM-ness:\n\n- No understanding of human intentions / zero DWIMness. E.g. a Task AGI that is focused on one task being communicated, where all the potential [2pf impacts] of that task need to be separately [2qq vetted]. If you tell this kind of AGI to 'cure cancer', you might need to veto plans which would remove the cancer but kill the patient as a side effect, because the AGI doesn't start out knowing that you'd prefer not to kill the patient.\n- Do What You Don't Know I Dislike. The Task AGI has a background understanding of some human goals or which parts of the world humans consider especially significant, so it can more quickly generate a plan likely to pass human [2qq vetting]. A Task AGI at this level, told to cure cancer, will take relatively fewer rounds of Q&A to generate a plan which carefully seals off any blood vessels cut by removing the cancer; because the AGI has a general notion of human health, knows that [2pf impacts] on human health are significant, and models that users will generally prefer plans which result in good human health as side effects rather than plans which result in poor human health.\n- Do What You Know I Understood. The Task AGI has a model of human *beliefs,* and can flag and report divergences between the AGI's model of what the humans expect to happen, and what the AGI expects to happen.\n- DWIKIM: Do What I Know I Mean. The Task AGI has an explicit psychological model of human preference - not just a list of things in the environment which are significant to users, but a predictive model of how users behave which is informative about their preferences. At this level, the AGI can read through a dump of online writing, build up a model of human psychology, and guess that you're telling it to cure a cancer because you altruistically want that person to be healthier.\n- DWIDKIM: Do What I Don't Know I Mean. The AGI can perform some basic [3c5 extrapolation] steps on its model of you and notice when you're trying to do something that, in the AGI's model, some further piece of knowledge might change your mind about. (Unless we trust the DWIDKIM model a *lot*, this scenario should imply "Warn the user about that" not "Do what you think the user would've told you.")\n- (Coherent) Extrapolated Volition. The AGI does what it thinks you (or everyone) would've told it to do if you were as smart as the AGI, i.e., your decision model is extrapolated toward improved knowledge, increased ability to consider arguments, improved reflectivity, or other transforms in the direction of a theory of normativity.\n\nRisks from pushing toward higher levels of DWIM might include:\n\n- To the extent that DWIM can originate plans, some portion of which are not fully supervised, then DWIM is a very complicated goal or preference system that would be harder to train and more likely to break. This failure mode may be less likely if some level of DWIM is *only* being used to *flag* potentially problematic plans that were generated by non-DWIM protocols, rather than generating plans on its own.\n- Accurate predictive psychological models of humans might push the system closer to the [10f programmer deception] failure mode being more accessible if something else goes wrong.\n- Sufficiently advanced psychological models might constitute [6v mindcrime].\n- The human-genie system might end up in the [2s8 Valley of Dangerous Complacency] where the genie *almost* always gets it right but occasionally gets it very wrong, and the human user is no longer alert to this possibility during the [2qq checking phase].\n - E.g. you might be tempted to skip the user checking phase, or just have the AI do whatever it thinks you meant, at a point where that trick only works 99% of the time and not 99.999999% of the time.\n- Computing sufficiently advanced DWIDKIM or [3c5 EV] possibilities for user querying might expose the human user to cognitive hazards. ("If you were sufficiently superhuman under scenario 32, you'd want yourself to stare really intently at this glowing spiral for 2 minutes, it might change your mind about some things... want to check and see if you think that's a valid argument?")\n- If the AGI was actually behaving like a safe genie, the sense of one's wishes being immediately fulfilled without effort or danger might expose the programmers to additional [2sb moral hazard].', metaText: '', isTextLoaded: 'true', isSubscribedToDiscussion: 'false', isSubscribedToUser: 'false', isSubscribedAsMaintainer: 'false', discussionSubscriberCount: '1', maintainerCount: '1', userSubscriberCount: '0', lastVisit: '', hasDraft: 'false', votes: [], voteSummary: 'null', muVoteSummary: '0', voteScaling: '0', currentUserVote: '-2', voteCount: '0', lockedVoteType: '', maxEditEver: '0', redLinkCount: '0', lockedBy: '', lockedUntil: '', nextPageId: '', prevPageId: '', usedAsMastery: 'false', proposalEditNum: '0', permissions: { edit: { has: 'false', reason: 'You don't have domain permission to edit this page' }, proposeEdit: { has: 'true', reason: '' }, delete: { has: 'false', reason: 'You don't have domain permission to delete this page' }, comment: { has: 'false', reason: 'You can't comment in this domain because you are not a member' }, proposeComment: { has: 'true', reason: '' } }, summaries: {}, creatorIds: [ 'EliezerYudkowsky', 'DamonPourtahmasebSasi' ], childIds: [], parentIds: [ 'safe_plan_identification' ], commentIds: [], questionIds: [], tagIds: [], relatedIds: [ 'cev' ], markIds: [], explanations: [], learnMore: [], requirements: [], subjects: [], lenses: [], lensParentId: '', pathPages: [], learnMoreTaughtMap: {}, learnMoreCoveredMap: {}, learnMoreRequiredMap: {}, editHistory: {}, domainSubmissions: {}, answers: [], answerCount: '0', commentCount: '0', newCommentCount: '0', linkedMarkCount: '0', changeLogs: [ { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '19830', pageId: 'dwim', userId: 'DamonPourtahmasebSasi', edit: '10', type: 'newEdit', createdAt: '2016-10-03 05:35:47', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '11830', pageId: 'dwim', userId: 'EliezerYudkowsky', edit: '9', type: 'newEdit', createdAt: '2016-06-06 19:38:45', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '11820', pageId: 'dwim', userId: 'EliezerYudkowsky', edit: '0', type: 'deleteTag', createdAt: '2016-06-06 18:08:54', auxPageId: 'stub_meta_tag', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '10912', pageId: 'dwim', userId: 'EliezerYudkowsky', edit: '8', type: 'newEdit', createdAt: '2016-05-25 20:39:42', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '9422', pageId: 'dwim', userId: 'EliezerYudkowsky', edit: '7', type: 'newUsedAsTag', createdAt: '2016-04-27 02:19:39', auxPageId: 'cev', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '9417', pageId: 'dwim', userId: 'EliezerYudkowsky', edit: '7', type: 'newEdit', createdAt: '2016-04-27 02:14:54', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '9416', pageId: 'dwim', userId: 'EliezerYudkowsky', edit: '6', type: 'newEdit', createdAt: '2016-04-27 02:13:03', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '9283', pageId: 'dwim', userId: 'EliezerYudkowsky', edit: '5', type: 'newEdit', createdAt: '2016-04-14 00:22:36', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '9282', pageId: 'dwim', userId: 'EliezerYudkowsky', edit: '4', type: 'newEdit', createdAt: '2016-04-14 00:21:10', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '9281', pageId: 'dwim', userId: 'EliezerYudkowsky', edit: '3', type: 'newEdit', createdAt: '2016-04-14 00:18:59', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '8983', pageId: 'dwim', userId: 'EliezerYudkowsky', edit: '2', type: 'newEdit', createdAt: '2016-03-23 22:38:14', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '8974', pageId: 'dwim', userId: 'EliezerYudkowsky', edit: '1', type: 'newParent', createdAt: '2016-03-23 22:17:54', auxPageId: 'safe_plan_identification', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '8972', pageId: 'dwim', userId: 'EliezerYudkowsky', edit: '0', type: 'deleteParent', createdAt: '2016-03-23 22:17:45', auxPageId: 'task_agi', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '8958', pageId: 'dwim', userId: 'EliezerYudkowsky', edit: '1', type: 'newEdit', createdAt: '2016-03-23 20:21:56', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '8957', pageId: 'dwim', userId: 'EliezerYudkowsky', edit: '0', type: 'newTag', createdAt: '2016-03-23 20:07:52', auxPageId: 'stub_meta_tag', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '8955', pageId: 'dwim', userId: 'EliezerYudkowsky', edit: '0', type: 'newParent', createdAt: '2016-03-23 20:06:51', auxPageId: 'task_agi', oldSettingsValue: '', newSettingsValue: '' } ], feedSubmissions: [], searchStrings: {}, hasChildren: 'false', hasParents: 'true', redAliases: {}, improvementTagIds: [], nonMetaTagIds: [], todos: [], slowDownMap: 'null', speedUpMap: 'null', arcPageIds: 'null', contentRequests: {} }