{ localUrl: '../page/uncontainability.html', arbitalUrl: 'https://arbital.com/p/uncontainability', rawJsonUrl: '../raw/9f.json', likeableId: '2449', likeableType: 'page', myLikeValue: '0', likeCount: '1', dislikeCount: '0', likeScore: '1', individualLikes: [ 'EliezerYudkowsky' ], pageId: 'uncontainability', edit: '2', editSummary: '', prevEdit: '1', currentEdit: '2', wasPublished: 'true', type: 'wiki', title: 'Cognitive uncontainability', clickbait: ''Cognitive uncontainability' is when we can't hold all of an agent's possibilities inside our own minds.', textLength: '3451', alias: 'uncontainability', externalUrl: '', sortChildrenBy: 'likes', hasVote: 'false', voteType: '', votesAnonymous: 'false', editCreatorId: 'AlexeiAndreev', editCreatedAt: '2015-12-16 16:13:34', pageCreatorId: 'EliezerYudkowsky', pageCreatedAt: '2015-07-01 19:47:16', seeDomainId: '0', editDomainId: 'EliezerYudkowsky', submitToDomainId: '0', isAutosave: 'false', isSnapshot: 'false', isLiveEdit: 'true', isMinorEdit: 'false', indirectTeacher: 'false', todoCount: '3', isEditorComment: 'false', isApprovedComment: 'true', isResolved: 'false', snapshotText: '', anchorContext: '', anchorText: '', anchorOffset: '0', mergedInto: '', isDeleted: 'false', viewCount: '781', text: '[summary: An [2s agent] is cognitively uncontainable in a [ domain], relative to us, when we can't hold all of the agent's possible strategies inside our own minds, and we can't make sharp predictions about what it can and can't do. For example, the most powerful modern chess programs, playing a human novice, would be cognitively uncontainable on the chess board (the novice can't imagine everything the chess program might do), but easily cognitively containable in the context of the larger world (the novice knows the chess program won't suddenly reach out and upset the board). One of the most critical [2c advanced agent properties] is if an agent is cognitively uncontainable in the real world.]\n\n[9g Vingean unpredictability] is when an agent is cognitively uncontainable because it is smarter than us: if you could predict in advance exactly where [ Deep Blue] would move, you could play chess at least as well as Deep Blue yourself by doing whatever you predicted Deep Blue would do in your shoes.\n\nAlthough Vingean unpredictability is the classic way in which cognitive uncontainability can arise, other possibilities are [9d imaginable]. For instance, the AI could be operating in a [9j rich domain] and searching a different part of the search space that humans have difficulty handling, while still being dumber or less competent overall than a human. In this case the AI's strategies might still be unpredictable to us, even while it was less effective or competent overall. Most [ anecdotes about AI algorithms doing surprising things] can be viewed from this angle.\n\nAn extremely narrow, exhaustibly searchable domain may yield cognitive containability even for intelligence locally superior to a human's. Even a perfect Tic-Tac-Toe player can only draw against a human who knows the basic strategies, because humans can also play perfect Tic-Tac-Toe. Of course this is only true so long as the agent can't modulate some transistors to form a wireless radio, escape onto the Internet, and offer a nearby bystander twenty thousand dollars to punch the human in the face - in which case the agent's strategic options would have included, in retrospect, things that affected the real world; and the real world is a much more complicated domain than Tic-Tac-Toe. There's some sense in which [9j richer domains] seem likely to feed into increased cognitive uncontainability, but it's worth remembering that every game and every computer is embedded into the extremely complicated real world.\n\n[2j Strong cognitive uncontainability] is when the agent knows some facts we don't, that it can use to formulate strategies that we wouldn't be able to recognize in advance as successful. From the perspective of e.g. the 11th century C.E. trying to cool their house, bringing in cool water from the nearby river to run over some nearby surfaces might be an understandable solution; but if you showed them the sketch of an air conditioner, without running the air conditioner or explaining how it worked, they wouldn't recognize this sketch as a smart solution because they wouldn't know the further facts required to see why it would work. When an agent can win using options that we didn't imagine, couldn't invent, and wouldn't understand even if we caught a glimpse of them in advance, it is strongly cognitively uncontainable in the same way that the 21st century is strongly uncontainable from the standpoint of the 11th century.', metaText: '', isTextLoaded: 'true', isSubscribedToDiscussion: 'false', isSubscribedToUser: 'false', isSubscribedAsMaintainer: 'false', discussionSubscriberCount: '1', maintainerCount: '1', userSubscriberCount: '0', lastVisit: '2016-02-23 21:10:15', hasDraft: 'false', votes: [], voteSummary: 'null', muVoteSummary: '0', voteScaling: '0', currentUserVote: '-2', voteCount: '0', lockedVoteType: '', maxEditEver: '0', redLinkCount: '0', lockedBy: '', lockedUntil: '', nextPageId: '', prevPageId: '', usedAsMastery: 'false', proposalEditNum: '0', permissions: { edit: { has: 'false', reason: 'You don't have domain permission to edit this page' }, proposeEdit: { has: 'true', reason: '' }, delete: { has: 'false', reason: 'You don't have domain permission to delete this page' }, comment: { has: 'false', reason: 'You can't comment in this domain because you are not a member' }, proposeComment: { has: 'true', reason: '' } }, summaries: {}, creatorIds: [ 'EliezerYudkowsky', 'AlexeiAndreev' ], childIds: [ 'rich_domain' ], parentIds: [ 'advanced_agent' ], commentIds: [], questionIds: [], tagIds: [], relatedIds: [], markIds: [], explanations: [], learnMore: [], requirements: [], subjects: [], lenses: [], lensParentId: '', pathPages: [], learnMoreTaughtMap: {}, learnMoreCoveredMap: {}, learnMoreRequiredMap: {}, editHistory: {}, domainSubmissions: {}, answers: [], answerCount: '0', commentCount: '0', newCommentCount: '0', linkedMarkCount: '0', changeLogs: [ { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '3909', pageId: 'uncontainability', userId: 'AlexeiAndreev', edit: '0', type: 'newAlias', createdAt: '2015-12-16 16:13:34', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '3910', pageId: 'uncontainability', userId: 'AlexeiAndreev', edit: '2', type: 'newEdit', createdAt: '2015-12-16 16:13:34', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '585', pageId: 'uncontainability', userId: 'AlexeiAndreev', edit: '1', type: 'newChild', createdAt: '2015-10-28 03:46:58', auxPageId: 'rich_domain', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '142', pageId: 'uncontainability', userId: 'AlexeiAndreev', edit: '1', type: 'newParent', createdAt: '2015-10-28 03:46:51', auxPageId: 'advanced_agent', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '1370', pageId: 'uncontainability', userId: 'EliezerYudkowsky', edit: '1', type: 'newEdit', createdAt: '2015-07-01 19:47:16', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' } ], feedSubmissions: [], searchStrings: {}, hasChildren: 'true', hasParents: 'true', redAliases: {}, improvementTagIds: [], nonMetaTagIds: [], todos: [], slowDownMap: 'null', speedUpMap: 'null', arcPageIds: 'null', contentRequests: {} }