{
localUrl: '../page/nonperson_predicate.html',
arbitalUrl: 'https://arbital.com/p/nonperson_predicate',
rawJsonUrl: '../raw/1fv.json',
likeableId: '404',
likeableType: 'page',
myLikeValue: '0',
likeCount: '3',
dislikeCount: '0',
likeScore: '3',
individualLikes: [
'AlexeiAndreev',
'EricBruylant',
'NathanFish'
],
pageId: 'nonperson_predicate',
edit: '1',
editSummary: '',
prevEdit: '0',
currentEdit: '1',
wasPublished: 'true',
type: 'wiki',
title: 'Nonperson predicate',
clickbait: 'If we knew which computations were definitely not people, we could tell AIs which programs they were definitely allowed to compute.',
textLength: '1317',
alias: 'nonperson_predicate',
externalUrl: '',
sortChildrenBy: 'likes',
hasVote: 'false',
voteType: '',
votesAnonymous: 'false',
editCreatorId: 'EliezerYudkowsky',
editCreatedAt: '2015-12-28 19:49:00',
pageCreatorId: 'EliezerYudkowsky',
pageCreatedAt: '2015-12-28 19:49:00',
seeDomainId: '0',
editDomainId: 'EliezerYudkowsky',
submitToDomainId: '0',
isAutosave: 'false',
isSnapshot: 'false',
isLiveEdit: 'true',
isMinorEdit: 'false',
indirectTeacher: 'false',
todoCount: '0',
isEditorComment: 'false',
isApprovedComment: 'true',
isResolved: 'false',
snapshotText: '',
anchorContext: '',
anchorText: '',
anchorOffset: '0',
mergedInto: '',
isDeleted: 'false',
viewCount: '168',
text: 'A "nonperson predicate" is a possible method for preventing an [2c advanced AI] from [6v accidentally running sapient computations] (it would be a potentially huge moral catastrophe if an AI created, ran, and discarded a large number of sapient programs inside itself). A nonperson predicate looks at potential computations and returns one of two possible answers, "Don't know" and "Definitely not a person". A successful nonperson predicate may (very often) return "Don't know" for computations that aren't in fact people, but it never returns "Definitely not a person" for something that *is* a person. In other words, to solve this problem, we don't need to know what consciousness *is* so much as we need to know what it *isn't* - we don't need to be sure what *is* a person, we need to be sure what *isn't* a person. For a nonperson predicate to be useful, however, it must still pass enough useful computations that we can build a working, capable AI out of them. (Otherwise "Rocks are okay, everything else might be a person" would be an adequate nonperson predicate.) The [6r foreseeable difficulty] of a nonperson predicate is that [10k instrumental pressures] to model humans accurately might tend to [42 seek out flaws and loopholes] in any attempted predicate. See the page on [6v] for more detail.',
metaText: '',
isTextLoaded: 'true',
isSubscribedToDiscussion: 'false',
isSubscribedToUser: 'false',
isSubscribedAsMaintainer: 'false',
discussionSubscriberCount: '1',
maintainerCount: '1',
userSubscriberCount: '0',
lastVisit: '2016-02-09 15:44:59',
hasDraft: 'false',
votes: [],
voteSummary: 'null',
muVoteSummary: '0',
voteScaling: '0',
currentUserVote: '-2',
voteCount: '0',
lockedVoteType: '',
maxEditEver: '0',
redLinkCount: '0',
lockedBy: '',
lockedUntil: '',
nextPageId: '',
prevPageId: '',
usedAsMastery: 'false',
proposalEditNum: '0',
permissions: {
edit: {
has: 'false',
reason: 'You don't have domain permission to edit this page'
},
proposeEdit: {
has: 'true',
reason: ''
},
delete: {
has: 'false',
reason: 'You don't have domain permission to delete this page'
},
comment: {
has: 'false',
reason: 'You can't comment in this domain because you are not a member'
},
proposeComment: {
has: 'true',
reason: ''
}
},
summaries: {},
creatorIds: [
'EliezerYudkowsky'
],
childIds: [],
parentIds: [
'mindcrime'
],
commentIds: [],
questionIds: [],
tagIds: [],
relatedIds: [],
markIds: [],
explanations: [],
learnMore: [],
requirements: [],
subjects: [],
lenses: [],
lensParentId: '',
pathPages: [],
learnMoreTaughtMap: {},
learnMoreCoveredMap: {},
learnMoreRequiredMap: {},
editHistory: {},
domainSubmissions: {},
answers: [],
answerCount: '0',
commentCount: '0',
newCommentCount: '0',
linkedMarkCount: '0',
changeLogs: [
{
likeableId: '0',
likeableType: 'changeLog',
myLikeValue: '0',
likeCount: '0',
dislikeCount: '0',
likeScore: '0',
individualLikes: [],
id: '4507',
pageId: 'nonperson_predicate',
userId: 'EliezerYudkowsky',
edit: '1',
type: 'newEdit',
createdAt: '2015-12-28 19:49:00',
auxPageId: '',
oldSettingsValue: '',
newSettingsValue: ''
},
{
likeableId: '0',
likeableType: 'changeLog',
myLikeValue: '0',
likeCount: '0',
dislikeCount: '0',
likeScore: '0',
individualLikes: [],
id: '4499',
pageId: 'nonperson_predicate',
userId: 'EliezerYudkowsky',
edit: '0',
type: 'newParent',
createdAt: '2015-12-28 19:34:43',
auxPageId: 'mindcrime',
oldSettingsValue: '',
newSettingsValue: ''
}
],
feedSubmissions: [],
searchStrings: {},
hasChildren: 'false',
hasParents: 'true',
redAliases: {},
improvementTagIds: [],
nonMetaTagIds: [],
todos: [],
slowDownMap: 'null',
speedUpMap: 'null',
arcPageIds: 'null',
contentRequests: {}
}