{
  localUrl: '../page/5s2.html',
  arbitalUrl: 'https://arbital.com/p/5s2',
  rawJsonUrl: '../raw/5s2.json',
  likeableId: '0',
  likeableType: 'page',
  myLikeValue: '0',
  likeCount: '0',
  dislikeCount: '0',
  likeScore: '0',
  individualLikes: [],
  pageId: '5s2',
  edit: '1',
  editSummary: '',
  prevEdit: '0',
  currentEdit: '1',
  wasPublished: 'true',
  type: 'comment',
  title: '"hm, do you actually need that discussion? In no..."',
  clickbait: '',
  textLength: '1648',
  alias: '5s2',
  externalUrl: '',
  sortChildrenBy: 'recentFirst',
  hasVote: 'false',
  voteType: '',
  votesAnonymous: 'false',
  editCreatorId: 'EricBruylant',
  editCreatedAt: '2016-08-05 13:26:47',
  pageCreatorId: 'EricBruylant',
  pageCreatedAt: '2016-08-05 13:26:47',
  seeDomainId: '0',
  editDomainId: '116',
  submitToDomainId: '0',
  isAutosave: 'false',
  isSnapshot: 'false',
  isLiveEdit: 'true',
  isMinorEdit: 'false',
  indirectTeacher: 'false',
  todoCount: '0',
  isEditorComment: 'true',
  isApprovedComment: 'true',
  isResolved: 'false',
  snapshotText: '',
  anchorContext: '',
  anchorText: '',
  anchorOffset: '0',
  mergedInto: '',
  isDeleted: 'false',
  viewCount: '136',
  text: 'hm, do you actually need that discussion? In no case does an agent know in advance that their vote will decide the election, just that there is some (usually extraordinarily slim) chance that they will. A situation where all agents have the impossible piece of information (the election is close enough that my actions can tip it, and, importantly that their tipping won't be undone by others who are in identical positions) seems not the right situation to be looking at, and would unsurprisingly lead to crazy outputs. Sure, in retrospect all the agents can go "damn, I should've put massive effort into acquiring more votes" if the election was close enough that they could have tipped it in a way they expect would have large positive EV, but that seems like a correct and reasonable conclusion in hindsight, just not one which is foreseeable.\n\nEV calc feels like a system I could actually use to weigh up the pros and cons, by looking at the statistics of closeness of various elections and estimating the value of tipping with maybe a few tens of hours of research, whereas estimating the correlation between my voting habits and various possible reference classes of voter seems in practice hopeless%%note: without, perhaps, having enough data to reconstruct key parts of large numbers of people's decision processes and massive effort classifying them, at which point you're not really running a process other people are likely to (unless you make your results publicly available, and things get recursive!)%%.\n\nMaybe explaining this is more of a detour than you want, though, since it's less interesting from a decision theory perspective?',
  metaText: '',
  isTextLoaded: 'true',
  isSubscribedToDiscussion: 'false',
  isSubscribedToUser: 'false',
  isSubscribedAsMaintainer: 'false',
  discussionSubscriberCount: '0',
  maintainerCount: '0',
  userSubscriberCount: '0',
  lastVisit: '',
  hasDraft: 'false',
  votes: [],
  voteSummary: 'null',
  muVoteSummary: '0',
  voteScaling: '0',
  currentUserVote: '-2',
  voteCount: '0',
  lockedVoteType: '',
  maxEditEver: '0',
  redLinkCount: '0',
  lockedBy: '',
  lockedUntil: '',
  nextPageId: '',
  prevPageId: '',
  usedAsMastery: 'false',
  proposalEditNum: '0',
  permissions: {
    edit: {
      has: 'false',
      reason: 'You don't have domain permission to edit this page'
    },
    proposeEdit: {
      has: 'true',
      reason: ''
    },
    delete: {
      has: 'false',
      reason: 'You don't have domain permission to delete this page'
    },
    comment: {
      has: 'false',
      reason: 'You can't comment in this domain because you are not a member'
    },
    proposeComment: {
      has: 'true',
      reason: ''
    }
  },
  summaries: {},
  creatorIds: [
    'EricBruylant'
  ],
  childIds: [],
  parentIds: [
    'rationality_of_voting',
    '5rq'
  ],
  commentIds: [],
  questionIds: [],
  tagIds: [],
  relatedIds: [],
  markIds: [],
  explanations: [],
  learnMore: [],
  requirements: [],
  subjects: [],
  lenses: [],
  lensParentId: '',
  pathPages: [],
  learnMoreTaughtMap: {},
  learnMoreCoveredMap: {},
  learnMoreRequiredMap: {},
  editHistory: {},
  domainSubmissions: {},
  answers: [],
  answerCount: '0',
  commentCount: '0',
  newCommentCount: '0',
  linkedMarkCount: '0',
  changeLogs: [
    {
      likeableId: '0',
      likeableType: 'changeLog',
      myLikeValue: '0',
      likeCount: '0',
      dislikeCount: '0',
      likeScore: '0',
      individualLikes: [],
      id: '18438',
      pageId: '5s2',
      userId: 'EricBruylant',
      edit: '1',
      type: 'newEdit',
      createdAt: '2016-08-05 13:26:47',
      auxPageId: '',
      oldSettingsValue: '',
      newSettingsValue: ''
    }
  ],
  feedSubmissions: [],
  searchStrings: {},
  hasChildren: 'false',
  hasParents: 'true',
  redAliases: {},
  improvementTagIds: [],
  nonMetaTagIds: [],
  todos: [],
  slowDownMap: 'null',
  speedUpMap: 'null',
  arcPageIds: 'null',
  contentRequests: {}
}