{ localUrl: '../page/9j2.html', arbitalUrl: 'https://arbital.com/p/9j2', rawJsonUrl: '../raw/9j2.json', likeableId: '0', likeableType: 'page', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], pageId: '9j2', edit: '1', editSummary: '', prevEdit: '0', currentEdit: '1', wasPublished: 'true', type: 'comment', title: '"I'm going to take the role of the "undergrad" h..."', clickbait: '', textLength: '612', alias: '9j2', externalUrl: '', sortChildrenBy: 'recentFirst', hasVote: 'false', voteType: '', votesAnonymous: 'false', editCreatorId: 'EyalRoth', editCreatedAt: '2019-03-27 15:33:15', pageCreatorId: 'EyalRoth', pageCreatedAt: '2019-03-27 15:33:15', seeDomainId: '0', editDomainId: '3306', submitToDomainId: '0', isAutosave: 'false', isSnapshot: 'false', isLiveEdit: 'true', isMinorEdit: 'false', indirectTeacher: 'false', todoCount: '0', isEditorComment: 'false', isApprovedComment: 'false', isResolved: 'false', snapshotText: '', anchorContext: 'Bayesian: I wrote pretty much the same Python program when I was first converting to Bayesianism and finding out about likelihood ratios and feeling skeptical about the system maybe being abusable in some way, and then a friend of mine found out about likelihood ratios and he wrote essentially the same program, also in Python\\. And lo, he found that false evidence of 20:1 for the coin being 55% biased was found at least once, somewhere along the way\\.\\.\\. 1\\.4% of the time\\. If you asked for more extreme likelihood ratios, the chances of finding them dropped off even faster\\.', anchorText: 'And lo, he found that false evidence of 20:1 for the coin being 55% biased was found at least once, somewhere along the way\\.\\.\\. 1\\.4% of the time', anchorOffset: '332', mergedInto: '', isDeleted: 'false', viewCount: '1116', text: 'I'm going to take the role of the "undergrad" here and try to interpret this in the following way:\n\nGiven that a hypothesis is true -- but it is unknown to be true -- it is far more likely to come by a "statistically significant" result indicating it is wrong, than it is likely to come by a result indicating that another hypothesis is significantly more likely.\n\nIn simpler words - it is far easier to "prove" a true hypothesis is wrong by accident, than it is to "prove" that an alternative hypothesis is superior (a better estimator of reality) by accident.\n\nWould you consider this interpretation accurate? ', metaText: '', isTextLoaded: 'true', isSubscribedToDiscussion: 'false', isSubscribedToUser: 'false', isSubscribedAsMaintainer: 'false', discussionSubscriberCount: '1', maintainerCount: '1', userSubscriberCount: '0', lastVisit: '', hasDraft: 'false', votes: [], voteSummary: 'null', muVoteSummary: '0', voteScaling: '0', currentUserVote: '-2', voteCount: '0', lockedVoteType: '', maxEditEver: '0', redLinkCount: '0', lockedBy: '', lockedUntil: '', nextPageId: '', prevPageId: '', usedAsMastery: 'false', proposalEditNum: '0', permissions: { edit: { has: 'false', reason: 'You don't have domain permission to edit this page' }, proposeEdit: { has: 'true', reason: '' }, delete: { has: 'false', reason: 'You don't have domain permission to delete this page' }, comment: { has: 'false', reason: 'You can't comment in this domain because you are not a member' }, proposeComment: { has: 'true', reason: '' } }, summaries: {}, creatorIds: [ 'EyalRoth' ], childIds: [], parentIds: [ 'likelihood_vs_pvalue' ], commentIds: [], questionIds: [], tagIds: [], relatedIds: [], markIds: [], explanations: [], learnMore: [], requirements: [], subjects: [], lenses: [], lensParentId: '', pathPages: [], learnMoreTaughtMap: {}, learnMoreCoveredMap: {}, learnMoreRequiredMap: {}, editHistory: {}, domainSubmissions: {}, answers: [], answerCount: '0', commentCount: '0', newCommentCount: '0', linkedMarkCount: '0', changeLogs: [ { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '23187', pageId: '9j2', userId: 'EyalRoth', edit: '1', type: 'newEdit', createdAt: '2019-03-27 15:33:15', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' } ], feedSubmissions: [], searchStrings: {}, hasChildren: 'false', hasParents: 'true', redAliases: {}, improvementTagIds: [], nonMetaTagIds: [], todos: [], slowDownMap: 'null', speedUpMap: 'null', arcPageIds: 'null', contentRequests: {} }