{ localUrl: '../page/strictly_confused.html', arbitalUrl: 'https://arbital.com/p/strictly_confused', rawJsonUrl: '../raw/227.json', likeableId: '1005', likeableType: 'page', myLikeValue: '0', likeCount: '3', dislikeCount: '0', likeScore: '3', individualLikes: [ 'EricBruylant', 'LawrenceChan3', 'MalcolmMcCrimmon' ], pageId: 'strictly_confused', edit: '11', editSummary: '', prevEdit: '10', currentEdit: '11', wasPublished: 'true', type: 'wiki', title: 'Strictly confused', clickbait: 'A hypothesis is strictly confused by the raw data, if the hypothesis did much worse in predicting it than the hypothesis itself expected.', textLength: '8722', alias: 'strictly_confused', externalUrl: '', sortChildrenBy: 'likes', hasVote: 'false', voteType: '', votesAnonymous: 'false', editCreatorId: 'NateSoares', editCreatedAt: '2016-07-04 04:08:40', pageCreatorId: 'EliezerYudkowsky', pageCreatedAt: '2016-02-21 04:48:53', seeDomainId: '0', editDomainId: 'AlexeiAndreev', submitToDomainId: '0', isAutosave: 'false', isSnapshot: 'false', isLiveEdit: 'true', isMinorEdit: 'false', indirectTeacher: 'false', todoCount: '2', isEditorComment: 'false', isApprovedComment: 'true', isResolved: 'false', snapshotText: '', anchorContext: '', anchorText: '', anchorOffset: '0', mergedInto: '', isDeleted: 'false', viewCount: '477', text: '[summary: A hypothesis is "strictly confused" by the data if the hypothesis does much worse at predicting the data than it expected to do. If, on average, you expect to assign around 1% likelihood to the exact observation you see, and you actually see something to which you assigned 0.000001% likelihood, you are strictly confused.]\n\n\nA hypothesis is "strictly confused" by the data if the hypothesis does much worse at predicting the data than it expected to do. If, on average, you expect to assign around 1% likelihood to the exact observation you see, and you actually see something to which you assigned 0.000001% likelihood, you are strictly confused.\n\n%%knows-requisite([1r6]):\nI.e., letting $H$ be a hypothesis and $e_0$ be the data observed from some set $E$ of possible observations, we say that $H$ is "strictly confused" when\n\n$$ \\log \\mathbb P(e_0 \\mid H) \\ll \\sum_{e \\in E} \\mathbb P(e \\mid H) \\cdot \\log \\mathbb P(e \\mid H)$$\n\n%%\n\n# Motivation and examples\n\nIn Bayesian reasoning, the main reason to reject a hypothesis is when we find a better hypothesis. Suppose we think a coin is fair, and we flip it 100 times, and we see that the coin comes up "HHHHHHH..." or all heads. After doing this 100 times, the hypothesis "This is a double-headed coin" has a likelihood ratio of $2^{100} : 1$ favoring it over the "fair coin" hypothesis, and the "double-headed coin" hypothesis isn't *more* improbable than $2^{-100}$ a priori.\n\nBut this relies on the insight that there's a simple / a priori plausible *alternative* hypothesis that does better. What if the coin is producing TTHHTTHHTTHH and we just never happen to think of 'alternating pairs of tails and heads' as a hypothesis? It's possible to do better by thinking of a better hypothesis, but so far as the 'fair coin' hypothesis sees the world, TTHHTTHH... is no more or less likely than any other possible sequence it could encounter; the first eight coinflips have a probability of $2^{-8}$ and this would have been true no matter which eight coinflips were observed. After observing 100 coinflips, the fair coin hypothesis will assign them a collective probability of $2^{-100},$ and in this sense, no sequence of 100 coinflips is any more 'surprising' or 'confusing' than any other from *within* the perspective of the fair coin hypothesis.\n\nWe can't say that we're 'confused' or 'surprised' on seeing a long sequence of coinflips to which we assigned some very low probability on the order of $2^{-100} \\approx 10^{-30},$ because we expected to assign a probability that low.\n\nOn the other hand, suppose we think that a coin is biased to produce 90% heads and 10% tails, and we flip it 100 times and get some fair-looking sequence like "THHTTTHTTTTHTHTHHH..." (courtesy of random.org). Then we *expected* to assign the observed sequence a probability in the range of $0.9^{90} \\cdot 0.1^{10} \\approx 7\\cdot 10^{-15},$ but we *actually* saw a sequence we assigned probability around $0.9^{50} \\cdot 0.1^{50} \\approx 5 \\cdot 10^{-53}.$ We don't need to consider any other hypotheses to realize that we are very confused. We don't need to have *invented* the concept of a 'fair coin', or know that the 'fair coin' hypothesis would have assigned a much higher likelihood in the region of $7 \\cdot 10^{-31},$ to realize that there's something wrong with the current hypothesis.\n\nIn the case of the supposed fair coin that produces HHHHHHH, we only do poorly relative to a better hypothesis 'all heads' that makes a superior prediction. In the case of the supposed 90%-heads coin that produces a random-looking sequence, we do poorly than we expected to do from inside the 90%-heads hypothesis, so we are doing poorly in an absolute, non-relative sense.\n\nBeing strictly confused is a sign that tells us to look for *some* alternative hypothesis in advance of our having any idea whatsoever what that alternative hypothesis might be.\n\n# Distinction from frequentist p-values\n\nThe classical frequentist test for rejecting the null hypothesis involves considering the probability assigned to particular 'obvious'-seeming partitions of the data, and asking if we ended up inside a low-probability partition.\n\nSuppose you think some coin is fair, and you flip the coin 100 times and see a random-looking sequence "THHTTTHTT..."\n\nSomeone comes along and says, "You know, this result is very surprising, given your 'fair coin' theory. You really didn't expect that to happen."\n\n"How so?" you reply.\n\nThey say, "Well, among all sequences of 1000 coins, only 1 in 16 such sequences start with a string like THHT TTHTT, a palindromic quartet followed by a palindromic quintet. You confidently predicted that had a 15/16 chance of *not* happening, and then you were surprised."\n\n"Okay, look," you reply, "if you'd written down that *particular* prediction in advance and not a lot of others, I might be interested. Like, if I'd already thought that way of partitioning the data — namely, 'palindrome quartet followed by palindrome quintet' vs. '*not* palindrome quartet followed by palindrome quintet' — was a specially interesting and distinguished one, I might notice that I'd assigned the second partition 15/16 probability and then it failed to actually happen. As it is, it seems like you're really reaching."\n\nWe can think of the frequentist tests for rejecting the fair-coin hypothesis as a *small* set of 'interesting partitions' that were written down in advance, which are supposed to have low probability given the fair coin. For example, if a coin produces HHHHH HTHHH HHTHH, a frequentist says, "*Partitioning by number of heads*, the fair coin hypothesis says that on 15 flips we should get between 12 and 3 heads, inclusive, with a probability of 98.6%. You are therefore surprised because this event you assigned 98.6% probability failed to happen. And yes, we're just checking the number of heads and a few other obvious things, not for palindromic quartets followed by palindromic quintets."\n\nPart of the point of being a Bayesian, however, is that we try to only reason on the data we actually observed, and not put that data into particular partitions and reason about those partitions. The partitioning process introduces potential subjectivity, especially in an academic setting fraught with powerful incentives to produce 'statistically significant' data - the equivalent of somebody insisting that palindromic quartets and quintets are special, or that counting heads isn't special.\n\nE.g., if we flip a coin six times and get HHHHHT, this is "statistically significant p < 0.05" if the researcher decided to flip coins until they got at least one T and then stop, in which case a fair coin has only a 1/32 probability of requiring six or more steps to produce a T. If on the other hand the researcher decided to flip the coin six times and then count the number of tails, the probability of getting 1 or fewer T in six flips is 7/64 which is not 'statistically significant'.\n\nThe Bayesian says, "If I use [21c the Rule of Succession] to denote the hypothesis that the coin has an unknown bias between 0 and 1, then the sequence HHHHHT is assigned 1/30 probability by the Rule of Succession and 1/64 probability by 'fair coin', so this is evidence with a likelihood ratio of ~ 2 : 1 favoring the hypothesis that the coin is biased - not enough to [ overcome] any significant [ prior improbability]."\n\nThe Bayesian arrives at this judgment by only considering the particular, exact data that was observed, and not any larger partitions of data. To compute the probability flow between two hypotheses $H_1$ and $H_2$ we only need to know the likelihoods of our *exact* observation given those two hypotheses, not the likelihoods the hypotheses assign to any partitions into which that observation can be put, etcetera.\n\nSimilarly, the Bayesian looks at the sequence HHHHH HTHHH HHTHH and says: this specific, exact data that we observed gives us a likelihood ratio of (1/1680 : 1/32768) ~ (19.5 : 1) favoring "[21c The coin has an unknown bias] between 0 and 1" over "The coin is fair". With that already said, the Bayesian doesn't see any need to talk about the total probability of the fair coin hypothesis producing data inside a partition of similar results that could have been observed but weren't.\n\nBut even though Bayesians usually try avoid thinking in terms of rejecting a null hypothesis using partitions, saying "I'm strictly confused!" gives a Bayesian a way of saying "Well, I know *something's* wrong..." that doesn't require already having the insight to propose a better alternative, or even the insight to realize that some particular partitioning of the data is worth special attention.', metaText: '', isTextLoaded: 'true', isSubscribedToDiscussion: 'false', isSubscribedToUser: 'false', isSubscribedAsMaintainer: 'false', discussionSubscriberCount: '3', maintainerCount: '1', userSubscriberCount: '0', lastVisit: '2016-02-21 08:50:29', hasDraft: 'false', votes: [], voteSummary: [ '0', '0', '0', '0', '0', '0', '0', '0', '0', '0' ], muVoteSummary: '0', voteScaling: '0', currentUserVote: '-2', voteCount: '0', lockedVoteType: '', maxEditEver: '0', redLinkCount: '0', lockedBy: '', lockedUntil: '', nextPageId: '', prevPageId: '', usedAsMastery: 'false', proposalEditNum: '0', permissions: { edit: { has: 'false', reason: 'You don't have domain permission to edit this page' }, proposeEdit: { has: 'true', reason: '' }, delete: { has: 'false', reason: 'You don't have domain permission to delete this page' }, comment: { has: 'false', reason: 'You can't comment in this domain because you are not a member' }, proposeComment: { has: 'true', reason: '' } }, summaries: { Summary: 'A hypothesis is "strictly confused" by the data if the hypothesis does much worse at predicting the data than it expected to do. If, on average, you expect to assign around 1% likelihood to the exact observation you see, and you actually see something to which you assigned 0.000001% likelihood, you are strictly confused.' }, creatorIds: [ 'EliezerYudkowsky', 'NateSoares' ], childIds: [], parentIds: [ 'bayes_reasoning' ], commentIds: [ '7qs', '7xp' ], questionIds: [], tagIds: [ 'nonstandard_terminology_meta_tag', 'b_class_meta_tag' ], relatedIds: [], markIds: [], explanations: [ { id: '5863', parentId: 'strictly_confused', childId: 'strictly_confused', type: 'subject', creatorId: 'AlexeiAndreev', createdAt: '2016-08-02 17:36:36', level: '2', isStrong: 'true', everPublished: 'true' } ], learnMore: [], requirements: [ { id: '2284', parentId: 'bayes_probability_notation', childId: 'strictly_confused', type: 'requirement', creatorId: 'AlexeiAndreev', createdAt: '2016-06-17 21:58:56', level: '2', isStrong: 'false', everPublished: 'true' }, { id: '2285', parentId: 'math1', childId: 'strictly_confused', type: 'requirement', creatorId: 'AlexeiAndreev', createdAt: '2016-06-17 21:58:56', level: '3', isStrong: 'true', everPublished: 'true' }, { id: '5865', parentId: 'probability', childId: 'strictly_confused', type: 'requirement', creatorId: 'AlexeiAndreev', createdAt: '2016-08-02 17:38:40', level: '2', isStrong: 'true', everPublished: 'true' } ], subjects: [ { id: '5863', parentId: 'strictly_confused', childId: 'strictly_confused', type: 'subject', creatorId: 'AlexeiAndreev', createdAt: '2016-08-02 17:36:36', level: '2', isStrong: 'true', everPublished: 'true' }, { id: '5864', parentId: 'bayes_reasoning', childId: 'strictly_confused', type: 'subject', creatorId: 'AlexeiAndreev', createdAt: '2016-08-02 17:37:06', level: '2', isStrong: 'false', everPublished: 'true' } ], lenses: [], lensParentId: '', pathPages: [], learnMoreTaughtMap: {}, learnMoreCoveredMap: {}, learnMoreRequiredMap: {}, editHistory: {}, domainSubmissions: {}, answers: [], answerCount: '0', commentCount: '0', newCommentCount: '0', linkedMarkCount: '0', changeLogs: [ { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '18255', pageId: 'strictly_confused', userId: 'EricBruylant', edit: '0', type: 'newTag', createdAt: '2016-08-03 17:55:16', auxPageId: 'b_class_meta_tag', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '18169', pageId: 'strictly_confused', userId: 'AlexeiAndreev', edit: '0', type: 'newRequirement', createdAt: '2016-08-02 17:38:41', auxPageId: 'probability', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '18168', pageId: 'strictly_confused', userId: 'AlexeiAndreev', edit: '0', type: 'newSubject', createdAt: '2016-08-02 17:37:06', auxPageId: 'bayes_reasoning', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '18165', pageId: 'strictly_confused', userId: 'AlexeiAndreev', edit: '0', type: 'newTeacher', createdAt: '2016-08-02 17:36:37', auxPageId: 'strictly_confused', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '18166', pageId: 'strictly_confused', userId: 'AlexeiAndreev', edit: '0', type: 'newSubject', createdAt: '2016-08-02 17:36:37', auxPageId: 'strictly_confused', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '15200', pageId: 'strictly_confused', userId: 'NateSoares', edit: '11', type: 'newEdit', createdAt: '2016-07-04 04:08:40', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '15199', pageId: 'strictly_confused', userId: 'NateSoares', edit: '0', type: 'newTag', createdAt: '2016-07-04 04:03:31', auxPageId: 'nonstandard_terminology_meta_tag', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '15003', pageId: 'strictly_confused', userId: 'NateSoares', edit: '10', type: 'newEdit', createdAt: '2016-07-01 01:20:04', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '7518', pageId: 'strictly_confused', userId: 'EliezerYudkowsky', edit: '9', type: 'newEdit', createdAt: '2016-02-21 05:13:16', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '7517', pageId: 'strictly_confused', userId: 'EliezerYudkowsky', edit: '8', type: 'newRequirement', createdAt: '2016-02-21 05:12:48', auxPageId: 'math1', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '7507', pageId: 'strictly_confused', userId: 'EliezerYudkowsky', edit: '8', type: 'newEdit', createdAt: '2016-02-21 04:59:38', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '7506', pageId: 'strictly_confused', userId: 'EliezerYudkowsky', edit: '7', type: 'newEdit', createdAt: '2016-02-21 04:58:53', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '7504', pageId: 'strictly_confused', userId: 'EliezerYudkowsky', edit: '6', type: 'newEdit', createdAt: '2016-02-21 04:55:16', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '7503', pageId: 'strictly_confused', userId: 'EliezerYudkowsky', edit: '5', type: 'newEdit', createdAt: '2016-02-21 04:50:48', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '7502', pageId: 'strictly_confused', userId: 'EliezerYudkowsky', edit: '4', type: 'newEdit', createdAt: '2016-02-21 04:50:28', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '7501', pageId: 'strictly_confused', userId: 'EliezerYudkowsky', edit: '3', type: 'newEdit', createdAt: '2016-02-21 04:49:56', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '7500', pageId: 'strictly_confused', userId: 'EliezerYudkowsky', edit: '2', type: 'newEdit', createdAt: '2016-02-21 04:49:41', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '7499', pageId: 'strictly_confused', userId: 'EliezerYudkowsky', edit: '1', type: 'newEdit', createdAt: '2016-02-21 04:48:53', auxPageId: '', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '7498', pageId: 'strictly_confused', userId: 'EliezerYudkowsky', edit: '0', type: 'newRequirement', createdAt: '2016-02-21 03:45:29', auxPageId: 'bayes_probability_notation', oldSettingsValue: '', newSettingsValue: '' }, { likeableId: '0', likeableType: 'changeLog', myLikeValue: '0', likeCount: '0', dislikeCount: '0', likeScore: '0', individualLikes: [], id: '7496', pageId: 'strictly_confused', userId: 'EliezerYudkowsky', edit: '0', type: 'newParent', createdAt: '2016-02-21 03:45:17', auxPageId: 'bayes_reasoning', oldSettingsValue: '', newSettingsValue: '' } ], feedSubmissions: [], searchStrings: {}, hasChildren: 'false', hasParents: 'true', redAliases: {}, improvementTagIds: [], nonMetaTagIds: [], todos: [], slowDownMap: 'null', speedUpMap: 'null', arcPageIds: 'null', contentRequests: {} }