{"id":16,"date":"2021-07-31T17:48:10","date_gmt":"2021-07-31T21:48:10","guid":{"rendered":"https:\/\/health.uconn.edu\/causality\/?page_id=16"},"modified":"2026-03-14T21:38:39","modified_gmt":"2026-03-15T01:38:39","slug":"firststeps","status":"publish","type":"page","link":"https:\/\/health.uconn.edu\/causality\/firststeps\/","title":{"rendered":"Causality and causal modeling"},"content":{"rendered":"<p><strong>General definitional and notational setup<\/strong><\/p>\n<p>*** Causality is <em>not directly detectable<\/em> or observable, much like forces<a href=\"#_edn1\" name=\"_ednref1\"><span>[i]<\/span><\/a> or energy aren\u2019t either: it needs to be discovered\/uncovered &#8211; \u2018detected\u2019 through observational consequences<a href=\"#_edn2\" name=\"_ednref2\"><span>[ii]<\/span><\/a>; on the other hand, it\u2019s a pretty natural intuitive human notion<a href=\"#_edn3\" name=\"_ednref3\"><span>[iii]<\/span><\/a>, yet is considered a worthy \u2018problem\u2019<a href=\"#_edn4\" name=\"_ednref4\"><span>[iv]<\/span><\/a>.<\/p>\n<p>*** Causality is hardly a statistical matter, and it shows up at the very roots of scientific inquiries: the basic notions of substance and time, e.g. are defined in excessive proximity to \u00a0causal relations: e.g. abstract entities (like \u2018number 7\u2019) are those that cannot cause anything<a href=\"#_edn5\" name=\"_ednref5\"><span>[v]<\/span><\/a>, unlike \u2018concrete substances\u2019, and time itself has definitions overlapping with causality [2-6].<\/p>\n<p>*** Causal language<a href=\"#_edn6\" name=\"_ednref6\"><span>[vi]<\/span><\/a> differs drastically by scientific field<a href=\"#_edn7\" name=\"_ednref7\"><span>[vii]<\/span><\/a>, and some are more shy than others in using the \u201cC\u201d word. In statistics it has been kept at bay for many decades, and epidemiologists have been to first advocates among statisticians at large (<a href=\"https:\/\/hsph.harvard.edu\/profile\/tyler-j-vanderweele\/\">Tyler VanderWeele<\/a>, e.g., but also Jay Kaufman [8] and [9] in health disparities, and many other nowadays).<\/p>\n<p>+++ This follows my \u2018let\u2019s talk (just talk!, no math!) about causality\u2019 recent philosophy. The old adagio \u201ccorrelation does not imply causation\u201d is a structural misunderstanding of constructs: while each are a 2-objects relation, causation is not symmetric, as correlation is, and \u2018proceeding from one to the other\u2019 requires more than the 2 objects of interest. The simple solution was offered &gt;100 years ago by Sewall Wright, and I rephrase to modern language: \u201ca correlation can come from: either variable causing the other one, or another variable causing both\u201d. This makes clear the \u2018logical tension\u2019 and the proper solution: we need to step above the 2 ingredients to make causal statements (otherwise we might just reply: \u201ccorrelation implies causation, but only sometimes\u201d, when there is a causal component of the relation in there).<\/p>\n<p>To rephrase still: causal inference involves decomposing correlations into their causal and non-causal components. Simple, right? Incidentally, a parallel view uses a similar decomposition maneuver. [0]<\/p>\n<p>*** Even \u2018simple\u2019 questions like \u2018What causes the universe to expand?\u2019 rely on deeply unobservable constructs, like dark energy.<a href=\"#_edn8\" name=\"_ednref8\"><span>[viii]<\/span><\/a><\/p>\n<p>*** In medicine, causal questions are broadly aimed at \u2018changing the current course of events\u2019<a href=\"#_edn9\" name=\"_ednref9\"><span>[ix]<\/span><\/a> in the future, whereas in astronomy e.g. mere deep descriptive understanding of \u2018how things work\u2019 is a goal in itself: finding the laws governing the Universe.<a href=\"#_edn10\" name=\"_ednref10\"><span>[x]<\/span><\/a><\/p>\n<p>*** The task then is to <em>discover<\/em> \u201cwhich variable \u2018listens to\u2019 which others\u201c (\u2018Book of Why\u2019 <a href=\"https:\/\/bayes.cs.ucla.edu\/WHY\/\">BoW<\/a> p. 7), e.g. whether blood glucose (say measured by hemoglobin A1c) \u2018listens to\u2019 one\u2019s body weight (say measured by body mass index, BMI), and responds to it: this metaphor invites deeper delving into how the \u2018listening\u2019 works<a href=\"#_edn11\" name=\"_ednref11\"><span>[xi]<\/span><\/a>: is an asteroid \u2018listening\u2019 to the Earth\u2019s gravity when heading to it? <em>Does it have a choice<\/em> not to listen, like biological bodies do? Since biological entities are the first ones to collect information and use it in choosing to respond differently, to benefit themselves, the \u2018listening\u2019 appears to make less sense before the first living organisms appeared: all natural bodies and substances before that blindly (and strictly, without choice) followed the tight grip of the natural laws, gravity among them being the central one.<a href=\"#_edn12\" name=\"_ednref12\"><span>[xii]<\/span><\/a><\/p>\n<p>*** Note that the way BMI relates to A1c is not merely direct and linear, e.g. [48]:<\/p>\n<p><a href=\"FIGURE%201%20Dynamic%20phase%20of%20obesity%20development\"><em>FIGURE 1 Dynamic phase of obesity development in the carbohydrate-insulin model<\/em><\/a><a href=\"#_edn13\" name=\"_ednref13\"><span><em><strong>[xiii]<\/strong><\/em><\/span><\/a>.<\/p>\n<p><a href=\"https:\/\/health.uconn.edu\/causality\/wp-content\/uploads\/sites\/264\/2025\/04\/FIGURE-1-Dynamic-phase-of-obesity-development.jpg\"><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/health.uconn.edu\/causality\/wp-content\/uploads\/sites\/264\/2025\/04\/FIGURE-1-Dynamic-phase-of-obesity-development-300x142.jpg\" alt=\"\" width=\"579\" height=\"274\" class=\"wp-image-85 aligncenter\" srcset=\"https:\/\/health.uconn.edu\/causality\/wp-content\/uploads\/sites\/264\/2025\/04\/FIGURE-1-Dynamic-phase-of-obesity-development-300x142.jpg 300w, https:\/\/health.uconn.edu\/causality\/wp-content\/uploads\/sites\/264\/2025\/04\/FIGURE-1-Dynamic-phase-of-obesity-development-768x363.jpg 768w, https:\/\/health.uconn.edu\/causality\/wp-content\/uploads\/sites\/264\/2025\/04\/FIGURE-1-Dynamic-phase-of-obesity-development.jpg 960w\" sizes=\"(max-width: 579px) 100vw, 579px\" \/><\/a><\/p>\n<p>*** Intuitively grasping how causal effects are \u2018extracted\u2019 from correlational-kind data, a bit of notational setup will be decisive: for the A1c(BMI) relation of interest, or BMI -&gt; A1c, or probability(A1c | BMI), where \u201c|\u201d stands for \u2018given\u2019, we want to find out whether lowering BMI would cause a A1c drop, and then by how much. To answer the \u2018cause-effect\u2019 research question (RQ), we simply need to compare for each patient <em>i <\/em>his\/her A1c\u2019s under the current (present state)<strong><em><sub> CurrentBMI<\/sub><\/em><\/strong> vs. a \u2018what if\u2019 potential state <strong><sup>If.LowerBMI<\/sup><\/strong>:<\/p>\n<p style=\"text-align: center\">A1c<strong><em><sub> i<\/sub><\/em><\/strong><sup> <\/sup><strong><em><sub>CurrentBMI<\/sub><\/em><\/strong> vs. A1c<sub> <\/sub><strong><em><sub>i<\/sub><\/em><\/strong><strong><sup> If.LowerBMI<\/sup><\/strong><\/p>\n<p>or with classic \u2018treatment\u2019 (Tx) as the cause: A1c<sub> <\/sub><strong><em><sub>i<\/sub><\/em><\/strong><strong><sup> If.Treated<\/sup><\/strong> vs. A1c<strong><em><sub> i<\/sub><\/em><\/strong><strong><sup> If.NotTreated<\/sup><\/strong><\/p>\n<p>where superscripts mean \u2018it could happen up there\u2019 [<a href=\"#_edn1\" name=\"_ednref1\"><span>xiv]<\/span><\/a> A1c<sub> <\/sub><strong><em><sub>i<\/sub><\/em><\/strong><strong><sup> If.LowerBMI<\/sup><\/strong> , vs. what actually happened \u2018here on the ground\u2019, in this reality A1c<sub> <\/sub><strong><em><sub>i<\/sub><\/em><\/strong><sup> <\/sup><strong><em><sub>CurrentBMI<\/sub><\/em><\/strong>. More details in the footnote, suffice to say that there is a causal effect of BMI on A1c for patient <em>i<\/em> if: A1c<sub> <\/sub><strong><em><sub>i<\/sub><\/em><\/strong><strong><sup> If.LowerBMI<\/sup><\/strong> \u2260 A1c<sub> <\/sub><strong><em><sub>i<\/sub><\/em><\/strong><sup> <\/sup><strong><em><sub>CurrentBMI<\/sub><\/em><\/strong>! One of these two quantities however can never be \u2018seen\u2019, once reality kicks in: either one has his\/her current BMI, or a lower BMI!<\/p>\n<p>Note though that when <em>realized<\/em>, potential values come \u2018down here on the ground\u2019 (become visible):<\/p>\n<p style=\"text-align: center\">A1c<strong><em><sub> i<\/sub><\/em><sup> <\/sup>*<sup>If.NotTreated<\/sup><em><sub> NotTreated<\/sub><\/em><\/strong> \u00a0= A1c<strong><em><sub> i<\/sub><\/em><sup> <\/sup><em><sub>NotTreated<\/sub><\/em><\/strong>\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0 and<\/p>\n<p style=\"text-align: center\">A1c<strong><em><sub> i<\/sub><\/em><sup> <\/sup>*<sup>If.Treated<\/sup><em><sub> Treated<\/sub><\/em><\/strong> \u00a0= A1c<strong><em><sub> i<\/sub><\/em><sup> <\/sup><em><sub>Treated<\/sub><\/em><\/strong> <sub><\/sub><\/p>\n<p>in which I added the <strong>*<\/strong>, for <strong>\u2018never visible\u2019: what would have happened to these folks, but\u2026 didn\u2019t (couldn\u2019t)!<\/strong> The fundamental causal inference problem is that one of these 2 in each condition is \u2018missing\u2019:<\/p>\n<p style=\"text-align: center\">A1c<strong><em><sub> i<\/sub><\/em><sup> <\/sup>*<sup>If.Treated<\/sup><\/strong> <strong><em><sub>NotTreated<\/sub><\/em><\/strong> \u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0 &amp;\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0 A1c<strong><em><sub> i<\/sub><\/em><sup> <\/sup>*<sup>If.NotTreated<\/sup><\/strong> <strong><em><sub>Treated <\/sub><\/em><\/strong><\/p>\n<p>***The statistical trickery to still derive a causal effect from such \u2018observational\u2019 data (i.e. not from a randomized controlled research design) is to \u2018identify\u2019 the average causal effect among those who say lowered their BMI, Average(A1c<sub> <\/sub><strong><em><sub>i<\/sub><\/em><\/strong><strong><sup> If.LowerBMI<\/sup><\/strong><strong><sub> LoweredBMI<\/sub><\/strong> &#8211; A1c<sub> <\/sub><strong><em><sub>i<\/sub><\/em><\/strong><strong> *<\/strong><em><sup>If.SameBMI<\/sup><\/em><strong><sub> LoweredBMI<\/sub><\/strong>), in which we added the <strong>*<\/strong>, for <strong>\u2018never visible\u2019: what would have happened to these folks, but\u2026 didn\u2019t!<\/strong><\/p>\n<p>We do this by \u2018wiping off\u2019 a bias quantity, the \u2018selection bias\u2019<a href=\"#_edn15\" name=\"_ednref15\"><span>[xv]<\/span><\/a>. All this math is drastically simplified if one uses Judea Pearl\u2019s notation for \u2018setting\u2019 the predictor to a value, so Average(A1c<sub> <\/sub><strong><em><sub>i<\/sub><\/em><\/strong><strong><sup> If.LowerBMI<\/sup><\/strong>) becomes simply Average(A1c<sub> <\/sub><strong><em><sub>i<\/sub><\/em><\/strong> | (<strong>set<\/strong> (BMI=Lower))<a href=\"#_edn16\" name=\"_ednref16\"><span>[xvi]<\/span><\/a>, [55, 56], where \u201cset\u201d symbolizes an actual intervention that sets to BMI to a lower value, and E means \u2018expected (average) value of\u2019: the challenge of \u2018causal inference\u2019 then is \u2018simply\u2019 deriving the causal quantity Average(A1c | (<strong>set<\/strong> (BMI=Lower)) from the observational one Average(A1c | \u2018seeing\u2019 BMI=Lower). Observing BMI=Low is NOT the same as \u2018setting\u2019 it to \u2018Low\u2019, or \u2018intervening\u2019 to \u2018do\u2019 that:<\/p>\n<p>Average(A1c<sub> <\/sub><strong><em><sub>i<\/sub><\/em><\/strong> | \u2018<strong><u>see<\/u><\/strong>\u2019 BMI=Lower) \u2260 Average(A1c<sub> <\/sub><strong><em><sub>i<\/sub><\/em><\/strong> | \u2018<strong><u>set<\/u><\/strong>\u2019 BMI=Lower)<\/p>\n<p>or Average(A1c<sub> <\/sub><strong><em><sub>i<\/sub><\/em><\/strong> | \u2018see\u2019 BMI=Lower) \u2260 Average(A1c<sub> <\/sub><strong><em><sub>i<\/sub><\/em><\/strong> | \u2018do\u2019 BMI=Lower)<\/p>\n<p>***Again, statistics can \u2018recover\u2019 the causal effects from observational (non-interventional) data when it can turn all \u2018set\u2019 (or \u2018do\u2019) into \u2018see\u2019 in the formulas: then, the math (or software, or AI) can directly yield the actual causal effect: even a simple t-test or a chi-squared test would thus become a \u2018causal analysis\u2019.<\/p>\n<p><strong>\u00a0<\/strong><strong>*** <\/strong>I start with an illustration of the simplest \u2018data\u2019, the yes\/no, presence\/absence, or 1\/0, like diabetic\/not (which can carry some error in it too<a href=\"#_edn17\" name=\"_ednref17\"><span>[xvii]<\/span><\/a>). This is a quick overview of several topics that all rely on this simplest possible data setup: a 2by2 binary fields\/variables, all using the simple 4-cell setup (0,0), (1,0), (0,1), and (1,1). They make use of the information derived from such a setup completely differently:<\/p>\n<p><strong>***.1. <\/strong>Finding \u2018it\u2019 vs. missing \u2018it\u2019: the medical sensitivity vs. specificity (&amp; ROC curve \u2013 to be added)<\/p>\n<p>The setup here is slightly different than most 2by2 tables(as chi-squared test examples, e.g.), and it has Test(for a disease), on the left, and Disease, on top, and inside:\u00a0 (1,1) (1,0), first row, then (0,1), (0,0), or the a,b,c, and d, as they are better known (an image is shown). They mean sequentially:<\/p>\n<p>a <strong>(1,1)<\/strong> Test is positive, Disease present, so we have (+,+) <strong>= Good! <\/strong><\/p>\n<p>b <em>(1,0)<\/em>,Test is positive, Disease not there, so (+,-) <em>= Bad!<\/em> first row, then<\/p>\n<p>c <em>(0,1)<\/em>, Test is negative, Disease present, so (-,+) <em>= Bad!<\/em><\/p>\n<p>d <strong>(0,0)<\/strong>, Test is negative, Disease not there, so (-,-) <strong>= Good!<\/strong><\/p>\n<p>Sensitivity reads the 1<sup>st<\/sup> column <strong>downwards<\/strong>: <strong>n(1,1)<\/strong> \/ [<strong>n(1,1) + <\/strong><em>n(0,1)<\/em><strong>] or <\/strong>with | meaning \u2018given\u2019<\/p>\n<p>P(Test + | Disease +) \u2013 this has a more intuitive write-up: percent of all disease+ patients with test+<\/p>\n<p>Specificity reads the 2<sup>nd<\/sup> column <strong>upwards<\/strong>: <strong>n(0,0)<\/strong> \/ [<strong>n(0,0) + <\/strong><em>n(1,0)<\/em><strong>] or <\/strong><\/p>\n<p>P(Test &#8211; | Disease -) : percent of all disease- patients with test-<\/p>\n<p>*** Note that it is different to talk about P(Disease + | Test +), than about P(Test + | Disease +), which has a more natural direction; the \u2018direction\u2019 differs: we \u2018condition on\u2018 (know) one, or the other.<\/p>\n<p>*** Note that this 2&#215;2 setup is <em>not<\/em> direction-less<a href=\"#_edn18\" name=\"_ednref18\"><span>[xviii]<\/span><\/a>, yet here the 1<sup>st<\/sup> # talks about the LEFT label (test), second about the TOP label (disease, so (1, 0) means (Test + , Disease -); 2by2 tables read more naturally from left-&gt;right and top-down. However, sensitivity and specificity read P(Test | Disease), so the \u2018given\u2019 (or \u2018knowing\u2019) is the Disease, which \u2018happens first. This setup becomes more meaningful when time is incorporated, as in RQ.2. in part 2, which has the before &amp; after of a binary category: here P(Post | Pre) makes sense, but P(Pre | Post) does not.<\/p>\n<p>*** A similar setup is the by-product of a logistic regression analysis, when the predictors of a binary outcome are more\/less useful in \u2018correctly\u2019 assigning 1\u2019s and 0\u2019s to their category; the \u2018classification table\u2019 shows the correct assignments <strong>(1,1) &amp;<\/strong> <strong>(0,0)<\/strong>, and the incorrect ones <em>(1,0)<\/em>,<em> (0,1)<\/em>, and provide an average of percent cases incorrectly classified.<\/p>\n<p><strong>***2. <\/strong>In formal logic \u201cIf Premise &#8211; Then Conclusion\u201d or simply \u201cPremise -&gt; Conclusion\u201d follows formal rules, which simply say that the entire statement \u201cPremise -&gt; Conclusion\u201d is false in relation to the 2 ingredients truth value only in 1 circumstance: When \u201cTrue -&gt; False\u201d: truth cannot lead to falsehood! A\u00a0 peculiar one however is \u201cFalse -&gt; True\u201d, this whole statement is \u2018True\u2019, (along with the other 2 possibilities \u201cTrue -&gt; True\u201d and \u201cFalse -&gt; False\u201d), because, as Claude.ai says \u201cSince the condition is never met, the implication can&#8217;t be falsified.\u201d, meaning because we start in an \u2018impossible world\u2019, we cannot test its truth value\u2026 \u201cIf pigs could fly, then anything goes\u201d pretty much\u2026<\/p>\n<p>If we follow the four-cell setup known in the sensitivity\/specificity context below (1,1) (1,0), (0,1), (0,0), the \u2018first\u2019 is logically the \u2018If\u2019 (vertical, from up to down, column), the second the \u2018Then\u2019 (horizontal, left to right, row) follow<\/p>\n<p>a (1,1), IF is True, THEN is True too, so we have (+,+) or<\/p>\n<p>b (1,0), IF is True, THEN is False, so (+,-) : this argument only of the 4 is FALSE<\/p>\n<p>c (0,1), IF is False, THEN is True, so (-,+)<\/p>\n<p>d (0,0), IF is False, THEN is False too, so (-,-)<\/p>\n<p><strong>***.3. <\/strong>H0 (null hypothesis) vs. H1 and what statistics says: rejecting H0 &#8211; seeing \u2018the effect\u2019 vs. not; unfortunately, the Null Hypothesis language leads to double (and more) negations (fail to reject H0\u2026 meaning \u201cwe see no effect), so the following emphasizes the \u2018<em>Effect seeing\u2019 (or not)<\/em> phrasing.<\/p>\n<p>If we follow <strong><em>the same<\/em><\/strong> four-cell setup known in the sensitivity\/specificity context below (1,1) (1,0), (0,1), (0,0), the \u2018first\u2019 category is logically \u2018there is<strong><em>\/not<\/em><\/strong> an <strong>Effect\u2019<\/strong> (= \u2018H0 (null hypothesis) if true\/false\u2019, vertical, from up to down, column), the second the \u2018Statistics finds<strong><em>\/not <\/em><\/strong>an effect\u2019 (= \u2018we reject\/<strong>fail to reject<\/strong> H0 t, horizontal, left to right, row) follow<\/p>\n<p>a: (1,1) or (+ <strong>| <\/strong>+): There is an Effect (H0=False), and Statistics finds<strong><em> it<\/em><\/strong> (rejects H0 true too); we correctly say \u2018There IS an Effect\u2019! This is <strong>(1-\u03b2<\/strong>) = \u2018statistical power\u2019 that we often we need estimated <strong><em>before<\/em><\/strong> a study is started: we want some solid\/large value here, a good likelihood we will FIND and Effect when it\u2019s there! This is ~= (akin to) the <strong><em>SENSITIVITY<\/em><\/strong> of a test, from above!<\/p>\n<p>b: (1,0) or (+ <strong>| <\/strong>-): There is NO Effect (H0=True), and Statistics finds<strong><em> it<\/em><\/strong> (rejects H0 true); = <strong>we<\/strong> make a Type I error: \u201cTrue null hypothesis is incorrectly rejected\u201d; Type I <em>because<\/em> it is most serious, we \u2018find something that doesn\u2019t exist\u2019 (i.e. \u2018make up stuff\u2019)! This is <strong>\u03b1<\/strong>, the customary .05 (5%) threshold value: we set this upfront, to <strong><em>as small<\/em><\/strong> an uncertainty value as we are <strong><em>willing to tolerate<\/em><\/strong>.<\/p>\n<p>c: (0,1) or (- <strong>| <\/strong>+): There is an Effect (H0=False), BUT Statistics does NOT find<strong><em> it<\/em><\/strong> (\u2018<strong>we<\/strong> <em>fail to reject<\/em>\u2019 H0); we make a make a Type II error: \u201ctrue null hypothesis is incorrectly rejected\u201d This is <strong>\u03b2<\/strong>.<\/p>\n<p>d: (0,0) or (- <strong>| <\/strong>-): There is NO Effect (H0=True), and Statistics does NOT find<strong><em> it <\/em><\/strong>either: we correctly say \u2018There if NO effect\u2019! This is<strong>(1-\u03b1)<\/strong>.<\/p>\n<p>***It is advisable to set up expectations based on prior knowledge in the form of a Hypothesis: Overweight -&gt; Diabetes, (a causal statement), or a RQ: \u201cOverweight -&gt; Diabetes?\u201d. The hypothesis testing procedure is a flexibly \u2018sharp\u2019 decision process, can be e.g. 2-sigma\u2019, rejecting H0 based on the \u20182 standard errors away\u2019 criterion (\u2018p &lt; .05\u2019), common in medicine and health research, or more conservatively based on the <a href=\"https:\/\/home.cern\/resources\/faqs\/five-sigma\">5-sigma rule<\/a> in physics (\u2018p &lt; .00006\u2019): the\u00a0 = .05 rule is a \u2018rule of thumb\u2019, but thumbs differ widely in size\u2026<\/p>\n<p>***I will demonstrate in part 2 several statistical tests and analyses, with an emphasis on the mechanics of each, the \u2018how to\u2019, all of them done \u2018by hand\u2019 and in Excel alone (no \u2018sophisticated\u2019 statistical software needed).<\/p>\n<p><span style=\"text-decoration: underline\"><strong>FOOTnotes:<\/strong><\/span><\/p>\n<p><a href=\"#_ednref1\" name=\"_edn1\"><span><\/span><\/a><a href=\"#_ednref1\" name=\"_edn1\">[0]<\/a><\/p>\n<p>Judea Pearl said in his \u2018Book of Why\u2019 (ch.1 <a href=\"https:\/\/bayes.cs.ucla.edu\/WHY\/why-ch1.pdf\">here<\/a>; books seems to be online <a href=\"http:\/\/repo.darmajaya.ac.id\/5342\/1\/The%20book%20of%20why_%20the%20new%20science%20of%20cause%20and%20effect%20%28%20PDFDrive%20%29.pdf\" class=\"broken_link\">here<\/a>): &#8220;Flip two coins simultaneously one hundred times and write down the results only when at least one of them comes up heads. Looking at your table, which will probably contain roughly seventy-five entries, you will see that the outcomes of the two simultaneous coin flips are not independent. Every time Coin 1 landed tails, Coin 2 landed heads. How is this possible? Did the coins somehow communicate with each other at light speed? Of course not. In reality you conditioned on a collider by censoring all the tails-tails outcomes. &#8220;Why:202)<\/p>\n<p>The format of the book does not allow \u2018spelling out\u2019 of every detail\u2026 (some <a href=\"https:\/\/smus.com\/books\/book-of-why-by-judea-pearl\/\">readers<\/a> reported this too), so I will try. This becomes \u2018visible\u2019 when one \u2018does\u2019 it (! a \u2018do\u2019 operation\u2026), say in Excel\/GSheets directly. I will explain it \u2018verbally\u2019 (a separate post will walk through the Excel exercise). So:<\/p>\n<p>If one generates a series or random yes\/no (1\/0) events, say flipping a coin 100 times: we list the results in a 1st column; repeat this process, list results in a 2nd column: the 2 cannot be correlated: they are both random occurrences. A third column can be built off the first two, by coding 0 when both ingredients were 0, and 1 otherwise (which is: 3rd = 1st+2nd\u22121st<strong>\u2219<\/strong>2nd): this 3rd is \u2018determined\u2019 causally (or \u2018caused deterministically\u2019\u2026) by the first 2. What one can verify directly is:<a href=\"#_ednref1\" name=\"_edn1\"><span><\/span><\/a><\/p>\n<ol>\n<li><strong><\/strong> Regressing 2nd on 1st will yield a null (near-zero, statistically) relation always (nearly so); BUT, controlling for the 3rd will always yield a non-null (conditional) relation: this is the artificial \u2018creation\u2019 of a correlation between two predictors of the same outcome, when conditioning on that outcome: THIS is the collider problem!<\/li>\n<li><strong><\/strong> Following the example as phrased in the book, but now giving extra-meaning to the 1st and the 2nd: say 1st is head injuries and 2nd is heart \u2018problems\u2019 (as a symptom, palpitations of a more serious nature, e.g., that would make one go to the hospital): these 2 should not relate for any natural reason, but if we restrict our data collection to those who had either, by selecting data from a hospital database, all of a sudden we will see a correlation between them. Granted, this particular one will be negative, patients in the hospital either have one, and not the other one (and vice versa). This is pretty much almost any medical study\u2026 which excludes healthy folks, because\u2026 there is \u2018nothing to see there\u2019.<\/li>\n<\/ol>\n<p>&#8211; There are other implications to be aware of: depending on the sign of the two effects between 1st and 2nd, and the 3rd, this \u2018artifact\u2019 can be &gt;0 or &lt;0: this has the consequence of \u2018shifting\u2019 the \u2018Null\u2019 point of our hypothesis testing process, technically. We start off \u2018crooked\u2019, and we should correct for this!<\/p>\n<p>+++ \u201cwe can decompose the observed outcome of a treatment into two effects:<\/p>\n<p>Outcome for treated \u2212Outcome for untreated<\/p>\n<p>= (Outcome for treated \u2212Outcome for treated if not treated) +<\/p>\n<p>+ (Outcome for treated if not treated\u2212 Outcome for untreated) =<\/p>\n<p>= Impact of treatment on treated + selection bias.\u201d {Varian, 2016 #15819} \u00a0p. 7311<\/p>\n<p><a href=\"#_ednref1\" name=\"_edn1\"><span>[i]<\/span><\/a> \u201cWhat are the merits of these fictitious variables called causes that make them worthy of such relentless human pursuit, and what makes causal explanations so pleasing and comforting once they are found? We take the position that human obsession with causation, like many other psychological compulsions, is computationally motivated. Causal models are attractive mainly because they provide effective data structures for representing empirical knowledge-they can be queried and updated at\u00a0 high speed with minimal external supervision. The effectiveness of causal models stems from their modular architecture, i.e., an architecture in which dependencies among variables are mediated by a\u00a0 few central mechanisms. \u201c p. 383 \u2018 8. Learning Structure from Data\\ 8.1. Causality, Modularity, and Tree Structures\u2019; \u201cthe construct of causality is merely a\u00a0 tentative, expedient device for encoding complex structures of dependencies in the closed world of a predefined set of variables. It serves to highlight useful independencies at a given level of abstraction, but causal relationships undergo drastic change upon the introduction of new variables.\u201d P. 397<\/p>\n<p>\u201cWith its dual role, summarizing and decomposing, a causal variable is analogous to an orchestra conductor: It achieves coordinated behavior through central communication and thereby relieves the players of having to communicate directly with one another. In the physical sciences, a\u00a0 classic example of such coordination is afield (e.g., gravitational, electric, or magnetic). Although there is a one-to-one mathematical correspondence between the electric field and the electric charges in terms of which it is defined, nearly every physicist takes the next step and ascribes physical reality to the electric field, imagining that in every point of space there is some real physical phenomenon taking place which determines both the magnitude and direction assigned to that point. This psychological construct had a huge impact on the historical development of electrical science. It decomposed the complex phenomena associated with interacting electrical charges into two independent processes: The creation of the field at a given point by the surrounding charges and the conversion of the field into a physical force once another charge passes near that point.\u201d {Pearl, 1988 #5884} P. 384<\/p>\n<p><a href=\"#_ednref2\" name=\"_edn2\"><span>[ii]<\/span><\/a> \u201cAn interesting philosophical question is whether any method based on probabilities can identify causal directionality among observable variables.\u201d {Pearl, 1988 #5884} p. 396, Learning Structure from Data \u201cFormally speaking, probabilistic analysis is indeed sensitive only to covariations, so it can never distinguish genuine causal dependencies from spurious correlations, i.e., variations coordinated by some common, often unknown causal mechanism. But what is\u00a0 the operational meaning of a\u00a0 genuine causal influence? How do humans distinguish it from spurious correlation? Our instinct is to invoke the notion of control; e.g.,\u00a0 we can cause the ice to melt by lighting a fire but we cannot cause fire by melting the ice. Yet the element of control is\u00a0 often missing from causal schemas. For example, we say that the rain caused the\u00a0 grass to become wet despite our being unable to control the rain. The only way to tell that wet grass does not cause rain is to find some other means of getting the grass wet, clearly\u00a0 distinct from rain, and to verify that when the other means is activated, the ground surely gets wet while the rain refuses to fall. Thus, the perception of voluntary control is in itself merely a by-product of covariation observed on a\u00a0 larger set of variables [Simon 1980], including, for example, the mechanism of turning one&#8217;s sprinkler on. In other words, whether X causes Y or Y causes X is not something that can be determined by running experiments on the pair (X, Y), in isolation from the rest of the world. The test for causal directionality must involve at least one additional variable, say Z, to test if by activating Z we can create variations in Y and\u00a0 none in X, or alternatively, if variations in Z are accompanied by variations in X while Y remains unaltered. This is exactly the meaning of causality that the polytree recovery algorithm attributes to the arrows on the branches. The meaning relates not to a\u00a0 single branch in isolation but to an assembly of branches heading toward a\u00a0 single node. <strong>An arrow is drawn from X to Y and not the other way around when a variable Z is found that correlates with Y but not with X<\/strong> (see Figure 8.2a). The discovery of the third variable Z, however, does not necessarily mean that X and Z are the ultimate causes of Y. The relationship among the three could change entirely upon the discovery of new variables.\u201d p. 396-7<\/p>\n<p><a href=\"#_ednref3\" name=\"_edn3\"><span>[iii]<\/span><\/a> \u201cAfter all, if you and I share the same understanding of physics, you should be able to figure out for yourself which mechanism it is that must be perturbed in order to realize the specified new event, and this should enable you to predict the rest of the scenario.<\/p>\n<p>This linguistic abbreviation defines a new relation among events, a relation we normally call &#8220;causation&#8221;: Event A causes B, if the perturbation needed for realizing A entails the realization of B.<strong><sup>2<\/sup><\/strong> (<strong><sup>2<\/sup><\/strong> The word &#8220;needed&#8221; connotes minimality and can be translated to: &#8221; .. .if every minimal perturbation realizing A, entails B&#8221;.) \u201d {Pearl, 1997 #15498} p. 55<\/p>\n<p><a href=\"#_ednref4\" name=\"_edn4\"><span>[iv]<\/span><\/a> \u201cWhen we ask, for example, what causes this fire, it is not its being this but its being fire that we are seeking to account for.\u201d {Anderson, 1938 #15495} p. 128<\/p>\n<p><a href=\"#_ednref5\" name=\"_edn5\"><span>[v]<\/span><\/a> \u201cI shall understand by a \u2018substance\u2019 a particular thing, capable of causing or being caused, such as Richard Swinburne, or a particular table, or a particular electron, or the planet Venus.\u201d [1] <a href=\"https:\/\/drive.google.com\/file\/d\/1kJHzNMGkrAZRSo9N2K6WT4JmHsdzkTeZ\/view?usp=sharing\">Drive<\/a><\/p>\n<p><a href=\"#_ednref6\" name=\"_edn6\"><span>[vi]<\/span><\/a> \u201cDespite heroic efforts by the geneticist Sewall Wright (1889\u20131988), causal vocabulary was virtually prohibited for more than half a century. And when you prohibit speech, you prohibit thought and stifle principles, methods, and tools\u201d <a href=\"https:\/\/bayes.cs.ucla.edu\/WHY\/why-ch1.pdf\">Pearl\u2019s BoW<\/a> p. 5.<\/p>\n<p><a href=\"#_ednref7\" name=\"_edn7\"><span>[vii]<\/span><\/a> What \u2018scientific research\u2019 is requires a long detour; a simpler approach is to delineate from \u2018non-scientific\u2019 research, which some deem to be the \u2018farthest away\u2019, e.g. theology, philosophy or knowledge through art [7] (it is rather paradoxical to consider philosophy non-scientific, as long as it provides all sciences the main argumentation tool, formal logic).<\/p>\n<p><a href=\"#_ednref8\" name=\"_edn8\"><span>[viii]<\/span><\/a> And this only talks about the \u2018observable\u2019 Universe, which is to the entire Universe like an atom is to the observable Universe\u2026<\/p>\n<p><a href=\"#_ednref9\" name=\"_edn9\"><span>[ix]<\/span><\/a> \u201cThis chapter deals with the problem of con-structing a network automatically from direct empirical observations, thus bypassing the human link in the process known as knowledge acquisition. [\u2026] it is more convenient to execute the learning process in two separate phases: structure learning and parameter learning. [ \u2026] Our focus in this chapter will be on learning structures rather than parameters. [\u2026] e shall focus on causal structures and in particular on causal trees and polytrees, where the computational role of causality as a modularizer of knowledge achieves its fullest realization. \u201c &amp; \u201cCausal labeling creates modularity not only by separating the past from the future, but also by decoupling events occurring at the same time. Knowing the set of immediate causes \u041f<strong><sub>B <\/sub><\/strong>renders X independent of all other variables except X&#8217;s descendant; many of these variables may occur at the same time as X, or even later. In fact, this sort of independence is causality&#8217;s most universal and distinctive characteristic. In medical diagnosis, for example, a group of co-occurring symptoms often become independent of each other once we identify the disease that caused them. When some of the symptoms directly influence each other, the medical profession invents a name for that interaction (e.g., syndrome, complication, or clinical state) and treats it as a\u00a0 new auxiliary variable, which again assumes the modularization role characteristic of causal agents-knowing the state of the auxiliary variable renders the interacting symptoms independent of each other.\u201d {Pearl, 1988 #5884} p. 385 \u20188. Learning Structure from Data\\ 8.1. Causality, Modularity, and Tree Structures\u2019;<\/p>\n<p><a href=\"#_ednref10\" name=\"_edn10\"><span>[x]<\/span><\/a> The graphical causal models in medicine are gaining ground: [10] [11]\u00a0\u00a0 [12] [12-33] [11, 27, 34-45]<\/p>\n<p><a href=\"#_ednref11\" name=\"_edn11\"><span>[xi]<\/span><\/a> \u201cThe calculus of causation consists of two languages: causal diagrams, to express what we know, and a symbolic language, resembling algebra, to express what we want to know. The causal diagrams are simply dot-and-arrow pictures that summarize our existing scientific knowledge. The dots represent quantities of interest, called \u201cvariables,\u201d and the arrows represent known or suspected causal relationships between those variables\u2014namely, which variable \u201c<strong>listens<\/strong>\u201d to which others. \u201c <a href=\"https:\/\/bayes.cs.ucla.edu\/WHY\/\">BoW<\/a> [46], p. 7<\/p>\n<p><a href=\"#_ednref12\" name=\"_edn12\"><span>[xii]<\/span><\/a> Listening from a communication (science) standpoint is also an active and deliberate process, not a mere \u2018reception of data\u2019, but a meaning-seeking activity (\u2018sensemaking\u2019 is another such label, see e.g. in the AI context [47]). Note that a large part of the causality thinking in the 1980\u2019s happened on what was then labeled AI research, or \u2018device behavior\u2019 {Iwasaki, 1986 #15488}, and in \u2018decision science\u2019, see e.g. {G\u00e4rdenfors, 2003 #15502} or {Georgeff, 1988 #15503}.<\/p>\n<p><a href=\"#_ednref13\" name=\"_edn13\"><span>[xiii]<\/span><\/a> \u201cThe relation of energy intake and expenditure to obesity is congruent with the conventional model. However, these components of energy balance are proximate, not root, causes of weight gain. In the compensatory phase (not depicted), insulin resistance increases, and weight gain slows, as circulating fuel concentration rises. (Circulating fuels, as measured in blood, are a proxy for fuel sensing and substrate oxidation in key organs.) Other hormones with effects on adipocytes include sex steroids and cortisol. Fructose may promote hepatic de novo lipogenesis and affect intestinal function, among other actions, through mechanisms independent of, and synergistic with, glucose. Solid red arrows indicate sequential steps in the central causal pathway; associated numbers indicate testable hypotheses as considered in the text. Interrupted red arrows and associated numbers indicate testable hypotheses comprising multiple causal steps. Black arrows indicate other relations. ANS, autonomic nervous\u00a0 system; GIP, glucose-dependent insulinotropic peptide.\u201d Medical diagnosing has been the focus or early \u2018decision science\u2019 works, see the \u2018Causal Understanding\u2019 [49] figure e.g.<\/p>\n<p><a href=\"#_ednref14\" name=\"_edn14\"><span>[xiv]<\/span><\/a> This notation was also used by [50] and by [51] p. 244, and by Lok [52]; Reichenbach also used a close-by symbol : \u201cwhere the events that show a slight variation [from E] are designated by E*:\u201d p.137 [53].<\/p>\n<p><a href=\"#_ednref15\" name=\"_edn15\"><span>[xv]<\/span><\/a> The process is more detailed: we subtract and add one same term to the causal effect, and then \u2018take expectations\u2019, i.e. averages the quantities across patients:<\/p>\n<p>E(CausalEffect<strong><em><sub>i<\/sub><\/em><\/strong>) = E (A1c<strong><em><sub>i<\/sub><\/em><\/strong><sup>If.<strong>LowerBMI<\/strong><\/sup><strong><sub> <\/sub><sub>LoweredBMI<\/sub><\/strong> &#8211; A1c<strong><em><sub>i<\/sub><\/em><\/strong><sup>If.<em>CurrentBMI<\/em><\/sup> <em><sub>Current.BMI<\/sub><\/em>) =<\/p>\n<p>E (A1c<strong><em><sub>i<\/sub><\/em><\/strong><sup>If.<strong>LowerBMI<\/strong><\/sup><strong><sub> <\/sub><sub>LoweredBMI<\/sub><\/strong> &#8211; A1c<strong>*<em><sub>i<\/sub><\/em><\/strong><sup>If.<em>CurrentBMI<\/em><\/sup><strong><sub> <\/sub><sub>LoweredBMI<\/sub><\/strong> ) + E (A1c<strong>*<em><sub>i<\/sub><\/em><\/strong><sup>If.<em>CurrentBMI<\/em><\/sup><strong><sub> <\/sub><sub>LoweredBMI<\/sub><\/strong> &#8211; A1c<strong><em><sub>i<\/sub><\/em><\/strong><sup>If.<em>CurrentBMI<\/em><\/sup> <em><sub>Current.BMI<\/sub><\/em>) =<\/p>\n<p>ATT \u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0 + \u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0 Selection.Bias<\/p>\n<p>Where ATT\u00a0 = the average treatment effect <em><u>on the treated<\/u><\/em>: how much the A1c would \u2026 not drop among those who did lower their BMI, if\u2026 they didn\u2019t! Note, Eric Brunner from UConn stated it using econometric language and labels (\ud835\udc37 is Treatment) : \u201cWe suspect we won\u2019t be able to learn about the causal effect of college education simply by comparing the average levels of earnings by education status because of selection bias:<\/p>\n<p>\ud835\udc38[\ud835\udc4c<strong><sub>\ud835\udc56<\/sub><\/strong>|\ud835\udc37<strong><sub>\ud835\udc56<\/sub><\/strong>=1]\u2212\ud835\udc38[\ud835\udc4c<strong><sub>\ud835\udc56<\/sub><\/strong>|\ud835\udc37<strong><sub>\ud835\udc56<\/sub><\/strong>=0]= \ud835\udc38[\ud835\udc4c<strong><sub>1<\/sub><\/strong><strong><sub>\ud835\udc56<\/sub><\/strong>|\ud835\udc37<strong><sub>\ud835\udc56<\/sub><\/strong>=1]\u2212\ud835\udc38[\ud835\udc4c<strong><sub>0<\/sub><\/strong><strong><sub>\ud835\udc56<\/sub><\/strong>|\ud835\udc37<strong><sub>\ud835\udc56<\/sub><\/strong>=1] (ATT)<\/p>\n<p>+\ud835\udc38[\ud835\udc4c<strong><sub>0<\/sub><\/strong><strong><sub>\ud835\udc56<\/sub><\/strong>|\ud835\udc37<strong><sub>\ud835\udc56<\/sub><\/strong>=1]\u2212\ud835\udc38[\ud835\udc4c<strong><sub>0<\/sub><\/strong><strong><sub>\ud835\udc56<\/sub><\/strong>|\ud835\udc37<strong><sub>\ud835\udc56<\/sub><\/strong>=0] (Selection bias). We suspect that potential outcomes under non-college status are better for those that went to college than for those that did not; i.e. there is positive selection bias\u201d\u00a0 in his \u2018Causal Evaluation\u2019 graduate course handout]; an alternative notation brings both real&amp;potential worlds in one formula, see <a href=\"https:\/\/www.academia.edu\/121704916\/Causality_for_Artificial_Intelligence_From_a_Philosophical_Perspective\">Jordi Vallverd\u00fa&#8217;s<\/a> book (but common among causality writers) \u2018Causality for Artificial Intelligence. From a Philosophical Perspective\u2019: \u201cLet Y(T) represent the potential outcome for a patient with treatment assignment Y (where T is either 0 or 1). We can express the observed outcome Y as follows:<\/p>\n<p>Y = T \u00d7 Y(1) + (1 &#8211; T) \u00d7 Y(0)\u201d or with my \u2018re-labeling\u2019, similar to David A. <a href=\"https:\/\/old.maa.org\/press\/maa-reviews\/statistical-models-and-causal-inference-a-dialogue-with-the-social-sciences\">Freedman<\/a>&#8216;s in fact (Y<sub>i<\/sub> = X<sub>i<\/sub>Y<sub>i<\/sub><sup>T<\/sup> + (1 \u2212 X<sub>i<\/sub>)Y<sub>i<\/sub><sup>C<\/sup>\u00a0:<\/p>\n<p>Y<strong><sub>Realized<\/sub><\/strong> = T \u00d7 Y<strong><sup>If.(T=1) <\/sup><\/strong>+ (1 &#8211; T) \u00d7 Y<strong><sup>If(T=0)<\/sup><\/strong><\/p>\n<p><a href=\"#_ednref16\" name=\"_edn16\"><span>[xvi]<\/span><\/a> We use Average for better intuition; mathematically inclined analysts use instead E(A1c), for \u2018expectation\u2019, which is the proper way to define a \u2018typical patient\u2019s value\u2019, or the mean of a variable. For a variable X with discrete values, the mean is \u03bc<sub>X <\/sub>= E(X) = \u03a3<strong><sub>i<\/sub><\/strong> [x<strong><sub>i<\/sub><\/strong>\u00b7p(x<strong><sub>i<\/sub><\/strong>)], while for one with continuous values \u03bc<sub>X <\/sub>= E(X) =\u00a0 \u00a0([54], p. 152).<\/p>\n<p><a href=\"#_ednref17\" name=\"_edn17\"><span>[xvii]<\/span><\/a> Even presence vs. absence can be sometimes difficult to separate: \u201cWhoever studied newspapers in the last decades came across the difficulties of defining when a person is dead, allowing the organs to be transplanted. Our hair and nails still grow long after our hearts stop beating;\u201d [57] p. 2. This topic of measurement error of binary measures is rarely talked about, in psychometric sense.<\/p>\n<p><a href=\"#_ednref18\" name=\"_edn18\"><span>[xviii]<\/span><\/a> The confounding of \u2018causality\u2019 and \u2018time\u2019 is a known definitional problem: e.g. VanFraasen cites Reichenbach defining time in terms of causality ([58], p, 190), and many define time on the basis of the time construct.<\/p>\n<p>*** Reichenbach himself said e.g. \u201cwe must now make sure that our definition of &#8220;later than&#8221; does not involve circular reasoning. Can we actually recognize what is a cause and what is an effect without knowing their temporal order?<\/p>\n<p>Should we not argue, rather, that of two causally connected events the effect is the later one?<\/p>\n<p>This objection proceeds from the assumption that causality indicates a connection between two events, but does not assign a direction to them. This assumption, however, is erroneous. Causality establishes not a symmetrical but an asymmetrical relation between two events.<\/p>\n<p>If we represent the cause-effect relation by the symbol C, the two cases C(E1, E2) and C(E2, E1) can be distinguished; experience tells us which of the two cases actually occurs. We can state this distinction as follows:<\/p>\n<p>If E1 is the cause of E2. then a small variation (a mark)&#8221; in E1 is associated with a small variation in E2. whereas small variations in E2 are not associated with variations in E1.\u201d [53] P. 136<\/p>\n<p style=\"text-align: center\"><a href=\"https:\/\/health.uconn.edu\/causality\/wp-content\/uploads\/sites\/264\/2025\/04\/FIGURE-1-Dynamic-phase-of-obesity-development.jpg\"><\/a><a href=\"https:\/\/health.uconn.edu\/causality\/wp-content\/uploads\/sites\/264\/2025\/05\/sesitiv.specific.png\"><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/health.uconn.edu\/causality\/wp-content\/uploads\/sites\/264\/2025\/05\/sesitiv.specific-300x66.png\" alt=\"\" width=\"663\" height=\"146\" class=\"alignnone wp-image-98\" srcset=\"https:\/\/health.uconn.edu\/causality\/wp-content\/uploads\/sites\/264\/2025\/05\/sesitiv.specific-300x66.png 300w, https:\/\/health.uconn.edu\/causality\/wp-content\/uploads\/sites\/264\/2025\/05\/sesitiv.specific-1024x224.png 1024w, https:\/\/health.uconn.edu\/causality\/wp-content\/uploads\/sites\/264\/2025\/05\/sesitiv.specific-768x168.png 768w, https:\/\/health.uconn.edu\/causality\/wp-content\/uploads\/sites\/264\/2025\/05\/sesitiv.specific.png 1131w\" sizes=\"(max-width: 663px) 100vw, 663px\" \/><\/a><a href=\"https:\/\/health.uconn.edu\/causality\/wp-content\/uploads\/sites\/264\/2025\/04\/FIGURE-1-Dynamic-phase-of-obesity-development.jpg\">\u00a0<\/a><\/p>\n<p><em>References<\/em><\/p>\n<ol>\n<li>Swinburne, R., <em>The coherence of theism<\/em>. 2016: Oxford University Press.<\/li>\n<li>Hall, E.W., <em>Time and Causality.<\/em> The Philosophical Review, 1934. <strong>43<\/strong>(4): p. 333-350.<\/li>\n<li>Davidson, R., <em>Time and Causality.<\/em> Annals of Economics and Statistics, 2013(109\/110): p. 7-22.<\/li>\n<li>Buehner, M.J., <em>Time and causality<\/em>. 2014, Frontiers Media SA. p. 228.<\/li>\n<li>Kleinberg, S., <em>Causality, probability, and time<\/em>. 2013: Cambridge University Press.<\/li>\n<li>Swinburne, R., <em>Space, Time and Causality.<\/em> 1983.<\/li>\n<li>David, D.O., <em>Metodologia cercet\u0103rii clinice: fundamente<\/em>. 2006: Polirom.<\/li>\n<li>Kaufman, J.S. and S. Kaufman, <em>Assessment of Structured Socioeconomic Effects on Health.<\/em> Epidemiology, 2001. <strong>12<\/strong>(2): p. 157-167.<\/li>\n<li>Kaufman, J.S. and R.S. Cooper, <em>Seeking causal explanations in social epidemiology.<\/em> American journal of epidemiology, 1999. <strong>150<\/strong>(2): p. 113-120.<\/li>\n<li>Caniglia, E.C., et al., <em>Does Reducing Drinking in Patients with Unhealthy Alcohol Use Improve Pain Interference, Use of Other Substances, and Psychiatric Symptoms?<\/em> Alcoholism: Clinical and Experimental Research, 2020. <strong>n\/a<\/strong>(n\/a).<\/li>\n<li>Pearce, N. and D.A. Lawlor, <em>Causal inference\u2014so much more than statistics.<\/em> International journal of epidemiology, 2016. <strong>45<\/strong>(6): p. 1895-1903.<\/li>\n<li>Staplin, N., et al., <em>Use of causal diagrams to inform the design and interpretation of observational studies: an example from the Study of Heart and Renal Protection (SHARP).<\/em> Clinical Journal of the American Society of Nephrology, 2017. <strong>12<\/strong>(3): p. 546-552.<\/li>\n<li>Burden, A.F. and N. Timpson, <em>Ethnicity, heart failure, atrial fibrillation and diabetes: collider bias<\/em>. 2019, BMJ Publishing Group Ltd and British Cardiovascular Society.<\/li>\n<li>Lamprea-Montealegre, J.A., et al., <em>Apolipoprotein B, Triglyceride-Rich Lipoproteins, and Risk of Cardiovascular Events in Persons with CKD.<\/em> Clinical Journal of the American Society of Nephrology, 2020. <strong>15<\/strong>(1): p. 47-60.<\/li>\n<li>Vo, T.-T., et al., <em>The conduct and reporting of mediation analysis in recently published randomized controlled trials: results from a methodological systematic review.<\/em> Journal of clinical epidemiology, 2020. <strong>117<\/strong>: p. 78-88.<\/li>\n<li>Fotheringham, J., et al., <em>The association between longer haemodialysis treatment times and hospitalization and mortality after the two-day break in individuals receiving three times a week haemodialysis.<\/em> Nephrology Dialysis Transplantation, 2019. <strong>34<\/strong>(9): p. 1577-1584.<\/li>\n<li>Jhee, J.H., et al., <em>Secondhand smoke and CKD.<\/em> Clinical Journal of the American Society of Nephrology, 2019. <strong>14<\/strong>(4): p. 515-522.<\/li>\n<li>Woods, G.N., et al., <em>Chronic kidney disease is associated with greater bone marrow adiposity.<\/em> Journal of Bone and Mineral Research, 2018. <strong>33<\/strong>(12): p. 2158-2164.<\/li>\n<li>Storey, B.C., et al., <em>Lowering LDL cholesterol reduces cardiovascular risk independently of presence of inflammation.<\/em> Kidney international, 2018. <strong>93<\/strong>(4): p. 1000-1007.<\/li>\n<li>Waikar, S.S., et al., <em>Association of urinary oxalate excretion with the risk of chronic kidney disease progression.<\/em> JAMA internal medicine, 2019. <strong>179<\/strong>(4): p. 542-551.<\/li>\n<li>Watkins, T.R., <em>Understanding uncertainty and bias to improve causal inference in health intervention research <\/em><a href=\"https:\/\/ses.library.usyd.edu.au\/bitstream\/handle\/2123\/20772\/watkins_tr_thesis.pdf\" class=\"broken_link\"><em>https:\/\/ses.library.usyd.edu.au\/bitstream\/handle\/2123\/20772\/watkins_tr_thesis.pdf<\/em><\/a>. 2018.<\/li>\n<li>Bjornstad, E.C., et al., <em>Racial and health insurance disparities in pediatric acute kidney injury in the USA.<\/em> Pediatric Nephrology, 2020: p. 1-12.<\/li>\n<li>Leite, T.T., et al., <em>Respiratory parameters and acute kidney injury in acute respiratory distress syndrome: a causal inference study.<\/em> Annals of Translational Medicine, 2019. <strong>7<\/strong>(23).<\/li>\n<li>Groenwold, R.H., T.M. Palmer, and K. Tilling, <em>Conditioning on a mediator to adjust for unmeasured confounding <\/em><a href=\"https:\/\/osf.io\/sj7ch\/\"><em>https:\/\/osf.io\/sj7ch\/<\/em><\/a><em>.<\/em> 2019.<\/li>\n<li>T\u00f8rnes, M., et al., <em>Hospital-Level Variations in Rates of Inpatient Urinary Tract Infections in Stroke.<\/em> Frontiers in neurology, 2019. <strong>10<\/strong>: p. 827.<\/li>\n<li>Benedum, C.M., <em>Applications of big data approaches to topics in infectious disease epidemiology<\/em>. 2019, Boston University Available from ProQuest 2241610916.<\/li>\n<li>Gaskell, A.L. and J.W. Sleigh, <em>An Introduction to Causal Diagrams for Anesthesiology Research.<\/em> Anesthesiology: The Journal of the American Society of Anesthesiologists, 2020. <strong>132<\/strong>(5): p. 951-967.<\/li>\n<li>Fotheringham, J., et al., <em>Hospitalization and mortality following non-attendance for hemodialysis according to dialysis day of the week: a European cohort study.<\/em> BMC Nephrology, 2020. <strong>21<\/strong>(1): p. 1-10.<\/li>\n<li>Sawchik, J., J. Hamdani, and M. Vanhaeverbeek, <em>Randomized clinical trials and observational studies in the assessment of drug safety.<\/em> Revue d&#8217;epidemiologie et de sante publique, 2018. <strong>66<\/strong>(3): p. 217-225.<\/li>\n<li>Banack, H.R., et al., <em>Investigating and Remediating Selection Bias in Geriatrics Research: The Selection Bias Toolkit.<\/em> Journal of the American Geriatrics Society, 2019. <strong>67<\/strong>(9): p. 1970-1976.<\/li>\n<li>T\u00f8rnes, M., et al., <em>Does service heterogeneity have an impact on acute hospital length of stay in stroke? A UK-based multicentre prospective cohort study.<\/em> BMJ open, 2019. <strong>9<\/strong>(4): p. e024506.<\/li>\n<li>Ferguson, T.W. and N. Tangri, <em>Identifying Potential Biases in the Nephrology Literature.<\/em> Advances in Chronic Kidney Disease, 2016. <strong>23<\/strong>(6): p. 373-376.<\/li>\n<li>Zeraatkar, D., et al., <em>Methods for the Selection of Covariates in Nutritional Epidemiology Studies: A Meta-Epidemiological Review.<\/em> Current developments in nutrition, 2019. <strong>3<\/strong>(10): p. nzz104.<\/li>\n<li>Birungi, N., et al., <em>Assessing causal effects of early life<\/em><em>\u2010course factors on early childhood caries in 5<\/em><em>\u2010year<\/em><em>\u2010old Ugandan children using directed acyclic graphs (DAG s): A prospective cohort study.<\/em> Community Dentistry and Oral Epidemiology, 2017. <strong>45<\/strong>(6): p. 512-521.<\/li>\n<li>Henckel, L., E. Perkovi\u0107, and M.H. Maathuis, <em>Graphical criteria for efficient total effect estimation via adjustment in causal linear models.<\/em> arXiv preprint arXiv:1907.02435, 2019.<\/li>\n<li>Sourial, N., et al., <em>Testing group differences for confounder selection in nonrandomized studies: flawed practice.<\/em> Canadian Medical Association Journal, 2019. <strong>191<\/strong>(43): p. E1189-E1193.<\/li>\n<li>Shapiro, D.D. and P. Msaouel, <em>Causal Diagram Techniques for Urologic Oncology Research.<\/em> Clinical Genitourinary Cancer, 2020.<\/li>\n<li>Hancox, R.J., <em>When is a confounder not a confounder?<\/em> Respirology, 2019. <strong>24<\/strong>(2): p. 105-106.<\/li>\n<li>Kuehne, F., et al., <em>Guidance for a causal comparative effectiveness analysis emulating a target trial based on big real world evidence: when to start statin treatment.<\/em> Journal of comparative effectiveness research, 2019. <strong>8<\/strong>(12): p. 1013-1025.<\/li>\n<li>Etminan, M., G.S. Collins, and M.A. Mansournia, <em>Using causal diagrams to improve the design and interpretation of medical research.<\/em> Chest, 2020. <strong>158<\/strong>(1): p. S21-S28.<\/li>\n<li>Oates, C.J., et al., <em>Repair of partly misspecified causal diagrams.<\/em> Epidemiology, 2017. <strong>28<\/strong>(4): p. 548-552.<\/li>\n<li>Wang, T., et al., <em>Sensitivity analysis for mistakenly adjusting for mediators in estimating total effect in observational studies.<\/em> BMJ open, 2017. <strong>7<\/strong>(11).<\/li>\n<li>Andreano, A., et al., <em>Adherence to guidelines and breast cancer patients survival: a population-based cohort study analyzed with a causal inference approach.<\/em> Breast Cancer Research and Treatment, 2017. <strong>164<\/strong>(1): p. 119-131.<\/li>\n<li>Lederer, D.J., et al., <em>Control of confounding and reporting of results in causal inference studies. Guidance for authors from editors of respiratory, sleep, and critical care journals.<\/em> Annals of the American Thoracic Society, 2019. <strong>16<\/strong>(1): p. 22-28.<\/li>\n<li>Williamson, E.J., et al., <em>Introduction to causal diagrams for confounder selection.<\/em> Respirology, 2014. <strong>19<\/strong>(3): p. 303-311.<\/li>\n<li>Pearl, J. and D. Mackenzie, <em>The Book of Why: The New Science of Cause and Effect<\/em>. 2018, London, UK: Basic Books, Penguin Books Limited.<\/li>\n<li>Madsbjerg, C., <em>Sensemaking: The Power of the Humanities in the Age of the Algorithm<\/em>. 2017: Grand Central Publishing.<\/li>\n<li>Ludwig, D.S., et al., <em>The carbohydrate-insulin model: a physiological perspective on the obesity pandemic.<\/em> The American Journal of Clinical Nutrition, 2021.<\/li>\n<li>Patil, R., P. Szolovitz, and W. Schwartz. <em>Causal understanding of patient illness in patient diagnosis<\/em>. in <em>Proceedings of AAAI-82<\/em>. 1982.<\/li>\n<li>Joffe, M.M., W.P. Yang, and H.I. Feldman, <em>Selective ignorability assumptions in causal inference.<\/em> The International Journal of Biostatistics, 2010. <strong>6<\/strong>(2).<\/li>\n<li>Kohler, U. and F. Kreuter, <em>Data analysis using Stata<\/em>. 3 ed. 2012: Stata press.<\/li>\n<li>Lok, J.J. and R.J. Bosch, <em>Causal organic indirect and direct effects: Closer to the original approach to mediation analysis, with a product method for binary mediators.<\/em> Epidemiology, 2021. <strong>32<\/strong>(3): p. 412-420.<\/li>\n<li>Reichenbach, H., <em>The philosophy of space and time<\/em>. 1957: Courier Corporation.<\/li>\n<li>Devore, J.L., <em>Probability and statistics for Engineering and the Sciences.<\/em> Pacific Grove: Brooks\/Cole, 2016.<\/li>\n<li>Pearl, J. <em>On the Statistical Interpretation of Structural Equations<\/em>. UCLA Cognitive Systems Laboratory, Technical Report (R-200), 2011.<\/li>\n<li>Pearl, J., <em>[Bayesian Analysis in Expert Systems]: Comment: Graphical Models, Causality and Intervention.<\/em> Statistical Science, 1993. <strong>8<\/strong>(3): p. 266-269.<\/li>\n<li>Stauffer, D., <em>Introduction to statistical physics outside physics.<\/em> Physica A: Statistical Mechanics and Its Applications, 2004. <strong>336<\/strong>(1-2): p. 1-5.<\/li>\n<li>Van Fraassen, B.C., <em>An introduction to the philosophy of time and space.<\/em> 1970.<\/li>\n<li>Varian, H. R. (2016). Causal inference in economics and marketing. Proceedings of the National Academy of Sciences, 113(27), 7310-7315.<\/li>\n<\/ol>\n<p>&nbsp;<\/p>\n","protected":false},"excerpt":{"rendered":"<p>General definitional and notational setup *** Causality is not directly detectable or observable, much like forces[i] or energy aren\u2019t either: it needs to be discovered\/uncovered &#8211; \u2018detected\u2019 through observational consequences[ii]; on the other hand, it\u2019s a pretty natural intuitive human notion[iii], yet is considered a worthy \u2018problem\u2019[iv]. *** Causality is hardly a statistical matter, and [&hellip;]<\/p>\n","protected":false},"author":2514,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"","meta":{"_acf_changed":false,"footnotes":""},"acf":[],"publishpress_future_action":{"enabled":false,"date":"2026-04-12 14:15:05","action":"change-status","newStatus":"draft","terms":[],"taxonomy":""},"_links":{"self":[{"href":"https:\/\/health.uconn.edu\/causality\/wp-json\/wp\/v2\/pages\/16"}],"collection":[{"href":"https:\/\/health.uconn.edu\/causality\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/health.uconn.edu\/causality\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/health.uconn.edu\/causality\/wp-json\/wp\/v2\/users\/2514"}],"replies":[{"embeddable":true,"href":"https:\/\/health.uconn.edu\/causality\/wp-json\/wp\/v2\/comments?post=16"}],"version-history":[{"count":31,"href":"https:\/\/health.uconn.edu\/causality\/wp-json\/wp\/v2\/pages\/16\/revisions"}],"predecessor-version":[{"id":134,"href":"https:\/\/health.uconn.edu\/causality\/wp-json\/wp\/v2\/pages\/16\/revisions\/134"}],"wp:attachment":[{"href":"https:\/\/health.uconn.edu\/causality\/wp-json\/wp\/v2\/media?parent=16"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}