<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
  <channel>
    <title>Keane Hauck</title>
    <description>The hub of failed CSS</description>
    <link>https://keanehauck.com/</link>
    <atom:link href="https://keanehauck.com/feed.xml" rel="self" type="application/rss+xml" />
    
      <item>
        <title>Why the F-statistic is equivalent to the squared t-statistic for a two-group comparison</title>
        <description>&lt;p&gt;Possibly the most horrifying seven words you can encounter when reading a textbook are the following: “The proof is left to the reader.” When learning ANOVA (Analysis of Variance), it is common to encounter the fact that the F-statistic is equal to the squared t-statistic when you are only comparing two groups. Unfortunately, this sentence is often followed by the previously-mentioned proof hand-waving, which can leave readers confused, myself included. Thus, I am going to clearly and concisely demonstrate how the F-test reduces to a t-test under certain conditions.&lt;/p&gt;

&lt;p&gt;Interestingly, this proof is quite elusive. In fact, while researching how to write this up, the only published proof I stumbled across was another blog post from a data scientist named &lt;a href=&quot;https://canovasjm.netlify.app/about/&quot;&gt;Juan Manuel&lt;/a&gt;. Big shoutout to him for his post, which helped me double-check my results.&lt;/p&gt;

&lt;p&gt;This proof assumes familiarity with the ANOVA and t-test framework.&lt;/p&gt;

&lt;h3 id=&quot;setup&quot;&gt;Setup&lt;/h3&gt;

&lt;p&gt;In the notation we’ll be using,&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;$SS_B$: Sum of Squares between groups due to treatment&lt;/li&gt;
  &lt;li&gt;$SS_W$: Sum of Squares within groups due to error&lt;/li&gt;
  &lt;li&gt;$MS_B$: Mean Sum of Squares between groups $=SS_B / k-1$&lt;/li&gt;
  &lt;li&gt;$MS_W$: Mean Sum of Squares within groups $=SS_W / N-k$&lt;/li&gt;
  &lt;li&gt;$N$: Total number of observations&lt;/li&gt;
  &lt;li&gt;$n_i$: Number of observations in group $i$&lt;/li&gt;
  &lt;li&gt;$k$: Number of groups&lt;/li&gt;
  &lt;li&gt;$j$: Index of a single observation&lt;/li&gt;
  &lt;li&gt;$\overline{y}$: The grand mean&lt;/li&gt;
  &lt;li&gt;$\overline{y}_i$: Mean of group $i$&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;A few formulae are important for the following results:&lt;/p&gt;

&lt;p&gt;$t=\frac{\overline{y_1}-\overline{y_2}}{SE_{\overline{y_1}-\overline{y_2}}}$&lt;/p&gt;

&lt;p&gt;where $SE_{\overline{y_1}-\overline{y_2}}=s_p \sqrt{\frac{1}{n_1}+\frac{1}{n_2}}$.&lt;/p&gt;

&lt;p&gt;Sample variance: $s_i^2=\frac{\sum_{j=1}^{n_i}(y_{ij}-\overline{y}_i)^2}{n_i-1}$ for group $i$.&lt;/p&gt;

&lt;p&gt;Pooled sample variance for two groups: $s^2_p = \frac{s^2_1(n_1-1) \ + \ s^2_2(n_2-1)}{N-2}$.&lt;/p&gt;

&lt;p&gt;$F = \frac{MS_{\text{B}}}{MS_{\text{W}}}$&lt;/p&gt;

&lt;p&gt;with $MS_B= \frac{\sum_{i=1}^{k} n_i (\overline{y}_i - \overline{y})^2}{k-1}$&lt;/p&gt;

&lt;p&gt;and $MS_{\text{W}} = \frac{ \sum_{i=1}^{k} \sum_{j=1}^{n_i} \left(y_{ij} - \overline{y}_i \right)^2 }{N-k}$.&lt;/p&gt;

&lt;p&gt;This final $F$ equation is the one we need to reduce to obtain $t^2$.&lt;/p&gt;

&lt;h3 id=&quot;simplifying-denominator&quot;&gt;Simplifying Denominator&lt;/h3&gt;

&lt;p&gt;The first step is to figure out what the denominator of the expression simplifies to. With two groups, $MS_{W}=\frac{SS_{W}}{N-2}$.&lt;/p&gt;

&lt;p&gt;Again, with two groups,&lt;/p&gt;

&lt;p&gt;$SS_W= \sum_{i=1}^{k} \sum_{j=1}^{n_i}(y_{ij} - \overline{y}_i)^2=\sum_{j=1}^{n_1}(y_{1j}-\overline{y}_1)^2 + \sum_{j=1}^{n_2}(y_{2j}-\overline{y}_2)^2$.&lt;/p&gt;

&lt;p&gt;Thus, $MS_{W}= \frac{\sum_{i=1}^{k} \sum_{j=1}^{n_i}(y_{ij} - \overline{y}_i)^2}{N-k}=\frac{\sum_{j=1}^{n_1}(y_{1j}-\overline{y}_1)^2\ +\ \sum_{j=1}^{n_2}(y_{2j}-\overline{y}_2)^2}{N-2}$ for two groups.&lt;/p&gt;

&lt;p&gt;The goal is to represent this equation in terms of something we recognize. We know that sample variance, $s_i^2$, is equal to $\frac{\sum_{j=1}^{n_i}(y_{ij}-\overline{y}_i)^2}{n_i-1}$ for group $i$. We might notice that the numerator of this equation shows up twice in our current $MS_W$ formula. We can make sample variance itself show up by multiplying both the numerator and denominator of $MS_W$ by $\frac{1}{(n_1-1)(n_2-1)}$.&lt;/p&gt;

&lt;p&gt;$MS_W \cdot \frac{\frac{1}{(n_1-1)(n_2-1)}}{\frac{1}{(n_1-1)(n_2-1)}}=\frac{\sum_{j=1}^{n_1}(y_{1j}-\overline{y}_1)^2\ +\ \sum_{j=1}^{n_2}(y_{2j}-\overline{y}_2)^2}{N-2} \cdot \frac{\frac{1}{(n_1-1)(n_2-1)}}{\frac{1}{(n_1-1)(n_2-1)}}$&lt;/p&gt;

&lt;p&gt;$= \frac{\sum_{j=1}^{n_1}(y_{1j}-\overline{y}_1)^2/(n_1-1)(n_2-1)\ +\ \sum_{j=1}^{n_2}(y_{2j}-\overline{y}_2)^2/(n_1-1)(n_2-1)}{(N-2)/(n_1-1)(n_2-1)}=\frac{s^2_1 / (n_2-1) \ + \ s^2_2 / (n_1-1)}{(N-2)/(n_1-1)(n_2-1)}$.&lt;/p&gt;

&lt;p&gt;This is starting to look a lot like the pooled variance estimator for two groups, $s^2_p = \frac{s^2_1(n_1-1) \ + \ s^2_2(n_2-1)}{N-2}$. If we multiply our current equation by $\frac{(n_1-1)(n_2-1)}{(n_1-1)(n_2-1)}$ we obtain exactly this formula:&lt;/p&gt;

&lt;p&gt;$\frac{s^2_1 / (n_2-1) \ + \ s^2_2 / (n_1-1)}{(N-2)/(n_1-1)(n_2-1)} \cdot \frac{(n_1-1)(n_2-1)}{(n_1-1)(n_2-1)} = \frac{s^2_1(n_1-1) \ + \ s^2_2(n_2-1)}{N-2} = s^2_p$.&lt;/p&gt;

&lt;p&gt;Thus, we have proven the denominator of the two-group $F$ formula is equal to the pooled variance estimator.&lt;/p&gt;

&lt;h3 id=&quot;simplifying-numerator&quot;&gt;Simplifying Numerator&lt;/h3&gt;

&lt;p&gt;The numerator, $MS_B$, is a little trickier. We know that $t=\frac{\overline{y_1}-\overline{y_2}}{s_p \sqrt{\frac{1}{n_1}+\frac{1}{n_2}}}$ and $t^2= \frac{(\overline{y_1}-\overline{y_2})^2}{s^2_p (\frac{1}{n_1}+\frac{1}{n_2})}$. If our denominator is equivalent to $s^2_p$, the numerator must equal $\frac{(\overline{y_1}-\overline{y_2})^2}{ (\frac{1}{n_1}+\frac{1}{n_2})}$ for our overall formula to equal $t^2$. Let’s prove it.&lt;/p&gt;

&lt;p&gt;$MS_B= \frac{\sum_{i=1}^{k} n_i (\overline{y}_i - \overline{y})^2}{k-1}$. For two groups, $k-1=1$ and this equation simplifies to $MS_B= n_1(\overline{y}_1 - \overline{y})^2\ + \ n_2(\overline{y}_2 - \overline{y})^2$ by expanding the sum for $k=2$.&lt;/p&gt;

&lt;p&gt;We can represent the grand mean $\overline{y}$ as a weighted average $\overline{y}= \frac{n_1 \overline{y}_1 + n_2 \overline{y}_2}{N}$ with $N=n_1 + n_2$.&lt;/p&gt;

&lt;p&gt;If we substitute it into our equation for two groups, we get $MS_B= n_1(\overline{y}_1 - \frac{n_1 \overline{y}_1 + n_2 \overline{y}_2}{N})^2 + n_2(\overline{y}_2 - \frac{n_1 \overline{y}_1 + n_2 \overline{y}_2}{N})^2$.&lt;/p&gt;

&lt;p&gt;This equation consists of two similar effect terms, one for each group. We will now focus on one of these terms to simplify it, with the logic that the other term will simplify in the same manner. Let’s examine $n_1(\overline{y}_1 - \frac{n_1 \overline{y}_1 + n_2 \overline{y}_2}{N})^2$.&lt;/p&gt;

&lt;p&gt;We will now make a few algebraic steps. Each should be clear and followable.&lt;/p&gt;

\[\begin{align*}
    n_1(\overline{y}_1 - \frac{n_1 \overline{y}_1 + n_2 \overline{y}_2}{N})^2 &amp;amp;= n_1(\frac{\overline{y}_1N}{N} - \frac{n_1 \overline{y}_1 + n_2 \overline{y}_2}{N})^2 \\
            &amp;amp;= n_1(\frac{n_1 \overline{y}_1+n_2 \overline{y}_1}{N} - \frac{n_1 \overline{y}_1 + n_2 \overline{y}_2}{N})^2 \ \ \ (\text{since} \ N=n_1+n_2) \\
            &amp;amp;= n_1(\frac{n_2 \overline{y}_1 - n_2 \overline{y}_2}{N})^2 \\
            &amp;amp;= n_1(\frac{n_2 (\overline{y}_1 - \overline{y}_2)}{N})^2 \\
            &amp;amp;= n_1(\frac{n_2^2 (\overline{y}_1 - \overline{y}_2)^2}{N^2}) \\
            &amp;amp;= \frac{n_1n_2^2}{N^2}(\overline{y}_1 - \overline{y}_2)^2.
\end{align*}\]

&lt;p&gt;Thus, our first term $n_1(\overline{y}_1 - \frac{n_1 \overline{y}_1 + n_2 \overline{y}_2}{N})^2$ is equivalent to $\frac{n_1n_2^2}{N^2}(\overline{y}_1 - \overline{y}_2)^2$. The same process can be repeated for the second term (switching the $n$’s and $\overline{y}’s$ as appropriate), which obtains $n_2(\overline{y}_2 - \frac{n_1 \overline{y}_1 + n_2 \overline{y}_2}{N})^2 = \frac{n_2n_1^2}{N^2}(\overline{y}_2 - \overline{y}_1)^2$.&lt;/p&gt;

&lt;p&gt;So, our whole equation for $MS_B$ can be rewritten as $MS_B = n_1(\overline{y}_1 - \frac{n_1 \overline{y}_1 + n_2 \overline{y}_2}{N})^2 + n_2(\overline{y}_2 - \frac{n_1 \overline{y}_1 + n_2 \overline{y}_2}{N})^2 = \frac{n_1n_2^2}{N^2}(\overline{y}_1 - \overline{y}_2)^2 + \frac{n_2n_1^2}{N^2}(\overline{y}_2 - \overline{y}_1)^2$.&lt;/p&gt;

&lt;p&gt;By definition, $(\overline{y}_1 - \overline{y}_2)^2 = (\overline{y}_2 - \overline{y}_1)^2$, so the entire equation simplifies to $[\frac{n_1n_2^2}{N^2} + \frac{n_2n_1^2}{N^2}](\overline{y}_1 - \overline{y}_2)^2$.&lt;/p&gt;

&lt;p&gt;We will conduct one final sequence of algebraic steps:&lt;/p&gt;

\[\begin{align*}
[\frac{n_1n_2^2}{N^2} + \frac{n_2n_1^2}{N^2}]
(\overline{y}_1 - \overline{y}_2)^2
&amp;amp;= [\frac{n_1n_2^2+n_2n_1^2}{N^2}]
(\overline{y}_1 - \overline{y}_2)^2 \\

&amp;amp;= [\frac{n_1n_2n_2+n_2n_1n_1}{N^2}]
(\overline{y}_1 - \overline{y}_2)^2 \\

&amp;amp;= [\frac{n_1n_2(n_1+n_2)}{N^2}]
(\overline{y}_1 - \overline{y}_2)^2 \\

&amp;amp;= [\frac{n_1n_2(n_1+n_2)}{(n_1+n_2)^2}]
(\overline{y}_1 - \overline{y}_2)^2 \\

&amp;amp;= [\frac{n_1n_2}{n_1+n_2}]
(\overline{y}_1 - \overline{y}_2)^2 \\

&amp;amp;= [\frac{1}{\frac{n_1+n_2}{n_1n_2}}]
(\overline{y}_1 - \overline{y}_2)^2 \\

&amp;amp;= [\frac{1}{\frac{n_1}{n_1n_2}+\frac{n_2}{n_1n_2}}]
(\overline{y}_1 - \overline{y}_2)^2 \\

&amp;amp;= [\frac{1}{\frac{1}{n_2}+\frac{1}{n_1}}]
(\overline{y}_1 - \overline{y}_2)^2 \\

&amp;amp;= \frac{(\overline{y}_1 - \overline{y}_2)^2}
{\frac{1}{n_2}+\frac{1}{n_1}}.
\end{align*}\]

&lt;p&gt;Thus, for two groups our numerator $MS_B$ is equivalent to $\frac{(\overline{y}_1 - \overline{y}_2)^2}{\frac{1}{n_2}+\frac{1}{n_1}}$.&lt;/p&gt;

&lt;h3 id=&quot;conclusion&quot;&gt;Conclusion&lt;/h3&gt;

&lt;p&gt;The F-statistic is the ratio of $\frac{MS_B}{MS_W}$. We have shown that for two groups ($k=2$) this is equivalent to the ratio $\frac{(\overline{y}_1 - \overline{y}_2)^2}{s^2_p(\frac{1}{n_2}+\frac{1}{n_1})}$, which is exactly equal to $t^2$.&lt;/p&gt;

&lt;p&gt;And there you have it.&lt;/p&gt;

&lt;h3 id=&quot;ps&quot;&gt;P.S.&lt;/h3&gt;

&lt;p&gt;There are obviously multiple ways of proving this result. A much simpler version is in the general linear model framework where you can show that the test of the coefficient for a single-predictor (like group membership) model follows a $t$ distribution with $t_{obs}=\frac{b_1}{s_{b_1}}$. It can then be shown that squaring this is equivalent to the observed $F$ value. However, I wanted to highlight a method of proof that only requires familiarity with the ANOVA framework and some solid algebra skills. I hope this is helpful for anyone like me who was curious about this often-stated fact about hypothesis testing.&lt;/p&gt;
</description>
        <pubDate>Mon, 09 Mar 2026 00:00:00 +0000</pubDate>
        <link>https://keanehauck.com/statistics/proof/anova/2026/03/09/ftsquared/</link>
        <guid isPermaLink="true">https://keanehauck.com/statistics/proof/anova/2026/03/09/ftsquared/</guid>
      </item>
    
      <item>
        <title>24 Hour Planet Survival Challenge</title>
        <description>&lt;p&gt;&lt;img src=&quot;/images/posts/planet-survival/planets.jpg&quot; alt=&quot;planets&quot; /&gt;&lt;/p&gt;

&lt;p&gt;So I was thinking about something as I fell asleep last night. I was reading about the environmental conditions of Venus—surprise: it’s incredibly hazardous and unsuitable for life. At sea level, the atmospheric pressure is about 93x that of Earth, and the temperature averages around &lt;a href=&quot;https://en.wikipedia.org/wiki/Venus#Atmosphere_and_climate&quot;&gt;850 °F&lt;/a&gt;. Obviously any human that found themselves transported to the surface of Venus would be immediately obliterated. But this begs the question—what planets in the solar system are the most hospitable to human life? The least?&lt;/p&gt;

&lt;p&gt;Thus I came up with an idea for a challenge. If I hypothetically had access to unlimited materials and a year’s worth of time, how difficult would it be for me to build a contraption that would allow me to survive on each planet in the solar system for 24 hours? I’m going to step through my process of hypothetical survival and rank each planet in terms of its survivability.&lt;/p&gt;

&lt;h3 id=&quot;ground-rules&quot;&gt;Ground rules&lt;/h3&gt;

&lt;p&gt;A few basic rules are in order.&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;It’s assumed that I &lt;em&gt;can&lt;/em&gt; eventually build each contraption.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Obviously I know nothing about building random machines. However, with unlimited access to materials (including money) and a full calendar year, I would be able to contract with different manufacturing companies to theoretically construct whatever I wanted. We could also just assume that I had a giant 3D printer that could spit out whatever machine I wanted. Either way, this is the least interesting part of the challenge so I’m just going to make it a given.&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;The contraptions need to be as simple as possible.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;It’s totally uninteresting to just surround myself with a 1000-foot thick ball of steel and plop it on the surface of some random planet (According to the &lt;a href=&quot;https://en.wikipedia.org/wiki/Thermal_diffusivity&quot;&gt;heat equation&lt;/a&gt; a 1000-foot thick ball of steel will definitely maintain a stable interior temperature for 24 hours. However, we can do better. The thermal diffusivity chart on wikipedia suggests that steel is a rather poor medium for our giant ball. Rubber actually fares really well, followed by glass). Therefore, the challenge is as interesting as possible if our goal is to make a simple, exposed contraption.&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;My contraption and I will be teleported to surface level of the planet (if it has a surface)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Similarly to the last point, it’s kind of cheating to say I can just float 5 miles above the surface, where no storms/gravity/temperatures can reach me. At the logical extreme, what’s to stop me from just orbiting a planet for 24h? This rule ensures I have to at least survive at a reasonable altitude. If the planet is a gas giant like Jupiter, we’ll discuss what seems fair as we get to it.&lt;/p&gt;

&lt;p&gt;Ok, those are the rules! I’ll go through the ranking from most survivable to least survivable.&lt;/p&gt;

&lt;h3 id=&quot;1-mars&quot;&gt;1. Mars&lt;/h3&gt;

&lt;p&gt;Perhaps the biggest shock of this entire ranking is that Earth isn’t the most survivable planet. The crux of the matter boils down to this: Earth has surface water, while &lt;a href=&quot;https://en.wikipedia.org/wiki/Mars&quot;&gt;Mars does not&lt;/a&gt;. I have to account for the fact that about &lt;a href=&quot;https://www.usgs.gov/water-science-school/science/how-much-water-there-earth&quot;&gt;71% of Earth’s surface is covered with water&lt;/a&gt;, so if I’m teleported to a random location at surface level there’s a good chance I end up in one of the oceans. Mars, on the other hand, is predictable. Most of its surface is dry, cold, and rocky, which makes it much simpler to figure out how to survive. NASA has &lt;a href=&quot;https://en.wikipedia.org/wiki/Mars_suit&quot;&gt;developed theoretical suits for walking on Mars&lt;/a&gt; and they have oxygen tanks rated for longer than 24 hours, so all I would need to do is contact NASA to build one for me and train me how to use it. Not only is it totally survivable (barring suit malfunctions), but this scenario might end up being the coolest trip of my entire life: walking around Mars and exploring the red landscape for 24 hours. Unbelievably, the biggest issue might be how I would use the bathroom.&lt;/p&gt;

&lt;h3 id=&quot;2-earth&quot;&gt;2. Earth&lt;/h3&gt;

&lt;p&gt;Mother Earth really let us down here. If I could pick and choose the location I’m being teleported to, there would be &lt;a href=&quot;https://www.daveandbusters.com/us/en/about/locations/canada-vaughan/&quot;&gt;no problem at all&lt;/a&gt;. Factoring for the oceans creates some problems. My contraption could be a cruise ship, but is this necessary? I think we can do this without even needing a boat of any sort. The &lt;a href=&quot;https://www.xingdunmarine.net/life-jackets/solas-inflatable-life-jacket/&quot;&gt;SOLAS standards&lt;/a&gt; for life jacket buoyancy require them to lose less than 5% buoyancy over the span of 24 hours, so we’re all good there. The bigger threat would be the temperature. A &lt;a href=&quot;https://www.ospo.noaa.gov/data/sst/contour/global_small.c.gif&quot;&gt;large percentage&lt;/a&gt; of the ocean has temperatures below freezing, so I would need to bundle up. We could get technical with the temperature ratings of different wetsuits, but I think there’s an easier solution.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/planet-survival/divingsuit.jpg&quot; alt=&quot;divingsuit&quot; /&gt;&lt;/p&gt;

&lt;p&gt;An &lt;a href=&quot;https://en.wikipedia.org/wiki/Atmospheric_diving_suit#Current_suits&quot;&gt;atmospheric diving suit&lt;/a&gt; has just about everything I’d need to survive 24h in the ocean: thruster maneuverability, a maintained internal atmospheric pressure/temperature, and a 24-hour emergency backup oxygen tank. So, I’d just secure a bunch of life jackets to one of those suits. If the lifejackets fail, I should be able to jet around with the thrusters to avoid getting too deep. I can also use my year of time to learn how to maneuver my suit. And in the off-chance I get teleported to land, I can just walk around in my modern suit of armor to avoid any natural predators. Easy survival.&lt;/p&gt;

&lt;h3 id=&quot;3-mercury&quot;&gt;3. Mercury&lt;/h3&gt;

&lt;p&gt;The &lt;a href=&quot;https://en.wikipedia.org/wiki/World_Sauna_Championships&quot;&gt;World Sauna Championships&lt;/a&gt; was an annual “heat-endurance” event held in Finland from 1999 to 2010. Each year, contestants would sit in an extremely hot sauna until they gave up, with the winner being the last person to be able to walk out of the sauna without outside assistance. Temperatures typically started at 110°C. In 2010, the event was permanently cancelled after a competitor unfortunately passed away due to complications from extreme heat. He had sat in the sauna for only 6 minutes. The moral? Surviving extreme heat is really difficult.&lt;/p&gt;

&lt;p&gt;This is the problem we run into with Mercury. Thanks to its nonexistent atmosphere and slow spin, the side of the planet facing away from the sun is extremely cold (-170°C) while the side facing the sun is extremely hot (420°C). A simple space suit isn’t going to cut it. Most of the suits designed for Mars or other space exploration are designed with insulation from cold in mind, not insulation from pure heat.&lt;/p&gt;

&lt;p&gt;Could a Honda Pilot be insulated enough to withstand 420°C for 24 hours? If so I could be teleported to Mercury in my car. I stumbled across a &lt;a href=&quot;https://www.modusadvanced.com/resources/blog/azw-la-ii-white-thermal-control-paint?utm_source=chatgpt.com&quot;&gt;website&lt;/a&gt; advertising &lt;em&gt;thermal control paint for spacecraft and satellite applications&lt;/em&gt;, which has to have the smallest target audience of any marketing website I’ve ever seen. They report that their thermal control paint, AZW/LA-II, operates in temperatures ranging from -180°C to 1400°C, which is perfect for our Mercury use! On the website, they also describe the important components of an overall temperature control system, including radiator surfaces, heat pipes, and thermal isolation systems. This company seems like they know what they’re talking about. So, all I’d have to do is contract with them to get my Pilot outfitted with the necessary components of an interstellar heat reduction system, and I’d be good to go! I could sit in my space suit for oxygen and for protection from the near-vacuum atmospheric conditions, and hang tight.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/planet-survival/pilotv1.png&quot; alt=&quot;pilotv1&quot; /&gt;
&lt;em&gt;My Pilot, complete with thermal shields, reflective coating, and a heat sink&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Also, when reading the wikipedia page for Mercury I saw this:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/planet-survival/badnews.png&quot; alt=&quot;badnews&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Better start selling those stocks.&lt;/p&gt;

&lt;h3 id=&quot;4-venus&quot;&gt;4. Venus&lt;/h3&gt;

&lt;p&gt;Now we’re getting really tough. Venus is the planet I mentioned in the introduction. As previously stated, Venus is very hot, albeit not as bad as Mercury. Surface temperatures are around 460°C, which is theoretically survivable in our heat-resistant Pilot. The problem(s) are twofold: the atmospheric pressure and the weather conditions.&lt;/p&gt;

&lt;p&gt;Venus’ atmosphere is very dense—its pressure is about 93 times higher thatn Earth’s atmosphere at &lt;a href=&quot;https://en.wikipedia.org/wiki/Venus&quot;&gt;surface level&lt;/a&gt;. At these pressures, our current contraption will be instantly crushed. We might need to abandon the Pilot.&lt;/p&gt;

&lt;p&gt;The &lt;a href=&quot;https://deepseachallenge.com/&quot;&gt;Deepsea Challenger&lt;/a&gt; is a submersible designed (with notably fewer critics than &lt;em&gt;Titan&lt;/em&gt; had) to withstand the pressure at the deepest parts of the ocean (depths of around 11km). According to Wikipedia, the pressure on Venus is equatable to the pressure of a water depth on Earth of about 1km. So, this submersible is well above the threshold for pressure survivability. As a bonus, it comes with air tanks, so I don’t need to worry about sitting in a space suit for the whole day. We could equip it with all the aforementioned heat shielding and be good.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/planet-survival/challengerv1.png&quot; alt=&quot;challengerv1&quot; /&gt;&lt;/p&gt;

&lt;p&gt;The weather might be the bigger problem. According to the website for our thermal control coating, the coating is susceptible to degradation from radiation, atomic oxygen, thermal cycling, and micrometeoroid impacts. The first three aren’t really a problem on Venus but the last one is an issue. Any surface exposed to the Venus weather will be constantly buffeted by rocks and other debris thanks to wind speeds of up to 300km/h. Also, it rains acid. Venus is a lovely place.&lt;/p&gt;

&lt;p&gt;One solution is to encase our contraption in a weather resistant dome shield. I have a cooler solution in mind.&lt;/p&gt;

&lt;p&gt;A &lt;a href=&quot;https://en.wikipedia.org/wiki/Tunnel_boring_machine&quot;&gt;tunnel boring machine&lt;/a&gt; is a large machine used for quickly drilling large tunnels. They’re commonly used to excavate large amounts of earth for subway systems or traffic tunnels.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/planet-survival/tbm.png&quot; alt=&quot;tbm&quot; /&gt;&lt;/p&gt;

&lt;p&gt;If we put one on the bottom of the Deepsea Challenger, could we dig ourselves into the ground quickly enough to escape the weather? Soil-tunneling machines can excavate at a rate of around 200m/week. This equates to 28.5m/day, and about 1.2 meters per hour. The dimensions of the submersible are 2.3m (7.7 ft.) X 1.7m (5.5 ft.) X 8.1m (26.6 ft.). If we laid it on its side, it would be dug underground in less than two hours with an industrial boring machine. Of course, the top would still be exposed, but we could just limit our weather shield to a disk at the top. I think this idea’s a winner.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/planet-survival/challengerv2.png&quot; alt=&quot;challengerv2&quot; /&gt;&lt;/p&gt;

&lt;h3 id=&quot;5-uranus&quot;&gt;5. Uranus&lt;/h3&gt;

&lt;p&gt;To no one’s surprise, the Gas Giants rank last. The final four planets have no defined surface, so it’s hard to determine what altitude I’d get teleported to. For each, I think it’s fair to teleport to the part of the atmosphere which corresponds to a pressure of about 1 Earth atmosphere. You may kvetch and say “well, that’s ignoring pressure! No fair!” and to that I say “wait and see how bad these planets get.” We’ll speed through these, as the contraption is very similar for each.&lt;/p&gt;

&lt;p&gt;All of these planets will kill you. All feature astronomical wind speeds, massive storms, extreme radiation, huge temperature fluctuations, and no stable surface. These rankings are pretty much determined by how easy it would be to fly a plane in their atmosphere. In researching for this post, I found an &lt;a href=&quot;https://what-if.xkcd.com/30/&quot;&gt;XKCD&lt;/a&gt; that helpfully described the flight conditions for all the planets (there’s always a relevant XKCD). I’ll be agreeing with some of the rankings, but not all.&lt;/p&gt;

&lt;p&gt;Uranus is the calmest of the bunch, relatively speaking. At 1 atm (atmosphere), wind speeds range from 360 to 180 km/h, which is very very fast. There is relatively little radiation compared to the other gas giants—about 100 times weaker than Jupiter, but we still need the lead walls. The gravity of Uranus is around .9g, so flying/floating is still difficult.&lt;/p&gt;

&lt;p&gt;The Lockheed WC-130 is an American military plane designed to &lt;a href=&quot;https://en.wikipedia.org/wiki/Lockheed_WC-130&quot;&gt;fly into the center of a hurricane&lt;/a&gt;. This will be our base plane for surviving the planets. We’ll redesign the walls to be lined with lead to block potential radiation rays, and add the thermal coating previously discussed.
Here’s me on Uranus:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/planet-survival/uranus.png&quot; alt=&quot;uranus&quot; /&gt;&lt;/p&gt;

&lt;h3 id=&quot;6-saturn&quot;&gt;6. Saturn&lt;/h3&gt;

&lt;p&gt;Saturn is somehow orders of magnitude worse. The winds achieve speeds of up to &lt;em&gt;1,800 km/h&lt;/em&gt;, which is just absurd. We can try and super-reinforce our Lockheed, but it’s not looking good. Our only chance of survival is if we are teleported to a patch of the atmosphere where the storms are relatively tame. If we can fly with one long jet stream and not hit any vertical shear, we might be able to fly until we hit turbulence.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/planet-survival/saturn.png&quot; alt=&quot;saturn&quot; /&gt;&lt;/p&gt;

&lt;h3 id=&quot;7-neptune&quot;&gt;7. Neptune&lt;/h3&gt;

&lt;p&gt;Not looking too good so far. Neptune has the strongest winds (2,200 km/h) and largest storms of any planet in the solar system. The good news is that gravity is only about 1.1g, which means we can still theoretically fly a plane. The bad news is that we don’t know of any potential plane wing material strong enough to withstand these atmospheric conditions. Winds on Saturn are almost as fast as Neptunes, but Saturn’s storms and cloud patterns are much less volatile. Neptune storms have constantly-shifting wind patterns which make flying much more difficult. Our plane will get torn apart within minutes of trying to exist in such extreme storm conditions.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/planet-survival/neptune.png&quot; alt=&quot;neptune&quot; /&gt;&lt;/p&gt;

&lt;h3 id=&quot;8-jupiter&quot;&gt;8. Jupiter&lt;/h3&gt;

&lt;p&gt;We’re dead. Not only does Jupiter have immense storms with &lt;a href=&quot;https://web.archive.org/web/20111008010724/http://www.nasa.gov/vision/universe/solarsystem/galileo_end.html&quot;&gt;lightning as strong as atomic bombs&lt;/a&gt; that strikes around &lt;a href=&quot;https://www.jpl.nasa.gov/news/juno-solves-39-year-old-mystery-of-jupiter-lightning/&quot;&gt;four times per second in some thunderheads&lt;/a&gt;, but the gravity is 2.4g: our plane isn’t staying in the air. The only possible solution is to try and add massive jet engines to our Lockheed to try and offset the massive differential. Oh, did I also mention the constant bombardment of extremely lethal radiation? Teleporting to the atmosphere of Jupiter would result in our contraption being immediately shredded by unbelievably turbulent winds, blasted by giga-lightning, and rapidly pulled to crush depths, probably in that order.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/planet-survival/jupiter.png&quot; alt=&quot;jupiter&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Don’t mess with Jupiter.&lt;/p&gt;

&lt;h3 id=&quot;conclusion&quot;&gt;Conclusion&lt;/h3&gt;

&lt;p&gt;And there you have it. A definitive ranking of how easy each planet would be to survive for 24 hours. My main takeaway from all of this? Space is scary. Good thing we have Mother Earth.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/planet-survival/boot2.png&quot; alt=&quot;boot2&quot; /&gt;&lt;/p&gt;
</description>
        <pubDate>Sat, 31 Jan 2026 00:00:00 +0000</pubDate>
        <link>https://keanehauck.com/hypothetical/challenge/2026/01/31/planet-survival/</link>
        <guid isPermaLink="true">https://keanehauck.com/hypothetical/challenge/2026/01/31/planet-survival/</guid>
      </item>
    
      <item>
        <title>How to Write Greek Letters</title>
        <description>&lt;p&gt;&lt;img src=&quot;/images/posts/greek-letters/spikedmath.jpg&quot; alt=&quot;spikedmath&quot; /&gt;&lt;/p&gt;

&lt;p&gt;What does every great statistician need the ability to do? If you answered &lt;em&gt;easily explain complicated theorems&lt;/em&gt; or &lt;em&gt;derive complex formulae by hand&lt;/em&gt;, you are incorrect. The true answer is that every stats person needs to be able to draw Greek letters on a whiteboard without them looking like scribbly blobs.&lt;/p&gt;

&lt;p&gt;This is something I am unsurprisingly bad at. To help myself (and hopefully share this knowledge with others), I’ve compiled a list of Greek letters and how to draw them. With the hefty assistance of some LLMs, I created a rainbow sketchpad that shows the direction of a line through its color (red $\rightarrow$ purple). So, as you browse through the list of letters, you can tell in which direction the pen stroke should start when handwriting. You can also have fun experimenting with the sketchpad below:&lt;/p&gt;

&lt;iframe id=&quot;rainbowFrame&quot; src=&quot;/images/posts/greek-letters/rainbow-draw.html&quot; width=&quot;100%&quot; style=&quot;border:none; overflow:hidden;&quot; scrolling=&quot;no&quot;&gt;
&lt;/iframe&gt;

&lt;script&gt;
window.addEventListener(&quot;message&quot;, (ev) =&gt; {
  if (ev.data &amp;&amp; ev.data.type === &quot;resize-iframe&quot;) {
    const iframe = document.getElementById(&quot;rainbowFrame&quot;);
    if (iframe) iframe.style.height = ev.data.height + &quot;px&quot;;
  }
});
&lt;/script&gt;

&lt;h3 id=&quot;the-greek-alphabet&quot;&gt;The Greek Alphabet&lt;/h3&gt;

&lt;p&gt;If we get the term &lt;em&gt;alphabet&lt;/em&gt; from the Greek system, do we really need to clarify that we’re using the Greek alphabet? Technically, we should be calling our system the “English Alphabet.”&lt;/p&gt;

&lt;p&gt;Here’s a list of all the letters and how I write them (starting on red). As a disclaimer, these depictions are gotten from use in mathematics or statistics, not everyday handwriting. So, some of them might be needlessly complex or not follow convention.&lt;/p&gt;

&lt;table&gt;
  &lt;thead&gt;
    &lt;tr&gt;
      &lt;th&gt;Uppercase&lt;/th&gt;
      &lt;th&gt;Lowercase&lt;/th&gt;
    &lt;/tr&gt;
  &lt;/thead&gt;
  &lt;tbody&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Alpha&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/ualpha.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Alpha&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/lalpha.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;I used to draw it starting from the bottom, but starting from the top seems to give it a more full circle.&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Beta&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/ubeta.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Beta&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/lbeta.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;The slant is important.&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Gamma&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/ugamma.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Gamma&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/lgamma.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Delta&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/udelta.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Delta&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/ldelta.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Epsilon&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/uepsilon.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Epsilon&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/lepsilon.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Zeta&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/uzeta.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Zeta&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/lzeta.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt; I have no clue how to write this one, really. I don’t think I’ve ever used it.&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Eta&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/ueta.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt; I think the slant makes it look more “Greek.”&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Eta&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/leta.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Theta&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/utheta.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Theta&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/ltheta.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Iota&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/uiota.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt; Tough one.&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Iota&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/liota.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Kappa&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/ukappa.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Kappa&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/lkappa.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt; I like to put a little tail at the top of the main stroke to distinguish it from a “k.”&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Lambda&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/ulambda.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Lambda&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/llambda.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt; Half-Life reference.&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Mu&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/umu.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Mu&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/lmu.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Nu&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/unu.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Nu&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/lnu.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt; Again, the slight curve distinguishes it from a “v.”&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Xi&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/uxi.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt; What in the world is this thing?&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Xi&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/lxi.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt; The most feared of them all. Known far and wide by matrix algebraists. Not too bad if you think of it as a lowercase zeta with an extra loop.&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Omicron&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/uomicron.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt; Yippee! Easy.&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Omicron&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/lomicron.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Pi&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/upi.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt; Looks like a &lt;a href=&quot;https://en.wikipedia.org/wiki/Torii&quot;&gt;Torii&lt;/a&gt;.&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Pi&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/lpi.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Rho&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/urho.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Rho&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/lrho.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Sigma&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/usigma.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Sigma&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/lsigma.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Tau&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/utau.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Tau&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/ltau.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt; Another little curve to make it look more Greek than a “t.”&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Upsilon&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/uupsilon.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Upsilon&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/lupsilon.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Phi&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/uphi.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt; Sometimes serifs are used at the top and bottom to distinguish it from the lowercase phi.&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Phi&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/lphi.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Chi&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/uchi.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Chi&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/lchi.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Psi&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/upsi.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt; The serif on the right side (and size of the symbol) differentiate it from the lowercase version.&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Psi&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/lpsi.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Omega&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/uomega.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Omega&lt;/strong&gt;&lt;br /&gt;&lt;img src=&quot;/images/posts/greek-letters/lomega.png&quot; alt=&quot;&quot; /&gt;&lt;br /&gt;&lt;/td&gt;
    &lt;/tr&gt;
  &lt;/tbody&gt;
&lt;/table&gt;

&lt;p&gt;And there you have it. Realistically, most of these aren’t very common in stats, but I figured it was a good idea to list them all anyway. It was definitely an adventure to try and write them all with my computer mouse instead of actually being able to draw them on paper, but I thought the rainbow line was too fun to pass up. Hopefully, some of these depictions will be a little useful to those of us who are still stats novices.&lt;/p&gt;

&lt;p&gt;P.S. If you make any cool drawings with the sketchpad, send them to me!&lt;/p&gt;
</description>
        <pubDate>Mon, 13 Oct 2025 00:00:00 +0000</pubDate>
        <link>https://keanehauck.com/statistics/math/art/2025/10/13/greek-letters/</link>
        <guid isPermaLink="true">https://keanehauck.com/statistics/math/art/2025/10/13/greek-letters/</guid>
      </item>
    
      <item>
        <title>Grad School Scramble III: Inter-rater Reliability</title>
        <description>&lt;p&gt;&lt;img src=&quot;/images/posts/interrater-reliability/tens.png&quot; alt=&quot;tens&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Well, here we are! Summer is over, the new students are flocking to campus, and I find myself in the beautiful and hot state of Arizona. Classes technically started on Thursday, but they mostly consisted of looking over syllabi and introducing yourself to your classmates. The ASU program is great so far: my fellow grad students are awesome, the faculty are trememdous, and I’ve already begun immersing myself in local communities like choir and card shops.&lt;/p&gt;

&lt;p&gt;During one of my Friday classes, Professional Issues in Psychology (or, as it’s informally known, Intro to Being a Grad Student), the instructor Cheryl Conrad told an anecdotal story about inter-rater reliability. As she was telling it, I realized that I really don’t know much about the concept, and decided that it would be a good installment of Grad School Scramble. I have not been nearly as prolific with this series as I had hoped, but I wanted to finish one last post before school gets heavily underway. Who knows—maybe I’ll continue it into the semester.&lt;/p&gt;

&lt;h3 id=&quot;inter-rater-reliability-definitions-and-concepts&quot;&gt;Inter-rater reliability: Definitions and concepts&lt;/h3&gt;

&lt;p&gt;Psychological studies and tests can often require responses that are long, involved, or otherwise unable to be directly assessed numerically. When such responses are present in a test, they are often interpreted by an individual (or individuals) to convert them to a format that can be analyzed. For example, essays on AP exams are sent to independent readers that rate the essays on a scale from 0-6. In AP’s case, two independent readers will score each essay, so that an individual score isn’t biased by one single reader. For more general tests and studies, these readers are generally known as “raters,” and it’s critically important that they are accurately and reliably assessing each response.&lt;/p&gt;

&lt;p&gt;How can we determine if raters are doing a good job? One method is by evaluating the &lt;em&gt;inter-rater reliability&lt;/em&gt;: a measure of how much the raters are agreeing with each other. If they display a large amount of agreement, we have reason to believe that their judgments are valid; conversely, if they disagree frequently, there might be reason to doubt the validity of their ratings. With that said, how do we measure their agreement?&lt;/p&gt;

&lt;p&gt;Before diving into agreement measures, it’s important to discuss what we even mean by “agreement.” When scoring is categorical, agreement is simple: the raters either agree or they don’t. When scoring is &lt;a href=&quot;https://en.wikipedia.org/wiki/Ordinal_data&quot;&gt;ordinal or better&lt;/a&gt;, however, the question of “scoring distance” comes into play. Raters who score an essay 5 and 6 respectively seem to agree more than raters who score an essay 1 and 6, but they still disagreed. How do we account for that? Furthermore, there are different types of rating. There exists the aforementioned scoring where raters are agreeing/disagreeing with a single rating, but what about if they’re both confirming an official score? What if they’re ranking different options? What if there are more than 2 raters? There are a variety of different measures to quantify agreement in each of these situations.&lt;/p&gt;

&lt;h3 id=&quot;simple-probability-of-agreement&quot;&gt;Simple probability of agreement&lt;/h3&gt;

&lt;p&gt;The first and easiest example of an agreement measure is the probability of agreement. This measure is only applicable for nominal data, but it was the standard method of determining inter-rater reliability for a lengthy era of quantitative methodology. When there are 2 or more raters scoring a response, it is easy to calculate the percentage of the time that they agree. We simply divide the number of identical scores by the total number of scores and obtain the percent agreement. When there are more than two raters, we can set up a matrix in which each row represents a variable of interest and each column is a rater score.&lt;/p&gt;

\[\begin{array}{c|ccc}
  VAR &amp;amp; Rater 1 &amp;amp; Rater 2 &amp;amp; Rater 3 &amp;amp; Agreement \\
  1 &amp;amp; 1 &amp;amp; 1 &amp;amp; 1 &amp;amp; 1.00 \\
  2 &amp;amp; 1 &amp;amp; 1 &amp;amp; 0 &amp;amp; 0.66 \\
  3 &amp;amp; 0 &amp;amp; 0 &amp;amp; 0 &amp;amp; 1.00 \\
  4 &amp;amp; 0 &amp;amp; 1 &amp;amp; 0 &amp;amp; 0.66 \\
  &amp;amp; &amp;amp; &amp;amp; Total &amp;amp; 0.83
\end{array}\]

&lt;p&gt;According to &lt;a href=&quot;https://pmc.ncbi.nlm.nih.gov/articles/PMC3900052/&quot;&gt;McHugh (2012)&lt;/a&gt; (who I shamelessly stole the matrix example from), the direct interpretation of the probability of agreement is the percent of valid data. That is, if you have 90% percent agreement, then 10% of the data in your study are erroneous.&lt;/p&gt;

&lt;h3 id=&quot;cohens-kappa&quot;&gt;Cohen’s kappa&lt;/h3&gt;

&lt;p&gt;As we can see from the previous example, the joint probability of agreement has some issues as a measure of inter-rater reliability. For one, it is only applicable for nominal data. Furthermore, its interpretation is overly simplistic and doesn’t tell you much about the direction of erroneous data. The biggest issue, however, is the one that statistician &lt;a href=&quot;https://en.wikipedia.org/wiki/Jacob_Cohen_(statistician)&quot;&gt;Jacob Cohen&lt;/a&gt; pointed out in the ’60s: it doesn’t account for agreement that arises due to chance. For example, if two professors were to rate student performance on a test as either pass/fail, and we asked them to give scores &lt;em&gt;completely randomly&lt;/em&gt;, we would still expect them to coincidentally give the same rating about half of the time. On paper, they would seem to exhibit 50% congruence, but in reality there would be no systematic agreement between the two raters.&lt;/p&gt;

&lt;p&gt;To account for this, Cohen developed &lt;em&gt;Cohen’s kappa&lt;/em&gt;, stylized as $\kappa$, as a correlational coefficient of rater agreement. It ranges from $-1$ to $1$, with $-1$ representing perfect disagreement, $1$ representing perfect agreement, and $0$ representing the agreement we would expect due to random chance.&lt;/p&gt;

&lt;p&gt;The formula is as follows:&lt;/p&gt;

&lt;p&gt;$\kappa=\frac{p_o-p_e}{1-p_e}$&lt;/p&gt;

&lt;p&gt;where $p_o$ is the observed percent agreement and $p_e$ is the expected percent agreement due to chance.&lt;/p&gt;

&lt;p&gt;The intuition behind the formula is that $p_o-p_e$ represents subtracting the expected chance percentage of agreement from the observed percentage, leaving the “non-chance” percentage of agreement left. Dividing by $1-p_e$ rescales the result so that $1$ is perfect agreement and $0$ is random chance. For example, if you expect 50% of your agreement to be due to chance and your raters perfectly agree, then you would still get a kappa coefficient of $1=\frac{1-.5}{1-.5}$. If you expected 50% of your agreement to be due to chance and your raters agreed 50% of the time, then you would get a kappa coefficient of $0=\frac{.5-.5}{1-.5}$.&lt;/p&gt;

&lt;p&gt;For 2 raters, $p_e$ can be calculated by a chi-square table where each class is a rater. For example:&lt;/p&gt;

\[\begin{array}{cccc}
  100 &amp;amp; 12 \\
  3 &amp;amp; 90 
\end{array}\]

&lt;p&gt;If this is our chi-square table (forgive the poor $ \LaTeX $), and rater 1 is the columns (rating of pass/fail) while rater 2 is the rows (rating of pass/fail), then we can calculate the expected agreement due to chance as follows. We can find the sum total of each row and column as $112$ and $93$ for the rows, and $103$ and $102$ for the columns, respectively. The total $n$ of the ratings is 205. According to the rule of independent probability, if our data were all independent then knowing the row tells you nothing about which column some data might be in. Let’s take the example of the top-left cell, which is the intersection of row 1 and column 1. Formally, if $P(A)$ is the probability that a variable is in row 1 and $P(B)$ is the probability that a variable is in column 1, then $P(A\space and\space B)=P(A)\cdot P(B)$. By the definition of a chi-square table, $P(A)=\frac{100+12}{205}=0.55$ and $P(B)=\frac{100+3}{205}=0.50$. Thus, $P(A)\cdot P(B)=.28$. So, if all of our data were due to random chance we would expect the top-left-most cell to contain 28% of our data, or a total of about 57. We can repeat this process for the other cells and sum up the totals of the congruent cells to determine the percent agreement we would expect due to chance. At the end of this process, it turns out that this chi-square table results in a $p_e=.50$, which is an expected percentage of agreement due to chance around 50% of the time. (For more information on probability, feel free to read my &lt;a href=&quot;https://keanehauck.com/quantitative/methodology/statistics/data/science/2025/06/09/probability/&quot;&gt;blog post&lt;/a&gt; about basic concepts in probability)&lt;/p&gt;

&lt;p&gt;Circling back to Cohen’s kappa: Once we know the expected probability $p_e$, we can finish our formula. Our total agreement $p_o$ is $\frac{190}{205}\approx.93$. When we substitute these numbers back into our kappa formula we get $\kappa=\frac{p_o-p_e}{1-p_e}=\frac{.93-.50}{1-.50}=0.85$.&lt;/p&gt;

&lt;p&gt;Can anyone see an issue with Cohen’s kappa from this example? I have one: It makes the assumption that ratings due to chance would follow the distribution of rows and columns exactly equal to the numbers we observe. Due to this, Cohen’s kappa can sometimes be &lt;em&gt;too&lt;/em&gt; critical of inter-rater reliability because it assumes that random chance ratings would look similar to the marginal proportions observed. It also assumes that raters are always independent. Despite this, $\kappa$ remains a more robust estimation of rater agreement than percent agreement.&lt;/p&gt;

&lt;p&gt;Because $\kappa$ is a correlation coefficient, it cannot be directly interpreted.&lt;/p&gt;

&lt;blockquote&gt;
  &lt;p&gt;Quick note: while I was researching and reading stuff for this post, I came across the aforementioned &lt;a href=&quot;https://pmc.ncbi.nlm.nih.gov/articles/PMC3900052/&quot;&gt;McHugh (2012)&lt;/a&gt; paper which stated that &lt;em&gt;squaring&lt;/em&gt; kappa results in the “amount of accuracy” due to agreement between the raters. This is a foundational article, with almost 22,000 citations. Regardless, I could not find any information to back this idea up, and plugging some example numbers into a quick Cohen’s kappa calculator seemed to disprove it. Their justification was that squaring any correlation coefficient results in an estimate of the variance explained, but I am not knowledgeable enough to confirm nor deny this. Can anyone smarter than me chime in and help with this?&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Regardless of the interpretation of kappa squared, we have accurate and verifiable standards for Cohen’s kappa. The following table shows which values are &lt;a href=&quot;https://www.jstor.org/stable/2529310?origin=crossref&quot;&gt;deemed acceptable&lt;/a&gt;:&lt;/p&gt;

\[\begin{array}{c|ccc}
  0.00-0.20 &amp;amp; Slight\space agreement \\
  0.21 - 0.40 &amp;amp; Fair\space agreement \\
  0.41 - 0.60 &amp;amp; Moderate\space agreement \\
  0.61 - 0.80 &amp;amp; Substantial\space agreement \\
  0.81 - 1.00 &amp;amp; Almost\space perfect\space agreement
\end{array}\]

&lt;h3 id=&quot;weighted-cohens-kappa&quot;&gt;Weighted Cohen’s kappa&lt;/h3&gt;

&lt;p&gt;In situations where raters need to make ordinal scorings with more than 2 categories—such as ranking patient pain from mild/moderate/severe—Cohen’s kappa has another issue: It does not differentiate between “levels” of disagreement. In the example of patient pain, we might want to quantify the disagreement between a doctor recording “mild pain” and a second doctor recording “severe pain” as a more extreme disagreement than “mild pain” and “moderate pain.” However, if we set up our 3x3 contingency table and calculate the expected percentages of agreement/disagreement, all disagreements are weighted equally.&lt;/p&gt;

&lt;p&gt;To solve this problem, in 1968 Cohen published &lt;a href=&quot;https://doi.org/10.1037/h0026256&quot;&gt;an extension to kappa&lt;/a&gt; known as weighted kappa (although, arguably, &lt;em&gt;normal&lt;/em&gt; kappa is a special extension of weighted kappa). Using weighted kappa allows for a researcher to assign weights to disagreements such that some can be viewed as more extreme than others.&lt;/p&gt;

&lt;p&gt;How does it work? Well, the formula for weighted kappa is as follows:&lt;/p&gt;

&lt;p&gt;$\kappa _w{}=1-\frac{\Sigma v_{ij}p_{oij}}{\Sigma v_{ij}p_{eij}}$&lt;/p&gt;

&lt;p&gt;To obtain this formula, Cohen began with some simple algebra from the formula for $\kappa$. As a reminder,&lt;/p&gt;

&lt;p&gt;$\kappa=\frac{p_o-p_e}{1-p_e}$.&lt;/p&gt;

&lt;p&gt;If we let $q$ represent the proportion of disagreement and define $q=1-p$, then we can do the following algebra:&lt;br /&gt;
$q+p=1$,&lt;br /&gt;
$p=1-q$.&lt;/p&gt;

&lt;p&gt;From here, we can represent our observed agreement and expected agreement in terms of $q$:&lt;br /&gt;
$p_o=1-q_o$,&lt;br /&gt;
$p_e=1-q_e$.&lt;/p&gt;

&lt;p&gt;Now, we can substitute these values into the original formula for $\kappa$:&lt;/p&gt;

&lt;p&gt;$\kappa=\frac{p_o-p_e}{1-p_e}=\frac{(1-q_0)-(1-q_e)}{1-(1-q_e)}$.&lt;/p&gt;

&lt;p&gt;After some simple algebra, we get $\kappa=\frac{q_e-q_o}{q_e}$. If we then divide both sides of the fraction by $q_e$, we get $\frac{1-\frac{q_o}{q_e}}{1}$, which quickly simplifies to:&lt;/p&gt;

&lt;p&gt;$\kappa=1-\frac{q_o}{q_e}$.&lt;/p&gt;

&lt;p&gt;Now we have our original kappa formula in terms of the proportion of &lt;em&gt;disagreement&lt;/em&gt; between the raters. Cohen next introduces a weighting concept based on the strength of disagreements. If we imagine a contingency table set up as so:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/interrater-reliability/contingency1.png&quot; alt=&quot;contingency1&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Then we want to weight the disagreements so that more severe disagreements are more costly to $\kappa$. The way in which this is done is up to the researcher; but, as Cohen stresses, it must be done prior to the collection of data. One common weighting method is linear weighting, where the weight assigned to each disagreement is simply the distance between the two cells:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/interrater-reliability/contingency2.png&quot; alt=&quot;contingency1&quot; /&gt;&lt;br /&gt;
&lt;em&gt;The weights are the numbers in purple&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Another alternative is to use quadratic weighting, which is when the distance between the incongruent cells is squared. Finally, as can be shown algebraically, when all disagreeing cells are weighted equally it recovers the original definition of $\kappa$.&lt;/p&gt;

&lt;p&gt;$\kappa{_w}$ simply represents the original kappa formula with $p_o$ and $p_e$ replaced by weighted proportions of disagreement. So, for each cell, we multiply the proportion (or frequency) by the weight, and get our weighted proportion. When we add this weighting factor back into our modified kappa formula we get the final $\kappa{_w}$ formula, which is:&lt;/p&gt;

&lt;p&gt;$\kappa _w=1-\frac{\Sigma v_{ij}p_{oij}}{\Sigma v_{ij}p_{eij}}$&lt;/p&gt;

&lt;p&gt;where $v_{ij}$ represents the weighting factor for cell $i,j$.&lt;/p&gt;

&lt;p&gt;Finally, we want to prove that this recovers the original kappa formula when all weights are equal. If we assume that all cells are weighted equally, then $v_{ij}$ becomes some non-zero constant, which can then be pulled out of the summation. After that, the constant can be cancelled out of the numerator and denominator, giving the formula:&lt;/p&gt;

&lt;p&gt;$\kappa _w=1-\frac{p_{oij}}{p_{eij}}$,&lt;/p&gt;

&lt;p&gt;which can be understood as our original kappa proportion of disagreement.&lt;/p&gt;

&lt;p&gt;For the purposes of space and time, I won’t be going through a full hand-calculated example of weighted kappa, but I hope that this run-through of the formulae was enough to serve as a decent introduction.&lt;/p&gt;

&lt;h3 id=&quot;pearsons-r&quot;&gt;Pearson’s r&lt;/h3&gt;

&lt;p&gt;If ratings are continuous, then Pearson’s r can be used pairwise as a measure of inter-rater reliability. However, this is mostly a poor choice, because it fails to take into account systematic biases between the two raters. For example, if one rater scores a measure (1, 2, 3, 4, 5) while another rater scores pairwise the same measure (2, 3, 4, 5, 6), then their correlation will be 1, but there will be difference in their agreement. So, I’m not going to go very in-depth into the usage of r as a measure of inter-rater reliability. Just know that it’s technically an option.&lt;/p&gt;

&lt;h3 id=&quot;intraclass-correlation-coefficient&quot;&gt;Intraclass Correlation Coefficient&lt;/h3&gt;

&lt;p&gt;So, what should we use if ratings are continuous (interval or ratio)? One option is to use the &lt;a href=&quot;https://en.wikipedia.org/wiki/Intraclass_correlation&quot;&gt;Intraclass Correlation Coefficient (ICC)&lt;/a&gt;. From a mixed-models standpoint (as was explained to me by the amazing Catherine Bain), the ICC is the amount of variance of the intercepts divided by the total variance (which comprises the variance of the intercepts + the unwanted variance). So, if the unwanted variance is higher than the variance of the intercepts, then the ICC is lower than $.5$ (which is a poor result). On the other hand, if the variance of the intercepts composes a large amount of the variance than the ICC will be higher.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/interrater-reliability/icc.png&quot; alt=&quot;icc&quot; /&gt;&lt;/p&gt;

&lt;p&gt;We can extend this idea to inter-rater reliability by considering the items as the clusters and determining how much variance in score is due to item difference vs. rater difference. If we look at the previous chart, if we let “Group Number” represent the different items and the blue dots represent the scores that raters assign to the item, then we can observe how the individual item is much more predictive of a score than the distinctive rater. This is what we want to see—and displays a good ICC of $.91$. Conversely, if there was more of a difference between raters than between items, then there would be more “unwanted variance” than intercept variance, and the ICC would be much lower.&lt;/p&gt;

&lt;p&gt;Unlike Cohen’s kappa, the ICC compares to a theoretical null model of “no intercept variance” which is more straightforward than the kappa assumption about what chance agreement would look like. This can make it more useful under certain conditions. However, according to &lt;a href=&quot;https://en.wikipedia.org/wiki/Intraclass_correlation&quot;&gt;Wikipedia&lt;/a&gt; the ICC does make the assumption that raters are exchangable for a certain item. In other words, there is no meaningful way to order the raters. This assumption is violated if there is systematic bias between different raters (which would result in some raters being naturally more critical than others), which is likely in real-world scenarios, so debate exists about whether other measures should still be used in place of the ICC.&lt;/p&gt;

&lt;p&gt;As always, I am not a professional, so I feel the need to give the disclaimer that I don’t have a full mathematical understanding of how the ICC works. Nor do I have an understanding of mixed models past what you can get from a couple of YouTube videos. Still, I hope that most of the information presented is at least somewhat accurate.&lt;/p&gt;

&lt;h3 id=&quot;concluding-thoughts&quot;&gt;Concluding thoughts&lt;/h3&gt;

&lt;p&gt;Well, there you have it. The results of 3 days of reading information about inter-rater reliability. Is it super, &lt;em&gt;incredibly&lt;/em&gt; useful? I’m not sure—I’ve never been a part of a study with multiple raters. In any case, it certainly was interesting. One of my big take-aways is how amazing Jacob Cohen’s work was, especially for the time. I learned about kappa before investigating weighted kappa, and the weighted version seemed a very powerful yet natural extension of the previous work. I tried for almost an hour to sit down and derive the formlua for weighted kappa from the unweighted version, and couldn’t figure it out (even with the weighted formula sitting in front of me). Lo and behold, Cohen does it in like three lines on the third page of his paper. Sometimes it almost feels like we’re falling off the shoulders of giants.&lt;/p&gt;
</description>
        <pubDate>Sat, 23 Aug 2025 00:00:00 +0000</pubDate>
        <link>https://keanehauck.com/quantitative/statistics/school/reliability/2025/08/23/interrater-reliability/</link>
        <guid isPermaLink="true">https://keanehauck.com/quantitative/statistics/school/reliability/2025/08/23/interrater-reliability/</guid>
      </item>
    
      <item>
        <title>Grimmsnarl to the Rescue</title>
        <description>&lt;p&gt;&lt;img src=&quot;/images/posts/naic2025/holding.jpg&quot; alt=&quot;holding&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Aaaand we’re back! I am proud to announce that I recently competed at the 2025 Pokémon North America International Championships (NAIC), placing 136th out of 3,812 competitors. This tournament took place in New Orleans, LA, and the whole trip was a delight. I traveled down with a few friends and my partner, and we got to see &lt;a href=&quot;https://en.wikipedia.org/wiki/Bourbon_Street&quot;&gt;Bourbon Street&lt;/a&gt; and the French Quarter. However, that’s not really what we’re concerned with. Time to run back the tournament and see how I did.&lt;/p&gt;

&lt;h3 id=&quot;deck-choice-and-meta-calls&quot;&gt;Deck Choice and Meta Calls&lt;/h3&gt;

&lt;p&gt;For this tournament, I chose to run a new deck: Grimmsnarl Froslass. NAIC was a tournament with some relatively special circumstances surrounding it: a new set of Pokémon cards released the same day as the tournament. This meant that there was an influx of potential new decks and strategies, and I wanted to bring one of the new decks in the hope that my opponents wouldn’t be extremely prepared for the matchup. Unfortunately, I made this decision about three weeks prior to the tournament, and in the days leading up to NAIC the popularity of Grimmsnarl exploded. The whole community was talking about it as a top meta contender, and everyone was preparing/teching for it. Still, just because a deck is well-known doesn’t mean it’s a bad choice, and I certainly believed in the power of the deck. Here’s the list I cooked up:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/naic2025/decklist.jpg&quot; alt=&quot;decklist&quot; title=&quot;Notice the Tatsugiri&quot; /&gt;
&lt;em&gt;Notice the Tatsugiri&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Grimmsnarl is a slow deck: a stark contrast to the Gholdengo deck I brought to Milwaukee. At my last regionals, I wanted a quick deck to be able to finish all 3 games in a 50-minute match. Why did I make the switch? Well, one of the big things that annoyed me about Gholdengo was the matchup into 2-2-2 decks like Raging Bolt. Aggressive decks like these want to KO big 2-prize Pokémon each turn as fast as possible to close the game out in just 3 attacks. If two decks like these are facing each other, falling behind in the prize race is a death sentence that is almost impossible to comeback from. (Side note: the Gholdengo-Bolt matchup is indeed a lot more nuanced, with Gholdengo having the opportunity to present a single-prize board state, but Bolt being able to counter that by only playing down one 2-prizer to enable noctowls and setting up a baby bolt to even the trade with a single-prize board state of their own. The prize map overall favors Bolt when both sides are playing optimally. Still, the point stands that falling behind in the map means you lose). Slow, set-up decks like Grimmsnarl have &lt;em&gt;many&lt;/em&gt; more comeback mechanics, and as a result provide you many more opportunities to outplay your opponent. As I was gaining confidence as a player going into this tournament, I wanted to bring a deck that would allow me to consistently outplay my opponent.&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;
&lt;img src=&quot;/images/posts/naic2025/grimmsnarl.jpg&quot; alt=&quot;Grimmsnarl&quot; width=&quot;200&quot; class=&quot;center-image&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;Also, I was a big believer in Grimmsnarl’s matchup spread. I believed that it was favored into both Dragapult and Gardevoir: the two decks that had defined the meta in the months leading up to this tournament. I also believed that it had an OK-ish Bolt matchup, which was projected to be a popular deck as well. The only matchup I was worried about was Joltik Box, but I didn’t think I’d see too many of them. Spoiler alert: I did.&lt;/p&gt;

&lt;p&gt;The overall strategy of the deck is to utilize Grimmsnarl ex alongside Froslass and Munkidori to spread massive amounts of damage on your opponent’s board. Alongside cards like &lt;a href=&quot;https://www.tcgplayer.com/product/523858/pokemon-sv04-paradox-rift-technical-machine-devolution&quot;&gt;TM: Devo&lt;/a&gt;, you can take huge swing turns with multiple KO’s to win games.&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;
&lt;img src=&quot;/images/posts/naic2025/froslass.jpg&quot; alt=&quot;Froslass&quot; width=&quot;200&quot; class=&quot;center-image&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;Let’s take a look at my matches and see how I did.&lt;/p&gt;

&lt;h3 id=&quot;round-1-win-against-poison-roaring-moon&quot;&gt;Round 1: Win against Poison Roaring Moon&lt;/h3&gt;

&lt;p&gt;My first round went swimmingly. My opponent was running a deck centered around Roaring Moon ex with a poison package intended to “donk” the other player. In the Pokémon TCG, if your active Pokémon is knocked out and you don’t have any other benched Pokémon to replace it with, you lose. With the help of cards like &lt;a href=&quot;https://www.tcgplayer.com/product/562025/pokemon-sv-scarlet-and-violet-promo-cards-pecharunt-149&quot;&gt;baby Pecharunt&lt;/a&gt; and &lt;a href=&quot;https://www.tcgplayer.com/product/542673/pokemon-sv05-temporal-forces-perilous-jungle&quot;&gt;Perilous Jungle&lt;/a&gt;, the deck aims to take out the opponent’s starting Pokémon before they can even play a single card.&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;
&lt;img src=&quot;/images/posts/naic2025/moon.jpg&quot; alt=&quot;Moon&quot; width=&quot;200&quot; class=&quot;center-image&quot; /&gt;
&lt;/p&gt;
&lt;p&gt;&lt;em&gt;The deck’s main attacker if the donk doesn’t work&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Fortunately, my Impidimps are immune to the effect of Perilous Jungle, and as a result are unable to be donked. Unfortunately, I cold-started Snorunt on my first game. All I could do was sit and watch as my opponent attempted to piece together the donk combo. He missed it by a single nest ball, and I was luckily able to take down the 1st game after he whiffed a supporter on the next turn. The second game went well - I started an Impidimp in the active and a Snorunt on the bench. He punished my misplay of benching Snorunt when he immediately whipped out the donk combo + an &lt;a href=&quot;https://www.tcgplayer.com/product/523666/pokemon-sv04-paradox-rift-iron-bundle&quot;&gt;Iron Bundle&lt;/a&gt; to KO my Snorunt. However, using a Defiance Band and some residual poison damage on Impidimps, I was able to target down his Roaring Moon ex’s with energy attached and prevent him from ever pulling off a Frenzied Gouging. 1-0-0.&lt;/p&gt;

&lt;h3 id=&quot;round-2-win-against-raging-bolt&quot;&gt;Round 2: Win against Raging Bolt&lt;/h3&gt;

&lt;p&gt;This round went well too. I was a bit nervous to see my opponent flip over a &lt;a href=&quot;https://www.tcgplayer.com/product/550069/pokemon-sv06-twilight-masquerade-teal-mask-ogerpon-ex-025167&quot;&gt;Teal Mask Ogerpon&lt;/a&gt;; Bolt is one of the worse matchups for Grimmsnarl. They can always oneshot a Grimmsnarl with the Ogerpon and threaten huge amounts of pressure with the baby &lt;a href=&quot;https://www.tcgplayer.com/product/567337/pokemon-sv07-stellar-crown-raging-bolt&quot;&gt;Raging Bolt&lt;/a&gt;. However, this is the match that made me incredibly thankful I had chosen to run Maractus:&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;
&lt;img src=&quot;/images/posts/naic2025/maractus.jpg&quot; alt=&quot;Maractus&quot; width=&quot;200&quot; class=&quot;center-image&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;Leading up to the tournament, there was a huge meta discussion surrounding Maractus. Its attack prevents the opposing Pokémon from retreating, which is an incredibly powerful effect against decks that don’t run some sort of switching tech. A few weeks ago, a Dragapult list running Maractus won a big regional, so people started preparing for Maractus by putting switch cards in their decks. Due to this, many people chose to cut Maractus, as they felt that it was now redundant. I viewed the card a bit differently: not as a “lock one Pokémon and you win” card but rather as a tool to slowly run your opponent out of resources. Even though my opponent ran 2 switch effects, I still was able to force my opponent to use his resources suboptimally and buy enough time for my Froslass + Munkidoris to build up massive damage on his board. 2-0-0.&lt;/p&gt;

&lt;h3 id=&quot;round-3-win-against-dragapult-dusknoir&quot;&gt;Round 3: Win against Dragapult Dusknoir&lt;/h3&gt;

&lt;p&gt;My opponent for this round was a really cool guy—very friendly and we had a nice chat before the match started. He was on &lt;a href=&quot;https://limitlesstcg.com/decks/list/18506&quot;&gt;Dragapult Dusknoir&lt;/a&gt;, one of the top decks in the format. I felt confident in this matchup, and I took game 1 after a long back-and-forth. He played it very well, but he prized what I believe was his only Dusknoir, which really hurt his gameplan. I had a cool play where I swung into his Budew with an Impidimp for 10 to set up a 20 damage Munkidori KO the following turn, but other than that the game was uneventful. Our first game took a long time, and so we only had around 20 minutes left in the round. We ended up not finishing our second game after I Iono’ed him low and he was unable to escape Maractus on one pivotal turn. 3-0-0.&lt;/p&gt;

&lt;h3 id=&quot;round-4-loss-against-grimmsnarl-froslass&quot;&gt;Round 4: Loss against Grimmsnarl Froslass&lt;/h3&gt;

&lt;p&gt;This was one of the best matches of the tournament. My opponent was a very skilled player who placed in the top 64 of multiple regionals this season. He was on the same deck I was playing: Grimmsnarl Froslass. Our first game was an absolute slugfest that took around 45 minutes. At the very beginning of the game, I was tactically forced to &lt;a href=&quot;https://www.tcgplayer.com/product/623582/pokemon-sv09-journey-together-professors-research&quot;&gt;Professor’s Research&lt;/a&gt; away my only &lt;a href=&quot;https://www.tcgplayer.com/product/497560/pokemon-sv02-paldea-evolved-super-rod-188193&quot;&gt;Super Rod&lt;/a&gt;, which I knew would come back to bite me. The other thing that hurt is that he was playing four Munkidori as opposed to my three.&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;
&lt;img src=&quot;/images/posts/naic2025/munkidori.jpg&quot; alt=&quot;Munkidori&quot; width=&quot;200&quot; class=&quot;center-image&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;In the mirror match, whoever establishes the most Munkidoris will usually win. The ability to reflect huge amounts of damage back over to your opponent is incredibly powerful in this matchup, and the best strategy is normally to try and kill your opponents Munkidoris instead of their Grimmsnarls. I was able to maneuver him into a position where he had to retreat two energy off of his only Grimmsnarl in play, leaving him a few turns later with a board of only a no-energy Grimmsnarl and four Munkidoris. Unfortunately, I had only one Munkidori in play myself, and I was unable to use Super Rod to get any back. This left me in a position where if I were to swing into his Grimmsnarl, he would be able to move 120 damage back onto my Grimmsnarl, making it so that I couldn’t even KO his Grimmsnarl next turn and setting me up for a loss 2 turns later (prizes were 2-2). This put us in a situation where both of us were passing back-and-forth because swinging for any damage would be a suboptimal play.&lt;/p&gt;

&lt;p&gt;What my opponent eventually realized is that he was going to deck out before I was, so he had to take the risk of putting a Snorunt down to try and establish Froslass. I was able to Iono, Counter Catcher up a Munkidori, and swing into it putting 30 on the Snorunt, with the logic that if he whiffed establishing the Froslass next turn I would immediately be able to do 30 more to the Snorunt and win the game. Unfortunately, he drew into the only Froslass he had left in deck, which was around a 3/15 chance. Because of this, I unfortunately lost the game on the following turn when he drew into an energy as well to attack with Grimmsnarl and win.&lt;/p&gt;

&lt;p&gt;With only 4 minutes left in the match, there’s no way I can manage a win in game 2: a loss for me. Even though I lost, I thought this was an incredibly strategic match, and I was happy with how I played. 3-1-0.&lt;/p&gt;

&lt;h3 id=&quot;round-5-loss-against-gardevoir&quot;&gt;Round 5: Loss against Gardevoir&lt;/h3&gt;

&lt;p&gt;In contrast to the previous round, this was one of the most frustrating moments of the tournament. My opponent was on &lt;a href=&quot;https://limitlesstcg.com/decks/list/18498&quot;&gt;Gardevoir&lt;/a&gt;: a matchup that was theoretically favored for me. Because of this, I was a bit overconfident and complacent during our match, and a few suboptimal plays in game 1 put me in a losing position. I tried my absolute hardest to squeak out a win, but my opponent was victorious. There was only 25 minutes left in the round, and I didn’t want to settle for a draw in a favored matchup. So, I tried to turbo through the next game and take a strategically weaker line for the sake of trying to end the game early, and was punished with an 0-2 for the round. 3-2-0.&lt;/p&gt;

&lt;p&gt;Before my next round, I go into Zen Mode. I sit in a bathroom stall for a few minutes and listen to dubstep while hyping myself up. I &lt;em&gt;have&lt;/em&gt; to lock in—if I approach my games with the misguided and foolhardy confidence I entered this round with, I will certainly lose. My next rounds will have to be played to perfection. I check my phone: pairings are up.&lt;/p&gt;

&lt;h3 id=&quot;round-6-spoler-blocker-against-joltik-box&quot;&gt;Round 6: SPOLER BLOCKER against Joltik Box&lt;/h3&gt;

&lt;p&gt;Well, this is horrifying. Round 6 of 9 and I’m already super close to elimination. Any record below 6-2-1 (6 wins, 2 losses, 1 tie) means you can’t make day 2, so I was in a position where I had to win rounds 6, 7, and 8 (and potentially 9 if my opponent didn’t want to intentionally tie) to make day 2. I’m riding the lightning.&lt;/p&gt;

&lt;p&gt;My next opponent is quite good: I checked his Limitless page before the round and he has multiple Top 32’s at major regional tournaments. To my absolute horror, he chooses second and starts Miriadon. Looks like we’ve got a Joltik Box.&lt;/p&gt;

&lt;p&gt;As I previously mentioned, this matchup is atrocious for Grimmsnarl. They run &lt;a href=&quot;https://www.tcgplayer.com/product/523706/pokemon-sv04-paradox-rift-iron-hands-ex-070182&quot;&gt;Iron Hands&lt;/a&gt; to take 2 prizes on your small Pokémon if you play it slow, and run Iron Leaves to get around Maractus, pivot energy off the Iron Hands, and oneshot your Grimmsnarl.&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;
&lt;img src=&quot;/images/posts/naic2025/ironleaves.jpg&quot; alt=&quot;Iron Leaves&quot; width=&quot;200&quot; class=&quot;center-image&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;I quickly lose the first game, and my dreams of making day 2 are flashing before my eyes. Game 2 goes more in my favor: I am able to choose to go 2nd, which severely hinders Joltik Box’s ability to turbo out an Iron Hands. Due to some solid disruption and a quickly established Froslass, my opponent scoops up the game fairly quickly. Game 3 gets off to a horrible start: I have Munkidori in the active and no way to search out any other Pokémon. I attach an energy and pass. My opponent immediately starts turboing out cards, playing Miriadon, Lillie’s Clefairy, and two Iron Hands. He benches a Mew ex, and then Crispin’s to the Lillie’s Clefairy on the bench. He retreats the Joltik in the active, promotes the Lillie’s Clefairy, and announces Full Moon Rondo for 120 damage, KOing my Munkidori and donking me to win the game. It’s over: I’m out of the tournament.&lt;/p&gt;

&lt;p&gt;But wait—something’s wrong. How did he retreat the Joltik? It has a retreat cost of 1. I ask him to clarify how he retreated the Joltik, and he looks at his board for a few seconds, looks back up and me, and goes “You know, that’s a really good point.” Apparently he missed something, or didn’t consider the retreat cost, or was mentally operating as if he already had a &lt;a href=&quot;https://www.tcgplayer.com/product/589983/pokemon-sv08-surging-sparks-latias-ex-076-191&quot;&gt;Latias ex&lt;/a&gt; in play. Whatever the case, this is an absolutely egregious misplay against Grimmsnarl. Because he was trying to max out the damage of Full Moon Rondo (which does more damage the more Pokémon you have in play), he had bench-locked himself with two Iron Hands and no Latias ex. This is a death sentence. I am able to prevent him from ever attacking with Joltik by stalling on one Iron Hands and switching to stall on the other once the first one was almost powered up. I don’t ever attack until I have a fully set-up board, and he quickly loses the 3rd game.&lt;/p&gt;

&lt;p&gt;I had mixed emotions about this match. My opponent was a cool guy: he was certainly not trying to pull a quick one on me and seeing such a bad misplay occur is unfortunate. On the other hand, I played perfectly to take advantage of the misplay, and avoiding mental lapses is a crucial part of competing in tournaments of these stakes. At the end of the day, a win is a win. Onto the next round 4-2-0.&lt;/p&gt;

&lt;h3 id=&quot;round-7-win-against-joltik-box&quot;&gt;Round 7: Win against Joltik Box&lt;/h3&gt;

&lt;p&gt;This round, I feel good. My opponent again starts Miriadon and chooses to go second, but this time I am filled with the confidence of just having beaten a Joltik Box. If I did it once, I can do it again.&lt;/p&gt;

&lt;p&gt;I quickly take game 1 after my opponent gets off to a slow start. I am able to rare candy Grimmsnarl on turn 2 and prevent him from ever swinging with a Joltik. Game 2 is going my way again, when my opponent’s nerves get the better of him and he makes a massive mistake: he accidentally shuffles his hand back into his deck when performing some other action like a Nest Ball. I was already on pace to win that game, but that mistake was the final nail in the coffin. My opponent scoops it up, and I take the victory. Despite such a big error, my opponent seems very happy with how he did in the tournament, and wishes me luck.&lt;/p&gt;

&lt;p&gt;5-2-0.&lt;/p&gt;

&lt;h3 id=&quot;round-8-win-against-dragapult-dusknoir&quot;&gt;Round 8: Win against Dragapult Dusknoir&lt;/h3&gt;

&lt;p&gt;I’m in shock heading into the next round. My previous two rounds (both of which I &lt;em&gt;had&lt;/em&gt; to win) were against my worst matchup, and somehow I’m surviving. Both of the wins felt like gifts: they were due to some huge, avoidable error made by my opponents. Nervously, I check my pairing for round 8. I’m against the guy who got 10th at the Mérida Regional Championships and 5th at the Special Event Bogotá. No more free wins against this guy.&lt;/p&gt;

&lt;p&gt;He’s playing Dragapult Dusknoir, much like my round 3 opponent. Our first game is a doozy: I am able to take a little bit of a lead by KO’ing his Budew while he’s still under itemlock from me. Unfortunately, he is later able to piece together the perfect Dragapult turn and wipe out all of my Impidimps while establishing a Dragapult. I try and slow him down by TM: Devo’ing his rare candy’d pult (the only one with energy on it), but he top decks a second rare candy to get right back in the game (he even showed me and we chuckled about it). Eventually, he gets down to 2 prizes with two powered-up Dragapults on the board. I am down to 4 prizes thanks to his Dusknoirs, but I’m in bad shape. My only play is to Counter Catcher up his Fezandipiti ex and stall on it with Maractus. I do this for two turns, and somehow it seems as it he has no way of escaping it. He uses Arven but doesn’t grab a switch, and even benches a Duskull. Once I see that he’s trying to build up a Dusknoir to KO my Maractus and escape the lock, I know I am still very much in this game. I start building back up my board, and eventually establish a Froslass. I piece together a four-turn plan which should hopefully end with me KO’ing his Fez and a benched Drakloak on the same turn, and he’s attempting to build up a Dusknoir to escape the lock. The pinnacle of my strategy occurs when he KO’s my Maractus, freeing his Fez the next turn, but he goes down to 1 prize card. I iono him, Counter Catcher up the only Pokémon (Drakloak) he has in play with no energy, and hope that he whiffs an energy to retreat. He digs with his recon directive, but fails to get an energy, and passes it back over to me. I’m able to perfectly move damage onto his Fez and swing into his Drakloak with &lt;em&gt;Budew&lt;/em&gt; to take my last 3 prizes.&lt;/p&gt;

&lt;p&gt;Our second game, my deck runs very well, and I’m able to spike &lt;a href=&quot;https://www.tcgplayer.com/product/550207/pokemon-sv06-twilight-masquerade-secret-box&quot;&gt;Secret Box&lt;/a&gt; off of my prize cards at the perfect moment to secure a win. 6-2-0.&lt;/p&gt;

&lt;h3 id=&quot;round-9-tie-against-dragapult-dusknoir&quot;&gt;Round 9: Tie against Dragapult Dusknoir&lt;/h3&gt;

&lt;p&gt;If I tie this round, I go 6-2-1 and make day 2. The issue with offering a draw (mutually confirming a spot in day 2) is that a tie in this round means that my opponent and I are likely out of the running for top cut. Still, I don’t mind. My goal was to day 2, and I will be super happy with a tie.&lt;/p&gt;

&lt;p&gt;As I walk to the table, I see my opponent already has his hand extended, and I know I’ve done it. We shake hands, sign the match slip as a tie, and congratulate each other on making day 2 at the hardest and largest Pokémon tournament of the entire year. It has been DONE.&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;
&lt;img src=&quot;/images/posts/naic2025/pikachu.jpg&quot; alt=&quot;Pikachu&quot; width=&quot;400&quot; class=&quot;center-image&quot; /&gt;
&lt;/p&gt;

&lt;h3 id=&quot;round-10-tie-against-joltik-box&quot;&gt;Round 10: Tie against Joltik Box&lt;/h3&gt;

&lt;p&gt;Honestly, there’s not much to say about my Day 2 matches. I’m on a victory lap and feeling good. Not even the fact that I’m going first against a Joltik Box can dampen my spirits. I actually manage to take game 1, but lose game 2 due to some minor misplays I could have avoided. We don’t have enough time to complete game 3, so I start the day off with a tie. 6-2-2.&lt;/p&gt;

&lt;h3 id=&quot;round-11-tie-against-raging-bolt&quot;&gt;Round 11: Tie against Raging Bolt&lt;/h3&gt;

&lt;p&gt;Another tie. My opponent gets an extremely powerful start during our first game, and blows me off the board. I manage to fend off the aggression during our second game, and take it as time is called. 6-2-3.&lt;/p&gt;

&lt;h3 id=&quot;round-12-win-against-raging-bolt&quot;&gt;Round 12: Win against Raging Bolt&lt;/h3&gt;

&lt;p&gt;At this point, I’m getting a bit disappointed. Games against Joltik Box and Raging Bolt are not only unfavorable matchups, but they’re also not very fun. Both decks are designed to simply apply maximum pressure as quickly as possible, so a lot of strategic maneuvering (like of the kind I would get against other slow decks) isn’t present. My opponent is also playing a list absolutely tech’d to the max against Grimmsnarl: He has Switch, Professor Turo’s Scenario, and Picnic Basket.&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;
&lt;img src=&quot;/images/posts/naic2025/picnic.jpg&quot; alt=&quot;Picnic Basket&quot; width=&quot;200&quot; class=&quot;center-image&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;Still, he makes a few minor overbenches and I am able to capitalize on his mistakes and take down both of our games.&lt;/p&gt;

&lt;h3 id=&quot;round-13-win-against-eevee-box&quot;&gt;Round 13: Win against Eevee Box&lt;/h3&gt;

&lt;p&gt;Round 13 was an absolutely fantastic round to finish the tournament on. My opponent was a great guy, and we were chit-chatting back and forth throughout the game. I totally put my foot in my mouth by asking him to speed up his actions and then immediately taking a 5-minute turn myself, but he just poked fun at me and we laughed it off. I was able to win the first game thanks to Budew, taking down all of his Eevees on one pivotal turn.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/naic2025/lastround.png&quot; alt=&quot;lastround&quot; /&gt;
&lt;em&gt;The final moments&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Our second game was a close one, and it ended up going to time. It came down to a situation where he had no Hoothoots in play, and I was able to stall his Latias ex with Maractus and Iono him to 1. He told me that he had a switch in deck, and was able to dig 2 cards with Mew ex. Multiple judges were watching, so we went into full theatric mode and flipped both of the cards on the table: no switch. I take down a fun albeit slightly undeserved final round. GGs Donny.&lt;/p&gt;

&lt;h3 id=&quot;conclusion&quot;&gt;Conclusion&lt;/h3&gt;

&lt;p&gt;Overall, I placed 136th / 3,812 participants. Damn, I kinda crushed it. I was super, super happy with how I did, and felt very confident about my deck choice and preparation. There were a few minor details in the Grimmsnarl list that the top players changed for NAIC, but I was happy with my overall list and tech choices. I believe wholeheartedly that Maractus was the correct play, alongside the Tatsugiri that I was so confident in. I was also very proud of my play. I was able to recognize minor errors (such as those in rounds 5 and 10), and I believe that recognizing the error is the important thing. The big thing, for me, is that I made day 2: something that many professional players failed to do at this tournament.&lt;/p&gt;

&lt;p&gt;What will happen to my Pokémon hobby once I start grad school? I’m not completely sure. NAIC was, for me, the culmination of all the play and preparation that has been occurring since I picked up the hobby back in October. I wanted to achieve a high placement at this tournament, and I was successful. I’m sure I’ll keep playing, but I probably won’t be able to commit as much time as I have been these past few months. Still, as long as I’m a part of the Pokémon community, I will be able to brag about the time I placed in the top 3% at NAIC.&lt;/p&gt;
</description>
        <pubDate>Tue, 17 Jun 2025 00:00:00 +0000</pubDate>
        <link>https://keanehauck.com/pokemon/gaming/cards/tournament/2025/06/17/naic2025/</link>
        <guid isPermaLink="true">https://keanehauck.com/pokemon/gaming/cards/tournament/2025/06/17/naic2025/</guid>
      </item>
    
      <item>
        <title>Grad School Scramble II: Much Ado About Probability
</title>
        <description>&lt;p&gt;Recently, I woke up to a phone call from my grandfather asking me to explain the &lt;a href=&quot;https://en.wikipedia.org/wiki/Monty_Hall_problem&quot;&gt;Monty Hall Problem&lt;/a&gt; in simple terms. For those of you who are unfamiliar, this problem is famous in the field of probability for confounding even highly-training individuals (like mathematicians and academics) due to its unintuitive nature.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/probability/monty.png&quot; alt=&quot;monty hall&quot; title=&quot;This problem always makes me feel bad for the goats&quot; /&gt;&lt;/p&gt;

&lt;p&gt;The problem is thus: Imagine that you’re on a TV show, and the host asks you to pick randomly from three doors. Behind two of the doors is a goat, and behind one of the doors is a brand-spanking-new car. After you select one of the doors, the host then intentionally selects one of the other doors to reveal a goat. He then asks if you want to switch your guess to the other door or stay on your current guess. Is it to your advantage to switch?&lt;/p&gt;

&lt;p&gt;Many people say that it doesn’t matter: your choice is a 50/50 between your door having a goat or a car. However, the true solution is that it is definitely to your advantage to switch! Swapping doors gives you a 2/3 chance to win the car.&lt;/p&gt;

&lt;p&gt;The reason for this has a lot to do with probability theory and other unintuitive math, but it boils down to the idea that if you started off picking a goat, you’ll win the car by swapping after the host reveals the other goat. Only if you start on the car would swapping win you a goat. You can also consider the three different possibilities in the chart below:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/probability/goats.png&quot; alt=&quot;goats&quot; /&gt;&lt;/p&gt;

&lt;p&gt;As we can see, 2/3 scenarios win the car by swapping.&lt;/p&gt;

&lt;p&gt;After (unsuccessfully) trying to explain this to my grandfather and other family, it got me thinking about the topic of probability in general. I have actually never taken a probability course, and it’s one of the areas of quantitative methodology that I’m probably weakest in. So, I figured that it would be a good idea to get a basic intro to some of the ideas in probability theory. Thus the second installment of Grad School Scramble™ was born.&lt;/p&gt;

&lt;h3 id=&quot;preparation&quot;&gt;Preparation&lt;/h3&gt;

&lt;p&gt;To see what topics I didn’t already have a good grasp of, I took an open-source final from an MIT probability course and graded myself using the provided key. Holy smokes, I have a &lt;em&gt;long&lt;/em&gt; way to go. I had a solid understanding of most of the counting/set theory stuff just because I have taken a course on discrete mathematics, but most of the other material I was lost on. After my self-confidence was ground into pieces, I realized that a blog post covering an entire probability course was probably unfeasible, so I came up with a list of easier concepts to learn (and saved the multivariate calculus stuff for another time).&lt;/p&gt;

&lt;h2 id=&quot;conditional-probability&quot;&gt;Conditional Probability&lt;/h2&gt;

&lt;p&gt;Conditional probability was the first concept that I didn’t have much prior experience with. I had worked with various concepts related to probability and counting in my discrete math course, including the &lt;a href=&quot;https://en.wikipedia.org/wiki/Inclusion%E2%80%93exclusion_principle&quot;&gt;principle of inclusion/exclusion&lt;/a&gt; and different methods of counting with combination/permutation rules. However, we never explicitly forayed into the world of conditional probability.&lt;/p&gt;

&lt;p&gt;At its core, the idea of conditional probability answers the question “how does the likelihood of an event change given additional information?” When I was first reading about it, my intuition told me that you could answer this question with simple counting techniques: for example, what is the probability of a card from a standard deck having a suit of hearts given that its color is red? It seemed as if you could simply use logic to determine the sample space (red heart, red diamond, black club, black spade) and only use the subset containing the necessary event. However, this method is a bit constrained and only applicable for simple problems.&lt;/p&gt;

&lt;p&gt;A formal definition of conditional probability is as follows:&lt;/p&gt;

&lt;p&gt;$P(A\mid B)=\frac{P(A\cap B)}{P(B)}$&lt;/p&gt;

&lt;p&gt;This reads as “the probability of $A$ given $B$ is equal to the probability of $A$ &lt;em&gt;and&lt;/em&gt; $B$ divided by the probability of $B$.”&lt;/p&gt;

&lt;p&gt;I had never seen this formula before. Intuitively, it makes sense, though. If $A$ and $B$ are dependent events, then dividing out the probability of $B$ from the intersection of $A$ and $B$ seems to leave you with only the probability of $A$ (given that $B$ has already occurred).&lt;/p&gt;

&lt;p&gt;Another way to interpret this is by considering venn diagrams.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/probability/conditional.png&quot; alt=&quot;conditional&quot; /&gt;&lt;/p&gt;

&lt;p&gt;When we state that $B$ has already occurred, we can discard the entire region outside of $B$ (everything in pink). Thus, we are left with $\frac{P(A\cap B)}{P(B)}$ as the total probability of $A$ occurring.&lt;/p&gt;

&lt;p&gt;One important consideration to make is that we can rewrite the formula as $P(A\mid B)P(B)=P(A\cap B)$ by simple multiplication. Thus, if we know the conditional probability of something we can use this formula to find the probability of event intersections.&lt;/p&gt;

&lt;p&gt;Let’s use this formula to solve a problem. Approximately 10% of cars are red. What is the probability that two red cars get into an accident on the highway given that one of the cars is red? We can define $P(A)$ as the probability that the crash has two red cars and $P(B)$ as the probability that at least one car is red. Because $A$ is necessarily a subset of $B$ (for both cars to be red, at least one must be red), we can see that $P(A\cap B)=A=.01$. $P(B)$ is the complement of neither car being red, so $P(B)=1-(.9\cdot.9)=.19$. Because $\frac{.01}{.19}\approx.053$, the overall probability $P(A\mid B)\approx.053$.&lt;/p&gt;

&lt;p&gt;It seems a bit counterintuitive that the answer isn’t just $.1$. After all, it seems like we know one of the cars is red, so the probability of the other car being red as well should just be $.1$. However, we don’t specifically know &lt;em&gt;which&lt;/em&gt; car is red. Imagine that we specify the cars as Car 1 and Car 2. Our event space is as follows (for $R=red$ and $N=not red$): ${(R,N), (N,R), (R,R), (N,N)}$. When we use the logic of “one of the cars is red, so the probability of the other one being red as well should just be $.1$”, we implicitly make the assumption that we’re talking about Car 1 being the red one. However, the event that at least one of the cars is red comprises the set ${(R,N), (N,R), (R,R)}$ while the event that Car 1 is red comprises ${(R,N),(R,R)}$. Because $(N,R)$ is non-empty (we know that it could be the case that Car 2 is the red one), we know that the probability of both cars being red is &lt;em&gt;smaller&lt;/em&gt; when we only know one car is red compared to when we know Car 1 is red (because in the first scenario $(R,R)$ is selected from a larger event space than in the second).&lt;/p&gt;

&lt;blockquote&gt;
  &lt;p&gt;Another note: Two events can be considered formally independent if $P(A\mid B)=P(A)$. This is because they are independent if knowing that $B$ has occurred does not change the probability of $A$. It then follows that if two events are independent then $P(A\cap B)=P(A\mid B)\cdot P(B)=P(A)\cdot P(B)$, which follows our prior understanding of probability.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2 id=&quot;bayes-theorem&quot;&gt;Bayes’ Theorem&lt;/h2&gt;

&lt;p&gt;Throughout the murky, dark waters of statistics and probability, one beast lies dormant, slumbering, waiting for an unfortunate soul to stumble across its gaping maw. This beast is none other than Bayes’ Theorem, striking fear into the hearts of frequentists everywhere (before you write an angry email—I know that Bayes’ Theorem itself is not wholly relegated to Bayesian statistics).&lt;/p&gt;

&lt;p&gt;Much like conditional probability, I hadn’t much experience with Bayes’ Theorem before looking into it. In one of my undergraduate psych classes with Dr. Terry, we discussed the famous Kahneman and Tversky base rate fallacy in terms of Bayes’ Theorem: where individuals fail to incorporate the base rate of an event and obtain incorrect probabilistic deductions. Here is the theorem in all its glory:&lt;/p&gt;

&lt;p&gt;$P(A\mid B)=\frac{P(B\mid A)\cdot P(A)}{P(B)}$.&lt;/p&gt;

&lt;p&gt;This theorem lets us find the inverse of a conditional probability: If we know the probability of $A$ given $B$, we can now find the probability of $B$ given $A$. The proof follows from our previous conditional probability formula:&lt;/p&gt;

&lt;p&gt;$P(A\cap B)$ is symmetric with respect to $A$ and $B$, so&lt;/p&gt;

&lt;p&gt;$P(B\mid A)\cdot P(A)=P(A\cap B)=P(A\mid B)\cdot P(B)$.&lt;/p&gt;

&lt;p&gt;Then, if we divide by $P(A)$ we obtain Bayes’ Theorem.&lt;/p&gt;

&lt;p&gt;We can use this tool to answer the Monty Hall problem discussed earlier. If event $A$ is that the car is behind our door (door 1), and event $B$ is that the host opened door 3, we find the probabilities and substitute them in:&lt;/p&gt;

&lt;p&gt;$P(B\mid A)$ will be $.5$ because the probability of the host opening door 3 &lt;em&gt;given&lt;/em&gt; that the car is behind our door is $.5$ (he could have chosen either door 2 or door 3).&lt;/p&gt;

&lt;p&gt;$P(A)=\frac{1}{3}$&lt;/p&gt;

&lt;p&gt;$P(B)=.5$ (There are three scenarios: the car is behind our door, which gives Monty a .5 probability of picking door 3, the car is behind door 2, which gives Monty a 1 probability of picking door 3, and the car is behind door 3, which gives Monty a 0 probability of picking door 3. Thus, the total probability [$(.5 + 1 + 0) * \frac{1}{3}$] of Monty picking door 3 is .5)&lt;/p&gt;

&lt;p&gt;So:&lt;/p&gt;

&lt;p&gt;$P(A\mid B)=\frac{.5 \cdot \frac{1}{3}}{.5}=\frac{1}{3}$.&lt;/p&gt;

&lt;p&gt;In other words, the probability of the car being behind our door &lt;em&gt;given&lt;/em&gt; that Monty opened door 3 is $\frac{1}{3}$. Thus, it is in our interest to switch doors.&lt;/p&gt;

&lt;p&gt;Bayes’ Theorem can also be stated as $Posterior=Prior \cdot Likelihood$, where each part of the statement is part of our original formula.&lt;/p&gt;

&lt;h2 id=&quot;common-discrete-probability-distributions&quot;&gt;Common (Discrete) Probability Distributions&lt;/h2&gt;

&lt;p&gt;Next, I’d like to cover a few distributions that I was a little shaky on. It’s easy to fixate on the normal distribution as the “big one” of important distributions, but many data-generating processes are non-normal.  Also, this gave me a chance to practice my plot-making skills in R.&lt;/p&gt;

&lt;p&gt;For discrete distributions (which all of the following distributions are), the &lt;strong&gt;probability mass function&lt;/strong&gt;, or &lt;strong&gt;pmf&lt;/strong&gt;, is the function which gives the probability of a random variable from the distribution being equal to a given variable. This function is written $p(a)$ where $a$ is the variable of interest. This will be important as we examine different distributions.&lt;/p&gt;

&lt;h4 id=&quot;bernoulli-distribution&quot;&gt;Bernoulli Distribution&lt;/h4&gt;

&lt;p&gt;The Bernoulli distribution is a simple one: it models one trial in an event with a binary outcome. So, for example, if we’re observationally assessing the behavior of dogs being walked in a park (and want to determine whether or not they will urinate on a fire hydrant), the behavior of one specific dog would be modeled by a Bernoulli distribution.&lt;/p&gt;

&lt;p&gt;The Bernoulli distribution takes one parameter, $p$, which represents the underlying probability of a $1$ for the outcome. If 70% of dogs will pee on the fire hydrant, then $p=.7$. If $X$ is the variable of interest, by definition $P(X=1)=p$.&lt;/p&gt;

&lt;p&gt;Here’s an example of a Bernoulli distribution with $p=.7$:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/probability/bernoulli.png&quot; alt=&quot;bernoulli&quot; /&gt;&lt;/p&gt;

&lt;h4 id=&quot;binomial-distribution&quot;&gt;Binomial Distribution&lt;/h4&gt;

&lt;p&gt;The binomial distribution is an extension of the Bernoulli distribution. It models the probability of an event (binary, represented by a Bernoulli distribution) occurring a certain number of times in a set number of trials. It has two parameters: $n$ (the number of trials) and $p$ (the probability of a success).&lt;/p&gt;

&lt;p&gt;For example, if we wanted to determine the probability of flipping a certain number of heads in ten total flips of a coin, we might use the following binomial distribution:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/probability/binomial.png&quot; alt=&quot;binomial&quot; /&gt;&lt;/p&gt;

&lt;p&gt;By looking at this distribution, we can see the approximate probability of flipping a certain number of heads out of 10 flips with a fair coin ($p=.5$). For example, the chance of flipping exactly 3 heads is around $12\%$. It looks Gaussian, but this is only because our probability of a heads is $.5$. If we want to imagine that the coin we are flipping is unfair, we can assess the following distributions with $p=.2$ and $p=.9$, respectively:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/probability/binomial2.png&quot; alt=&quot;binomial2&quot; /&gt;&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/probability/binomial3.png&quot; alt=&quot;binomial3&quot; /&gt;&lt;/p&gt;

&lt;p&gt;How do we obtain these distributions? The pmf for a binomial distribution with probability $p$ and # of trials $n$ is defined as&lt;/p&gt;

&lt;p&gt;$P(X=k)=\binom{n}{k}p^k(1-p)^{n-k}$&lt;/p&gt;

&lt;p&gt;From this, we can see why it’s called the binomial distribution. The $n$ choose $k$ part of the formula is a binomial coefficient: coefficients of binomial expansion that follow &lt;a href=&quot;https://en.wikipedia.org/wiki/Pascal%27s_triangle&quot;&gt;Pascal’s Triangle&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/probability/pascal.png&quot; alt=&quot;pascal&quot; /&gt;&lt;/p&gt;

&lt;h4 id=&quot;geometric-distribution&quot;&gt;Geometric Distribution&lt;/h4&gt;

&lt;p&gt;Geometric distributions model the probability of a certain number of failures before a success in a sequence of Bernoulli trials. In other words, they show how many tails you can expect to flip before you get a heads.&lt;/p&gt;

&lt;p&gt;If $X$ is the number of tails, the pmf for a geometric distribution is defined as:&lt;/p&gt;

&lt;p&gt;$P(X=k)=(1-p)^kp$&lt;/p&gt;

&lt;p&gt;where $p$ is the probability of a success. The intuition behind this formula is that it’s simply a representation of the odds of getting the desired sequence. For example, if you want to find the odds of flipping 4 heads before a tails, you would find the odds of getting HHHHT, which is simply $.5^5$.&lt;/p&gt;

&lt;p&gt;Take a look at the following distribution:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/probability/geometric.png&quot; alt=&quot;geometric&quot; /&gt;&lt;/p&gt;

&lt;p&gt;From this, we can see the probability behind each number of tails before getting a heads. The interpretation of $0$ is that the first flip was a heads. Again, if we were using an unfair coin, the distribution might look something like this:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/probability/geometric2.png&quot; alt=&quot;geometric2&quot; /&gt;&lt;/p&gt;

&lt;h4 id=&quot;poisson-distribution&quot;&gt;Poisson Distribution&lt;/h4&gt;

&lt;p&gt;The Poisson distribution models the number of fish you can expect to catch in a given period of time. Just kidding. Kind of. This distribution was named after the French mathematician Simeon Poisson, whose last name translates to “Fish.” It models the expected number of events in a certain period of time given the average number of observed events. For example, if you know that you pass 3 gas stations every half-hour on a road trip, and you wanted to determine the probability of passing 5 in the next half-hour, you might use a Poisson distribution.&lt;/p&gt;

&lt;p&gt;It has one parameter: $\lambda$. This paramter is the average number of successes for the specified time period. The pmf is defined as&lt;/p&gt;

&lt;p&gt;$P(\lambda, k)=\frac{\lambda^ke^{-\lambda}}{k!}$&lt;/p&gt;

&lt;p&gt;where $k$ is the desired number of successes per period.&lt;/p&gt;

&lt;p&gt;We’ve officially passed into the realm of pmf’s that I don’t understand. Intuitively, you can derive this from the binomial distribution by setting up a formula which represents a number of discrete probability chunks, each with probability $\frac{\lambda}{n}$ (where $\lambda$ is the average # of events per time period and $n$ is the number of discrete chunks). A binomial distribution models the expected number of successes given a certain probability and number of trials, so we can substitute this probability into a binomial limit as $n\rightarrow \infty$. After some complicated math with limits and $e$, we get the pmf of a Poisson distribution.&lt;/p&gt;

&lt;p&gt;Below you can see a Poisson distribution for $\lambda=1$, $\lambda=2$, and $\lambda=6$.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/probability/poisson.png&quot; alt=&quot;poisson&quot; /&gt;&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/probability/poisson2.png&quot; alt=&quot;poisson2&quot; /&gt;&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/probability/poisson3.png&quot; alt=&quot;poisson3&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Many of these discrete distributions have analogues in continuous distributions (for example, the geometric distribution and the exponential distribution). I chose not to cover the continuous distributions because I was already at least conceptually familiar with them, but maybe I’ll dive into their probability functions in a later post.&lt;/p&gt;

&lt;h2 id=&quot;concluding-thoughts&quot;&gt;Concluding Thoughts&lt;/h2&gt;

&lt;p&gt;Well, that covers a lot of the things I’ve been learning related to probability. I wanted to cover some other interesting stuff like continuous distributions and the &lt;a href=&quot;https://en.wikipedia.org/wiki/Central_limit_theorem&quot;&gt;Central Limit Theorem&lt;/a&gt;, but I didn’t think I would do them justice. This post was mostly the product of a day and a half, so I could only get so detailed with some of the formulae. I might revisit this topic in a future blog post or after I’ve actually taken a class on the matter. Anyway, thanks for reading. I need to go—I have an upcoming Pokémon tournament to prepare for 😈&lt;/p&gt;
</description>
        <pubDate>Mon, 09 Jun 2025 00:00:00 +0000</pubDate>
        <link>https://keanehauck.com/quantitative/methodology/statistics/data/science/2025/06/09/probability/</link>
        <guid isPermaLink="true">https://keanehauck.com/quantitative/methodology/statistics/data/science/2025/06/09/probability/</guid>
      </item>
    
      <item>
        <title>The Gholdengo Experience</title>
        <description>&lt;p&gt;&lt;img src=&quot;/images/posts/milwaukee/selfie.jpg&quot; alt=&quot;selfie&quot; title=&quot;Top tables, baby&quot; /&gt;&lt;/p&gt;

&lt;p&gt;I recently attended the 2025 Milwaukee Regional Championship for the Pokémon trading card game. Going into this tournament, I had one goal: to &lt;strong&gt;day 2&lt;/strong&gt;. Large Pokémon tournaments (regional or international) are organized such that the first day of play consists of 8 or 9 rounds of swiss bracket: players with the same win record play each other in best-of-3 matches. At the end of the first day, the top players (typically the top 10% or so) are invited to day 2, where the tournament is concluded via additional rounds of swiss and a final top cut.&lt;/p&gt;

&lt;p&gt;In wanting to day 2, I am wanting to place in the top 10% of a group of serious players who all have traveled from large distances to be at this regional tournament—not an easy task. Still, I think it’s achievable. I had only played in one prior regional, but I was one win away from day 2-ing—and I am a much more experienced player now than I was at the previous event.&lt;/p&gt;

&lt;h3 id=&quot;deck-choice--meta-calls&quot;&gt;Deck choice &amp;amp; meta calls&lt;/h3&gt;

&lt;p&gt;The first step to doing well at a big tournament like this is selecting a deck that you believe will give you the best chance of success. For a few reasons which I will outline later, I chose Gholdengo Dragapult as the deck for the regional.&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;
&lt;img src=&quot;/images/posts/milwaukee/gholdengo.jpg&quot; alt=&quot;Gholdengo&quot; width=&quot;200&quot; class=&quot;center-image&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;The strategy behind this deck is to run 1 of every type of basic energy card, search them out with &lt;a href=&quot;https://www.tcgplayer.com/product/589933/pokemon-sv08-surging-sparks-energy-search-pro?srsltid=AfmBOopRA95sGx9fqVtOSaPt7OFTTnmV3a0gcbD06E1PTqbscyEddto8&quot;&gt;Energy Search Pro&lt;/a&gt;, and swing for huge damage numbers using Gholdengo ex’s Make It Rain attack. This deck mainly attacks with Gholdengo, but also runs a thin line of &lt;a href=&quot;https://www.tcgplayer.com/product/550174/pokemon-sv06-twilight-masquerade-dragapult-ex-130167?country=US&amp;amp;utm_campaign=18099737719&amp;amp;utm_source=google&amp;amp;utm_medium=cpc&amp;amp;utm_content=&amp;amp;utm_term=&amp;amp;adgroupid=&amp;amp;gad_source=1&amp;amp;gad_campaignid=17428625972&amp;amp;gbraid=0AAAAADHLWY30i0iftExF3hs_AmD-0ez_S&amp;amp;gclid=CjwKCAjwiezABhBZEiwAEbTPGAY1gkX0q14Hz3GQ8cFRchrBrA-amLPcJ8DePqmZbixn94HunA8LLBoCcNsQAvD_BwE&quot;&gt;Dragapult ex&lt;/a&gt; for draw power and as an additional attacker. Pokémon like &lt;a href=&quot;https://www.tcgplayer.com/product/550156/pokemon-sv06-twilight-masquerade-cornerstone-mask-ogerpon-ex-112-167?srsltid=AfmBOop8-vog8pOyxVKhUxJVgOJD8d2fXLC2Q2B20XVWyymUvB39g6xG&amp;amp;Language=English&quot;&gt;Cornerstone Mask Ogerpon&lt;/a&gt; are a huge issue for Gholdengo, and having the option to attack with an additional Pokémon can make a huge difference in some matchups. Dragapult only needs 1 psychic and 1 fire to attack—both of which are already being run in the deck.&lt;/p&gt;

&lt;p&gt;To me, this deck felt like a good meta call for the following reasons: I could finish out games quickly, and the new up-and-coming deck Joltik Box has a pretty bad matchup into it. The other deck I was considering for the event was pure Dragapult: a deck which slowly locks out your opponent and tries to make a comeback by disrupting them and swinging with the huge health attacker of Dragapult. Joltik box was a deck created in response to this strategy. It aims to power up an &lt;a href=&quot;https://www.tcgplayer.com/product/523706/pokemon-sv04-paradox-rift-iron-hands-ex-070-182?srsltid=AfmBOoqKcXwBrHoRR1htWUXxdJgem8DGMR0Wd_mxjC_MvPVbK4ypf1T5&amp;amp;Language=English&quot;&gt;Iron Hands ex&lt;/a&gt; and start KO’ing your small Pokémon for 2 prizes each. I didn’t want to face down these hyper-turbo aggressive decks with Dragapult. In addition, Dragapult games can take a long time, and in a tournament where you only get 50 minutes to play 3 games in a match, I wanted the ability to close out games quickly. Thus, I landed on Gholdengo.&lt;/p&gt;

&lt;p&gt;Gholdengo had a good matchup against Dragapult (one of the most popular decks) but didn’t have the best matchup against some of the decks in the format like &lt;a href=&quot;https://www.tcgplayer.com/product/542867/pokemon-sv05-temporal-forces-raging-bolt-ex-123-162?Language=English&amp;amp;page=1&quot;&gt;Raging Bolt&lt;/a&gt;. I figured I wouldn’t see many of them. Spoiler: this was not the case…&lt;/p&gt;

&lt;h3 id=&quot;round-1-tie-against-charizard&quot;&gt;Round 1: Tie against Charizard&lt;/h3&gt;

&lt;p align=&quot;center&quot;&gt;
&lt;img src=&quot;/images/posts/milwaukee/charizard.jpg&quot; alt=&quot;Charizard&quot; width=&quot;200&quot; class=&quot;center-image&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;Let’s walk through each round of the tournament and see how I did! The first round was against a pretty new player on Charizard—one of the worst matchups for my deck. Not a good start. This deck aims to utilize cards like &lt;a href=&quot;https://www.tcgplayer.com/product/610392/pokemon-sv-prismatic-evolutions-dusknoir?country=US&amp;amp;srsltid=AfmBOoqF2xhAE7ZWOt8XbvxeuYNsEtP4BGipg_HTHOVpuWi_cK2l24dI_bM&amp;amp;gQT=1&amp;amp;Language=English&quot;&gt;Dusknoir&lt;/a&gt; and &lt;a href=&quot;https://www.tcgplayer.com/product/610455/pokemon-sv-prismatic-evolutions-briar?Language=English&amp;amp;page=1&quot;&gt;Briar&lt;/a&gt; to manipulate prizes and win faster than your opponent is able to. Now, the guy playing this deck was obviously fairly new, so I wasn’t &lt;em&gt;too&lt;/em&gt; worried, but I still had to be careful.&lt;/p&gt;

&lt;p&gt;The first game was an easy win: he got off to a slow start and I was able to run him off the board. The second game was not. I prized &lt;a href=&quot;https://www.tcgplayer.com/product/589933/pokemon-sv08-surging-sparks-energy-search-pro?srsltid=AfmBOoqbegADDsoFhmMT0BWGI_grxDcXFo579oq8vDYx2LKU86X5Oyv2&amp;amp;Language=English&quot;&gt;Energy Search Pro&lt;/a&gt;. Remember how Gholdengo needs a lot of energy to discard in order to be able to swing for big damage? Without the energy search pro, there’s no way to aggressively search those energy out of the deck. So, I was in a predicament. Should I just scoop (forfeit) the second game of the match and try and win the third one? Or should I stick it out and attempt to mount a comeback? I opted for the latter—I didn’t want to scoop against a new player like this and I figured leaving the match to the last game was a scary risk. Unfortunately, I was unable to win, and we moved on to the third game. This was a heartbreaker—time was called during our game, and I was 1 turn away from winning when the tie was officially announced. In tournaments like these, if time is called during a game, each player gets an additional two turns, and if neither player has won at the end of the turns then it is officially a tie.&lt;/p&gt;

&lt;p&gt;I was quite frustrated with the tie: I knew that it would hurt my resistance later (the guy went 1-5 drop) and it certainly wasn’t the start I had been looking for. Still, it easily could have been a loss against Charizard, so I guess I have to count my blessings. On to the next.&lt;/p&gt;

&lt;h3 id=&quot;round-2-win-against-dragapult&quot;&gt;Round 2: Win against Dragapult&lt;/h3&gt;

&lt;p align=&quot;center&quot;&gt;
&lt;img src=&quot;/images/posts/milwaukee/dragapult.jpg&quot; alt=&quot;Dragapult&quot; width=&quot;200&quot; class=&quot;center-image&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;Round 2 I faced a really good player: we both sat down and immediately talked about how disappointed we were with the first round tie. Luckily, he was on Dragapult, which was one of the decks that my list was designed to beat. In the first game, I was able to quickly take aggro and start taking KO’s before he was set up. Dragapult, as a deck, relies on a lot of hand disruption with cards like &lt;a href=&quot;https://www.tcgplayer.com/product/497557/pokemon-sv02-paldea-evolved-iono-185193?country=US&amp;amp;utm_campaign=18099737719&amp;amp;utm_source=google&amp;amp;utm_medium=cpc&amp;amp;utm_content=&amp;amp;utm_term=&amp;amp;adgroupid=&amp;amp;gad_source=1&amp;amp;gad_campaignid=17428625972&amp;amp;gbraid=0AAAAADHLWY30i0iftExF3hs_AmD-0ez_S&amp;amp;gclid=CjwKCAjwiezABhBZEiwAEbTPGGy31VYzRdUaHYkvr99JQeBrNqRLnQqetYwmp17NoYQ5x3hUxJ25GhoCEJ8QAvD_BwE&amp;amp;Language=English&quot;&gt;Iono&lt;/a&gt; and &lt;a href=&quot;https://www.tcgplayer.com/product/550209/pokemon-sv06-twilight-masquerade-unfair-stamp?srsltid=AfmBOorK3aTcnhR-Q4l-OCeAqPzkqklJ-Hq69LKNeRamCLeJJX9BVdkN&amp;amp;Language=English&quot;&gt;Unfair Stamp&lt;/a&gt;. To combat this, I run a Genesect:&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;
&lt;img src=&quot;/images/posts/milwaukee/genesect.jpg&quot; alt=&quot;Genesect&quot; width=&quot;200&quot; class=&quot;center-image&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;This card allows me to quickly take aggressive KO’s against Dragapult without fearing an early unfair stamp, which makes a huge difference in how the matchup goes. It works like a charm, and I take quick 2-0. The second game was incredibly close, but a clutch &lt;a href=&quot;https://www.tcgplayer.com/product/523852/pokemon-sv04-paradox-rift-professor-turos-scenario?country=US&amp;amp;utm_campaign=18099737719&amp;amp;utm_source=google&amp;amp;utm_medium=cpc&amp;amp;utm_content=&amp;amp;utm_term=&amp;amp;adgroupid=&amp;amp;gad_source=1&amp;amp;gad_campaignid=17428625972&amp;amp;gbraid=0AAAAADHLWY30i0iftExF3hs_AmD-0ez_S&amp;amp;gclid=EAIaIQobChMInuXRseCRjQMVl3N_AB3zmSQeEAQYASABEgJQ1vD_BwE&amp;amp;Language=English&quot;&gt;Professor Turo’s Scenario&lt;/a&gt; allows me the turn I need to win the game. 1-0-1.&lt;/p&gt;

&lt;h3 id=&quot;round-3-loss-against-gholdengo&quot;&gt;Round 3: Loss against Gholdengo&lt;/h3&gt;

&lt;p&gt;This round absolutely destroyed my confidence. The player I was facing was great: I actually ran into him the next day and we had a nice conversation about our matches. I won the first game in convincing fashion, but bricked (had a horrible hand) the second game. In the third game I was again off to a good start. Remember the Genesect from earlier which shuts down ACE SPECs? In the mirror match, you’re normally good to evolve and start swinging if you have a Genesect down. I did, so I evolve to Gholdengo and KO a gimmi. However, he reveals an insane tech: Klefki.&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;
&lt;img src=&quot;/images/posts/milwaukee/klefki.jpg&quot; alt=&quot;Klefki&quot; width=&quot;200&quot; class=&quot;center-image&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;When this is in the active, basic Pokémon don’t have abilities. This means that with Klefki in the active, he is able to get around Genesect and play his Energy Search Pro. So, he takes the return KO on my Gholdengo, and the first two prizes. This means that I’m almost completely unable to win. I try to mount a comeback with Dragapult + Iono, but it doesn’t work. 1-1-1.&lt;/p&gt;

&lt;p&gt;Klefki &lt;em&gt;completely&lt;/em&gt; took me by surprise. I hadn’t seen this card being run in any Gholdengo list previously, so I really started to wonder if I was behind the meta for this tournament. A little worried, I headed into round 4.&lt;/p&gt;

&lt;h3 id=&quot;round-4-win-against-blissey&quot;&gt;Round 4: Win against Blissey&lt;/h3&gt;

&lt;p&gt;I was in the pits this game. I wasn’t feeling good about the tournament, and starting off the day at 1-1-1 was not putting me on track to day 2. So, I needed an easy win. Luckily, this is exactly what I hit.&lt;/p&gt;

&lt;p&gt;I was facing a deck which used &lt;a href=&quot;https://www.tcgplayer.com/product/550178/pokemon-sv06-twilight-masquerade-blissey-ex-134-167?srsltid=AfmBOorOuS6aLRvBIcEJalJdgmvUfqXRwrqXuoPXAYplowTNt9Ugs6zU&amp;amp;Language=English&quot;&gt;Blissey ex&lt;/a&gt; alongside cards like &lt;a href=&quot;https://www.tcgplayer.com/product/550139/pokemon-sv06-twilight-masquerade-munkidori?srsltid=AfmBOorsNEskicvkRe4GUfjhy5DfKZ2UcQGRjH3BPgzGch8ZAZlSt9tD&amp;amp;Language=English&quot;&gt;Munkidori&lt;/a&gt; to outlive and take KO’s on the opponent. Gholdengo doesn’t leave any damage on the board for Munkidori to manipulate, so theoretically I have a good matchup into this list. The one problem was Cornerstone Mask Ogerpon:&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;
&lt;img src=&quot;/images/posts/milwaukee/cornerstone.jpg&quot; alt=&quot;Cornerstone&quot; width=&quot;200&quot; class=&quot;center-image&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;Gholdengo is unable to hit this Pokémon. So, if they leave this guy in the active and don’t play any Blissey, I am pretty screwed. However, this is where the Dragapult part of my deck comes into play: It can hit Cornerstone. However (again), I prized my Dragapult on the first game. However (again again), my opponent misplayed and benched two Blissey. This allowed me to gust and KO both for 4 prizes, and take snipe KO’s on Munkidoris with &lt;a href=&quot;https://www.tcgplayer.com/product/560348/pokemon-sv-shrouded-fable-fezandipiti-ex-038-064?srsltid=AfmBOoqkdEdDzd13KWGIMb2r27oF0_qb_i30K7ojHXgixWUeDDugZVnF&quot;&gt;Fezandipiti ex&lt;/a&gt; for the last 2. Dragapult was the last prize…&lt;/p&gt;

&lt;p&gt;Game 2 was a stomp. I didn’t prize any of the pieces for Dragapult, so he was unable to do anything with Cornerstone and I won handily. 2-1-1.&lt;/p&gt;

&lt;h3 id=&quot;round-5-win-against-dragapult&quot;&gt;Round 5: Win against Dragapult&lt;/h3&gt;

&lt;p&gt;This round was scary for 1 main reason: The clock. Remember how I mentioned that each round is 50 minutes for a best-of-3? Well, in this round my opponent was playing &lt;em&gt;incredibly&lt;/em&gt; slowly on Dragapult Dusknoir. I won the first game, but we only had 20 minutes for the last 2 games after this.&lt;/p&gt;

&lt;p&gt;We start the 2nd game, and I notice that my Energy Search Pro is prized, along with &lt;em&gt;4&lt;/em&gt; energy. This is a huge problem. Beating Dragapult often requires being able to piece together a KO against the first Dragapult they swing with, after they disrupt you. No Energy Search Pro means that accomplishing this is much, &lt;em&gt;much&lt;/em&gt; harder. I was in another predicament: Do I scoop now and try to take a quick win in game 3? Or do I stick it out and try to win an unlikely comeback? I opted for the latter, which was probably a mistake. I was able to win, but only because my opponent missed a turn of attaching energy. I had to swing into a Dragapult for 150 with Gholdengo, which is never something you want to do. However, they were unable to find another energy to attach to a benched Drakloak, so after I KO’d the Dragapult the next turn, they were unable to piece together an attack. All I needed was this 1 turn of leeway, and I was able to roll the advantage to a victory. 3-1-1.&lt;/p&gt;

&lt;h3 id=&quot;round-6-win-against-joltik-box&quot;&gt;Round 6: Win against Joltik Box&lt;/h3&gt;

&lt;p&gt;I’m finally seeing the benefits from my deck choice. He opened Joltik, and I knew I was favored. My list runs a &lt;a href=&quot;https://www.tcgplayer.com/product/510898/pokemon-sv03-obsidian-flames-vengeful-punch?country=US&amp;amp;srsltid=AfmBOorh8osrWJWV1ll942KGhY9qZSm0k2HFYf7AROApONE0N4DaVgm5zjI&amp;amp;gQT=2&quot;&gt;Vengeful Punch&lt;/a&gt; to get around Pikachu ex, and so I can normally take advantage in the prize race.&lt;/p&gt;

&lt;p&gt;Game 1 he bricked—not much to say. However, the second game was very interesting. My opponent started with a Mew ex, and benched Miriadon to find Joltik and Pikachu ex. He then benched Fez and another Pikachu. This means that he is in a board state where he has Mew, Fez, Joltik, Miriadon, and 2 Pikachu—and no room to bench any other Pokémon. This is normally &lt;em&gt;OK&lt;/em&gt;; you don’t want to bench more 2-prize liabilities than you have to against Gholdengo, but it’s unlikely that I will be able to gust up something 3 turns in a row, so it’s not the end of the world. However—he was bench locked. He was obviously anticipating me evolving into a Gholdengo, but while planning for this he did not set up an Iron Hands, just two Pikachu. This means that I am free to &lt;em&gt;never&lt;/em&gt; evolve into a 2-prizer until he swings into my active Gimmighoul with a Pikachu: Something you never want to do. Thanks to this, I was able to successfully stall him for turns while he pointlessly waited for me to evolve into Gholdengo. Eventually, he reluctantly took aggro and swung into me, taking 1 prize in the process and enabling my Counter Catcher. Perfect. I was able to gust as needed to win the 2-2-2 prize race. I was really proud of this game because it was a great punish of my opponents misplay of board-locking themselves. 4-1-1.&lt;/p&gt;

&lt;h3 id=&quot;round-7-win-again-poison-arch&quot;&gt;Round 7: WIN again Poison Arch&lt;/h3&gt;

&lt;p align=&quot;center&quot;&gt;
&lt;img src=&quot;/images/posts/milwaukee/pecharunt.jpg&quot; alt=&quot;Pecharunt&quot; width=&quot;200&quot; class=&quot;center-image&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;This round was horrifying. One reason is that the deck my opponent was playing, Poison Archaludon, has the ability to win before you can even play a card. With a combo involving poison damage-modifying cards and Pecharunt, the deck can take a KO on your only Pokémon in play if they have 80 health or less: winning the game on turn 1. As my Gimmighouls have 70 health, I am a prime target for this donk strategy. Luckily, I was able to start Genesect in game 1, which is unable to be donked. I took a win when he missed a gust effect on his second turn to take a 2-prize kill.&lt;/p&gt;

&lt;p&gt;The other reason this round was horrifying is that if I win, I officially qualify for day 2. I had clawed my win record back from the poor 1-1-1 at the start to the quite decent 4-1-1 I currently had. 5 wins and a tie makes day 2, so if I won this round I would achieve my goal.&lt;/p&gt;

&lt;p&gt;Heading into game 2, I know that all I need to do is bring this one home. He goes first—but I start Munkidori this time. No donk potential. He goes ahead in the prize race, taking the first 2 poison KO’s before I can piece together a Gholdengo + gust response. I’m in trouble. On the turn he goes down to 2 prizes, he Iono’s the both of us to 4 cards—disrupting the large hand I had amassed via Coin Bonus, but also disrupting himself in the process. He takes the KO, and I know that I am in trouble.&lt;/p&gt;

&lt;p&gt;I promote Gholdengo. He has an &lt;a href=&quot;https://www.tcgplayer.com/product/589867/pokemon-sv08-surging-sparks-archaludon-ex-130-191?srsltid=AfmBOoqISs9ZWPijr1jiV0w0suuWm5ZVJLNWfJ4IOg25zyOHwE18D3wS&amp;amp;Language=English&quot;&gt;Archaludon ex&lt;/a&gt; in the active with a Binding Mochi. I bench Genesect, attaching a Vengeful Punch. No ACE SPECs for him. I then take a KO on his active Archaludon—but I do something smart. He is currently poisoned, and taking 30 each turn from the &lt;a href=&quot;https://www.tcgplayer.com/product/542673/pokemon-sv05-temporal-forces-perilous-jungle?srsltid=AfmBOop8a-gZAxtIIqJYdAYsEXuA17EVHMxxKhyahy4QP5Cc7dTx2avT&amp;amp;Language=English&quot;&gt;Perilous Jungle&lt;/a&gt; in play. This means that if I do 250 damage, the poison will knock him out, preventing him from using Fezandipiti’s ability on the next turn. I perform this play, leaving him with 4 cards. He needs to piece together an Arch, a Binding Mochi, and an attach for turn.&lt;/p&gt;

&lt;p&gt;He immediately plays arch and attaches an energy for turn. All he needs is the Mochi. However: He stops to pick up my Genesect and asks “No pink cards, right?” I confirm. He sighs, and plays a Carmine discarding a Secret Box. This is huge: If I hadn’t played the Genesect on the previous turn, he would have won. He draws his 5 cards—&lt;em&gt;no Binding Mochi&lt;/em&gt;. He promotes a 1-prizer, poisons, and passes. I now have 1 goal: Find a Boss’s Orders. I have about 15 cards in deck, a Superior Energy Retrieval in hand, and 4 draw abilities enabled. In addition, time has been called. If I don’t find this win, it’s all over. Here we go.&lt;/p&gt;

&lt;p&gt;Recon directive 1: nothing. Recon directive 2: nothing. Coin bonus 1: nothing. At this point, I have seen 6 cards from my deck, and no Boss’s Orders. It’s down to the final Coin Bonus from the active Gholdengo. I slowly flip over the first card… nothing. I look at my opponent and flip the last card onto the table—not even bothering to hide it. &lt;strong&gt;Boss’s Orders&lt;/strong&gt;. I immediately Boss the Fez and take a KO to take my last two prizes.&lt;/p&gt;

&lt;p&gt;I have officially made day 2, and in the most cinematic way possible. Digging with Coin Bonus to find the 1 card I needed at the last possible instance: I can’t imagine anything more perfect.&lt;/p&gt;

&lt;h3 id=&quot;round-8-tie-against-tera-box&quot;&gt;Round 8: Tie against Tera Box&lt;/h3&gt;

&lt;p&gt;I’m facing Alex Schemanske this round: a professional player. He is absolutely amazing at the game, and I’m so frazzled from the adrenaline of the previous round that I immediately misplay game 1 and lose as a result. However, in game 2 I lock in: I perfectly manage the prize-state of my board and take a close game 2 after a long back-and-forth. In game 3, we’re both playing incredibly quickly. Time is almost up, and we’re both screaming through our decks. Unfortunately for him, time is called, and on his final turn he is 1 prize away from winning. I can’t take it either, and so we tie. Pretty happy to take a game off of one of the pros.&lt;/p&gt;

&lt;h3 id=&quot;round-9-win-against-ns-box&quot;&gt;Round 9: Win against N’s Box&lt;/h3&gt;

&lt;p align=&quot;center&quot;&gt;
&lt;img src=&quot;/images/posts/milwaukee/darmanitan.jpg&quot; alt=&quot;Darmanitan&quot; width=&quot;200&quot; class=&quot;center-image&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;This round was the first of day 2. I am pretty nervous—the players who qualify for day 2 are some of the best in the world, and I don’t know how well I’m prepared to hang with them. In addition, this deck has an absolutely brutal matchup against mine: Darmanitan can hit Gholdengo for weakness, spreading massive damage across the board and one-shotting if I discard too many energy.&lt;/p&gt;

&lt;p&gt;Game 1 goes well; he doesn’t get a quick start and I run him over. Game 2 is extremely close. We both play it very well—but he makes a major misplay on his final turn. The board state is such that if he is able to Boss and use flamebody cannon, I lose. So, I swing into a Zoroark for 250: leaving it susceptible to a Munkidori play the next turn, and turning off his Counter Catchers. I also attach a Vengeful Punch to the active Gholdengo, and he forgets about it. He swings back into me, KO’ing himself in the process, and letting me boss a 2-prizer for game. I believe he would have lost anyway to the Munki play, but his gaffe made the win easier. 6-1-2.&lt;/p&gt;

&lt;h3 id=&quot;round-10-loss-against-raging-bolt&quot;&gt;Round 10: Loss against Raging Bolt&lt;/h3&gt;

&lt;p&gt;No! I hit the deck I was scared of. Not only do a lot of things have to go right in order for me to win this matchup, but I only have a few turns to piece together the cards I need. I absolutely brick both games: no supporters or even ball search. I get donked the first game and run off the board the second. Unfortunate, but it happens. 6-2-2.&lt;/p&gt;

&lt;h3 id=&quot;round-11-win-against-tera-box&quot;&gt;Round 11: Win against Tera Box&lt;/h3&gt;

&lt;p&gt;Game 1 he bricks. Game 2 was interesting: He started Cornerstone Mask Ogerpon, which is fortunate for him against Gholdengo. Hoewver, I am able to start building up Drakloaks, and after he Boss’s one of them he is unable to piece together a KO on the second, leaving me with a powered up Dragapult. By now, he is down to 3 prizes, so I Iono the both of us and swing into his Cornerstone, damaging a Pikachu ex on the bench, a fan rotom, and a hoothoot. He whiffs a Noctowl from the 3 cards he drew, and is forced to swing back into my Dragapult with the Cornerstone. I am able to gust Pikachu, move damage with Munkidori, and Phantom Dive for a &lt;em&gt;6-prize turn&lt;/em&gt; with Gholdengo. How often does that happen?&lt;/p&gt;

&lt;p&gt;7-2-2.&lt;/p&gt;

&lt;h3 id=&quot;round-12-loss-against-raging-bolt&quot;&gt;Round 12: Loss against Raging Bolt&lt;/h3&gt;

&lt;p&gt;This was an anticlimactic ending to the tournament. Before the match, my opponent asked if I wanted to just tie to guarantee a spot in the top 100. I declined, because this was my first day 2 and I wanted to see how well I could do. This ended up being a mistake, as I megabricked the first game and got blown off the board. In the second game I actually had a chance to take the lead on the prize race with a well-timed Ciphermaniac’s Codebreaking, but my 1 Counter Catcher was prized. What a heartbreaker. I ended the tournament with a record of 7-3-2, placing 136th.&lt;/p&gt;

&lt;h3 id=&quot;concluding-thoughts&quot;&gt;Concluding thoughts&lt;/h3&gt;

&lt;p&gt;I was very happy with my performance this tournament. I didn’t make too many major misplays: the one that stands out is not playing around Klefki in game 3 of round 3. I was fairly happy with my deck and list—I wish that I had realized how much of the meta is decks that were not Dragapult. A large portion of my 60 was dedicated to giving me great odds against Pult, but this left me weaker against hyperaggressive decks like Raging Bolt.&lt;/p&gt;

&lt;p&gt;Still, I am proud of myself for making day 2. I think I’ll pick up a different deck for NAIC, so we’ll have to get to testing.&lt;/p&gt;
</description>
        <pubDate>Wed, 07 May 2025 00:00:00 +0000</pubDate>
        <link>https://keanehauck.com/pokemon/gaming/cards/strategy/2025/05/07/milwaukee/</link>
        <guid isPermaLink="true">https://keanehauck.com/pokemon/gaming/cards/strategy/2025/05/07/milwaukee/</guid>
      </item>
    
      <item>
        <title>Grad School Scramble I: Ordinary Least Squares Estimation
</title>
        <description>&lt;p&gt;&lt;img src=&quot;/images/posts/ordinary-least-squares/logo.png&quot; alt=&quot;logo&quot; /&gt;&lt;/p&gt;

&lt;p&gt;I am proud to announce that I will be attending ASU in the fall to pursue a PhD in quantitative psychology under the guidance of Kevin Grimm! This decision was arrived at after a long and arduous application process wherein I flew to 7 different programs to interview. After all of my traveling, ASU was the one that felt most like home.&lt;/p&gt;

&lt;p&gt;To prepare for grad school, I figured that it would be helpful and cathartic (and a little bit exposing) to use the blog to revisit some of the beginner topics in quantitative psychology that I am not incredibly well-versed in. Thus: Welcome to the birth of &lt;strong&gt;Grad School Scramble&lt;/strong&gt;!!! In each installment of this multi-part series, I will attempt to learn and explain some area, subject, or topic of quantitative psychology that I had previously not understood well. &lt;em&gt;Important disclaimer&lt;/em&gt;: this is a judgement-free zone! If I cover some concept that makes you say, “Hey, shouldn’t an incoming &lt;em&gt;PhD student&lt;/em&gt; have a better understanding of this subject matter?” then, well, too bad! In this pastel-gray and baby-blue blog page, we can all find solace together in the fact that we are amateurs.&lt;/p&gt;

&lt;h2 id=&quot;todays-topic-ordinary-least-squares-estimation-in-linear-regression-and-other-methods&quot;&gt;Today’s topic: Ordinary Least Squares Estimation in Linear Regression (And Other Methods)&lt;/h2&gt;

&lt;p&gt;Kevin has been doing this awesome thing where he sends me small coding assignments designed to introduce me to the field of data simulation in advance of my first semester. These assignments are typically in R, and they normally involve me simulating some data on which I perform some statistical technique. For the first one, here was the task: Generate $n=500$ where $x_{1i}$ follows a standard normal distribution (mean = 0, std = 1), and $y_i=.5⋅x_{1i}+e_i$ where $e_i$ is normally distributed with a mean of 0 and a variance of .75. It was a nice little foray into data simulation, and I also calculated an estimated slope coefficient each time the data were generated.&lt;/p&gt;

&lt;p&gt;However, this got me thinking. To determine the coefficient I used the &lt;code class=&quot;language-plaintext highlighter-rouge&quot;&gt;lm&lt;/code&gt; function, which states the following as one of its parameters:&lt;/p&gt;

&lt;blockquote&gt;
  &lt;p&gt;&lt;code class=&quot;language-plaintext highlighter-rouge&quot;&gt;weights&lt;/code&gt;&lt;/p&gt;

  &lt;p&gt;an optional vector of weights to be used in the fitting process. Should be NULL or a numeric vector. If non-NULL, weighted least squares is used with weights &lt;code class=&quot;language-plaintext highlighter-rouge&quot;&gt;weights&lt;/code&gt; (that is, minimizing &lt;code class=&quot;language-plaintext highlighter-rouge&quot;&gt;sum(w*e^2)&lt;/code&gt;); otherwise ordinary least squares is used.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Notice that? Ordinary least squares. I realized that I didn’t actually have a good handle on what the ordinary least squares estimation method was doing, so I spent some time googling it.&lt;/p&gt;

&lt;h2 id=&quot;background&quot;&gt;Background&lt;/h2&gt;

&lt;p&gt;To contextualize, let’s get a quick run-down on linear regression. (Feel free to skip this entire section if you are familiar with it.)&lt;/p&gt;

&lt;p&gt;Linear regression, at its core, aims to estimate the level of some outcome variable, say $y$, on the level of some predictor variable, say $x$. Imagine that you have some (ambiguous) set of data like so:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/ordinary-least-squares/data1.png&quot; alt=&quot;data1&quot; title=&quot;Nice and simple&quot; /&gt;&lt;/p&gt;

&lt;p&gt;You might rightly notice that there is a general relationship between variable $x$ and variable $y$: as the level of $x$ increases, so does $y$ (generally). You then might ask: is there a way, given $x$, to predict a corresponding value for $y$? This is exactly what linear regression aims to do—to create a non-deterministic formula which predicts outcomes from a given predictor. One way to do this is by fitting a line (&lt;em&gt;linear&lt;/em&gt; regression) that follows the trend of the data:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/ordinary-least-squares/data2.png&quot; alt=&quot;data2&quot; title=&quot;How&apos;s my MS paint?&quot; /&gt;&lt;/p&gt;

&lt;p&gt;As per our 8th grade math education, we know that this line would take the form of $y=mx+b$, where $m$ is the slope and $b$ is the intercept. If our line perfectly described the data, then we would be able to perfectly predict each level of $y$ from each observation of $x$. Unfortunately, as we can see from the fact that not all of our data points fall along the line, it is not a perfect prediction. We call this error, or the &lt;strong&gt;residuals&lt;/strong&gt;. So, linear regression aims to incorporate this by adding a third variable to the equation of the line: the variance of the residuals. Let’s dissect this.&lt;/p&gt;

&lt;p&gt;Generally, when we fit the line to the data, we expect the data points to be closer &lt;em&gt;to&lt;/em&gt; the line than &lt;em&gt;away&lt;/em&gt; from it. For the purposes of linear regression, we assume that they do so in a normal manner. Thus, for each level of the $x$ variable, there is a range of values we would expect for the $y$ variable.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/ordinary-least-squares/data3.png&quot; alt=&quot;data3&quot; title=&quot;Mmmm grimace milkshake&quot; /&gt;&lt;/p&gt;

&lt;p&gt;For example, in this picture the purple curve shows the range of values we would expect from $y$ at the given point. We would expect to see the blue dot more than we would expect to see the green dot. (This assumption is called the assumption of normality of the residuals and the assumption that this range is the same at all levels is homoscedasticity.)&lt;/p&gt;

&lt;p&gt;Regression accounts for this in the following manner: If we want to predict $Y$, we can do so by predicting it from two measures—a function of $X$, and an error term $ε$. If our function of $X$ is linear, we get the following form:&lt;/p&gt;

&lt;p&gt;$Y_i=β_0+β_1X_i+ε_i$&lt;/p&gt;

&lt;p&gt;where $β_1$ is basically the $m$ and $β_0$ is basically the $b$ in $y=mx+b$. The “spread” of data around the line is accounted for by $ε$.&lt;/p&gt;

&lt;p&gt;We tracking so far? Basically, once we have values for $β_0$ and $β_1$ and understand our error, we can regress $y$ on $x$ to understand their relation.&lt;/p&gt;

&lt;h2 id=&quot;ordinary-least-squares-estimation-theory&quot;&gt;Ordinary Least Squares Estimation Theory&lt;/h2&gt;

&lt;p&gt;The question then becomes: how do we estimate values for $β_0$ and $β_1$? Think back to our previous (data-ambiguous) example. If you wanted to draw a “line of best fit” through the data, you might come up with the line we drew previously:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/ordinary-least-squares/data2.png&quot; alt=&quot;data2&quot; /&gt;&lt;/p&gt;

&lt;p&gt;However, maybe your colleague disagrees, and says that a different line (orange) actually fits the data better:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/ordinary-least-squares/data4.png&quot; alt=&quot;data4&quot; title=&quot;They don&apos;t know what they&apos;re talking about.&quot; /&gt;&lt;/p&gt;

&lt;p&gt;We need to come up with some objective way to state that a line fits the data best. This is where Ordinary Least Squares Estimation comes in. One way to objectively state that a line fits “best” is to attempt to minimize the distance between each data point and the line of prediction. Recall that the distance is defined as the error:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/ordinary-least-squares/data5.png&quot; alt=&quot;data5&quot; /&gt;&lt;/p&gt;

&lt;p&gt;If our line fits reasonably well, we would expect the distance between the points and the line to be small. For a counterexample, a terrible line of prediction:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/ordinary-least-squares/data6.png&quot; alt=&quot;data6&quot; /&gt;&lt;/p&gt;

&lt;p&gt;See how big the distance (error) is? This is the basis of Ordinary Least Squares: the smaller the overall error, the better.&lt;/p&gt;

&lt;p&gt;Soooo… this is easy, right? Just minimize the total amount of error! Well, there’s one problem. If we have a large &lt;em&gt;positive&lt;/em&gt; distance between a point and the line, and we have a large &lt;em&gt;negative&lt;/em&gt; distance between another point and the line, they cancel out, and the overall error is 0. So, we instead &lt;em&gt;square&lt;/em&gt; the distances so that we can meaningfully compare them. If we can find values for $β_0$ and $β_1$ that minimize the squared error, we have our line.&lt;/p&gt;

&lt;h2 id=&quot;ordinary-least-squares-estimation-calculation&quot;&gt;Ordinary Least Squares Estimation Calculation&lt;/h2&gt;

&lt;p&gt;All the previous description of linear regression was assuming that we knew population values for our data. In reality, we always work with samples. So, we transform our previous population equation $Y_i=β_0+β_1X_i+ε_i$ to a sample-specific equation:&lt;/p&gt;

&lt;p&gt;$y=b_0+b_1x+e$.&lt;/p&gt;

&lt;p&gt;Well, we might notice something. In this equation, $y$ is the actual data point that we observe. But we also have access to another version of $y$: $ŷ$. This symbol represents the &lt;em&gt;predicted&lt;/em&gt; value of $y$ for a given data point—the value for $y$ we would expect without error. Thus, we also have the equation $ŷ=b_0+b_1x$ by definition.&lt;/p&gt;

&lt;p&gt;Yay! Now we can do some algebra.&lt;/p&gt;

&lt;p&gt;$y=b_0+b_1x+e$ and $ŷ=b_0+b_1x$. Thus,&lt;/p&gt;

&lt;p&gt;$y=ŷ+e$&lt;/p&gt;

&lt;p&gt;$e=y-ŷ$&lt;/p&gt;

&lt;p&gt;Well now! Recall that we’re trying to minimize the residuals for our line of prediction. Now we have a definition for them. According to our previous theory, we need to square this term to meaningfully compare observations:&lt;/p&gt;

&lt;p&gt;$e^2=(y-ŷ)^2$&lt;/p&gt;

&lt;p&gt;And all we need to do is sum these babies up and find the values for $b_0$ and $b_1$ that minimize that sum!&lt;/p&gt;

&lt;p&gt;$∑(y-ŷ)^2$&lt;/p&gt;

&lt;h2 id=&quot;example-ordinary-least-squares-estimation-with-three-data-points&quot;&gt;Example: Ordinary Least Squares Estimation with Three Data Points&lt;/h2&gt;

&lt;p&gt;Let’s illustrate this calculation with an example (which I came up with myself, so bear with me). Imagine that we have 3 data points, and we want to estimate the slope and intercept of a prediction line via our previously-discussed method. We’ll say these points are $(1,2)$, $(2,3)$, and $(4,6)$.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/ordinary-least-squares/data7.png&quot; alt=&quot;data7&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Recall our previous formulae: $y=b_0+b_1x+e$ and $ŷ=b_0+b_1x$. We want to solve for the values of $b_0$ and $b_1$ by minimizing the sum $∑(y-ŷ)^2$. So, let’s use some calculus.&lt;/p&gt;

&lt;p&gt;How can we find the local minimum of a function? By calculating its derivative and determining when the derivative is equal to $0$. Remember, the derivative of a function at a certain point represents the rate of change of the function at that point: in other words, the slope of the tangent line at that point. So, if we know that the slope of the tangent line is $0$ (and that we’re looking for a minimum), we know that the function is going from decreasing to increasing, and we have found a local minimum.&lt;/p&gt;

&lt;p&gt;So, let’s calculate the sum $∑(y-ŷ)^2$ and find its derivative!&lt;/p&gt;

&lt;p&gt;$∑(y-ŷ)^2$ will have 3 different components: one for each point. Let’s label these $(y_1-ŷ_1)^2$, $(y_2-ŷ_2)^2$, and $(y_3-ŷ_3)^2$ where $point 1 = (1,2)$, $point 2 = (2,3)$, and $point 1 = (4,6)$.&lt;/p&gt;

&lt;p&gt;Remember, $ŷ=b_0+b_1x$. So for each observation we can plug the actual point’s value of $y$ in for $y_i$ and plug in our variables of $b_0$ and $b_1$ for $ŷ_i$. This will let us get a function in terms of $b_0$ and $b_1$—the values we want to solve for.&lt;/p&gt;

&lt;p&gt;$(y_1-ŷ_1)^2=(2-(b_1+b_0))^2$&lt;/p&gt;

&lt;p&gt;$(y_2-ŷ_2)^2=(3-(2b_1+b_0))^2$&lt;/p&gt;

&lt;p&gt;$(y_3-ŷ_3)^2=(6-(4b_1+b_0))^2$&lt;/p&gt;

&lt;p&gt;In each equation, we put the x-value of the point alongside the variable $b_1$ because our predicted value $ŷ$ is a function of $b_0+b_1x$.&lt;/p&gt;

&lt;p&gt;Now, we mash all of these components together:&lt;/p&gt;

&lt;p&gt;$∑(y-ŷ)^2=(2-(b_1+b_0))^2+(3-(2b_1+b_0))^2+(6-(4b_1+b_0))^2$&lt;/p&gt;

&lt;p&gt;Next is finding the derivative. Let’s expand this equation so we’re not dealing with exponents around parentheses (I’m gonna gloss over a few intermediate steps).&lt;/p&gt;

&lt;p&gt;$∑(y-ŷ)^2=4-4b_1-4b_0+b_1^2+2b_1b_0+b_0^2+9-12b_1-6b_0+4b_1^2+4b_1b_0+b_0^2+36-48b_1-12b_0+16b_1^2+8b_1b_0+b_0$&lt;/p&gt;

&lt;p&gt;$∑(y-ŷ)^2=21b_1^2-64b_1+3b_0^2-22b_0+14b_1b_0+49$&lt;/p&gt;

&lt;p&gt;Awesome! Now we have a more manageable function with two variables. We want to find the minimum of this function with respect to both, so we take the partial derivative with respect to $b_1$ and $b_0$. (Again, skipping some intermediate steps.)&lt;/p&gt;

&lt;p&gt;$\frac{d}{db_1}=42b_1-64+14b_0$&lt;/p&gt;

&lt;p&gt;$\frac{d}{db_0}=6b_0-22+14b_1$&lt;/p&gt;

&lt;p&gt;Ok. Home stretch. Now we set these two equations equal to 0:&lt;/p&gt;

&lt;p&gt;$42b_1-64+14b_0=0$, and $6b_0-22+14b_1=0$. We can set up a system of equations:&lt;/p&gt;

&lt;p&gt;$42b_1+14b_0-64=0$&lt;/p&gt;

&lt;p&gt;$14b_1+6b_0-22=0$&lt;/p&gt;

&lt;p&gt;Which we can actually transform into a matrix by putting the constants on the right-side of the equation:&lt;/p&gt;

\[\begin{array}{cc|c}
  42 &amp;amp; 14 &amp;amp; 64 \\
  14 &amp;amp; 6 &amp;amp; 22 
\end{array}\]

&lt;p&gt;$\overset{3r_2}{\longrightarrow}$&lt;/p&gt;

\[\begin{array}{cc|c}
  42 &amp;amp; 14 &amp;amp; 64 \\
  42 &amp;amp; 18 &amp;amp; 66
\end{array}\]

&lt;p&gt;$\overset{r_2-r_1}{\longrightarrow}$&lt;/p&gt;

\[\begin{array}{cc|c}
  42 &amp;amp; 14 &amp;amp; 64 \\
  0 &amp;amp; 4 &amp;amp; 2
\end{array}\]

&lt;p&gt;So, the solution to the system of equations which sets both derivatives equal to 0 includes $b_0=\frac{1}{2}$. We can plug this value into the first equation to obtain:&lt;/p&gt;

&lt;p&gt;$42b_1+14(\frac{1}{2})-64=0$&lt;/p&gt;

&lt;p&gt;$42b_1+7-64=0$&lt;/p&gt;

&lt;p&gt;$42b_1-57=0$&lt;/p&gt;

&lt;p&gt;$42b_1=57$&lt;/p&gt;

&lt;p&gt;$b_1=\frac{57}{42}$&lt;/p&gt;

&lt;p&gt;$b_1=\frac{19}{14}$&lt;/p&gt;

&lt;p&gt;You get to see all the steps that time. Well, OK! Now we know the values for which our function $∑(y-ŷ)^2$ has a local minimum! So, if our calculations are correct, we should be able to plug those values back into the formula $y=b_0+b_1x+e$ to obtain a slope and intercept coefficient for our regression.&lt;/p&gt;

&lt;p&gt;Let’s try it:&lt;/p&gt;

&lt;p&gt;$y=\frac{1}{2}+\frac{19}{14}x$:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/ordinary-least-squares/data8.png&quot; alt=&quot;data8&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Hey that looks pretty good! Like, really good! I made the starting numbers up on the fly, so I’m happy to see that we got an equation with nice simple fractions and not complex decimals.&lt;/p&gt;

&lt;p&gt;The final part of calculating the linear regression is to get the variance of the residuals (I’m gonna &lt;em&gt;speed&lt;/em&gt; through this). Using our formula for variance we can see that $\hat{σ}^2=\frac{1}{n-p}∑\hat{ε_i}^2$ where we divide by $n-p$ as the number of degrees of freedom for $p$ parameters. We can recall that $\hat{ε_i}=y_i-ŷ_i$ by definition, and use our new slope and intercept to calculate $ŷ$ for each observation.&lt;/p&gt;

&lt;p&gt;For $(1,2)$:&lt;/p&gt;

&lt;p&gt;$y-ŷ=2-ŷ=2-(\frac{1}{2}+\frac{19}{14}(1))\approx0.143$&lt;/p&gt;

&lt;p&gt;For $(2,3)$:&lt;/p&gt;

&lt;p&gt;$y-ŷ=3-ŷ=3-(\frac{1}{2}+\frac{19}{14}(2))\approx-0.214$&lt;/p&gt;

&lt;p&gt;For $(4,6)$:&lt;/p&gt;

&lt;p&gt;$y-ŷ=6-ŷ=6-(\frac{1}{2}+\frac{19}{14}(4))\approx0.071$&lt;/p&gt;

&lt;p&gt;We can sum the squares of these $3$ values: $(.143)^{2}+(-.214)^{2}+(.071)^{2}\approx0.071$.&lt;/p&gt;

&lt;p&gt;Finally, we divide this value by $n-p=3-2=1$. Dividing by $1$ does not change the number, and so our $\hat{σ}^2=0.071$.&lt;/p&gt;

&lt;p&gt;So, all together, we have the following formula for our linear model: $y=\frac{1}{2}+\frac{19}{14}x+e$ with $\hat{σ}^2=0.071$.&lt;/p&gt;

&lt;h2 id=&quot;verification&quot;&gt;Verification&lt;/h2&gt;

&lt;p&gt;Let’s see how we did. Here is code which runs it in R:&lt;/p&gt;

&lt;p&gt;&lt;code class=&quot;language-plaintext highlighter-rouge&quot;&gt;x &amp;lt;- c(1, 2, 4)&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;&lt;code class=&quot;language-plaintext highlighter-rouge&quot;&gt;y &amp;lt;- c(2, 3, 6)&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;&lt;code class=&quot;language-plaintext highlighter-rouge&quot;&gt;d &amp;lt;- lm(y~x)&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;&lt;code class=&quot;language-plaintext highlighter-rouge&quot;&gt;summary(d)&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;&lt;code class=&quot;language-plaintext highlighter-rouge&quot;&gt;(summary(d)$sigma)**2&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;And here is the output:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/ordinary-least-squares/data9.png&quot; alt=&quot;data9&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Holy mackerel. That looks great. Notice that under “Coefficients” we can see the estimate for the intercept at $0.5000$, which is exactly what we predicted. Similarly, the estimate for the slope is $1.3571$, which is approx. $\frac{19}{14}$. Spot on.&lt;/p&gt;

&lt;p&gt;Also, notice the residuals: $0.14$, $-0.21$, $0.07$. Do those numbers look familiar? Finally, we can observe the &lt;code class=&quot;language-plaintext highlighter-rouge&quot;&gt;sigma**2&lt;/code&gt; value at the end, and notice that it is incredibly close to the value we predicted of $0.071$.&lt;/p&gt;

&lt;p&gt;WE DID IT!!!!!!!!&lt;/p&gt;

&lt;h2 id=&quot;conclusion&quot;&gt;Conclusion&lt;/h2&gt;

&lt;p&gt;So there you have it. How to hand calculate ordinary least squares estimation in a linear regression. Much like all my blog posts, this ran on &lt;em&gt;far&lt;/em&gt; longer than I originally meant it to. Congrats if you stuck around till the end. For the next installment of Grad School Scramble, I’ll try to keep it more theoretical and less concrete. I don’t like wrangling $\LaTeX$.&lt;/p&gt;
</description>
        <pubDate>Wed, 26 Feb 2025 00:00:00 +0000</pubDate>
        <link>https://keanehauck.com/quantitative/methodology/statistics/data/science/2025/02/26/ordinary-least-squares/</link>
        <guid isPermaLink="true">https://keanehauck.com/quantitative/methodology/statistics/data/science/2025/02/26/ordinary-least-squares/</guid>
      </item>
    
      <item>
        <title>College, in Retrospect</title>
        <description>&lt;p&gt;&lt;img src=&quot;/images/posts/graduation/jumping.jpg&quot; alt=&quot;Grad Pic&quot; title=&quot;Woohoo!&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Boomer! I recently graduated from the University of Oklahoma with my B.S. in Psychology with Honors, &lt;em&gt;Summa Cum Laude&lt;/em&gt;, with Minors in Computer Science and Music, and a certificate in Applied Statistics. Apparently, the longer it takes you to say your degree and all relevant qualifications, the more impressive it is, so that’s why I get to look really pretentious whenever I list it on an application. (No fair! Multidisciplinary Studies: Concentration in Public Promotion &amp;amp; Social Perception majors have &lt;em&gt;such&lt;/em&gt; an advantage!)&lt;/p&gt;

&lt;p&gt;As I recover from my typical post-semester mental collapse (read: playing the Pokémon video games for 11hrs/day all week), I thought it would be good to reflect on the different stages of my college journey and consider the events that led me to the place I’m at. Buckle up.&lt;/p&gt;

&lt;h3 id=&quot;the-beginnings&quot;&gt;The beginnings&lt;/h3&gt;

&lt;p&gt;What drew me to OU in the first place? Basically, money. The first big thing to realize about me is that I was not always the shining exemplary student that I obviously am today. In high school, I didn’t place much weight on the college admissions process. I didn’t totally understand the gravitas of accepting an offer, and I didn’t have a comprehensive grasp of the full cost of tuition, loans, and living fees. So, I took it easy. I tossed my hat in the ring with big name schools like Stanford and MIT (wait-listed, technically), but didn’t try extremely hard to get in. I didn’t even tour anywhere - other than briefly visiting OU for a piano audition. I wanted to both be a therapist and continue playing, so my listed degrees were psychology and piano.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/graduation/piano.png&quot; alt=&quot;Piano&quot; title=&quot;Practicing hard&quot; /&gt;&lt;/p&gt;

&lt;p&gt;I got an offer from OU, but I kept dragging my feet about accepting. The longer I put it off, the sweeter the deal got. They kept eliminating various costs like the application fee or first-semester fees, so I just kept waiting to see how much money I could save by procrastinating. Finally, the day that responses were due, I formally accepted the offer from OU due to the National Merit scholarship package and acceptance to their piano program.&lt;/p&gt;

&lt;h3 id=&quot;the-death-of-the-piano-dream&quot;&gt;The death of the piano dream&lt;/h3&gt;

&lt;p&gt;To be fair, I technically lasted more than 1 semester as a piano major. After my first juries, my teacher &lt;a href=&quot;https://www.ou.edu/finearts/music/faculty-staff/john-patrick-murphy&quot;&gt;Dr. John Patrick Murphy&lt;/a&gt; sat me down and asked if I fully understood the standards for daily practice expected of piano majors throughout college (Spoiler: I did not). By your junior/senior year, around 4 hours/day piano practice was expected. Some of my friends (read: &lt;a href=&quot;https://www.linkedin.com/in/aidan-sudler-8b57aa204/&quot;&gt;Aidan Sudler&lt;/a&gt;) may have been talented enough to exist in such an environment, but I who was struggling to even practice 1hr/day was not. So, I decided to relegate the study of music to a vocal minor, and focus on psychology.&lt;/p&gt;

&lt;p&gt;The first year of college also held many non-academic developments that were incredibly impactful. Notably, I discovered a wealth of social connections and amazing friends that uplifted and enriched my freshman year. I started college in 2020: right in the peak of COVID. During this time, most classes were online, and many in-person events were shut down. Luckily, I had groups such as:&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;The guys in my dorm hall (Couch 6W)&lt;/li&gt;
  &lt;li&gt;The girls on the other side of the dorm floor&lt;/li&gt;
  &lt;li&gt;The David L. Boren dorm people&lt;/li&gt;
  &lt;li&gt;The guys in Sigma Alpha Mu (Go SAMMY)&lt;/li&gt;
  &lt;li&gt;The other music students&lt;/li&gt;
  &lt;li&gt;The people in Campus Activities Council&lt;/li&gt;
  &lt;li&gt;The people who worked in the Gender + Equality Center&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;All of these groups were incredibly influential in shaping my first-year college development. Socially, many of the first groups listed helped me feel welcomed and accepted on campus. Professionally, my interaction with the last two groups paved the way for involvement that would occur my sophomore/junior year.&lt;/p&gt;

&lt;p&gt;Many of the first things I did on campus were either for or concerned with the frat &lt;a href=&quot;https://sam.org/&quot;&gt;Sigma Alpha Mu&lt;/a&gt;. As a freshman, I was adamant about not joining a fraternity. I did not have a positive perception of frat guys, and didn’t jive with the flavor of masculinity expressed in many greek spaces. However, I probably watched Monsters University too many times, and eventually became curious about greek life. What eventually convinced me to join was the fact that a new fraternity, Sigma Alpha Mu (or Sammy, for short) was getting founded on campus the year that I was rushing. The representative from SAM headquarters who was founding the chapter at OU marketed it as a frat for non-frat-guys, where the founding fathers would get to play a role in shaping a culture in the fraternity that was different from many other houses. I figured this sounded cool, so I joined as a founding father.&lt;/p&gt;

&lt;blockquote&gt;
  &lt;p&gt;(Slight post interruption! I’m actually writing this next part from a coffee shop in downtown Budapest while sipping the most delicious espresso. Everything prior was written in the London Heathrow airport during a layover. Ok, travel update done - back to my schooling)&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;One of my favorite things Sammy did together was OU Scandals. Scandals is the name for OU’s yearly student-produced musical production (no, the name Scandals didn’t result from all of the drama that theatre kids love to make). Each year, various groups (typically fraternities and sororities) self-produce a short, 15-minute musical which is then performed at a large showcase. The cast of each show consists of a pairing of two specific groups (e.g., the sorority Gamma Phi Beta and the fraternity Pi Kappa Alpha), and the directors and leads are individual members from each group. When Sammy did Scandals, we paired with OU Class Council - a student government organization that performs some function I still don’t totally understand. I was both one of the artistic directors and one of the leads for our show, so I was tasked with writing some of the music and performing the role of the villain (my favorite song was an evil rendition of “You’re Welcome” from Moana). Overall, it was an incredibly fun experience, and I met some great people during the production.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/graduation/sam.jpg&quot; alt=&quot;SAM&quot; title=&quot;SAMMYYY&quot; /&gt;&lt;/p&gt;

&lt;p&gt;That year, I also joined the leadership team of OUr Mental Health, a club dedicated to raising awareness and providing resources for mental health on campus. One of my personal highlights was facilitating a stress-relieving meditation night for a group of students during dead week. One girl fell asleep, so I consider my meditation facilitation a success.&lt;/p&gt;

&lt;blockquote&gt;
  &lt;p&gt;2020-2021 Notable Classes/Professors:&lt;/p&gt;
  &lt;ul&gt;
    &lt;li&gt;Piano: Fresh/Soph with Dr. John Patrick Murphy&lt;/li&gt;
    &lt;li&gt;Elements of Psychology with Dr. Jenel Cavazos&lt;/li&gt;
    &lt;li&gt;Aural Skills with Dr. Sarah Ellis&lt;/li&gt;
    &lt;li&gt;Gateway: Fine Arts with Dr. Johnnie-Margaret McConnell&lt;/li&gt;
    &lt;li&gt;Music in Culture with Dr. Zoe Sherinian&lt;/li&gt;
  &lt;/ul&gt;
&lt;/blockquote&gt;

&lt;h3 id=&quot;the-campus-life-explosion&quot;&gt;The campus life explosion&lt;/h3&gt;

&lt;p&gt;I still don’t know exactly what got into me, but the next two years of my college career were marked by my full immersion into any form of campus involvement I could find. There was a sort of snowball effect: the more things I got involved in, the more opportunities I would hear about. One of the things I love about OU is that many of the programs and events on campus are fully student-led, so there are nearly endless ways to get plugged in.&lt;/p&gt;

&lt;p&gt;One of the major organizations I got involved with this year was the Gender + Equality Center (GEC). Holy moly, these people were amazing. All incoming freshmen at OU are required to take a federally-mandated consent training, and I had a great experience with mine. I was a bit of a snot-nosed incoming student, and I went into the training expecting to poke fun at it and not take it too seriously. However, the peer-educator moderator (Abby Tow, who later became one of my good friends in the GEC) was phenomenal, and I learned a lot about consent and how it is communicated. So, the next year I interviewed to become a peer educator for the GEC and was eventually accepted.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/graduation/consent.jpg&quot; alt=&quot;Consent Convo&quot; title=&quot;Alongside amazing co-presenters&quot; /&gt;&lt;/p&gt;

&lt;p&gt;If I had to pick one organization that helped me develop professionally during my time at OU, it would be the GEC. All of the people I worked with and all of the bosses I had were amazing and really cared about their employees (Bliss Brown, Jerry Lessley, Quan Phan, Crissy Brown, Amber May, Erin Simpson, Ann Schaefer, to name a few). The team of peer educators was full of motivated students and I formed some lasting friendships. Overall, I can’t recommend it enough.&lt;/p&gt;

&lt;p&gt;Alongside the GEC, I also got involved with Campus Activities Council (CAC). Remember earlier how I mentioned that most of the organizations at OU are student-led? CAC is the main hub for many of the major events, like OU Film Series and OU Concert Series (OU Scandals is also based out of the CAC). During the fall semester, I was an exec member of CAC College Bowl, an organization that hosted a large trivia tournament each year. I got involved with College Bowl because many of the girls I was friends with in the dorms were members, and it was a blast. Much like the GEC, the team of College Bowl exec members was super fun and close-knit, and the process of setting up and hosting the tournament was full of fun ideas and events.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/graduation/elvis.jpg&quot; alt=&quot;Elvis&quot; title=&quot;Q: What was Elvis&apos; favorite sandwich?&quot; /&gt;&lt;/p&gt;

&lt;p&gt;In the spring, I was a vice chair for CAC Fall Family Weekend (FFW): the organization that hosts the yearly family visitation weekend at OU. I was in charge of reaching out to various other organizations to host events during the weekend, but the fantastic FFW chair Sarah Stringfield honestly did most of the work for me. It was fun, but I definitely gained a reputation for slacking off. 8/10.&lt;/p&gt;

&lt;p&gt;Finally, I also was a Teaching Assistant to Dr. Jenel Cavazos for Elements of Psychology 1113 and to Stephanie Oakes for First-Year Foundations of Psychology 1002. Being a TA was great fun, and helped me gain experience in the classroom that was not purely being a student. Highlight of TA’ing: when I planned to host a end-of-the-year party for my group in Elements of Psychology, and was told that if I actually hosted an event with alcohol for undergraduates I would get fired and barred from TA’ing again. #TeamMilgramForLyfe&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/graduation/ffw.jpg&quot; alt=&quot;FFW&quot; title=&quot;I was a bit of a &apos;class clown,&apos; I guess you could say.&quot; /&gt;&lt;/p&gt;

&lt;blockquote&gt;
  &lt;p&gt;2021-2022 Notable Classes/Professors:&lt;/p&gt;
  &lt;ul&gt;
    &lt;li&gt;Men’s Glee Club with Dr. David Howard&lt;/li&gt;
    &lt;li&gt;Research Methods with Dr. Scott Gronlund&lt;/li&gt;
    &lt;li&gt;Men and Masculinity with Dr. Rodney Bates&lt;/li&gt;
    &lt;li&gt;Introduction to Personality with Dr. Eugenia Fuenzalida&lt;/li&gt;
    &lt;li&gt;Beginning Spanish with Christian Doug&lt;/li&gt;
  &lt;/ul&gt;
&lt;/blockquote&gt;

&lt;h3 id=&quot;research-rumblings&quot;&gt;Research rumblings&lt;/h3&gt;

&lt;p&gt;Uh oh… you hear that? It’s the sound of the typical upperclassman realization that graduation is now sooner than later, and that you should probably figure out a semblance of a plan. Junior year is when I started reevaluating whether or not I actually wanted to be a therapist. It was sort of my “mid-life crisis” of college; I took cognitive psych and wanted to be a cognitive psychologist, I took virtue ethics and wanted to start a philosophy degree, I started working for my first research lab and REALLY wanted to do research. At the end of the day, this is the year that I first took Psych Test &amp;amp; Measurements with Dr. Robert Terry, and you only need read my graduate school application essays to know how pivotal this class was in getting me excited about quant psych.&lt;/p&gt;

&lt;p&gt;My first research experience was working with Dr. Scott Gronlund in the OU Memory Lab. It was very fun! I was in charge of running participants for an experiment that aimed to determine how object memories were stored in the brain. Participants were given a list of items (objects like chairs, tables, lamps, etc.) to memorize, and they were asked to recall if certain inages (other objects) had been present in the list. The catch is that the cue images were either exact replicas of the object or the same TYPE of object (e.g., a lamp) with subtle variations. The theory is that if there was no significant difference in levels of observed change then people were storing objects conceptually rather than as a flashbulb image memory. Cool stuff.&lt;/p&gt;

&lt;p&gt;My second research experience, during the spring semester, was working under Dr. Edward Cokely in the OU Decision Psychology Lab. The content was very cool, but I didn’t once interact with Dr. Cokely himself. He was completely hands-off with the undergraduates in the lab and tasked his grad students with managing us. I thought that was a little bit over-the-top, and I was sad that I never got to ask him questions about the material. The framework for this lab was that there exists a measure of intelligence–numeracy–that is more predictive of good decision making than any other measure of intelligence. This lab was an interesting experience. At the time, I didn’t have the vocabulary to express my questions, but I wresled with the concept of divergent validity with regard to numeracy and other measured forms of intelligence. I was also constantly curious about the methodology for how they determined that numeracy training makes someone a better decision maker overall. In the end, the experience turned sour as I accidentally missed showing up in the lab for a participant during dead week and was asked to step down as an assistant. Shhh! Don’t tell the people that read my CV!&lt;/p&gt;

&lt;p&gt;This was the point at which I discovered quantitative psychology. During the spring of this year, I took Psych Test &amp;amp; Measurements with Dr. Robert Terry, which introduced me to the field. I spent a lot of time after class asking him questions I had from the other research I was involved in, and he eventually invited me to participate in OU’s Academic Success Center Research Team. I’ll get into specific quant stuff later, but this was my first foray into psych data science.&lt;/p&gt;

&lt;p&gt;In terms of campus involvement, I was still plugged in. One of my favorite organizations I was a part of was Camp Crimson, for which I served as a Camp Guide. Camp Crimson is OU’s largely student-led freshman orientation camp, responsible for introducing new students to campus and hosting events before the semester such as Class Kick-off. Camp Guides are the individuals who run around making sure everything is running smoothly. We also get to do fun stuff such as being the hype men for competitions and leading excursions to attractions in OKC.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/graduation/crimson.jpg&quot; alt=&quot;Camp&quot; title=&quot;Camp Crimson!&quot; /&gt;&lt;/p&gt;

&lt;p&gt;I continued my involvement in CAC and the GEC. At this point, I also started trying to make professional use of my involvement, so I volunteered to be a facilitator for peer support groups that met on campus to discuss stress and other facets of being a college student. It was very fulfilling. The ironic part about these groups, however, is that I honestly got as much use out of them as the students among whom I was supposed to be facilitating discussion. At this point in my college career, the pressure of everything I was involved in really started to affect me stress-wise, and I made the conscious decision to dial down the involvement for my senior year.&lt;/p&gt;

&lt;blockquote&gt;
  &lt;p&gt;2022-2023 Notable Classes/Professors:&lt;/p&gt;
  &lt;ul&gt;
    &lt;li&gt;Biblical Literature with Dr. Jill Hicks-Keeton&lt;/li&gt;
    &lt;li&gt;Sociology of Family with Matthew Bejar&lt;/li&gt;
    &lt;li&gt;Virtue Ethics with Dr. Adam Green&lt;/li&gt;
    &lt;li&gt;Psych Test &amp;amp; Measurements with Dr. Robert Terry&lt;/li&gt;
    &lt;li&gt;Voice with Matthew Corcoran&lt;/li&gt;
  &lt;/ul&gt;
&lt;/blockquote&gt;

&lt;h3 id=&quot;honing-in&quot;&gt;Honing in&lt;/h3&gt;

&lt;p&gt;Senior year is where I really started to focus on pursuing the aspirations that had developed in the last few semesters. The spark having been lit by Dr. Terry, the field of quantitative psychology was calling to me like no subject had ever before. As I conducted more research with my professors, I realized how much I might enjoy doing it as a career. Being able to watch them operate in the academic environment slowly became something that embodied where I wanted to be someday.&lt;/p&gt;

&lt;p&gt;One funny story: As I started my senior year, I decided to try and attain a computer science minor during my last year (and a half) of college. I had been missing math and coding (shockingly) and I wanted to fold it back into my education. The funny thing about this ambition is that I was not even close to having the required math credits to start the CS track. CS starts with Calculus as a co-requisite for many coding classes, and the pre-requisites for Calc 1 are trig, geometry, alg, etc. The problem was: the psychology degree at OU only requires VERY basic math classes, and the only one I had taken was a “functions for the social sciences” class my first year that probably didn’t even count for credit in any program other than psychology. I had only gotten a 2 on the AP Calc BC exam in high school, which didn’t count for any credits either. So, I was stuck. I started emailing around, and the only way I was going to be able to take calc that semester was if I took the incoming student ALEKS math placement exam, which is the test given to freshmen who want to test out of intro-level math classes. The culmination of all of this context is that I ended up having to cram study the ENTIRE math sequence–alg, geometry, trig, and pre-calc–in a matter of 3 days to prepare for a placement exam. I am proud to announce that I successfully ended up relearning the entire high school math curriculum in 72 hours, and scored into the highest category of placement. So, I was able to start work on the CS minor.&lt;/p&gt;

&lt;p&gt;I did Camp Crimson during the summer (OK WHO? OKU). Other than that, I wasn’t active in any organization on campus other than the GEC. This allowed me to focus heavily on research.&lt;/p&gt;

&lt;p&gt;The bulk of my work focused on my honors thesis. At OU, in order to graduate with honors you must complete a thesis during your senior year over a topic of your choosing. I wanted my thesis to be in the form of a published research paper, so I reached out to and worked with Dr. Jenel Cavazos to plan, organize, and complete a full paper throughout the course of the year. The paper focused on AI use among college students; more information about it can be found &lt;a href=&quot;/previousresearch/chatgpt-goes-to-college&quot;&gt;here&lt;/a&gt;. More than anything, this project allowed me to realize that I wanted to continue my education into a PhD. The process of conducting review, developing a study, assessing results, and authoring a paper made me feel like I was at home, academically. I loved it–and learning &lt;em&gt;that exact experience&lt;/em&gt; constituted a majority of grad school was enough for me to fully commit to the idea.&lt;/p&gt;

&lt;p&gt;Alongside my senior thesis, I also conducted work on student retention data with Dr. Terry. I was tasked with assessing methods of handling missing data such as FIML and Multiple Imputations with Chained Equations. I recognize that the length of this post is starting to approach the upper limit of what I was hoping for, so I’ll forego explaining all the research specifics. Information about any particular project can be found &lt;a href=&quot;/research&quot;&gt;here&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/graduation/swpa.jpg&quot; alt=&quot;SWPA&quot; title=&quot;@ SWPA&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Overall, this year really was just a bunch of academics. It was SO, SO fun, though. More than anything, it opened my eyes to the opportunities that are available to individuals pursuing research. Some of my favorite memories from the year are from presenting my paper at conferences like SWPA or OPS (I won ‘Best Undergraduate Paper’ at one!). It helped me solidify my certainty that I was on a good path for me.&lt;/p&gt;

&lt;blockquote&gt;
  &lt;p&gt;2023-2024 Notable Classes/Professors:&lt;/p&gt;
  &lt;ul&gt;
    &lt;li&gt;Great Books of Western Civ with Dr. Robert Lifset&lt;/li&gt;
    &lt;li&gt;Psychology Capstone with Dr. Robert Terry&lt;/li&gt;
    &lt;li&gt;Discrete Structures with Dr. Lauren Grimley&lt;/li&gt;
    &lt;li&gt;Stat Models of Tests Scores with Dr. Robert Terry&lt;/li&gt;
    &lt;li&gt;Computer Organization with Dr. Mohammed Atiquzzaman&lt;/li&gt;
  &lt;/ul&gt;
&lt;/blockquote&gt;

&lt;h3 id=&quot;the-home-stretch&quot;&gt;The home stretch&lt;/h3&gt;

&lt;p&gt;After all of the previous exposition, my super-senior semester didn’t really bring any new revelations. It was much a continuation of the previous developments, both academically and personally. My victory lap.&lt;/p&gt;

&lt;p&gt;I realized this semester that I technically qualified for a certificate in Applied Statistics (which is basically a minor? Still not sure exactly what it means) due to all of the math/stats courses I had taken. One more qualification on my degree! Yay!&lt;/p&gt;

&lt;p&gt;Alongside continuing some of the research I had previously done such as my work with Dr. Cavazos and work with Dr. Terry, I got involved with some oddball projects you might not expect. I started working with Dr. Lauren Grimley, one of my math professors, on math research involving the game of &lt;a href=&quot;https://en.wikipedia.org/wiki/Peg_solitaire&quot;&gt;peg solitaire&lt;/a&gt;. The way I got involved on this project was as follows: I had Dr. Grimley for the class Discrete Structures the previous year, and I thought the class seemed really cool. I needed an honors credit still to graduate with honors, so I asked her if I could do an extra paper/project for the class to qualify it for honors credit. She seemed receptive to the idea, so we met to discuss potential ideas. After a while, she introduced me to this side project of hers looking at the solvability of various classic board games: peg solitaire being one of them. Through the project, we analyzed peg solitaire in 3 colors: the specifics of which can be found on my research tab. This was another project that resulted in a paper, which I am incredibly proud of. (One funny story: the Discrete class that Dr. Grimley taught was so difficult that the students staged a full revolution. The dean was so overwhelmed with complaints and with students visiting his office that he had to announce that something would be done and that they would investigate the class. For the record, I am completely on Dr. Grimley’s side, and think that the students were just complaining too much.)&lt;/p&gt;

&lt;p&gt;Finally, I also worked for Dr. Eric Day to help modify a video game that he used for a lab assessing exploration and exploitation patterns in workplace learning. Again, I don’t want to just blandly describe all the various projects I was working on, but if you want to try a version of the game for yourself you can visit the following link: &lt;a href=&quot;https://deeper-paths.firebaseapp.com/&quot;&gt;https://deeper-paths.firebaseapp.com/&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;I could not have been happier with the classes I took my final semester. I took Linear Algebra (incredibly engaging), and also TA’d for Advanced Undergraduate Statistics taught by Catherine Bain (one of the graduate students who helped with my honors thesis, among other things). Alongside those classes, I also took Chinese Religions with Dr. Geoffrey Goble. Oh boy. What an adventure. This class was structured such that the second half of the semester was purely dedicated to a massive roleplaying game that took place in-class. The main focus of the class was assessing the impact of the religions of Confucianism, Daoism, and Buddhism throughout Chinese history, and the game featured factions of representatives reenacting the Huichang Persecution of Buddhism court debates circa 841. I was the leader of the Confucian faction - wise, arrogant, and traditional. We argued for the restriction and abolishment of Buddhism throughout the Tang Dynasty, as such foreign ideas were obviously dangerous to the peace and stability of Our Kingdom. Not to brag too much, but our faction won the game, and I achieved the title of “winning character.” Long live Niu Sengru.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/graduation/sengru.jpg&quot; alt=&quot;Sengru&quot; title=&quot;The champion&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Overall, it was a great final semester to close the book on my college journey.&lt;/p&gt;

&lt;blockquote&gt;
  &lt;p&gt;Fall 2024 Notable Classes/Professors:&lt;/p&gt;
  &lt;ul&gt;
    &lt;li&gt;Linear Algebra with Dr. Elizabeth Kelley&lt;/li&gt;
    &lt;li&gt;Advanced Undergraduate Statistics with Catherine Bain&lt;/li&gt;
    &lt;li&gt;Chinese Religions with Dr. Geoffrey Goble&lt;/li&gt;
  &lt;/ul&gt;
&lt;/blockquote&gt;

&lt;h3 id=&quot;now-what&quot;&gt;Now what?&lt;/h3&gt;

&lt;p&gt;Well, dang. That’s my college experience.&lt;/p&gt;

&lt;p&gt;I’m not certain what the future will hold. I’m currently in the final chapter of the grad school application process: Next week I start flying out to schools that have invited me to visit. I currently have next-to-no idea on the heuristic I should be using to select a program, so I’m hoping these trips will bring some much-needed clarity. (Plus, the schools all pay for my flights/lodging. Nice!)&lt;/p&gt;

&lt;p&gt;Would I have done anything differently throughout college, in retrospect? Probably. I would have loved to study abroad for a semester—the price tag always kept me away. I think it would have been worth it, though. Other than that, I’m honestly very happy with how it went. OU was a great school—I met some great people. I grew a lot, and I found a passion in quantitative psychology. Not sure you can ask for more than that.&lt;/p&gt;

&lt;p&gt;Time to see what grad school is like!&lt;/p&gt;
</description>
        <pubDate>Mon, 30 Dec 2024 00:00:00 +0000</pubDate>
        <link>https://keanehauck.com/graduation/college/academics/life/2024/12/30/graduation/</link>
        <guid isPermaLink="true">https://keanehauck.com/graduation/college/academics/life/2024/12/30/graduation/</guid>
      </item>
    
      <item>
        <title>An Intuitive Method of Interpreting Quantile-Quantile Plots</title>
        <description>&lt;p&gt;This semester, I have the honor of being a teaching assistant for Advanced Undergraduate Statistics. Despite TA’ing for the course, I have never actually taken it myself. I had worked with the instructor (the fantastic Catherine Bain) on prior research, and she figured that I was already knowledgeable enough in the material to help teach it (Oh, if only she knew how inept I actually am).&lt;/p&gt;

&lt;p&gt;One of the consequences of her gracious assumption is that there is a lot of material in class that I have attemped to learn through guerrilla tactics, but have not had &lt;em&gt;formal&lt;/em&gt; training in. As such, we often cover material in class that fills in some basic gap of knowledge that I hadn’t realized existed.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/qq-plots/plural.png&quot; alt=&quot;Plural&quot; title=&quot;An example&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Other times, we cover material in a more structured manner than I had previously experienced, which is always helpful. Overall, it’s been a great way to more fully immerse myself in basic statistical methodology.&lt;/p&gt;

&lt;p&gt;During our last class period, we completed a lab where the students had to work with a dataset and clean it of missing values and outliers. As is common, testing for outliers involved checking the data for normality. For the purposes of undergraduate statistics, this consisted of generating a histogram of the data and visually assessing it.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/qq-plots/histogram.png&quot; alt=&quot;Histogram&quot; title=&quot;Looking good, R.&quot; /&gt;&lt;/p&gt;

&lt;p&gt;The data were the individual means of a 6-item 5-point scale measuring general parenting behaviors from 1 = coercive parenting to 5 = positive parenting. As we can see, the data seem to have a negative skew. While we discussed it in class, I wanted to introduce some of the students to a more formal way of assessing normality: Quantile-Quantile plots.&lt;/p&gt;

&lt;blockquote&gt;
  &lt;h2 id=&quot;keanes-quick-and-quality-quantile-quantile-plot-guide&quot;&gt;Keane’s quick-and-quality Quantile-Quantile plot guide&lt;/h2&gt;
  &lt;p&gt;A Quantile-Quantile plot (Q-Q plot, for short) is a relatively quick way to assess the normality of a distribution. A Q-Q plot is a scatterplot which plots two sets of quantiles against each other: theoretical vs. sample.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;blockquote&gt;
  &lt;p&gt;What is a quantile? Well, a quantile is basically another way to say “percentile.” When you plot quantiles, you are plotting the points at which a certain amount of data fall below (50%, 90%, etc).&lt;/p&gt;
&lt;/blockquote&gt;

&lt;blockquote&gt;
  &lt;p&gt;A z-score normalizes data by &lt;a href=&quot;https://www.statisticshowto.com/probability-and-statistics/z-score/&quot;&gt;expressing its distance from the mean in terms of standard deviations&lt;/a&gt;. 
When we plot quantiles, we plot them in terms of their z-score. As such, the x-axis of a Q-Q plot often ranges from around -3 to +3, which gives us a range of data of about 3 standard deviations from the mean. 
The y-axis of a Q-Q plot is often the numerical value of the data.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;blockquote&gt;
  &lt;p&gt;When we look at a Q-Q plot (shown below), it shows us the &lt;em&gt;expected&lt;/em&gt; distribution of quantiles (if our data were normal) plotted against our &lt;em&gt;actual&lt;/em&gt; sample of data. If our data were perfectly normal, they would follow the expected distribution perfectly linearly. The blue line on the plot below displays this expected relationship.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;blockquote&gt;
  &lt;p&gt;From this, we can evaluate any deviation from the expected line to mean that our data deviates from normality.
If this is confusing, don’t worry — I’m writing this because it confused me too.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p align=&quot;center&quot;&gt;
  &lt;img src=&quot;/images/posts/qq-plots/exampleqq.png&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;As we can tell from our quick-and-easy Q-Q plot guide, Q-Q plots are useful in assessing the normality of data. Prior to yesterday, this was about the extent of my knowledge of them. However, as I was thinking about them in class, I realized that I didn’t fully understand &lt;em&gt;why&lt;/em&gt; we interpreted them the way we do.&lt;/p&gt;

&lt;p&gt;For example, by running a Q-Q plot on the data previously observed (in the histogram), we obtain the following plot:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/qq-plots/qqplot.png&quot; alt=&quot;QQ Plot&quot; title=&quot;curvy curvy&quot; /&gt;&lt;/p&gt;

&lt;p&gt;As we can see, our data deviate from the expected distribution. Notably, our scatterplot shows a curve, which (apparently) means that our data are skewed. But, why?&lt;/p&gt;

&lt;p&gt;I realized that I had no real knowledge of the theoretical basis for Q-Q plot interpretation. Part of this problem originates from the fact that online materials for learning about Q-Q plots &lt;em&gt;suck&lt;/em&gt;. Most online sources do a great job of explaining what a Q-Q plot is, but fail to delve into proper interpretation. For example, &lt;a href=&quot;https://www.learningtree.com/blog/interpret-q-q-plot/&quot;&gt;this website&lt;/a&gt; explains very well what they are, but simply lists the different possibilities for Q-Q plot outcomes. Even UVA (a school I’m applying to) on &lt;a href=&quot;https://library.virginia.edu/data/articles/understanding-q-q-plots&quot;&gt;their website&lt;/a&gt; don’t explain why we see differences between curved plots and extremity-deviated plots.&lt;/p&gt;

&lt;p&gt;So, I asked Catherine, who ended up being a great help. One key part of intrepreting a Q-Q plot, I realized, is thinking about it in terms of individual data points. Take a closer look at the previous graph:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/qq-plots/qqhighlighted.png&quot; alt=&quot;QQ Plot&quot; title=&quot;this color scheme brought to you by Christmas&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Let’s think about what the data point circled in red is communicating. On the x-axis, we have our theorized z-scores for perfectly normal data. On the y-axis, we can see that the point’s value is around ~1.2. So, what the heck does this mean? Well, we observe that the point is distant from the line. Now consider what the line is telling us: the &lt;em&gt;expected&lt;/em&gt; relationship between our data and the normal distribution. So, we can imagine our line extending all the way past the left side of the graph, like so:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/qq-plots/qqextended.png&quot; alt=&quot;QQ Plot&quot; title=&quot;Not to scale&quot; /&gt;&lt;/p&gt;

&lt;p&gt;From this, we can see that if our line continued, we would expect to see the value of 1.2 at approximately a z-score of -5. Since our score of 1.2 occurs around a -3 z-score, it appears at a frequency about two standard deviations higher than expected. Thus, it (and all the scores under the curve around it) is observed more than we would expect from a typical normal distribution.&lt;/p&gt;

&lt;p&gt;How do we synthesize this with our understanding of normality? Well, this highlights another problem with how normal distributions and skew are communicated. Most diagrams of skewed data vs. normal data that I’ve seen look something like this:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/qq-plots/skew.png&quot; alt=&quot;Skew&quot; title=&quot;I&apos;m getting tired of MS paint&quot; /&gt;&lt;/p&gt;

&lt;p&gt;This isn’t &lt;em&gt;wrong&lt;/em&gt; per se, but it’s a little misleading. These distributions would no longer be matched on their respective means. For example, consider our previous example. If we take a data point at the extreme end of our negatively-skewed data (highlighted on the plot), we know that it is occurring at a frequency higher than we would expect from a typical normal curve. But wait, this appears inconsistent with the chart of skewed data we just saw! Doesn’t it look like that data point is actually &lt;em&gt;below&lt;/em&gt; where it should be on the normal distribution? Doesn’t that mean we are seeing it less than we should?&lt;/p&gt;

&lt;p&gt;Thus the problem. I propose conceptualizing skewed data in one of two different ways.&lt;/p&gt;

&lt;h3 id=&quot;the-first-way-drawing-distributions-accurately&quot;&gt;The first way: drawing distributions accurately&lt;/h3&gt;

&lt;p&gt;I made a slight modification on the previous graph which I think makes all the difference:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/qq-plots/method1.png&quot; alt=&quot;Method1&quot; title=&quot;Showing off MS paint&apos;s curvilinear function&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Now, if we think about a data point on the extreme end of the negative skew, we can see that it truly is falling above where we would expect it: something that our initial Q-Q plot was telling us. Most comparisons of skewed vs. normal data don’t show it to this detail, but it’s important to understand that the logistic slope of the left side of a negatively-skewed distribution (and right side of a positively-skewed distribution) decreases at a slower rate than that of a normal curve. To consider the ramifications for the right side of the graph, see the following example.&lt;/p&gt;

&lt;h3 id=&quot;the-second-way-thinking-about-where-the-deviation-from-normality-originates&quot;&gt;The second way: thinking about where the deviation from normality originates&lt;/h3&gt;

&lt;p&gt;Consider the following graph:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/images/posts/qq-plots/compare.png&quot; alt=&quot;Method2&quot; /&gt;&lt;/p&gt;

&lt;p&gt;In this graph, we can see a normal distribution compared to a negatively-skewed distribution. On the tail ends of the distributions (the parts we’re concerned with in many of our Q-Q plots), we can see exactly how the data differ from normality. On the left, we see the green (negative skew) sitting higher than the base normal curve. On the right, we can see how it falls below the normal curve: it is occurring less than we would expect.&lt;/p&gt;

&lt;h2 id=&quot;final-takeaways&quot;&gt;Final takeaways&lt;/h2&gt;

&lt;p&gt;That’s mostly it - I hope it’s been helpful. One additional point that needs to be made is that it’s important to remember that what it “means” for data to fall above/below the expected line in a Q-Q plot switches depending on whether the data is to the left or the right of the z-score of 0. When we considered our previous extreme data point, it was below the line on the left, meaning that it had a higher z-score than expected. However, if we consider an extreme data point on the positive end of the z-distribution, the opposite holds true: if a point is below the line, it means that it is more extreme than we would expect. This distinction comes from the fact that a z-score being “lower” means that it’s more extreme when it’s negative, but less extreme when it’s positive. Keep that in mind!&lt;/p&gt;

&lt;p&gt;The aforementioned point influences how we can interpret data that is leptokurtic or platykurtic. If our Q-Q plot looks like a squiggly s-shaped curve, it can mean either. We just have to think about the position of our data with regard to the expected line to see which quality of kurtosis the data have.&lt;/p&gt;

&lt;p&gt;Also, I do not claim to be an expert on Q-Q plots. I still have much to learn in my academic career. Still, I figured that I’d expound on some of the thoughts I’ve had concerning Q-Q plots these past few days.&lt;/p&gt;

&lt;p&gt;Time to go write some grad school application essays.&lt;/p&gt;
</description>
        <pubDate>Wed, 11 Sep 2024 00:00:00 +0000</pubDate>
        <link>https://keanehauck.com/quantitative/methodology/statistics/data/science/graphing/2024/09/11/qq-plots/</link>
        <guid isPermaLink="true">https://keanehauck.com/quantitative/methodology/statistics/data/science/graphing/2024/09/11/qq-plots/</guid>
      </item>
    
  </channel>
</rss>
