Tuesday Tip #23: Solve a medical mystery 🧪


Hi Reader,

Do you like a good mystery? I’ve got one for you below! 👇

But first, I thought it would be fun to include a “link of the week” at the top of each issue. I'll share something that I think is worth checking out!


🔗 Link of the week

Large language models, explained with a minimum of math and jargon

This is a LONG read — I spent 30+ minutes to read it carefully — but this is an excellent place to start if you want to understand the inner workings of LLMs such as ChatGPT. (Make sure to read the footnotes, and also the comments if you have time!)


👉 Tip #23: Solve a medical mystery with a confusion matrix

Throughout your life (and especially in the era of COVID! 🦠), you’ve probably taken a bunch of diagnostic tests to detect the presence or absence of a condition.

Have you ever wondered how much you should trust the test result? Since you’re reading a newsletter about Data Science, I’m going to guess the answer is “yes”!

Below, I’m going to pose a made-up testing scenario, and together we’ll figure out how to use a confusion matrix to solve this mystery!


Here’s the scenario:

Let’s pretend that there's a disease that is present in 2% of the population. (That's known as the "prevalence" of the disease.)

The disease can be detected by a nasal swab, and the test is 95% accurate for both positive and negative cases. (More formally, the test’s "True Positive Rate" and "True Negative Rate" are both 95%.)

You were tested for the disease and received a positive test result. What is the probability that you actually have the disease?


Why use a confusion matrix?

There are many methods you could use to solve this problem, but I’m going to create a confusion matrix since it helps me to visualize the problem.

If you’re not familiar with the confusion matrix, it’s a table that describes the performance of a classifier on a set of test data for which the true values are known.

In this case, the nasal swab is the classifier, and we will fill in the test data for which the true values are known based on the True Positive and True Negative Rates stated above.

(Side note: These rates would have been determined by the company that developed the nasal swab based on clinical trials they conducted prior to releasing the test.)


Step 1:

Let’s pretend there are 1000 people in the population. Because the stated prevalence of the disease is 2%, we believe that the disease is present in 2% of those 1000 people, or 20 people.

We can now begin to fill in the confusion matrix. There are two rows, “Actual YES” and “Actual NO”, meaning how many people actually have the disease:

  • The Actual YES row contains a total of 20 people.
  • The Actual NO row contains the remaining 980 people.

Step 2:

As stated above, when someone actually has the disease, the nasal swab will return a positive result 95% of the time.

In other words, it will predict YES for 95% of those people (which is the correct result), and it will predict NO for 5% of those people (which is the incorrect result).

  • The Actual YES / Predicted YES box contains 19 people, since 0.95 x 20 = 19. These are called True Positives (TP).
  • The Actual YES / Predicted NO box contains 1 person, since 0.05 x 20 = 1. These are called False Negatives (FN).

Step 3:

As stated above, when someone does not actually have the disease, the nasal swab will return a negative result 95% of the time.

In other words, it will predict NO for 95% of those people (which is the correct result), and it will predict YES for 5% of those people (which is the incorrect result).

  • The Actual NO / Predicted NO box contains 931 people, since 0.95 x 980 = 931. These are called True Negatives (TN).
  • The Actual NO / Predicted YES box contains 49 people, since 0.05 x 980 = 49. These are called False Positives (FP).

Now that we have filled in the confusion matrix, we can also calculate the column totals for Predicted NO (931 + 1 = 932) and Predicted YES (49 + 19 = 68).


Calculate the answer:

Finally, we’re ready to answer the question!

Given that you received a positive result from the nasal swab, what is the probability that you actually have the disease?

In our population of 1000 people, 68 tested positive (indicated by the Predicted YES column total). You represent one of those 68 people.

Of those 68 people, 19 actually have the disease (True Positives) and 49 do not (False Positives).

Thus if you test positive, the probability that you actually have the disease is 19 / 68 = 27.9%.


Wait, what???

27.9% might “feel” like the wrong answer, given that the test is 95% accurate. But here’s the key insight:

For a condition with low prevalence, most of the positive test results will be False Positives (even for a highly accurate test), since there are far more people taking the test who don’t have the condition than do have the condition.


Here’s my challenge to you:

One way to reduce the number of False Positives is to only test people who are symptomatic for the condition, since that significantly increases the prevalence within the population being tested.

To see an example of this, change the prevalence in the scenario above from 2% to 30% and then recreate the confusion matrix. In this new scenario, what is the probability that you actually have the disease, given that you tested positive?

Reply with your answer and I’ll let you know if you’re correct! ✅


Confusion matrix resources:

The confusion matrix is widely used in Machine Learning and many other fields. If you’re interested in learning more, I highly recommend reading my blog post:

🔗 Simple guide to confusion matrix terminology

I also created a 35-minute video that goes deeper into this topic:

🔗 Making sense of the confusion matrix

Finally, the Wikipedia page is also quite good:

🔗 Wikipedia: Confusion matrix

Keep in mind that confusion matrices are not always formatted the same way! For example, the Wikipedia page places the YES values instead of the NO values in the upper left, so it’s important to pay attention to the labels when interpreting a confusion matrix! 💡


If you enjoyed this week’s tip, please forward it to a friend! Takes only a few seconds, and it really helps me out! 🙌

See you next Tuesday!

- Kevin

P.S. You have a lot of measurements. Quite a few variables.

Did someone awesome forward you this email? Sign up here to receive data science tips every week!

Learn Data Science from Data School 📊

Join 25,000+ aspiring Data Scientists and receive Python & Data Science tips every Tuesday!

Read more from Learn Data Science from Data School 📊

Hi Reader, Next week, I’ll be offering a Black Friday sale on ALL of my courses. I’ll send you the details this Thursday! 🚨 👉 Tip #50: What is a "method" in Python? In Python, a method is a function that can be used on an object because of the object's type. For example, if you create a Python list, the "append" method can be used on that list. All lists have an "append" method simply because they are lists: If you create a Python string, the "upper" method can be used on that string simply...

Hi Reader, I appreciate everyone who has emailed to check on me and my family post-Helene! It has been more than 6 weeks since the hurricane, and most homes in Asheville (mine included) still don't have clean, running water. We're hopeful that water service will return within the next month. In the meantime, we're grateful for all of the aid agencies providing free bottled water, free meals, places to shower, and so much more. ❤️ Thanks for allowing me to share a bit of my personal life with...

Hi Reader, Regardless of whether you enrolled, thanks for sticking with me through the launch of my new course! 🚀 I've already started exploring topics for the next course... 😄 🔗 Link of the week git cheat sheet (PDF) A well-organized and highly readable cheat sheet from Julia Evans, the brilliant mind behind Wizard Zines! 👉 Tip #48: Three ways to set your environment variables in Python I was playing around with Mistral LLM this weekend (via LangChain in Python), and I needed to set an...