Circles · Session 1
Begins June 8, 2026
Enrollment is closed for Session 1. Current proposals remain available for reference.
Learning from Evidence: The Mathematics of Updating Beliefs
Probability Theory, Bayesian Statistics
Research Proposal
Learning from Evidence: The Mathematics of Updating Beliefs
Preceptor in Mathematics
Department of Mathematics
Harvard University
Tentative Tue/Thu 6:30-8:00 PM ET; final meeting times are subject to change based on overall group availability.
Code used: R
Technology needed: RStudio, a computer with internet access, and a working microphone/camera
Project overview
Surprisingly, the answer is often much smaller than one might expect; in fact, the probability may be so small that you most likely do not carry the disease at all. This scenario illustrates the importance of Bayes’ Theorem, one of the most fundamental results in probability theory.
In this project, students will develop a strong understanding of Bayes’ rule in both theoretical and practical settings. They will build probabilistic intuition and mathematical modeling skills by studying both the motivation behind the theorem and its formal mathematical statement. We will then transition to more complex scenarios involving medical diagnosis, machine learning classification, and statistical inference.
Students will also gain hands-on experience using R to analyze data, visualize distributions, and draw conclusions based on numerical computation and statistical summaries.
Background
For this project, we will use RStudio to write code in R. Before the first meeting, students should install this program or another environment for working with R. However, no prior coding experience is required; students will work with prewritten code and learn how to modify it for their own purposes.
Apart from this, no particular background knowledge is required.
Possible extension
Selected references
- Blitzstein, J., and Hwang, J. Introduction to Probability, CRC Press, 2019.
- Rice, J. Mathematical Statistics and Data Analysis, 3rd ed., Cengage Learning, 2006.
- James, G., Witten, D., Hastie, T., and Tibshirani, R. An Introduction to Statistical Learning: with Applications in R, Springer, 2021.
Learning from Uncertainty: A Bayesian View of Probability and Data
Probability, Bayesian Statistics, Uncertainty Quantification
Research Proposal
Learning from Uncertainty: A Bayesian View of Probability and Data
Ph.D. Candidate in Mathematics
Department of Mathematics
Florida State University
Tentative Mon/Wed 7:00-8:30 PM ET; final meeting times are subject to change based on overall group availability.
Code used: Python or spreadsheet-based simulation
Technology needed: Google Colab or Excel, a computer with internet access, and a working microphone/camera
Project overview
In this project, students will explore the Bayesian view of probability, where probability represents a degree of belief rather than a fixed long-run frequency. We will begin with intuitive examples such as coin flips, guessing unknown quantities, and predicting outcomes with incomplete information. Students will learn how initial beliefs (called priors) can be updated using observed data to form improved beliefs (called posteriors).
Through simulations and experiments, students will investigate how uncertainty changes as more data is collected, how prior assumptions influence conclusions, and how Bayesian reasoning differs from traditional deterministic thinking. Visual tools such as probability distributions, histograms, and simulation plots will be used throughout.
The project emphasizes intuition, experimentation, and explanation. By the end of the program, students will understand how Bayesian ideas help quantify uncertainty and support rational decision-making in science, data analysis, and everyday life.
Background
Students will engage with hands-on experiments and simulations using simple computational tools such as Excel or Python (no prior programming experience assumed). Emphasis will be placed on conceptual understanding, interpretation, and communication rather than technical formalism.
Possible extension
Selected references
- Martin O. Bayesian Analysis with Python. Birmingham, UK: Packt Publishing; 2016 Nov 25.
- Gelman, A., et al. Bayesian Data Analysis, 3rd ed., CRC Press, 2013.
When Can We Trust Simple Decisions?: Binary Classification with Imbalanced Data
Applied Mathematics, Mathematical Modeling, Data-Driven Decision Rules
Research Proposal
When Can We Trust Simple Decisions?: Binary Classification with Imbalanced Data
Ph.D. Candidate
Department of Mathematics and Statistics
Auburn University
Tentative Tue/Thu 6:00-7:30 PM CT; final meeting times are subject to change based on overall group availability.
Code used: Python or spreadsheet-based modeling
Technology needed: Google Colab or Excel, a computer with internet access, and a working microphone/camera
Project overview
In this project, students investigate simple decision rules for binary classification and explore how these rules behave when data are highly imbalanced, meaning that one outcome is much rarer than the other. Through hands-on experiments, students discover that a rule with very high accuracy can nevertheless perform poorly in practice by systematically missing rare but important cases.
Rather than introducing complex models, students focus on understanding why accuracy can be misleading and how evaluation criteria shape our conclusions. Using tables, graphs, and basic arithmetic, students compare different decision rules under varying levels of imbalance.
In the final stage of the project, students propose a simple remedy by redefining what it means for a rule to perform well. By examining multiple error rates and incorporating basic cost considerations, students develop a more nuanced framework for evaluating decisions. The project emphasizes the process of mathematical research—posing questions, designing experiments, interpreting results, and communicating findings—using tools accessible to high-school students.
Background
All necessary concepts—such as decision rules, evaluation metrics, and experimental comparison—will be introduced within the context of the research questions. Computational experiments will be conducted using Excel, R, or Python, with code templates provided as needed.
Possible extension
Selected references
- He, H. and Garcia, E. A. Learning from Imbalanced Data. IEEE Transactions on Knowledge and Data Engineering, 21(9), 2009.
- Saito, T. and Rehmsmeier, M. The Precision-Recall Plot Is More Informative than the ROC Plot When Evaluating Binary Classifiers on Imbalanced Datasets. PLoS ONE, 10(3), 2015.
Extensions of Colley’s Matrix and Ranking Methods
Applied Linear Algebra
Research Proposal
Extensions of Colley’s Matrix and Ranking Methods
PhD Candidate (ABD)
Department of Mathematics
Florida State University
Tentative Mon/Wed 6:30-8:00 PM ET; final meeting times are subject to change based on overall group availability.
Code used: MATLAB or Julia
Technology needed: MATLAB Online or Julia, Excel for simple computations, and presentation software such as PowerPoint or Beamer
Project overview
In this project, students will work toward replicating Colley’s Matrix for a small “toy” example using college football teams from the previous season. They will analyze these replications and investigate whether Colley’s method was able to identify potential major upsets that the College Football Committee did not anticipate. This implementation will require solving a linear system of the form:
Cr = b,
C ∈ ℝn×n,
b, r ∈ ℝn
where r is the ranking vector, b is one plus the average win rate for a team, and C is Colley’s matrix. This framework leads naturally to the central research question of the project: how can Colley’s Matrix be improved when additional information about the system is available?
The project emphasizes mathematical structure, computational experimentation, and interpretation. By the end of the program, students will understand how a linear algebra model can be used to construct rankings and how modifying that model may lead to more informative decision-making tools in sports and beyond.
Background
This project serves as a strong introduction to ideas from linear algebra, introductory computer science, statistics, and mathematical modeling. During the experimental stage, students will implement the mathematical theory computationally using tools such as Excel, Matlab, and/or Julia. During the final week, when results are compiled and presented, students may use tools such as PowerPoint or Beamer.
Possible extension
Selected references
- Boginski, V., Butenko, S., and Pardalos, P. M. Matrix-based methods for college football rankings. Economics, Management and Optimization in Sports, 2004, pp. 1–13.
- Colley, Wesley N. Colley’s bias free college football ranking method: The Colley matrix explained, 2002.