Projects
Summer 2023
Buffon Needle Problem
Supervisors: Matthew Dannenberg and Alex Iosevich
Prerequisites: Honors multi-variable calculus and strong
python programming skills
Participants: William Hagerstrom, Gabriel Hart, Tran Duy Anh
Le, Isaac Li, Nathan Skerett
Description: This problem is related to the material that
will be touched upon in probability and geometry
mini-courses. Consider a bounded convex set K in the plane of fixed
perimeter. Drop a needle of length r such that one end is inside K
with equal probability and the orientation of the needle is random.
Let p(K,r) denote the probability that the other end of the needle
also lands inside K. Our goal is to prove that if r is small enough,
then among all the convex sets of a given perimeter, p(K,r) is
maximized if K is a disk. Furthermore, we will conduct computer
experiments to estimate the value of r for which the disk maximizes
the Buffon probability. Chat GPT will be used as an auxiliary tool.
Suggested steps to understand this problem:
i) Compute p(K,r) when K is a disk of radius 1.
ii) Compute p(K,r) when K is a square of the same perimeter as the
disk of radius 1.
iii) Compute p(K,r) when K is an ellipse of the same perimeter as
the disk of radius 1. This calculation may need to be done
numerically.
iv) Write python code that estimates p(K,r) with the input of K
accomplished in some reasonable way.
v) Try to conceptualize the optimization of p(K,r) as a
reinforcement learning problem.
vi) Look up and understand the Fourier series proof (due to
Hurewicz) of the iso-perimetric inequality in the plane.
Systems of nonlinear equations: reinforcement learning versus
classical methods
Supervisors: Alex Iosevich and Azita Mayeli
Prerequisites: Multi-variable calculus (not necessarily
honors) and strong python programming skills
Participants: Xiaolu Chen, Joshua Iosevich, Stephanie Wang,
Tae Ho Yoo
Description: This problem is related to the material from
the geometry mini-course. There is a variety of classical
methods for estimating solutions of systems of non-linear equations.
For example, in this case of a single equation, the Newton method
taught in most first calculus sequences is an example of such an
approach. For systems of equations, the Gauss-Newton method is
extremely effective. When the functions involved in a nonlinear
system are very rough, or discrete, which could mean a variety of
different things, classical methods tend to break down and other
approached are needed. One possible approach is reinforcement
learning.
An example of a situation where reinforcement learning and classical
methods co-exist is the study of shortest paths in graphs. The
Dykstra algorithms yields the shortest path, but in many situations,
reinforcement learning based algorithms do better, and it would be
interesting to understand this phenomenon systematically.
In this project, we are going to create a reinforcement learning
code for a variety of situation, including approximating solutions
to systems of nonlinear equations, grid search, graph algorithms,
and others, and compare the results to the classical techniques in
cases of overlap. Chat GPT will be used as an auxiliary tool.
There is plenty of room for theoretical work on this problem as
well.
Suggested steps to understand this problem:
i) Remind yourself exactly how the Newton method works.
ii) Look up and understand related methods for estimating solutions
of systems of equations.
iii) Write your own python code pertaining to i) and ii).
iv) Read up on reinforcement learning. An excellent source is a
book, entitled "Reinforcement Learning" by Sutton and Barto.
v) Write reinforcement learning code to estimate solutions of
systems of nonlinear equations.
vi) Find a reasonable way to compare the performance of your code in
parts v) and iii).
Geometric combinatorcs and complexity:
Supervisors: Alex Iosevich and Azita Mayeli
Prerequisites: Strong Python programming skills
Participants: Joshua Iosevich, Alhussein Khalil, Anuurag
Kumar, Nathan Skerrett
Description: The main thrust of this research group is to
investigate two concepts that keep arising in geometric
combinatorics. The first is understanding zeroes of the Fourier
transform of an indicator function of a set in vector spaces over
integers modulo a prime. The second is to understand the complexity
of graphs that arise in the study of finite point configurations in
vector spaces over finite fields.
Suggested steps to understand this problem:
i) Read the paper where the triangle result is proved. It can be
found on arxiv.org under the designation arXiv: 1311.4788. If you
run into questions, which is almost inevitable, please contact me
immediately and I will point you in the right direction.
ii) Write a code that encodes triangles in a two-dimensional vector
spaces over integers modulo p, where p is a prime, as a python
class. Within that class, write a function that determines whether
two given triangles (with vertices as input) are or not are
congruent. Note that two triangles are congruent if one can map the
vertices of one onto the vertices of the other using a translation
and a rotation. What a translation is in this context is clear. For
the right notion of a rotation, read the paper mentioned above.
iii) Consider the following theoretical exploration. One can find an
infinite sequences of primes p such that the two dimensional vector
space over the integer modulo p does not contain an equilateral
triangle. By the way, what an equilateral triangle means in this
context is in itself an interesting question. By following the same
idea, you can construct a prime p that excludes other congruence
classes of triangles. An interesting question is just how many
congruence classes (relative to p) can we exclude by imposing
conditions on the prime p?
Modeling seizures using machine learning:
Supervisors: Svetlana Pack and Yujia Zhai
Prerequisites: Good knowledge of statistics and strong Python
programming skills.
Participants: Andrew Isaacson,
Ajax Benander, Xiangyi Chen, Wentao Jiang, Jenifer Kim, Svetlana
Pack, Yining Qian, Scott Sun, Gus Vietze, Matthew Xie, Jonathan
Zhang
Description: It is widely believed in the medical
community that epileptic seizures do not follow a particular daily
or weekly time patterns. We believe that the techniques of modern
data science have not yet been deployed in the study of this
problem in a systematic way. We are going to experiment with a
variety of techniques, including reinforcement learning, to look
for patterns in the commercially available data sets. Chat
GPT will be used as an auxiliary tool.
Suggested steps to understand this
problem:
i) Learn how to use TensorFlow
or similar to make time series forecasts.
ii) Read up on what it means for a time series to be
forecastable.
Time series and retail analytics:
Supervisors: Alex Iosevich, Azita Mayeli and Evan Witz
Prerequisites: Good knowledge of
statistics and strong Python programming skills.
Participants: Moez Boussarsar, Colin Hascup, Jennifer Kim,
Jiaming Lyu, Peter MacNeil, Cooper Orio, Jiamu
Tang, Yining Qian
Description: We are going to experiment with a
variety of methods of determining whether a time series can be
accurately predicted. We are also going set up retail forecasting
models using a variety of regressors.
Suggested steps to understand the problem: You
should study the initial notebook for the Python sessions this
week carefully. It contains a variety of forecasting mechanisms
that will serve as the the basis for our analysis.
Fractal structures in large data sets:
Supervisors: Alex Iosevich and Azita Mayelia
Prerequisites: Good knowledge of analysis at the level of
Rudin's book and strong python programming skills
Participants: Yuesong
Huang, Xinyi Liu, Peter MacNeil, Svetlana Pack, Zhelin Sheng,
Nathan Skerrett,
Description: This is a followup on one of the projects from
Tripods 2021 which resulted in a research paper which can be found
at https://arxiv.org/pdf/2209.12079.pdf. The paper describes an
approach to detecting fractal structures in large data sets using
discrete energy. In this project we are going to develop this
theory further and create synthetic data sets to test the results.
One of the applied avenues we may pursue is the idea of using the
Chaitin-Komogorov-Solomonoff complexity of a time series to
determined the degrrr yo which it can be accurately forecasted. Chat
GPT will be used as an auxiliary tool.
Suggested steps to understand this problem:
i) Read arxiv2209.12079 (the predecessor
paper from Tripods 2021) and start asking questions about anything
that is not clear.
ii) At the same time, read up on basic properties of
the Minkowski and Hausdorff dimension.
iii) Try to duplicate the codes used to produce the diagrams in
the paper above.
VC dimension and applications:
Supervisors: Alex Iosevich, Donovan Snyder and Emmett Wyman
Prerequisites: Linear algebra, real analysis and elementary
probability
Participants: Xiaolu
Chen, Lily Testa,Tran Duy Anh Le, Scott Sun, Tae Ho Yoo
Description: Given a collection of subsets in the
plane, what is the smallest set which contains an element of each?
We use the notion of the VC-dimension to find such witness sets in
both the finite and infinite settings. One of the main focus for
this project will be the setting of fractal sets, though other
settings of interested are welcome! Chat GPT will be used
as an auxiliary tool.
Suggested steps to understand this problem:
i) Become familiar with the definition of VC Dimension and the
theory of epsilon nets (see the enclosed pdf).
ii) Be able to calculate the VC dimension of common hypothesis
classes.
iii) Learn about fractal sets and the Minkowski dimension (example
notes are available).
Discrete Neural Nets
Supervisor: Charlotte
Aten
Prerequisites: Good programming skills and ability to
write rigorous arguments
Participants: Karam Aldahleh, Ajax
Benander, Alhussein Khalil, Jennifer Kim, Isaac Li, Lilian
Stolberg, TJ Weaver, Kevin Xue
Description: This
project is a follow-up of the discrete neural nets project from
the 2021 Tripods REU. That project, which is introduced in a video
linked below, was finally coded and tested (with promising
results) as part of Rachel Dennis's 2023 senior thesis, also
linked below. Discrete neural nets are neural nets whose
activation functions can only take on a finite set of values. This
prevents us from using established techniques like gradient
descent but opens up new avenues not available for continuous
functions. In this year's project, we will explore how well this
technique can be extended from classifying/transforming
black-and-white images to arbitrary data types (such as images
with color).
Suggested steps to understand this problem:
i) Watch the initial
video pitch (https://youtu.be/nr0KbcloYW4) for the discrete
neural nets project from the 2021 Tripods REU.
ii) Read Rachel Dennis's 2023 senior thesis
(https://www.sas.rochester.edu/mth/undergraduate/honorspaperspdfs/r_dennis23.pdf).
Start asking questions about any parts you don't understand.
iii) A preprint of a paper will be posted soon, and a
cleaned-up version of the code will be available to examine,
as well. Take a look at these and start playing around when
they appear.
Don't worry if some of the category theory or
universal/abstract algebra language is a bit intimidating.
All the necessary algebra will be introduced during the
abstract algebra mini-course.