Seminars and Colloquia by Series

Asymptotically half of binary words are shuffle squares

Series
Stochastics Seminar
Time
Thursday, April 16, 2026 - 15:30 for 1 hour (actually 50 minutes)
Location
Skiles 006
Speaker
Logan PostGeorgia Institute of Technology

A binary shuffle square is a binary word of even length that can be partitioned into two disjoint, identical subwords. While recognizing shuffle squares is NP-hard, we show that they are surprisingly ubiquitous. We prove that a uniformly random binary word $s$ of length $2n$ is a shuffle square with probability $\frac 12-o(n^{-1/15})$, verifying a conjecture of He, Huang, Nam, and Thaper. In particular, almost every binary word is at most two bit-deletions away from a shuffle square, giving the best possible average case for the “Longest Twin” problem.

 

By revealing the bits of $s$ sequentially,  we reformulate the problem as a discrete stochastic process. We track the evolution of a “buffer set”, a collection of suffixes produced by the revealed bits. In this setting, there is a simple greedy algorithm which behaves like a SSRW; we define a local optimization which creates a negative bias. We also show that the buffer set is robust enough to absorb small defects, yielding a perfect partition with high probability.

Minimax D-Optimal designs in generalized linear models: Nonasymptotic theory and efficient algorithms

Series
Stochastics Seminar
Time
Thursday, April 9, 2026 - 15:30 for 1 hour (actually 50 minutes)
Location
Skiles 006
Speaker
Jacob AguirreGeorgia Tech

We study robust D-optimal experiment design in generalized linear models, choosing design weights to maximize the worst-case determinant of the Fisher information matrix over a convex parameter uncertainty set. This can be thought of as the natural generalization of D-optimal design in linear regression, and the key challenge is that the information matrix depends on the parameter, so the resulting minimax problem is generally not convex-concave. We show that the desired convexity-concavity, in fact, reduces to a scalar curvature condition on the log-partition function of the exponential family, namely its second derivative h must satisfy the inequality h''h ≥ q(h')² for some q > 1. This insight is connected to the notion of Volumetric Barrier (VB) convexity for self-concordant functions, a result first introduced by Tseng et al (2025) in the context of online quantum state estimation. With self-concordant barriers on the design weights simplex and the parameter uncertainty set, the regularized saddle objective becomes a self-concordant convex-concave (SCCC) function, enabling efficient minimax interior-point methods developed by Nemirovski. We also consider the generalization of the framework, where convexity in the model parameter breaks, but the q-inequality holds up to a deficit proportional to h; in this case, our methods are just as applicable. It turns out that this class includes all canonical GLMs, and we identify logistic regression as the hardest model in the class.

This joint work with Dmitrii Ostrovskii.

Negative association of the Busemann functions in exactly solvable KPZ models

Series
Stochastics Seminar
Time
Thursday, April 2, 2026 - 15:30 for 1 hour (actually 50 minutes)
Location
Skiles 006
Speaker
Xiao ShenNorth Carolina State University

In the study of random growth models belonging to the Kardar--Parisi--Zhang (KPZ) universality class, a notably successful approach has been to analyze stationary initial conditions defined by the Busemann functions. Recently, this perspective has been extended to handle multiple asymptotic directions simultaneously, but the joint distribution of the Busemann process is more difficult to access, and many aspects of this process remain elusive. In particular, the remarkable independence property present in the exactly solvable setting fails when considering Busemann functions across different directions. In the corner growth model, also known as exponential last-passage percolation (LPP), we prove that, regardless of their different directions, Busemann functions along a down-right path are always negatively associated across each individual direction. In other words, increasing the value of Busemann functions in one direction tends to probabilistically decrease the values of neighboring ones. As an application, we obtain an exponential concentration inequality on the diffusive scale for Busemann functions along a down-right path, in the absence of independence. Joint work with Erik Bates.

TBA : Hung Nguyen

Series
Stochastics Seminar
Time
Thursday, March 19, 2026 - 15:30 for 1 hour (actually 50 minutes)
Location
Skiles 006
Speaker
Hung NguyenUniversity of Tennessee, Knoxville

Joint parameter estimation of spin glasses

Series
Stochastics Seminar
Time
Thursday, March 12, 2026 - 15:30 for 1 hour (actually 50 minutes)
Location
Skiles 006
Speaker
Qiang WuUniversity of Minnesota

Spin glasses are disordered statistical physics system with both ferromagnetic and anti-ferromagnetic spin interactions. The Gibbs measure belongs to the exponential family with parameters, such as inverse temperature $\beta>0$ and external field $h\in R$.  A fundamental statistical problem is to estimate the system parameters from a single sample of the ground truth. In 2007, Chatterjee first proved that under reasonable conditions, for spin glass models with $h=0$, the maximum pseudo-likelihood estimator for $\beta$ is $\sqrt{N}$-consistent. This is in contrast to the existing estimation results for classical non-disordered models. However, Chatterjee's approach has been restricted to the single parameter estimation setting.  The joint parameter estimation of $(\beta,h)$ for spin glasses has remained open since then. In this talk, I will introduce a new idea to show that under some easily verifiable conditions,  the bi-variate maximum pseudo-likelihood estimator is jointly $\sqrt{N}$-consistent for a large collection of spin glasses, including the Sherrington-Kirkpatrick model and its diluted variants. Based on joint work with Wei-Kuo Chen, Arnab Sen. 

Spectral gaps and measure decompositions

Series
Stochastics Seminar
Time
Thursday, March 5, 2026 - 15:30 for 1 hour (actually 50 minutes)
Location
Skiles 006
Speaker
March BoedihardjoMichigan State University

I'll introduce a new set of computable and orthogonally invariant quantities for a given probability measure on a Euclidean space. We show how these quantities can determine the extent to which the given probability measure can be decomposed as an equal weight mixture of two probability measures with significantly different second order statistics. Joint work with Joe Kileel and Vandy Tombs.

 

Continuous directed polymers in a Gaussian environment

Series
Stochastics Seminar
Time
Thursday, February 26, 2026 - 15:30 for 1 hour (actually 50 minutes)
Location
Skiles 006
Speaker
Le ChenAuburn University

We study a broad class of space–time continuous directed polymers in a Gaussian environment that is white in time and spatially correlated (Itô sense). Under Dalang’s condition, we prove key properties of the partition function—positivity, stationarity, scaling, homogeneity, and a Chapman–Kolmogorov relation—and establish pathwise regularity of the polymer (Hölder continuity and quadratic variation). We give a sharp singularity criterion: the polymer measure is singular w.r.t. Wiener measure iff the spectral measure has infinite total mass. Finally, for d≥3, we prove diffusive large-time behavior in the high-temperature regime, providing a unified framework for polymers driven by singular Gaussian noise.

 

Joint work with Cheng Ouyang (UIC), Samy Tindel (Purdue), and Panqiu Xia (Cardiff).

Beyond propagation of chaos: A stochastic algorithm for mean-field optimization

Series
Stochastics Seminar
Time
Thursday, February 5, 2026 - 15:30 for 1 hour (actually 50 minutes)
Location
Skiles 006
Speaker
Chandan TankalaUniversity of Oregon

Sampling and mean-field optimization can be viewed as optimization in the space of probability distributions. Stochastic optimization algorithms like stochastic gradient descent have been immensely successful for optimization over Euclidean spaces. However, the infinite-dimensional space of probability distributions poses unique challenges. In this talk, I will discuss my recent work on the design and analysis of a stochastic algorithm for mean-field optimization with applications to the increasingly studied area of mean-field neural networks.

Partial identification with Schrödinger bridges

Series
Stochastics Seminar
Time
Thursday, January 29, 2026 - 15:30 for 1 hour (actually 50 minutes)
Location
Skiles 006
Speaker
Florian GunsiliusEmory University

Partial identification provides an alternative to point identification: instead of pinning down a unique parameter estimate, the goal is to characterize a set guaranteed to contain the true parameter value. Many partial identification approaches take the form of linear optimization problems, which seek the "best- and worst-case scenarios" of a proposed model subject to the constraint that the model replicates correct observable information. However, such linear programs become intractable in settings with multivalued or continuous variables. This paper introduces a novel method to overcome this computational and statistical curse of cardinality: an entropy penalty transforms these potentially infinite-dimensional linear programs into general versions of multi-marginal Schrödinger bridges, enabling efficient approximation of their solutions. In the process, we establish novel statistical and mathematical properties of such multi-marginal Schrödinger bridges---including an analysis of the asymptotic distribution of entropic approximations to infinite-dimensional linear programs. We illustrate this approach by analyzing  instrumental variable models with continuous variables, a setting that has been out of reach for existing methods.

Similarities and Differences between the Longest Common and Longest Common and Increasing Subsequences in Random Words

Series
Stochastics Seminar
Time
Thursday, January 22, 2026 - 15:30 for 1 hour (actually 50 minutes)
Location
Skiles 006
Speaker
Christian HoudréGeorgia Institute of Technology

Let $LC_n$ be the length of the longest common subsequences of two independent random words whose letters are taken  in a finite alphabet and when the alphabet is totally ordered and let $LCI_n$ be the length of the longest common and increasing subsequences of the words.   Results on the asymptotic means, variances and limiting laws of these well-known random objects will be described and compared.

Pages