Measure-Theoretic Dvoretzky Theorem and Applications to Data Science
- Probability,Uploaded Videos ( 1451 Views )SEPC 2021 in honor of Elizabeth Meckes. Slides from the talks and more information are available <a href="https://services.math.duke.edu/~rtd/SEPC2021/SEPC2021.html">at this link (here).</a>
Roman Vershynin : Mathematics of synthetic data and privacy
- Probability,Uploaded Videos ( 1110 Views )An emerging way to protect privacy is to replace true data by synthetic data. Medical records of artificial patients, for example, could retain meaningful statistical information while preserving privacy of the true patients. But what is synthetic data, and what is privacy? How do we define these concepts mathematically? Is it possible to make synthetic data that is both useful and private? I will tie these questions to a simple-looking problem in probability theory: how much information about a random vector X is lost when we take conditional expectation of X with respect to some sigma-algebra? This talk is based on a series of papers with March Boedihardjo and Thomas Strohmer.
David Aldous: Probability Seminar
- Probability,Uploaded Videos ( 649 Views )David Aldous, Probability Seminar Sept 30, 2021 TITLE: Can one prove existence of an infectiousness threshold (for a pandemic) in very general models of disease spread? ABSTRACT: Intuitively, in any kind of disease transmission model with an infectiousness parameter, there should exist a critical value of the parameter separating a very likely from a very unlikely resulting pandemic. But even formulating a general conjecture is challenging. In the most simplistic model (SI) of transmission, one can prove this for an essentially arbitrary large weighted contact network. The proof for SI depends on a simple lemma concerning hitting times for increasing set-valued Markov processes. Can one extend to SIR or SIS models over similarly general networks, where the lemma is no longer applicable?
Sayan Banerjee : Singular Reflected Diffusions
- Probability ( 384 Views )I will talk about some models coming from Physics and Queueing Theory that give rise to singular reflected processes in their diffusion limit. Such diffusions are characterized by non-elliptic generators (which are not even hypoelliptic) in the interior, and ergodicity arises from non-trivial interactions between the diffusion, drift and reflection. I will introduce a regenerative process approach which identifies renewal times in diffusion paths and analyzes excursions between successive renewal times. This provides a detailed description of the stationary distribution even when closed form expressions are unavailable. Based on joint works with Chris Burdzy, Brendan Brown, Mauricio Duarte and Debankur Mukherjee.
David Sivakoff : Polluted Bootstrap Percolation in Three Dimensions
- Probability ( 231 Views )In r-neighbor bootstrap percolation, the vertices of Z^d are initially occupied independently with probability p and empty otherwise. Occupied vertices remain occupied forever, and empty vertices iteratively become occupied when they have at least r occupied neighbors. It is a classic result of van Enter (r=d=2) and Schonmann (d>2 and r between 2 and d) that every vertex in Z^d eventually becomes occupied for any initial density p>0. In the polluted bootstrap percolation model, vertices of Z^d are initially closed with probability q, occupied with probability p and empty otherwise. The r-neighbor bootstrap rule is the same, but now closed vertices act as obstacles, and remain closed forever. This model was introduced 20 years ago by Gravner and McDonald, who studied the case d=r=2 and proved a phase transition exists for this model as p and q tend to 0. We prove a similar phase transition occurs when d=r=3, and we identify the polynomial scaling between p and q at which this transition occurs for the modified bootstrap percolation model. For one direction, our proof relies on duality methods in Lipschitz percolation to find a blocking structure that prevents occupation of the origin. The other direction follows from a rescaling argument, and the recent results of Holroyd and Gravner for d>r=2. This is joint work with Holroyd and Gravner.
Stanislav Molchanov : On the random analytic functions
- Probability ( 230 Views )The talk will contain a review of several recent results on the analytic continuation of the random analytic functions. We will start from the classical theorem on the random Taylor series (going to Borel s school), but the main subject will be the random zeta function (which was introduced implicitly by Cramer) and its generalizations. We will show that true primes are not truly random , since zeta functions for the random pseudo-primes (in the spirit of Cramer) have no analytic continuation through the critical line Re (z) = 1/2.
David Herzog : Supports of Degenerate Diffusion Processes: The Case of Polynomial Drift and Additive Noise
- Probability ( 223 Views )We discuss methods for computing supports of degenerate diffusion processes. We assume throughout that the diffusion satisfies a stochastic differential equation on Rd whose drift vector field X0 is ``polynomial'' and whose noise coefficients are constant. The case when each component of X0 is of odd degree is well understood. Hence we focus our efforts on X0 having at least one or more components of even degree. After developing methods to handle such cases, we shall apply them to specific examples, e.g. the Galerkin truncations of the Stochastic Navier-Stokes equation, to help establish ergodic properties of the resulting diffusion. One benefit to our approach is that, to prove such consequences, all we must do is compute certain Lie brackets.
Ofer Zeitouni : Slowdown in Branching random walks and the inhomogeneous KPP equation
- Probability ( 207 Views )The classical result of Bramson gives a precise logarithmic correction to the speed of front propagation in one dimensional branching random walks and Brownian motions. I will discuss several variants of this model where the slowdown term is not classical.
Nicolas Zygouras : Pinning-depinning transition in Random Polymers
- Probability ( 206 Views )Random Polymers are modeled as a one dimensional random walk (S_n), with excursion length distribution P(S_1 = n) = \phi(n)/n^\alpha, \alpha > 1 and \phi(n) a slowly varying function. The polymer gets a random reward whenever it visits or crosses an interface. The random rewards are realised as a sequence of i.i.d. variables (\omega_n). Depending on the relation between the mean value of the disorder \omega_n and the temperature, the polymer might prefer to stick to the interface (pinnings) or undergo a long excursion away from it (depinning). In this talk we will review some aspects of random polymer models. We will also discuss in more detail the pinning-depinning transition of the `Pinning' model and prove its annealed and quenched critical points are distinct. This is joint work with Ken Alexander.
Erika Berenice Roldan Roa : Asymptotic behavior of the homology of random polyominoes
- Probability ( 197 Views )In this talk we study the rate of growth of the expectation of the number of holes (the rank of the first homology group) in a polyomino with uniform and percolation distributions. We prove the existence of linear bounds for the expected number of holes of a polyomino with respect to both the uniform and percolation distributions. Furthermore, we exhibit particular constants for the upper and lower bounds in the uniform distribution case. This results can be extend, using the same techniques, to other polyforms and higher dimensions.
Elizabeth Meckes : Projections of probability distributions: a measure-theoretic Dvoretzky theorem
- Probability ( 196 Views )Dvoretzky's theorem tells us that if we put an arbitrary norm on n-dimensional Euclidean space, no matter what that normed space is like, if we pass to subspaces of dimension about log(n), the space looks pretty much Euclidean. A related measure-theoretic phenomenon has long been observed: the (one-dimensional) marginals of many natural high-dimensional probability distributions look about Gaussian. A question which had received little attention until recently is whether this phenomenon persists for k-dimensional marginals for k growing with n, and if so, for how large a k? In this talk I will discuss recent work showing that the phenomenon does indeed persist if k less than 2log(n)/log(log(n)), and that this bound is sharp (even the 2!).
Leonid Koralov : An Inverse Problem for Gibbs Fields
- Probability ( 168 Views )It is well known that for a regular stable potential of pair interaction and a small value of activity one can define the corresponding Gibbs field (a measure on the space of configurations of points in $\mathbb{Z}^d$ or $\mathbb{R}^d$). We consider a converse problem. Namely, we show that for a sufficiently small constant $\overline{\rho}_1$ and a sufficiently small function $\overline{\rho}_2(x)$, $x \in \mathbb{Z}^d$ or $\mathbb{R}^d$, there exist a hard core pair potential, and a value of activity, such that $\overline{\rho}_1$ is the density and $\overline{\rho}_2$ is the pair correlation function of the corresponding Gibbs field.
Sebastien Roch : Cascade Processes in Social Networks
- Probability ( 161 Views )Social networks are often represented by directed graphs where the nodes are individuals and the edges indicate a form of social relationship. A simple way to model the diffusion of ideas, innovative behavior, or word-of-mouth effects on such a graph is to consider a stochastic process of ``infection'': each node becomes infected once an activation function of the set of its infected neighbors crosses a random threshold value. I will prove a conjecture of Kempe, Kleinberg, and Tardos which roughly states that if such a process is ``locally'' submodular then it must be ``globally'' submodular on average. The significance of this result is that it leads to a good algorithmic solution to the problem of maximizing the spread of influence in the network--a problem known in data mining as "viral marketing"'. This is joint work with Elchanan Mossel.
David Sivakoff : Random Site Subgraphs of the Hamming Torus
- Probability ( 159 Views )The critical threshold for the emergence of a giant component in the random site subgraph of a d-dimensional Hamming torus is given by the positive root of a polynomial. This value is distinct from the critical threshold for the random edge subgraph of the Hamming torus. The proof uses an intuitive application of multitype branching processes.
Carla Staver : Stable coexistence of savannah and forest in a spatial model
- Probability ( 157 Views )The goal of this talk is to further a joint project involving Carla Staver, Simon Levin, Rick Durrett, and Ruibo Ma. The puzzle is: why can savannah and forest display stable coexistence when this is not possible in a spatially homogeneous system.
James Johndrow : Error Bounds for Approximations of Markov Chains
- Probability ( 140 Views )We give results on the properties of Markov kernels that approximate another Markov kernel. The basic idea is that when the original kernel satisfies a contraction condition in some metric, the long-time dynamics of the two chains -- as well as the invariant measures, when they exist -- will be close in that metric, so long as the approximating kernel satisfies a suitable approximation error condition. We focus on weighted total variation and Wasserstein metrics, and motivate the results with applications to scalable Markov chain Monte Carlo algorithms. This is joint work with Jonathan Mattingly.
Kevin McGoff : An introduction to thermodynamic formalism in ergodic theory through (counter)examples
- Probability ( 129 Views )The goal of this talk is to give a self-contained introduction to some aspects of the thermodynamic formalism in ergodic theory that should be accessible to probabilists. In particular, the talk will focus on equilibrium states and Gibbs measures on the Z^d lattice. We'll present some basic examples in the theory, as well as some recent results that are joint with Christopher Hoffman.
Eric Foxall : The compulsive gambler with allowances
- Probability ( 127 Views )We consider a process in which a finite set of n agents continually receive a 1 dollar allowance and gamble their fortunes, all in, with one another at a constant rate. This is a variation on the existing compulsive gambler process; in that process, initial fortunes are prescribed and no further allowances are given out. For our process, we find that after some time the distribution of wealth settles into a pattern in which most people have only a few dollars, a few are very wealthy, and a single person possesses most of the cash currently present in the population. In addition, eventually the only way to attain first rank is by winning a bet against the current champion. Moreover, if agents play a fair game, i.e., the probability of winning a bet is proportional to the players' fortunes, the title of champion is assumed by every player infinitely often, although it changes less and less frequently as time goes on. Finally, by examining the process from both the perspective of typical fortune, and that of large fortune, we can go one step further and obtain two distinct limiting processes as n --> infty, with each one admitting a detailed description of its dynamics.
Matt Junge : Parking
- Probability ( 125 Views )Parking functions were introduced by combinatorialists in the 1960s, and have recently been studied by probabilists. When the parking lot is an infinite graph and cars drive around at random, we will look at how many parking spots are needed for every car to eventually find a spot.
Joe Jackson : The convergence problem in mean field control
- Probability ( 123 Views )This talk will be about the convergence problem in mean field control (MFC), i.e. the challenge of rigorously justifying the convergence of certain "symmetric" $N$-particle stochastic control problems towards their mean field counterparts. On the one hand, this convergence problem is already well-understood from a qualitative perspective, thanks to powerful probabilistic techniques based on compactness. On the other hand, quantitative results (i.e. rates of convergence) are more difficult to obtain, in large part because the value function of the mean field problem (which is also the solution to a certain Hamilton-Jacobi equation on the Wasserstein space) may fail to be $C^1$, even if all the data is smooth. After giving an overview of the convergence problem, I will discuss the results of two recent joint works with Cardaliaguet, Daudin, Delarue, and Souganidis, in which we use some ideas from the theory of viscosity solutions to overcome this lack of regularity and obtain rates of convergence of the $N$-particle value functions towards the value function of the corresponding MFC problem.
Xue-Mei Li : Stirring the geodesics
- Probability ( 122 Views )In this talk, we discuss stochastic homogeneization on the Hopf fibration. Let us consider Berger's metrics on the three sphere, obtained by shrinking the Hopf circle directions by a factor epsilon. So we think of three spheres as two spheres attached at each point a circle. We consider a particle that is moved by two vector fields: a unit speed vector field, with respect to Berger's metrices, along the Hopf circle; and also a non-zero vector field in` \(S^2\) direction' with speed given by a one dimensional Brownian motion. In the limit of epsilon goes to zero, we obtain a Brownian motion on \(S^2\). The effective motion is obtained by moving a particle along a fast rotating horizontal direction.
Lingjiong Zhu : Self-Exciting Point Processes
- Probability ( 121 Views )Self-exciting point processes are simple point processes that have been widely used in neuroscience, sociology, finance and many other fields. In many contexts, self-exciting point processes can model the complex systems in the real world better than the standard Poisson processes. We will discuss the Hawkes process, the most studied self-exciting point process in the literature. We will talk about the limit theorems and asymptotics in different regimes. Extensions to Hawkes processes and other self-exciting point processes will also be discussed.
Ivana Bozic : Dynamics of cancer in response to targeted combination therapy
- Probability ( 118 Views )A link to the paper can be found on her web page. In solid tumors, targeted treatments can lead to dramatic regressions, but responses are often short-lived because resistant cancer cells arise. The major strategy proposed for overcoming resistance is combination therapy. We present a mathematical model describing the evolutionary dynamics of lesions in response to treatment. We first studied 20 melanoma patients receiving vemurafenib. We then applied our model to an independent set of pancreatic, colorectal, and melanoma cancer patients with metastatic disease. We find that dual therapy results in long-term disease control for most patients, if there are no single mutations that cause cross-resistance to both drugs; in patients with large disease burden, triple therapy is needed. We also find that simultaneous therapy with two drugs is much more effective than sequential therapy. Our results provide realistic expectations for the efficacy of new drug combinations and inform the design of trials for new cancer therapeutics.
Philip Matchett Wood : Random doubly stochastic tridiagonal matrices
- Probability ( 115 Views )Let $T_n$ be the compact convex set of tridiagonal doubly stochastic matrices. These arise naturally as birth and death chains with a uniform stationary distribution. One can think of a typical matrix $T_n$ as one chosen uniformly at random, and this talk will present a simple algorithm to sample uniformly in $T_n$. Once we have our hands on a 'typical' element of $T_n$, there are many natural questions to ask: What are the eigenvalues? What is the mixing time? What is the distribution of the entries? This talk will explore these and other questions, with a focus on whether a random element of $T_n$ exhibits a cutoff in its approach to stationarity. Joint work with Persi Diaconis.
Dane Johnson : Large deviations, moderate deviations, and importance sampling
- Probability ( 115 Views )Importance sampling is an accelerated Monte Carlo algorithm that can reduce variance when estimating small probabilities. The design of the algorithm involves the choice of a change of measure, and based on this choice the performance can range from substantially better than standard Monte Carlo to substantially worse. One approach to choosing a change of measure involves embedding the problem of interest in a sequence of processes that satisfies a large deviations principle, and then basing the change of measure on subsolutions to the Hamilton-Jacobi-Bellman equation associated the large deviations rate function. This approach has the benefit of guaranteeing a certain level of asymptotic performance based on the subsolution, but different embeddings can lead to different rate functions, subsolutions, and consequently different algorithms. I will contrast the strengths and weaknesses of two different embeddings, one using a scaling commonly referred to as the standard large deviations scaling and the other using a scaling referred to as moderate deviations.
Jasmine Foo : Modeling diversity in tumor populations and implications for drug resistance
- Probability ( 115 Views )In this talk I will discuss a branching process model developed to study intra-tumor diversity (i.e. the variation amongst the cells within a single tumor). This variation arises due to heritable (epi)genetic alterations which can confer changes in cellular fitness and drug response. In the asymptotic (t-> infinity) regime, we study the growth rates of the population as well as some ecological measures of diversity in the tumor. In the second half of the talk I will discuss applications of this model to studying the emergence of drug resistant populations in Chronic Myeloid Leukemia (CML). (Joint work w/K. Leder, J. Mayberry, R. Durrett, F. Michor)
Sayan Banerjee : Coupling, geometry and hypoellipticity
- Probability ( 114 Views )Coupling is a way of constructing Markov processes with prescribed laws on the same space. The coupling is called Markovian if the coupled processes are co-adapted to the same filtration. We will first investigate Markovian couplings of elliptic diffusions and demonstrate how the rate of coupling (how fast you can make the coupled processes meet) is intimately connected to the geometry of the underlying space. Next, we will consider couplings of hypoelliptic diffusions (diffusions driven by vector fields whose Lie algebra span the whole tangent space). Constructing successful couplings (where the coupled processes meet almost surely) for these diffusions is a much more subtle question as these require simultaneous successful coupling of the driving Brownian motions as well as a collection of their path functionals. We will construct successful Markovian couplings for a large class of hypoelliptic diffusions. We will also investigate non-Markovian couplings for some hypoelliptic diffusions, namely the Kolmogorov diffusion and Brownian motion on the Heisenberg group, and demonstrate how these couplings yield sharp estimates for the total variation distance between the laws of the coupled diffusions when Markovian couplings fail. Furthermore, we will demonstrate how non-Markovian couplings can be used to furnish purely analytic gradient estimates of harmonic functions on the Heisenberg group by purely probabilistic means, providing yet another strong link between probability and geometric analysis. This talk is based on joint works with Wilfrid Kendall, Maria Gordina and Phanuel Mariano.