Search results
(1 - 3 of 3)
- Title
- Learning Stochastic Governing Laws from Noisy Data Using Normalizing Flows
- Creator
- McClure, William Jacob
- Date
- 2021
- Description
-
With the increasing availability of massive collections of data, researchers in all sciences need tools to synthesize useful and pertinent...
Show moreWith the increasing availability of massive collections of data, researchers in all sciences need tools to synthesize useful and pertinent descriptors of the systems they study. Perhaps the most fundamental knowledge of a dynamical system is its governing laws, which describe its evolution through time and can be lever-aged for a number of analyses about its behavior. We present a novel technique for learning the infinitesimal generator of a Markovian stochastic process from large, noisy datasets generated by a stochastic system. Knowledge of the generator in turn allows us to find the governing laws for the process. This technique relies on normalizing flows, neural networks that estimate probability densities, to learn the density of time-dependent stochastic processes. We establish the efficacy of this technique on multiple systems with Brownian noise, and use our learned governing laws to perform analysis on one system by solving for its mean exit time. Our approach also allows us to learn other dynamical behaviors such as escape probability and most probable pathways in a system. The potential impact of this technique is far-reaching, since most stochastic processes in various fields are assumed to be Markovian, and the only restriction for applying our method is available data from a time near the beginning of an experiment or recording.
Show less
- Title
- Choice-Distinguishing Colorings of Cartesian Products of Graphs
- Creator
- Tomlins, Christian James
- Date
- 2022
- Description
-
A coloring $f: V(G)\rightarrow \mathbb N$ of a graph $G$ is said to be \emph{distinguishing} if no non-identity automorphism preserves every...
Show moreA coloring $f: V(G)\rightarrow \mathbb N$ of a graph $G$ is said to be \emph{distinguishing} if no non-identity automorphism preserves every vertex color. The distinguishing number, $D(G)$, of a graph $G$ is the smallest positive integer $k$ such that there exists a distinguishing coloring $f: V(G)\rightarrow [k]$ and was introduced by Albertson and Collins in their paper ``Symmetry Breaking in Graphs.'' By restricting what kinds of colorings are considered, many variations of distinguishing numbers have been studied. In this paper, we study proper list-colorings of graphs which are also distinguishing and investigate the choice-distinguishing number $\text{ch}_D(G)$ of a graph $G$. Primarily, we focus on the choice-distinguishing number of Cartesian products of graphs. We determine the exact value of $\text{ch}_D(G)$ for lattice graphs and prism graphs and provide an upper bound on the choice-distinguishing number of the Cartesian products of two relatively prime graphs, assuming a sufficient condition is satisfied. We use this result to bound the choice distinguishing number of toroidal grids and the Cartesian product of a tree with a clique. We conclude with a discussion on how, depending on the graphs $G$ and $H$, we may weaken the sufficient condition needed to bound $\text{ch}_D(G\square H)$.
Show less
- Title
- Independence and Graphical Models for Fitting Real Data
- Creator
- Cho, Jason Y.
- Date
- 2023
- Description
-
Given some real life dataset where the attributes of the dataset take on categorical values, with corresponding r(1) × r(2) × … × r(m)...
Show moreGiven some real life dataset where the attributes of the dataset take on categorical values, with corresponding r(1) × r(2) × … × r(m) contingency table with nonzero rows or nonzero columns, we will be testing the goodness-of-fit of various independence models to the dataset using a variation of Metropolis-Hastings that uses Markov bases as a tool to get a Monte Carlo estimate of the p-value. This variation of Metropolis-Hastings can be found in Algorithm 3.1.1. Next we will consider the problem: ``out of all possible undirected graphical models each associated to some graph with m vertices that we test to fit on our dataset, which one best fits the dataset?" Here, the m attributes are labeled as vertices for the graph. We would have to conduct 2^(mC2) goodness-of-fit tests since there are 2^(mC2) possible undirected graphs on m vertices. Instead, we consider a backwards selection method likelihood-ratio test algorithm. We first start with the complete graph G = K(m), and call the corresponding undirected graphical model ℳ(G) as the parent model. Then for each edge e in E(G), we repeatedly apply the likelihood-ratio test to test the relative fit of the model ℳ(G-e), the child model, vs. ℳ(G), the parent model, where ℳ(G-e) ⊆ℳ(G). More details on this iterative process can be found in Algorithm 4.1.3. For our dataset, we will be using the alcohol dataset found in https://www.kaggle.com/datasets/sooyoungher/smoking-drinking-dataset, where the four attributes of the dataset we will use are ``Gender" (male, female), ``Age", ``Total cholesterol (mg/dL)", and ``Drinks alcohol or not?". After testing the goodness-of-fit of three independence models corresponding to the independence statements ``Gender vs Drink or not?", ``Age vs Drink or not?", and "Total cholesterol vs Drink or not?", we found that the data came from a distribution from the two independence models corresponding to``Age vs Drink or not?" and "Total cholesterol vs Drink or not?" And after applying the backwards selection likelihood-ratio method on the alcohol dataset, we found that the data came from a distribution from the undirected graphical model associated to the complete graph minus the edge {``Total cholesterol”, ``Drink or not?”}.
Show less