Stochastic Process

Bernoulli process : a sequence of independent and identically distributed (iid) random variables. A finite or infinite sequence of binary random variables, so it is a discrete-time stochastic process that takes only two values, canonically 0 and 1.

Random walk : a path that consists of sums of iid random variables or random vectors in Euclidean space. An example of Markov processes.

Wiener process (or Brownian motion process) $W_{t}$ : a stochastic process with stationary and independent increments that are normally distributed based on the size of the increments.

  • Properties:
    • $W_{0} = 0$ almost surely.
    • $W_{t}$ has independent increments.
    • $W_{t}$ has independent Gaussian increments. $(W_{t + u} - W_{t}) \sim \mathcal{N}(0, u)$
  • Brownian Motion with drift : $U_{t} = W_{t} + \mu t$.
  • Geometric Brownian Motion the logarithm of the randomly varying quantity follows a Brownian motion (usually used for non-negative process):
    • $G(t) = G(0)exp[(\mu - \frac{1}{2}\sigma^{2})t + \sigma W(t)]$
    • $log[G(t)] = log[G(0)] + (\mu - \frac{1}{2}\sigma^{2})t + \sigma W(t)$ a brownian motion.
    • $\partial G(t) = (\mu - \frac{1}{2}\sigma^{2})G(t) \partial t + \sigma \partial W(t) = \mu^{*}G(t) \partial t + \sigma \partial W(t)$ a stochastic process following it, is called to follow a GBM.
    • model fitting : estimate $\mu$ and $\sigma$.
  • python simulation.

Poisson process : consists of points randomly located on a mathematical space with the essential feature that the points occur independently of one another.

  • Used for modeling number of events and times at which events occur in a fixed interval of time or space.
    • The number of events occurring is modeled by Poisson distribution.
    • The time between events by exponential distribution.
  • python simulation

Causal Map

Designing Universal Causal Deep Learning Models: The Geometric (Hyper)Transformer 2022

Causal Map : given two metric spaces $\mathcal{X}$ and $\mathcal{Y}$. a causal map $F: \mathcal{X}^{\mathbb{Z}} \to \mathcal{Y}^{\mathbb{Z}}$ is a function which maps discrete-time paths in $\mathcal{X}$ to discrete-time paths in $\mathcal{Y}$ while respecting the causal forward-flow of information in time.

  • In stochastic process - a special case : the universal approximation of a stochastic process’ evolution, conditioned on its realized trajectory.
    • $\mathcal{X} = \mathbb{R}^{d}$ and $\mathcal{Y}$ being a space of laws of a process on a pre-specified number of future steps.
  • GHTs (geometric hyper transformers) can approximate any causal map over any time-horizon. GHTs are consist of :
    • A transformer network gives “encoder” parameters. since transformers has advantages over RNN : (1) avoid recursion; (2) can learn encode before decode into prediction.
    • A small hypernetwork to interpolate “encoder” parameters and predict future steps.