Posted in 2020

Biased stochastic approximation

  • 01 December 2020

We explore the convergence of continuous-time ordinary differential equations and their discrete-time analogs, such as stochastic approximation and gradient descent, through the lens of Lyapunov theory [B+98] [LR15]. From this perspective, we will study biased stochastic approximation [KMMW19] where the expectation of the stochastic updates conditioned on the past (which we call the conditional expectation) is not the same as the expectation of the stochastic updates under the stationary distribution (which we call the total expectation).

Read more ...


Machine learning with relative information

  • 23 October 2020

We will reframe some common machine learning paradigms, such as maximum likelihood, stochastic gradients, stochastic approximation and variational inference, in terms of relative information.

Read more ...


Path integrals and continuous-time Markov chains

  • 14 October 2020

We give an introduction to continuous-time Markov chains, and define path measures for these objects.

Read more ...


Motivic relative information

  • 07 October 2020

So far, our definition of relative information studies the divergence between real-valued measures. In this post, we will explore motivic measures which take values more generally in some ring \(R\), and have some fun applying motivic relative information to zeta functions.

Read more ...


Zeta functions, Mellin transforms and the Gelfand-Leray form

  • 05 October 2020

We outline the similarities between zeta functions appearing in number theory and in statistical learning.

Read more ...


Conditional relative information and its axiomatizations

  • 18 September 2020

In this post, we will study the conditional form of relative information. We will also look at how conditional relative information can be axiomatized and extended to non-real-valued measures.

Read more ...


Building foundations of information theory on relative information

  • 08 September 2020

The relative information [BF14] (also known as relative entropy or Kullback-Leibler divergence) is an important object in information theory for measuring how far a probability measure \(Q\) is from another probability measure \(P.\) Here, \(Q\) is usually the true distribution of some real phenomenon, and \(P\) is some model distribution.

Read more ...


Motivic information, path integrals and spiking networks

  • 28 August 2020

I’m writing a series of posts that will explore the connections between these topics. Here is a rough outline of the series, which I will fill in slowly over time.

Read more ...


Processes and variety maximization

  • 07 August 2020

“It’s a theory about processes, about the sequences and causal relations among things that happen, not the inherent properties of things that are. The fundamental ingredient is what we call an “event.” Events are things that happen at a single place and time; at each event there’s some momentum, energy, charge or other various physical quantity that’s measurable. The event has relations with the rest of the universe, and that set of relations constitutes its “view” of the universe. Rather than describing an isolated system in terms of things that are measured from the outside, we’re taking the universe as constituted of relations among events. The idea is to try to reformulate physics in terms of these views from the inside, what it looks like from inside the universe.”

Read more ...


Adjunctions

  • 23 July 2020

Read about the Curry-Howard-Lambek correspondence. Some call it the holy trinity of Logic, Computation and Categories. Lambek adds the “objects as propositions” and “arrows as proofs” part to the mix. You may need to learn some basic category theory.

Read more ...


Directed spaces and types

  • 26 May 2020

Joint work with Jin Xing Lim, Liang Ze Wong, Georgios Piliouras.

Read more ...


Logical frameworks

  • 21 May 2020

Logical frameworks are formal meta-languages for specifying different kinds of object theories (e.g. logical theories and type theories).

logical-framework-rules

Read more ...