Hermann G. Matthies
Abstract:
Probability theory was axiomatically built on the concept of measure by A. Kolmogorov in the early 1930s, giving the probability measure and the related integral as primary objects and random variables, i.e., measurable functions, as secondary. Not long after Kolmogorov ́s work, developments in operator algebras connected to quantum theory in the early 1940s lead to similar results in an approach where algebras of random variables and the expectation functional are the primary objects. Historically this picks up the view implicitly contained in the early probabilistic theory of the Bernoullis.
Institute of Scientific Computing, TU Braunschweig
Tue, 14/03/2017 - 3:05pm to 3:55pm
RC-4082, The Red Centre, UNSW
This algebraic approach allows extensions to more complicated concepts like non-commuting random variables and infinite dimensional function spaces, as it occurs, e.g., in quantum field theory, random matrices, and tensor-valued random fields. It not only fully recovers the measure-theoretic approach, but can extend it considerably. For much practical and numerical work, which is often primarily concerned with random variables, expectations, and conditioning, it offers an independent theoretical underpinning. In short words, it is “probability without measure theory”.
This functional analytic setting has also strong connections to the spectral theory of linear operators, where analogies to integration are apparent if they are looked for. These links extend in a twofold way to the concept of weak distribution, which describes probability on infinite dimensional vector spaces. Here the random elements are represented by linear mappings, and factorisations of linear maps are intimately connected with representations and tensor products, as they appear in numerical approximations.
Taking this conceptual basis of vector spaces, algebras, linear functionals, and operators gives a fresh view on the concepts of expectation and conditioning, as it occurs in applications of Bayes's theorem.