You're reading the public-facing archive of the Category Theory Zulip server.
To join the server you need an invite. Anybody can get an invite by contacting Matteo Capucci at name dot surname at gmail dot com.
For all things related to this archive refer to the same person.
I'm trying to unpack the definition of Stoch (the Kleisli category of the Giry monad) as a Markov category. My goal is to be able to take any string diagram and be able to express it explicitly in terms of measure-theoretic integrals, in order to prove properties of Stoch. I don't know much measure theory and am trying to learn what I need as I go along.
So far the best resource I've found for unpacking Stoch is Panangaden's 'The Category of Markov Kernels' (1999), which is really helpful but stops short of completely answering my question. The tricky part is in the details of how to reason about the interaction between the monoidal product and composition, as I explain below.
Here is what I know so far:
Objects are measurable spaces, which I'll write as . I'll write etc. for elements of .
Then a morphism is a map , where we write for . This must have the properties that
is a probability measure for every
is a measurable function for every .
Composition is given by
There is some work needed to see that this function has the needed properties and that composition is associative. Panangaden gives an explicit proof of associativity. Identities are Dirac deltas.
The monoidal product is given on objects by , where is the product -algebra. For morphisms and we define
This defines the measure only on elements of that are of the form , but this uniquely extends to a measure on by Carathéodory's extension theorem.
So far so good. The tricky part is in figuring out how composition and this monoidal product relate to each other.
For example, suppose I have morphisms , and , and I want to calculate . The morphism is as defined above, but in order to compose it with I need to be able to integrate against it - I can't see a way to write down as a single equation because of the step where we extend the partially defined measure to a full one.
Ultimately I want to be able to prove that certain morphisms are equal under certain conditions, and this sort of thing is causing a roadblock to that. So I guess my questions are
1) can the morphism be written down explicitly in a nice way?
2) is there somewhere where this explicit unpacking of Stoch is done and written down already?
If these questions get answers I might have more questions as well, but let's see how this goes.
It can't hurt to email Prakash: he's a friendly guy and he'd be happy to know someone is interested in this.
You could write it down as an integral. For all measurable
we have
Hence,
using the notation in https://link.springer.com/book/10.1007/978-0-387-87859-1
Sorry if this is a bit brief, but I think that the explicit formula should simply be
This is basically the same equation as calculating the expectation value of a measurable function with respect to a product measure. BTW I personally prefer writing this as
with the 's at the integration variables, which seems more intuitive.
Tobias Fritz said:
Sorry if this is a bit brief, but I think that the explicit formula should simply be
This is basically the same equation as calculating the expectation value of a measurable function with respect to a product measure.
Thank you, I'm sure this is right. (And brevity is good.) I was confused because I was actually trying to evaluate for some , and I was stuck at trying to integrate against . The latter is a measure defined on a product space but it isn't a product measure, so I couldn't do this.
But your comment made me realise I can just evaluate it as instead and then there's no problem. Maybe this will always work - I can always just evaluate things from back to front and then I'll always either be integrating against a product measure or against a measure that's fully defined instead of partially defined and then extended. I'll try it on my real problem and see if it works.
BTW I personally prefer writing this as
with the 's at the integration variables, which seems more intuitive.
I write it the way I do because I think of as a souped-up (applying a 'dual vector' to a 'vector'), where the is a dummy variable and the is really just the comma. But I can see the point of the other notation and will give it a try - it might be less confusing in the end.
Richard Samuelson said:
You could write it down as an integral. For all measurable
we have
Hence,
using the notation in https://link.springer.com/book/10.1007/978-0-387-87859-1
This makes me start to like the notation actually - I like being able to write it implicitly like this. But I think we can only get rid of one of the integral signs, because we still have to integrate over and . So I'd write this as
which is basically what Tobias Fritz wrote only with in place of .
I have one more question, which is as much a notation question as anything else. Suppose I have and . What's the right way to write down their composite?
It feels like I should be able to write it as
but I'm a bit uneasy about writing and , because the space being integrated over is , not and separately. Is there a formal argument that allows it to be written this way, or should it really be something like
instead?
Without a formal syntax in place, I'd say that both of those notations make sense since they express what you mean clearly enough. You could also write it as a double integral (by Fubini's theorem), and/or write in the argument of in order to express the idea that one sums up the measures of tiny rectangles.
I guess the semantics of these things is what I'm trying to grasp towards. Or at least I want to be sure I know what I mean when I write these things down.
Can I really use Fubini's theorem here? I thought I could only use that in the case where I'm integrating over a product measure. If I write
then it looks to me like I need to evaluate something like
and then integrate that over .
If that's right then what's the name of this kind of partial integral? I'm not getting much luck trying to google things like "integrating over one variable of a joint measure".
The same thing comes up in this situation: suppose I have and , and I want to calculate . I end up with
It seems like I should be able to integrate out to get
but then I'm back to integrating only over even though is a measure over , and I don't know what it means formally.
Urgh, sorry, I'm a bit out of it today -- please ignore my reference to Fubini, you're right that it's not that.
Nathaniel Virgo said:
Richard Samuelson said:
You could write it down as an integral. For all measurable
we have
Hence,
using the notation in https://link.springer.com/book/10.1007/978-0-387-87859-1
This makes me start to like the notation actually - I like being able to write it implicitly like this. But I think we can only get rid of one of the integral signs, because we still have to integrate over and . So I'd write this as
which is basically what Tobias Fritz wrote only with in place of .
My original post was incorrect (I just edited it). I believe that the expression can be written
without integral signs, and where is to the right of and . For example, see page 44 of Cinlar, Probability and Stochastics: