You're reading the public-facing archive of the Category Theory Zulip server.
To join the server you need an invite. Anybody can get an invite by contacting Matteo Capucci at name dot surname at gmail dot com.
For all things related to this archive refer to the same person.
Hi all, I am opening a new thread regarding the conversion of data and measurements into logical statements. One of the main problems I see is that data will not always provide consistent probabilities, which may give non-consistent conclusions for different measurements. This often forces statisticians to ask whether one has enough data.
The idea of p-value encompasses this aspect, in which we ask whether we can believe in the conclusions we inferred from the data, given that we could have continued to collect even more data and have the risk of changing our conclusions.
I think that if one wants to be able to have a sound (probabilistic/statistical/learning) logic for reasoning about the things one observes in the world, one needs to address data consistency within the logic itself (because logic is about assessing consistency too).
Can you clarify what you mean by
One of the main problems I see is that data will not always provide consistent probabilities, which may give non-consistent conclusions for different measurements.
Maybe some examples?
Oliver Shetler said:
Maybe some examples?
Suppose that the data follow a curve from time = 0 to time = 100 and a curve from time = 100 to time = 1000.
If you only collect data from 0 to 100, then your statement will be "data looks like on average", if you decide to collect more data until time = 1000, then your statement will be "data looks like on average" (because is only defined on a small interval). As a result, you can obtain two conclusions and that may be inconsistent with each other.
A more concrete example is when our perception of the data varies -- different beliefs can lead to inconsistencies: see Sleeping Beauty Problem.
This problem interests me a lot. It seems a pretty good use case for sheaves
Sheaves allow you to reason about local information. You are asking about translating these measurements into logical statements. As you point out in your example, these measurements can only carry a logical value that has something to do with the underlying structure of the space you are measuring on (time in your example)
There are two things I'd look into: The work on dynamical systems done by Spivak & co. using temporal logic and the work by Michael Robinson that uses sheaves todo data analysis.
So, my point is that the internal logic of a sheaf generalizes truth values from "true or false" to "true in this open set of your space". I don't know if this is what you want, but for sure it looks to me as the most appropriate setting to represent measurements as logical statements.
I'm using sheaves heavily lately, so I'm happy to collaborate on this if you want to. :slight_smile:
Another important thing: When you have incompatible measurements, you can use sheaf cohomology to get qualitative information about why your measurements are inconsistent with each other. I guess this could be useful as well. :slight_smile:
Rémy Tuyéras said:
Suppose that the data follow a curve from time = 0 to time = 100 and a curve from time = 100 to time = 1000.
If you only collect data from 0 to 100, then your statement will be "data looks like on average", if you decide to collect more data until time = 1000, then your statement will be "data looks like on average" (because is only defined on a small interval). As a result, you can obtain two conclusions and that may be inconsistent with each other.
Yes, clearly your estimator can change with more data, especially in a time series where the parameter may vary over time. Just as clearly, there is often not enough information in the data to anticipate the change, so your problem can't be to "solve" the fact that data evolves over time. I am still unclear on what the problem is, then. Can you try to state your idea in the form of a problem in need of a solution?
Fabrizio Genovese said:
Sheaves allow you to reason about local information. You are asking about translating these measurements into logical statements. As you point out in your example, these measurements can only carry a logical value that has something to do with the underlying structure of the space you are measuring on (time in your example)
So the goal is to look for ways to make "inconsistent" measurements into logically consistent statements?
Fabrizio Genovese said:
Another important thing: When you have incompatible measurements, you can use sheaf cohomology to get qualitative information about why your measurements are inconsistent with each other. I guess this could be useful as well. :)
This sounds very interesting. Do you have a resource that describes this in more detail?
Fabrizio Genovese said:
So, my point is that the internal logic of a sheaf generalizes truth values from "true or false" to "true in this open set of your space". I don't know if this is what you want, but for sure it looks to me as the most appropriate setting to represent measurements as logical statements.
Yes, as far as I uderstand, sheaves & cohomology can be used to detect missing data that can used to complete already existing data in a given sheaf. Any obstruction detected at the sheaf level can tell us about
Oliver Shetler said:
Rémy Tuyéras said:
Suppose that the data follow a curve from time = 0 to time = 100 and a curve from time = 100 to time = 1000.
If you only collect data from 0 to 100, then your statement will be "data looks like on average", if you decide to collect more data until time = 1000, then your statement will be "data looks like on average" (because is only defined on a small interval). As a result, you can obtain two conclusions and that may be inconsistent with each other.
Yes, clearly your estimator can change with more data, especially in a time series where the parameter may vary over time. Just as clearly, there is often not enough information in the data to anticipate the change, so your problem can't be to "solve" the fact that data evolves over time. I am still unclear on what the problem is, then. Can you try to state your idea in the form of a problem in need of a solution?
My point is: how do we incorporate the concept of "being enough (data)" in the logic. Such a concept would push us to be more caucious on the truth values of our logical deductions, and hopefully, allow us to avoid inconsistencies.
I feel like statisticians have tried to address this question through test statistics and p-values. My question is: what do we have on our side?
Did you try by introducing modalities in your logic?
The modalities reflect directly on the topologies you allow on the base space, which is also interesting in this setting
Oliver Shetler said:
Fabrizio Genovese said:
Another important thing: When you have incompatible measurements, you can use sheaf cohomology to get qualitative information about why your measurements are inconsistent with each other. I guess this could be useful as well. :slight_smile:
This sounds very interesting. Do you have a resource that describes this in more detail?
It's not interesting. It's plain dope. Sheaves are heavily used in ACT and yet very underrated considering how much they have to offer. An example of what I say can be found here: https://arxiv.org/abs/2005.12798
Fabrizio Genovese said:
The modalities reflect directly on the topologies you allow on the base space, which is also interesting in this setting
Yes, agree that modalities go in this direction. It would be nice to have a formalism comparing two modalities, so one can decide which is more appropriate to the context, a bit like how it is done in hypothesis testing, when one compares a null hypothesis to .
Well, the "comparison" could exactly be "comparing the underlying topologies given by the modalities" :slight_smile:
Fabrizio Genovese said:
Well, the "comparison" could exactly be "comparing the underlying topologies given by the modalities" :)
That idea sounds interesting. In your case, how do you compare two topologies? could you develop with an example?
I don't have any real example in mind, but regarding topologies as just distributive lattices there are quite a lot of tools we can throw at them
If we are thinking about Grothendieck topologies then I guess we could even order them by inclusion. Dunno how much sense this makes tho
For finite topologies (i.e. finite sets), we can always use the partitioning induced by the topology on the underlying space.
I would personally take it to the category of partitions, in which I would compare the two partitions with categorical tools. That is what I am getting at in this paper where I essentially take two partitions of a space , say and , and I try to compare them through a morphism or .
Of course, you cannot always do that, so one way to do it is to introduce a product term where the extra partition is a "pseudo-universal" object measuring the discrepancy between and .
I am not sure if this can be extended to the infinite case though
Why use the partitioning and not the topology itself?
Just for practical reasoning: I am thinking from a programming viewpoint too.
Note that, in the finite case, you can reduce your topology to a set of disjoint intersections. Then, you can retrieve the unions by taking sequences of partitions . Each arrow in this sequence joins parts of the partitions together to form coarser subdivisions.
Yeah, I am not sure this holds in general
Take the real numbers with the usual topology for instance
Yeah, this does not hold in general I believe. But do we really need this?
True that I'm more comfortable with algebra than with geometry, but I'd try to do comparisons by looking at the following things:
The second one is interesting since may actually have "bigger" open sets than , or the open sets of may be totally unrelated
I'm sure there's plenty of stuff in pointless topology one can check out, btw
Fabrizio Genovese said:
Oliver Shetler said:
Fabrizio Genovese said:
Another important thing: When you have incompatible measurements, you can use sheaf cohomology to get qualitative information about why your measurements are inconsistent with each other. I guess this could be useful as well. :)
This sounds very interesting. Do you have a resource that describes this in more detail?
It's not interesting. It's plain dope. Sheaves are heavily used in ACT and yet very underrated considering how much they have to offer. An example of what I say can be found here: https://arxiv.org/abs/2005.12798
"OPINION DYNAMICS ON DISCOURSE SHEAVES? Whaaaat? Dope really is the only way to describe this paper.
Oliver Shetler said:
Fabrizio Genovese said:
Another important thing: When you have incompatible measurements, you can use sheaf cohomology to get qualitative information about why your measurements are inconsistent with each other. I guess this could be useful as well. :)
This sounds very interesting. Do you have a resource that describes this in more detail?
If the goal is to use sheaf cohomolgy for asking whether different measurements are giving answers which glue together consistently or not, I'd look at https://arxiv.org/abs/1502.03097 and related works.
I may be biased here, but I'm never been a fan of the contextuality work. I feel like the underlying ideas are conceptually easy to understand, but at least to me those papers always obscured what the whole point was about
Actually, I've watched many lectures and presentations about this line of work
And those have been responsible for me constantly feeling too stupid to get sheaf theory.
I don't know, I used to think that "if it doesn't click for me after so many times maybe I'm too stupid to get it". Then I watched one talk by Michael Robinson and I got it immediately. This obviously left me with a strong bias towards contextuality. Even if it's a matter of personal experience, I wouldn't suggest anyone to dive into that stuff to understand how cohomology can help. :slight_smile:
Fabrizio Genovese said:
Then I watched one talk by Michael Robinson and I got it immediately.
This DARPA tutorial and the associated videos? Well, clearly he knew how to design the material for his target audience. :smiley:
In my specific case, I watched a presentation at NIST in 2018, and it was illuminating.
Well, he's got a series of slides and videos on that tutorial site, which looks promising.
Last time I checked some lectures had some audio problems tho
Fair enough. Those are usually relatively minor though. I should learn this material, though, so thanks for the tip.
Fabrizio Genovese said:
I may be biased here, but I'm never been a fan of the contextuality work. I feel like the underlying ideas are conceptually easy to understand, but at least to me those papers always obscured what the whole point was about
I might be biased too, having contributed to some more recent developments in the framework. It's certainly not the source to learn sheaf theory, and in a sense the story is precisely about probabilities failing to be a sheaf: contextuality in this pov is exactly about how you might have a consistent family of probability distributions (i.e. the marginals agree) that fails to admit any gluings at all to a global joint distribution.
Yes, now I get it. But imagine seeing the same damn diagram showing up in any Samson/Kohei talk (you know the diagram I'm talking about I guess, it's everywhere xD) and not getting it. Not. Once.
You mean something like the fibre-bundle picture of the PR box that admits no global sections?
Yes, that one xD
Every time it was like "No, I swear this time I have to understand it. It's almost criminal that I don't at this point" ...5 minutes later "WTFFFFFF"
Here's one big-picture pov on those pictures (not written out in those papers afaik): you have a simplicial complex that gives you measurements and specifies which of them are jointly possible. If you collect data over those measurements, and only remember whether particular (joint) outcomes were possible or not, you can view this as another simplicial complex , which comes with a fibration to - faces lying over a given face of are the possible joint outcomes over that joint measurement. Now, such a thing is strongly contextual if $$\Gamma\to\Sigma$ has no sections, i.e. no simplicial maps such that the composite is . The picture in question is an example of this when and lists the specific possible outcomes of the PR-box.
I mean, they do talk about the fibre-bundle picture, but afaik the published works don't define fibrations of simplicial complexes etc.
And concretely, that says that there is no assignment of values to each variable in a way that agrees with the predictions of the observed data (i.e. is deemed a possible outcome for every joint measurement).