91

In category-theoretic discussions, there is often the temptation to look at the category of all abelian groups, or of all categories, etc., which quickly leads to the usual set-theoretic problems. These are often avoided by using Grothendieck universes. In set-theoretic language, one fixes some strongly inaccessible cardinal $\kappa$ -- this means that $\kappa$ is some uncountable cardinal such that for all $\lambda<\kappa$, also $2^\lambda<\kappa$, and for any set of $<\kappa$ many sets $S_i$ of size $<\kappa$, also their union is of size $<\kappa$. This implies that the stage $V_\kappa\subset V$ of "sets of size $<\kappa$" is itself a model of ZFC -- by applying any of the operations on sets, like taking powersets or unions, you can never leave $V_\kappa$. These sets are then termed "small", and then the category of small abelian groups is definitely well-defined.

Historically, this approach was first used by Grothendieck; a more recent foundational text is Lurie's work on $\infty$-categories. However, their use has always created somewhat of a backlash, with some people unwilling to let axioms beyond ZFC slip into established literature. For example, I think at some point there was a long discussion whether Fermat's Last Theorem has been proved in ZFC, now settled by McLarty. More recently, I've seen similar arguments come up for theorems whose proofs refer to Lurie's work. (Personally, I do not have strong feelings about this and understand the arguments either way.)

On the other hand, it has also always been the case that a closer inspection revealed that any use of universes was in fact unnecessary. For example, the Stacks Project does not use universes. Instead, (see Tag 000H say) it effectively weakens the hypothesis that $\kappa$ is strongly inaccessible, to something like a strong limit cardinal of uncountable cofinality, i.e.: for all $\lambda<\kappa$, one has $2^\lambda<\kappa$, and whenever you have a countable collection of sets $S_i$ of size $<\kappa$, also the union of the $S_i$ has size $<\kappa$. ZFC easily proves the existence of such $\kappa$, and almost every argument one might envision doing in the category of abelian groups actually also works in the category of $\kappa$-small abelian groups for such $\kappa$. If one does more complicated arguments, one can accordingly strengthen the initial hypothesis on $\kappa$. I've had occasion to play this game myself, see Section 4 of www.math.uni-bonn.de/people/scholze/EtCohDiamonds.pdf for the result. From this experience, I am pretty sure that one can similarly rewrite Lurie's "Higher Topos Theory", or any other similar category-theoretic work, in a way to remove all strongly inaccessible cardinals, replacing them by carefully chosen $\kappa$ with properties such as the ones above.

In fact, there seems to be a theorem of ZFC, the reflection principle (discussed briefly in Tag 000F of the Stacks project, for example), that seems to guarantee that this is always possible. Namely, for any given finite set of formulas of set theory, there is some sufficiently large $\kappa$ such that, roughly speaking, these formulas hold in $V_\kappa$ if and only if they hold in $V$. This seems to say that for any given finite set of formulas, one can find some $\kappa$ such that $V_\kappa$ behaves like a universe with respect to these formulas, but please correct me in my very naive understanding of the reflection principle! (A related fact is that ZFC proves the consistency of any given finite fragment of the axioms of ZFC.)

On the other hand, any given mathematical text only contains finitely many formulas (unless it states a "theorem schema", which does not usually happen I believe). The question is thus, phrased slightly provocatively:

Does the reflection principle imply that it must be possible to rewrite Higher Topos Theory in a way that avoids the use of universes?

Edit (28.01.2021): Thanks a lot for all the very helpful answers! I think I have a much clearer picture of the situation now, but I am still not exactly sure what the answer to the question is.

From what I understand, (roughly) the best meta-theorem in this direction is the following (specialized to HTT). Recall that HTT fixes two strongly inaccessible cardinals $\kappa_0$ and $\kappa_1$, thus making room for small (in $V_{\kappa_0}$), large (in $V_{\kappa_1}$), and very large (in $V$) objects. One can then try to read HTT in the following axiom system (this is essentially the one of Feferman's article "Set-theoretic foundations of category theory", and has also been proposed in the answer of Rodrigo Freire below).

(i) The usual ZFC axioms

(ii) Two other symbols $\kappa_0$ and $\kappa_1$, with the axioms that they are cardinals, that the cofinality of $\kappa_0$ is uncountable, and that the cofinality of $\kappa_1$ is larger than $\kappa_0$.

(iii) An axiom schema, saying that for every formula $\phi$ of set theory, $\phi\leftrightarrow \phi^{V_{\kappa_0}}$ and $\phi\leftrightarrow \phi^{V_{\kappa_1}}$.

Then the reflection principle can be used to show (see Rodrigo Freire's answer below for a sketch of the proof):

Theorem. This axiom system is conservative over ZFC. In other words, any theorem in this formal system that does not refer to $\kappa_0$ and $\kappa_1$ is also a theorem of ZFC.

This is precisely the conclusion I'd like to have.

Note that $V_{\kappa_0}$ and $V_{\kappa_1}$ are models of ZFC, but (critically!) this cannot be proved inside the formal system, as ZFC is not finitely axiomatizable, and only each individual axiom of ZFC is posited by (iii).

One nice thing about this axiom system is that it explicitly allows the occasional arguments of the form "we proved this theorem for small categories, but then we can also apply it to large categories".

A more precise question is then:

Do the arguments of HTT work in this formal system?

Mike Shulman in Section 11 of https://arxiv.org/abs/0810.1279 gives a very clear exposition of what the potential trouble here is. Namely, if you have a set $I\in V_{\kappa_0}$ and sets $S_i\in V_{\kappa_0}$ for $i\in I$, you are not allowed to conclude that the union of the $S_i$ is in $V_{\kappa_0}$. This conclusion is only guaranteed if the function $i\mapsto S_i$ is also defined in $V_{\kappa_0}$ (or if $I$ is countable, by the extra assumption of uncountable cofinality). In practice, this means that when one wants to assert that something is "small" (i.e. in $V_{\kappa_0}$), this judgment pertains not only to objects, but also to morphisms etc. It is not clear to me now how much of a problem this actually is, I would have to think more about it; I might actually imagine that it is quite easy to read HTT to meet this formal system. Shulman does say that, with this caveat, the adjoint functor theorem can be proved, and as Lurie says in his answers, the arguments in HTT are of similar set-theoretic complexity. However, I'd still be interested in a judgment whether the answer to the question is "Yes, as written" or rather "Probably yes, but you have to put some effort in" or in fact "Not really". (I sincerely hope that the experts will be able to agree on roughly where the answer falls on this spectrum.)

A final remark: One may find the "uncountability" assumption above a bit arbitrary; why not allow some slightly larger unions? One way to take care of this is to add a symbol $\kappa_{-1}$ with the same properties, and ask instead that the cofinality of $\kappa_0$ is larger than $\kappa_{-1}$. Similarly, one might want to replace the bound $\mathrm{cf} \kappa_1>\kappa_0$ by a slightly stronger bound like $\mathrm{cf} \kappa_1>2^{\kappa_0}$ say. Again, if it simplifies things, one could then just squeeze another $\kappa_{1/2}$ in between, so that $\mathrm{cf} \kappa_{1/2}>\kappa_0$ and $\mathrm{cf} \kappa_1>\kappa_{1/2}$. This way one does not have to worry whether any of the "standard" objects that appear in some proofs stay of countable size, or whether one can still take colimits in $V_{\kappa_1}$ when index sets are not exactly of size bounded by $\kappa_0$ but have been manipulated a little.

PS: I'm only now finding all the relevant previous MO questions and answers. Some very relevant ones are Joel Hamkins' answers here and here.

Peter Scholze
  • 19,800
  • 3
  • 98
  • 118
  • 7
    I might be missing something, but the setting you are proposing seems to me a bit uncomfortable to do category theory (and higher category theory): with this type of definition of "small" the category of small set no longer have small colimits. If you want to have a category of small set whose class objects is a set, which has small limits and small colimits and which is indeed a full subcategory of the category of sets, then its set of object is an inaccessible cardinal. – Simon Henry Jan 27 '21 at 00:17
  • 3
    What could cause a problem is if HTT ever quantifies over arbitrary subsets of $V_{\kappa},$ e.g. if it quantifies over what $V_{\kappa}$ considers to be large categories. Note that $\kappa$ is inaccessible iff $V_{\kappa+1}$ satisfies the finitely axiomatizable theory NBG, so once you move up to the next layer of sets, there are sentences which distinguish inaccessible cardinals from those guaranteed by reflection. – Elliot Glazer Jan 27 '21 at 00:27
  • 22
    Have you read Feferman's paper “Set-theoretical foundations of category theory"? He describes a modification of ZFC that does this sort of "reflection" automatically in the background. I wrote a bit about this idea myself in https://arxiv.org/abs/0810.1279 and https://golem.ph.utexas.edu/category/2009/11/feferman_set_theory.html. – Mike Shulman Jan 27 '21 at 01:08
  • One point which might be relevant but which is unclear to me: is the statement "for every statement $\varphi$ in the language of set theory, there is an ordinal $\alpha$ such that $\varphi^{V_\alpha} \leftrightarrow \varphi$" a theorem of ZFC? Or is it only a theorem schema, i.e. for each statement $\varphi$ in the language of set theory "there is an ordinal $\alpha$ such that $\varphi^{V_\alpha} \leftrightarrow \varphi$" is a theorem of ZFC? – Tim Campion Jan 27 '21 at 03:00
  • 9
    @TimCampion: It's a theorem schema. If it were a theorem of ZFC, I think that would violate undefinability of truth. – Nate Eldredge Jan 27 '21 at 03:51
  • 1
    Maybe it's worth mentioning that the idea "the use of Grothendieck universes in category theory is always inessential" has been put to the test in a serious way in the study of the metamathematics of Fermat's Last Theorem. In "What does it take to prove Fermat's Last Theorem, a quick search doesn't seem to reveal McLarty appealing to the reflection principle. When McLarty actually carries out the reduction, he doesn't even bother stopping at ZFC, jumping straight to reducing to higher-order arithmetic. – Tim Campion Jan 27 '21 at 04:46
  • 1
    Is there anything particularly interesting about HTT? Consider the question "Does the reflection principle imply that X can be rewritten without universes?" for any of X∈{HTT,SGA,FLT}. There is one substantive difference I see -- HTT and SGA are foundational in character, and have plenty of open-ended statements into which all sorts of cardinal parameters would have to be inserted to be useful, whereas for FLT, one can just choose the minimal values of these parameters once and for all. But otherwise, I think the existing literature on FLT ought to be a pretty good guide. – Tim Campion Jan 27 '21 at 05:20
  • 5
    The discussion here may be relevant. – Noah Schweber Jan 27 '21 at 06:06
  • I think Timothy Chow's comment hits the nail on its head: 'if you want a meta-theorem that "automatically" eliminates universes, then my understanding is that there isn't one [...] because practitioners don't write in a way that makes it easy to formulate a suitable meta-theorem.' Consider, for example, Lefschetz's principle in algebraic geometry. The fact that the first-order theory of algebraically closed fields of characteristic 0 and infinite absolute transcendence degree is complete yields a precise form of this principle, but it is not user-friendly and, I think, not used in practice. – Zhen Lin Jan 27 '21 at 07:13
  • Thanks for all the relevant pointers! I agree that the setting is slightly less convenient, but actually not really much so: Reading the Stacks Project, you will not find that you are unnecessarily burdened with set-theoretic problems. I personally have tried to write my papers in a way that does not make explicit use of universes, and found that this was always rather easy to achieve (without invoking or actually knowing the reflection principle). Later I learned about the reflection principle, and then wondered whether this is, in a sense, a formalization of this experience. – Peter Scholze Jan 27 '21 at 09:20
  • To make the question more concrete: Theorem 6.3.5.13 is a concrete assertion, for whose proof however Lurie goes to some auxiliary larger universe. I found this application of universes very disconcerting when I read HTT, and I would very much guess that the argument can be rewritten to avoid the use of a universe. To convince myself of this, do I have to sit down and check line by line that it works for a suitable $\kappa$, the conditions on which I need to carefully note down; or does the reflection theorem tell me that, as long as Lurie does not invoke a whole axiom schema or so, it's good? – Peter Scholze Jan 27 '21 at 09:24
  • 1
    (There may be a subtle distinction here between giving a proof, and giving a (meta?)proof of the existence of a proof; I'm only asking for the latter, in case this makes a difference.) – Peter Scholze Jan 27 '21 at 10:13
  • 1
    I haven't read HTT in full detail, but it always seemed to me that it was written in such a way as to avoid universal statements such as "for all cardinals", and in particular statements such as "blah is co/complete" followed from specific cardinality statements, "blah is $\kappa$-cocomplete". In particular I think for category theory, since (afaik) co/completeness is the biggest place where relevant size issues appear, they can always be bypassed by analyzing the proof and seeing that "oh, actually if I only need $\kappa$-completeness, I don't have to use universes" – Maxime Ramzi Jan 27 '21 at 11:46
  • 2
    Here is also a very insightful answer of Joel Hamkins to basically the same question. For some reason I didn't find it before. So sorry for a duplicate question! And thanks for the many really helpful answers! – Peter Scholze Jan 27 '21 at 23:52
  • 1
    Another question worth linking to might be: https://mathoverflow.net/questions/365947/when-size-matters-in-category-theory-for-the-working-mathematician – Sam Hopkins Jan 28 '21 at 04:08
  • 4
    We use the real numbers and infinite dimensional Hilbert spaces in physics and other sciences and engineering not because it's really needed (the observations and measurements we make in real life are always very very very finite), but because it simplifies a lot of things. Mathematicians tend to shun annoying logic where we need to carefully make sure that all our formulas are suitable for our purposes (this is one of the reasons Quine's NF never fully caught on). This here is a real nice example for large cardinals alleviating this issue as well. It saves time and effort on bookkeeping. – Asaf Karagila Jan 28 '21 at 14:03
  • @AsafKaragila I'm very sympathetic to just assuming universes, it definitely makes for a cleaner reading. But it's also fair to keep track of the axioms used in the proofs, and trying to achieve some consistency (no pun intended ;-)). So I definitely appreciate that the Stacks project goes the extra mile to avoid universes. – Peter Scholze Jan 28 '21 at 14:46
  • 1
    @Peter: I agree that some effort to keep track of assumptions should be made. But it's a lot easier to have a statement like "every such and such statement can be proved in a bounded fragment", and not resorting to saying things like $\Sigma_{15}$ should be enough, that will then force everyone to start counting quantifiers all of a sudden. There's a reason we assume the axiom of choice, and there's a reason we don't really work in second-order arithmetic. To use the Dead Kennedys' album title, "give me convenience or give me death". – Asaf Karagila Jan 28 '21 at 14:50
  • @AsafKaragila Sure! But what I'm actively looking for is the meta-theorem "every such and such statement can be proved in (a bounded fragment of) ZFC". I don't care about $\Sigma_{15}$ and clearly nobody should compute that $15$, but I care whether the existence of $15$ implies that such a meta-theorem can be applied. Currently, it seems to me that the answer is no. This is different than for the axiom of choice, where Gödel proved that it is consistent, so we may as well assume it if we are trying to prove something about the integers, say. – Peter Scholze Jan 28 '21 at 14:55
  • 6
    Yes, I understand. I was merely trying to give motivation as to why we should be more relaxed (for lack of a better term) about large cardinals in our assumptions. I tend to agree with the idea that people mistrust LCA because they haven't really worked with them, and they don't understand them properly. This can be easily remedied, all it requires is that (1) people around the world stop saying that set theory is dead weight and will soon be over; (2) people start hiring more set theorists; (3) people learn about set theory from actual expert and from the very beginning. Everyone's a winner. – Asaf Karagila Jan 28 '21 at 15:18
  • PS: If moderators would rather like me to make the Edit into a separate question, please let me know. – Peter Scholze Jan 28 '21 at 23:03
  • 1
    If you're bringing more related questions into this, I think this might also be relevant. – Asaf Karagila Jan 29 '21 at 11:03
  • Given that the real numbers are idealisations, I really don't see the problem of postulating large cardinals such as an inaccessible cardinal, so long as it's shown consistent with set theory. – Mozibur Ullah Feb 03 '21 at 05:40
  • @MoziburUllah "as long as it's shown consistent with set theory." - No, it is even impossible (to prove in ZFC), unless ZFC is inconsistent: https://en.wikipedia.org/wiki/Inaccessible_cardinal#Models_and_consistency – Z. M Sep 20 '21 at 17:43
  • @Z.M.: I didn't mean consistency in the mathematical sense. – Mozibur Ullah Sep 20 '21 at 17:49

10 Answers10

32

I'm going to go out on a limb and suggest that the book HTT never uses anything stronger than replacement for $\Sigma_{15}$-formulas of set theory. (Here $15$ is a randomly chosen large number, and HTT is randomly chosen math book that is not specifically about set theory.)

Jacob Lurie
  • 17,538
  • 4
  • 74
  • 77
  • Thanks for the answer Jacob! And sorry for referencing HTT in the question ;-).

    At the set theorists around: would this assessment, if true, imply that the answer to the question is "Yes"?

    – Peter Scholze Jan 27 '21 at 12:29
  • Cheeky question: from what probability distribution was $\Sigma_{15}$ drawn? After all, if we bound the complexity of formulas in HTT by, say, the number of characters in the book, then we get an absurd bound like $\Sigma_{10^5}$, and we should pad by a few orders of magnitude to allow for formalization into ZFC. – Tim Campion Jan 27 '21 at 14:45
  • Now, I'd be willing to buy that after formalization into a finite definitional extension of ZFC, no formula of complexity higher than, say, $\Sigma_5$ appears in HTT. And I'm willing to buy that no definition in HTT has complexity higher than, say, $\Sigma_5$ (and probably $\Sigma_1$ or $\Sigma_2$) relative to previously-made definitions. So to be slightly more careful, my worry would be that these complexities might "stack", with complexities adding up to something bigger than (say) $\Sigma_{15}$ after compiling out definitions. How worried should I be about this? – Tim Campion Jan 27 '21 at 14:45
  • 2
    I'm reasonably sure that I can't even parse a $\Sigma_{15}$-sentence (due to human limitations), let alone use one in a proof. You could probably replace $15$ by a much smaller number (like $1$ or $2$), but I haven't thought about it. I did think at some point about how much you could do using just KP, and convinced myself that in addition to obvious problems (like inability to talk about categories of functors) you will also run into some non-obvious problems (having to do with the inability to close up inductive definitions). I think these mostly go away if you have power sets though. – Jacob Lurie Jan 27 '21 at 15:23
  • Thanks! I was worrying that if $\phi$ is a low-complexity formula in a definitional extension $ZFC[\psi_0,\dots,\psi_n]$, the translation $\bar \phi$ of $\phi$ into ZFC might have much higher complexity. But I guess your point is that when reasoning about $\phi$, the "complexity" (in the informal sense) of our considerations about $\phi$ is highly sensitive to the complexity (in the formal sense) of $\bar \phi$, and not just the complexity (in the formal sense) of $\phi$ (unless perhaps the definitions $\psi_0,\dots,\psi_n$ were poorly chosen and add "spurious" additional complexity). – Tim Campion Jan 27 '21 at 15:43
  • A side note: I assume that that the limb you're sitting on has a branch which says that HTT also doesn't use anything higher than $\Sigma_{15}$ separation, in addition to $\Sigma_{15}$ replacement? I understand that both need to be restricted for the reflection principle to apply. – Tim Campion Jan 27 '21 at 15:45
  • Outside of set theory, I don't think you will encounter many mathematical definitions that require quantifying over the entire universe of sets. (Almost all quantifiers that appear are bounded in the sense that you are quantifying over elements of some structure which is already under discussion.) – Jacob Lurie Jan 27 '21 at 15:47
  • 2
    (Corrected version of earlier comment): ZFC allows me to construct cardinals $\kappa$ with the property that, for every theorem $\varphi$ of "ZC + $\Sigma_{15}$-replacement", ZFC will prove that $V_{\kappa}$ satisfies $\varphi$. (With the caveat that the proof is not uniform until one bounds the complexity of comprehension instances used in the proof of $\varphi$.) – Jacob Lurie Jan 27 '21 at 16:23
  • 1
    So if I understand it right, the only thing to verify is that only a bounded amount of replacement is used (which for any given book is little extra effort to check when one actually reads the proofs). Given this, there is also necessarily only a bounded amount of comprehension used, and at this point one could go back to the beginning and replace "universe" with "large $\kappa$ depending on the complexity of the proofs I just read", and arrive at a ZFC proof. So a proof using universes is just a blueprint for a proof without universes. – Peter Scholze Jan 27 '21 at 16:40
  • 1
    Peter: With a little bit less confidence, I'll say that I think the amount of "essential" set theory in my book is roughly the same as the amount needed for Grothendieck's construction of injective resolutions/Quillen's small object argument (which involves transfinite constructions but doesn't actually need replacement). I think that's actually strictly less than the amount needed for your work with Dustin (probably some free resolutions which come up there can't be formalized in ETCS?) – Jacob Lurie Jan 27 '21 at 17:01
  • 14
    With, say, $\Sigma_2$-replacement plus the existence of a universe (itself a $\Sigma_2$ sentence), one can prove statements (of very low quantifier complexity) that are not provable in ZFC; for example, the statement "ZFC is consistent." So if you are looking for a precise metatheorem, it is not enough to know that only $\Sigma_{15}$-formulas appear in HTT. You need to know that instead of applying universes, you are really applying "$\Sigma_{15}$-universes." The point is, it's possible to invoke the second-order replacement axiom for universes without using high complexity formulas. – Gabe Goldberg Jan 27 '21 at 19:48
  • @GabeGoldberg Thanks!! I think I need to understand better what the "second-order" replacement axiom is. Is this implicitly about arbitrarily large "first-order" versions of replacement, but hidden in a quantification over formulas? If so, it would really be that quantification over formulas that's the problem? – Peter Scholze Jan 27 '21 at 20:25
  • 6
    Gabe: Yes, one does not need anything better than "$\Sigma_{15}$-universes" (and actually far far less). The whole "universe" business is just to avoid explicitly spelling out the (much weaker) closure properties that cardinals need to have to make certain arguments go through. – Jacob Lurie Jan 27 '21 at 20:40
  • 5
    The second-order replacement axiom for $U$ states that for all $I\in U$, the union of any $I$-indexed family of sets in $U$ is an element of $U$. This is $\Pi_1$ with parameter $U$. Quantifier complexity alone cannot tell you whether you exceed ZFC. One needs to be more careful about how universes are used. One option is to assert a proper class of "$\Sigma_{15}$-universes," where $U$ is a $\Sigma_{15}$-universe if it is closed under unions of $I$-indexed families ($I\in U$) that are $\Sigma_{15}$-definable with parameters in $U$ (plus the other universe axioms). This follows from ZFC. – Gabe Goldberg Jan 27 '21 at 22:34
  • 3
    Thanks a lot for the explanation! One further question (sorry for cluttering this comment thread...): If $V_\kappa$ is a $\Sigma_{15}$-universe, does it follow that $\kappa$ has uncountable cofinality? – Peter Scholze Jan 27 '21 at 23:38
  • 5
    No, that's a good point (the least such $\kappa$ has countable cofinality). Maybe a more useful principle is: for every set $I$, there is a $\Sigma_{15}$-universe $U$ over $I$, meaning that $I\in U$ and $U$ satisfies second-order replacement with index set $I$. If $I = \mathbb N$, one gets uncountable cofinality. This is also a consequence of ZFC. – Gabe Goldberg Jan 28 '21 at 00:01
  • "roughly the same as the amount needed for Grothendieck's construction of injective resolutions" <-- Grothendieck's requirements are actually less than ZFC, as it turns out: https://arxiv.org/abs/1102.1773 (added: oh, I see @Tim pointed this out in comments to the original question) – David Roberts Jan 28 '21 at 00:04
27

Reflecting on Gabe's comment on my original answer, I now think what I wrote is misleading because it conflates two separate (but related) assertions:

  1. The existence of strongly inaccessible cardinals is not really needed in category theory.

  2. The full strength of ZFC is not really needed in category theory.

I agree with both of these statements, but think that the best way to convince someone of 1) would not be to combine 2) with a reflection principle: that is, one shouldn't try to replace the use of a strongly inaccessible cardinal $\kappa$ by one for which $V_{\kappa}$ models a large part of ZFC.

As I see it, the "problem" which universes solve is to justify the combination of two types of reasoning:

A) It is sometimes useful to prove theorems about small categories $\mathcal{C}$ by embedding them into "big" categories (for example, using the Yoneda embedding) which have nice additional features: for example, the existence of limits and colimits.

B) Big categories are also categories, so any theorem which applies to categories in general should also apply to big categories.

If you were worried only about B), then a reflection principle could be relevant. Choosing a cardinal $\kappa$ such that $V_{\kappa}$ satisfies a big chunk of ZFC, you can redefine "small category" to mean "category belonging to $V_{\kappa}$" and "big category" to mean "category not necessarily belonging to $V_{\kappa}$", and you can feel confident that all the basic theorems you might want are valid in both cases.

But if you're also worried about A), then this isn't necessarily helpful. Say you start with a category $\mathcal{C}$ belonging to $V_{\kappa}$, and you want some version of the Yoneda embedding. A natural guess would be to embed into the category of functors from $\mathcal{C}^{\mathrm{op}}$ to the category of sets of size $<\tau$ (or some equivalent model of it), for some cardinal $\tau$. A first guess is that you should take $\tau = \kappa$, but I think this only makes sense $\kappa$ is strongly inaccessible (otherwise some Hom sets will be too big). In any case guarantee that this construction has good properties, you're going to want to demand different properties of the cardinal $\tau$. For example, if you want the this category of presheaves to have lots of colimits, then you're going to want $\tau$ to have large cofinality. And if you start thinking about what kinds of additional assumptions you might need to make, you're back where you started: thinking about what kind of cardinality estimates guarantee that "presheaves of sets of size $< \tau$" are a good approximation to the category of all presheaves of sets. So the reflection principle doesn't really help you to avoid those issues.

(Edit: I realized after writing that the text below is mostly reiterating Peter's original post. But I'll leave it here in case anyone finds it useful.)

If you want a rigorous formalization in something like ZFC, probably the best thing to do is to do away with big categories altogether. So then B) is a non-issue. To deal with A), let me remark that many of the "big" categories that one would like to talk about arise in a particular way: one starts with a small category $\mathcal{C}$ which has certain kinds of colimits already, and formally enlarges $\mathcal{C}$ to make a bigger category $\mathcal{C}^{+}$ which has arbitrary colimits (without changing the ones that you started with). Categories which arise this way are called locally presentable, and there's a simple formula for $\mathcal{C}^{+}$: it's the category of functors $F: \mathcal{C}^{\mathrm{op}} \rightarrow \mathrm{Set}$ which preserve the limits that you started with (that is, the colimits which you started with in $\mathcal{C}$).

Now, if you want to mimic this in the world of small categories, you could instead pick some cardinal $\kappa$ and instead contemplate functors $F: \mathcal{C}^{\mathrm{op}} \rightarrow \{ \text{Sets of size < $\kappa$} \}$, which is equivalent to a small category $\mathcal{C}^{+}_{\kappa}$. The question you meet is whether this is a good enough replacement for the big category $\mathcal{C}^{+}$ above. For example, does it have lots of limits and colimits? It's unreasonable to ask for it to have all colimits, but you could instead ask the following:

Q) Does the category $\mathcal{C}^{+}_{\kappa}$ have colimits indexed by diagrams of size $< \kappa$?

The answer to Q) is "no in general, but yes if $\kappa$ is nicely chosen". For example, if you have some infinite cardinal $\lambda$ bounding the size of $\mathcal{C}$ and the number of colimit diagrams that you start with, then I believe you can guarantee (i) by taking $\kappa = (2^{\lambda})^{+}$ (and the category $\mathcal{C}^{+}_{\kappa}$ can be characterized by the expected universal property). Moreover, to prove this you don't need any form of replacement.

Now you could also ask the following:

Q') Does the category $\mathcal{C}^{+}_{\kappa}$ have limits indexed by diagrams of size $< \kappa$?

Here the answer will usually be "no" unless $\kappa$ is strongly inaccessible. But if you're only interested in limits of a particular type (for example, if you're studying Grothendieck topoi, you might be especially interested in finite limits), then the answer will again "yes for $\kappa$ well chosen". And this is something you can prove using very little of ZFC.

Now my claim is that, based on my experience, the above discussion is representative of the kind of questions that you'll run into trying to navigate the distinction between "small" and "large" categories (certainly it's representative of the way these things come up in my book, which the original question asked about). In practice, you never need to talk about the entirety of a big category like $\mathcal{C}^{+}$; it's enough to build a large enough piece of it (like $\mathcal{C}^{+}_{\kappa}$) having the features that you want to see, which you can arrange by choosing $\kappa$ carefully.

I find it conceptually clearer to ignore the issue of how things are formalized in ZFC and to phrase things in terms of the "big" category $\mathcal{C}^{+}$, referring back to its "small" approximations $\mathcal{C}^{+}_{\kappa}$ only as auxiliaries in a proof (which will inevitably still turn up somewhere!). The invocation of "universes" is just a way to write like this while still paying lip service to the axiomatic framework of ZFC, and is definitely inessential.

Jacob Lurie
  • 17,538
  • 4
  • 74
  • 77
  • 1
    I think the related work of Zhen Lin Low is worth pointing out here: Universes for category theory https://arxiv.org/abs/1304.5227, which addresses the issue about what happens if you enlarge the cardinal $\kappa$ (you don't want things changing as you go up a universe). The paper works in Mac Lane set theory, which is weaker than ZFC, and does work with universes, but one could ask for similar results with a weaker assumption on $\kappa$ than it being strongly inaccessible. – David Roberts Jan 28 '21 at 05:50
  • When you say the full strength of ZFC is not needed, do you mean that any particular theorem or application can be formalized with less strength, or that the entirety of the theory, including all abstract lemmas and subsuming all possible applications, can actually be proved in a certain finite fragment of ZFC? – Monroe Eskew Jan 28 '21 at 08:46
  • 3
    Thanks Jacob! So the answer to the question posed seems to be "No", and the way to verify that things can be formalized without universes is to proceed as in the Stacks project, and as you suggest in your answer: Just pay a bit closer attention to the required assumptions. I'd still have imagined that there could be a good meta-theorem here that shows that this is always possible as long as you are "not actively looking for trouble", but it seems the reflection theorem per se does not do it, as it generally produces $\kappa$ of low cofinality, while sometimes it needs to be a bit bigger. – Peter Scholze Jan 28 '21 at 09:00
  • 3
    @Peter The reflection theorem gives you a closed unbounded class of such $\kappa$, so any particular cofinality you need is available. – Monroe Eskew Jan 28 '21 at 10:40
  • 3
    @MonroeEskew: Sure! I just meant that a blanket appeal to reflection can't be good enough, because such a blanket appeal wouldn't ensure any cofinality bound. So you need to also figure out which cofinality you need, and then you're back to where you started: Having to actually look at the arguments in some detail and specify which hypotheses you need. – Peter Scholze Jan 28 '21 at 11:10
  • 1
    @MonroeEskew: Maybe "category theory" is too broad, let's replace it by "the contents of my book". I feel 100% confident of the assertion of my first answer in the following form: every assertion in my book that concerns small $\infty$-categories can be proven in ZFC without anything stronger that $\Sigma_{15}$-replacement. Having thought about it a bit, I'm now tempted to guess that you wouldn't need any form of replacement, but I'm less sure of that. – Jacob Lurie Jan 28 '21 at 13:14
  • @PeterScholze To phrase it differently: maybe one should think there are two cardinal parameters: $\kappa$ (bounding the size/complexity of small categories from above) and $\tau$ (bounding the size of certain big categories from below). A reflection principle guarantees you that you can choose $\kappa$ so that the small categories have expected properties (closed under operations, etcetera). But for properties you care about this is probably obvious to begin with, and don't require a careful choice of $\kappa$. The cardinality estimates appearing in the stacks project (and elsewhere) are... – Jacob Lurie Jan 28 '21 at 14:29
  • 1
    @PeterScholze ... more about choosing $\tau$ appropriately; roughly speaking it needs to be "large enough compared to $\kappa$" (to have a tight dictionary between locally presentable categories and small categories that approximate them, say). If $\kappa$ is strongly inaccessible you can take $\tau=\kappa$, and then think only about one parameter. And I'd argue that this is the main feature that you want: the fact that this also guarantees that $V_{\kappa}$ models ZFC is a red herring. – Jacob Lurie Jan 28 '21 at 14:35
  • @JacobLurie Well, in my experience I bound $\kappa$ (bounding the size of the rings that I allow to build schemes later on, or of profinite sets to build condensed sets, or...), but I usually do not put any size bound on the sheaves: I allow objects in my categories to form a proper class (but morphisms are a set). This seems in slight conflict with what an $\infty$-category is, where objects and morphisms are less cleanly separated. But I guess I'd also be fine with just taking $\tau$-small sheaves for some $\tau$ with cofinality $>\kappa$ or so. – Peter Scholze Jan 28 '21 at 14:37
  • @JacobLurie However, in my experience I also want the cutoff $\kappa$ to be of uncountable cofinality, because occasionally I want to take countable colimits of rings to achieve certain closure properties, in cases where I can't quite estimate the cardinality of all of them (I think), which in itself is not guaranteed by $V_\kappa$ being a $\Sigma_{15}$-universe. But it can very well be that I was just not careful enough. – Peter Scholze Jan 28 '21 at 14:41
  • 1
    @PeterScholze I think this is related to my earlier remark, that certain maneuvers in the setting of condensed mathematics/use of the proetale topology involve strictly more set theory than is really needed in HTT (but still far less than the full strength of ZFC). Roughly speaking, for the latter one just needs to know that for any $\kappa$ one can choose an appropriately bigger $\tau$ once and for all, but the former sometimes requires iterating cardinality-raising constructions like Stone-Cech compactification infinitely many times. – Jacob Lurie Jan 28 '21 at 14:45
21

I'd like to mention something that I think hasn't been pointed out yet. The original question began with

In set-theoretic language, one fixes some strongly inaccessible cardinal $\kappa$... This implies that the stage $_\kappa\subset $ of "sets of size $<\kappa$" is itself a model of ZFC.

However, the statement that $V_\kappa$ is a model of ZFC is significantly weaker than saying that $\kappa$ is inaccessible. In fact, if $\kappa$ is inaccessible, then $\{ \lambda\mid V_\lambda$ is a model of ZFC $\}$ is stationary in $\kappa$. Hence, the smallest inaccessible (if there is one) is much bigger than the smallest $\kappa$ such that $V_\kappa$ models ZFC.

Insofar as the reflection principle is useful (which, as some other answers have pointed out, one may at least question), it is only directly helpful for arguments in which the relevant property of a Grothendieck universe is that it is a model of ZFC. However, at least when formulated naively, there are plenty of places where category theory uses more than this. Specifically, we use the fact that a Grothendieck universe satisfies second-order replacement, meaning that any function $f:A\to V_\kappa$, where $A \in V_\kappa$, has an image. Saying that $V_\kappa$ models ZFC only implies that it satisfies first-order replacement, which only allows us to conclude that such an $f$ has an image if $f$ is definable from $V_\kappa$ by a logical formula.

I believe that second-order replacement is ubiquitous in universe-based category theory as usually formulated. For instance, if ${\rm Set}_\kappa$ denotes the category of sets in $V_\kappa$, then to prove that ${\rm Set}_\kappa$ is "complete and cocomplete" in the naive sense that it admits a limit and colimit for any functor whose domain is small, we need second-order replacement to collect the images of such a functor into a single set.

Now, there are ways to reformulate category theory to avoid this. McLarty's paper does it in some set-theoretic ways. A categorically consistent approach is to replace naive "large categories" (meaning categories whose sets of objects and morphisms may not belong to $V_\kappa$) with large ${\rm Set}_\kappa$-indexed categories. But this is a much more substantial sort of reformulation to perform by hand.

Mike Shulman
  • 65,064
  • This reminds me of the very first collaboration I've engaged with Misha Gavrilovich on his model categories derived from a model of ZFC. We tried to understand something related to Quillen adjunctions in that context. On the first day I asked him what is a "class", is it a subset of the universe (so $V_{\kappa+1}$ when $\kappa$ is inaccessible), or is it a formula and variables interpreted in the particular model. He opted for the "easy way out", i.e. the former, and at the end what we tried to see turned out somehow trivial (although still interesting). I still believe, though ... (1/2) – Asaf Karagila Jan 28 '21 at 17:45
  • ... that if we consider a class as a functor from the category of formulas (even with one parameter), we would have proved the result we set out to prove. But it would have been an order of magnitude harder, if not more. It's not that what we ended up with isn't interesting, just not very, I think, compared to the hypothesis we set out on the first day. It was interesting over all. Anyway, I digress. (2/2) – Asaf Karagila Jan 28 '21 at 17:48
16

OK, I spent much of today trying to figure this out by actually looking in some detail at HTT. It's been quite a ride; I have definitely changed my perspective multiple times in the process. Currently, it seems to me that the answer is that HTT, as written, can be read in this formal system. (So this is like in the joke where after hours someone says "Yes, it is obvious." There are definitely points where the correct interpretation has to be chosen, but as in any mathematical text, that is already the case anyways.) So with this answer, I want to advance an argument that HTT can be read in this formal system, trying to explain a little how to interpret certain things in case ambiguity can arise, and why I think reading it this way everything should work. But it is fairly likely that I overlooked something important, so please correct me!

As Tim Campion notes, most of the early stuff works without problems -- in fact, it doesn't even mention universes. As long as it doesn't, everything works in $V_{\kappa_0}$, in $V_{\kappa_1}$, and in $V$, and the given axiom schema even guarantees that any constructions made will be compatible.

One has to pay closer attention when one reaches Chapters 5 and 6. Let me try to present some definitions and propositions from these chapters from three different points of view.

  1. The classical ZFC point of view, or (equiconsistently) the one of von Neumann--Bernays--Gödel (NBG) theory, which allows classes in addition to sets, so we can talk about the (class-sized) category of all sets $\mathrm{Set}$.

  2. The point of view of HTT, which is ZFC + Grothendieck universes.

  3. The point of view of Feferman's set theory, in the form stated in the question. (Actually, I am no longer sure whether I really need these cofinality bounds. But it's nice to know that they can be assumed.)

Note that the question asked presumes that one is really interested in the first point of view, and in the others only insofar as they are conveniences to prove something about the first setting. This aligns with the contents of Chapter 5 and 6: the whole theory of presentable categories fits nicely into the first setting, also philosophically.

OK, so recall that a presentable category -- let me just stick with categories instead of $\infty$-categories, the difference is inessential for our concerns -- is a (class-sized) category $C$ that admits all small colimits, and such that for some regular cardinal $\kappa$, there is some small category $C_0$ and an equivalence $C\cong \mathrm{Ind}_{\kappa}(C_0)$,

i.e. $C$ is obtained by freely adjoining $\kappa$-filtered colimits to $C_0$. (In particular, $C_0$ is necessarily equivalent to the full subcategory of $\kappa$-compact objects of $C$.) In particular, presentable categories are determined by a small amount of data. Also, the idea is that $C$ is really the category of all objects (sets, groups, whatever). This point of view is really most clearly articulated in 1), while in 2) and 3) the notion of presentability suddenly depends again on the universe, and suddenly they are again only containing small sets/groups/...; let me then accordingly call them small-presentable. Note that this notion makes sense in both 2) and 3), and it depends only on $V_{\kappa_0}$. A small-presentable category is then in particular small-definable, so lives in $\mathrm{Def}(V_{\kappa_0})\subset V_{\kappa_0+1}$, where this inclusion is an equality in 2) (but not in 3)).

In 2), one would usually define a small-presentable category as a special kind of large category, which is the approach of HTT. But here I'm actually alreading getting a bit confused: There seem to be two notions of functors $F: C\to D$: The ones definable in $V_{\kappa_0}$, equivalently $F\in V_{\kappa_0+1}$ (namely, $V_{\kappa_0+1}$ are exactly the classes of $V_{\kappa_0}$), or all functors in $V_{\kappa_1}$. It does not seem obvious to me that any functor $F: C\to D$ in $V_{\kappa_1}$ lies in $V_{\kappa_0+1}$, as $C$ and $D$ themselves only live in $V_{\kappa_0+1}$. The difference between these two notions disappears when one restricts to accessible functors, all of which are definable. Note that 1) says it's really the first notion we should care about! (Before writing this post, I was not aware of the difference.)

In 3), the proper way to proceed is to use the perspective dictated by 1), which is that of "$V_{\kappa_0}$-definable categories", so they live in $\mathrm{Def}(V_{\kappa_0})\subset V_{\kappa_0+1}$. One can again consider these as $\kappa_1$-small categories. At first I thought that there would be a substantial difference here between the approaches of 2) and 3), but actually it seems that in both cases one arrives at two different notions of functors, which are reconciled once one restricts to accessible functors.

One of the main theorems is the adjoint functor theorem: If $F: C\to D$ is a functor of presentable categories that preserves all small colimits, then it admits a right adjoint. What does this theorem actually mean?

In 1), it means that there is a functor $G: D\to C$ -- which in particular means that it must be definable by formulas, as this is what functors between class-sized categories are -- together with (definable!) unit and counit transformations satisfying the usual conditions.

In 2), one is simply regarding $C$ and $D$ as being small when considered in $V_{\kappa_1}$ and then asserts the existence of the right adjoint there. Without further information, this actually does not seem to give what we wanted in 1), as a priori $G$ (and the unit and counit transformations) all lie in the larger universe. But this information can be obtained by remembering that $G$ is actually accessible (a part of the adjoint functor theorem I omitted to state above, but should be included), and so everything is determined on a set.

In 3), one would again like to get to the result of 1), but can try to do this as in 2) by first proving the existence of such data in $V_{\kappa_1}$ and then proving accessibility, thus yielding that everything lies in $\mathrm{Def}(V_{\kappa_0})$.

Let us see how this plays out in a couple of early places in Chapter 5 where universes are used.

Definition 5.1.6.2: Let $C$ be a category which admits all small colimits. An object $X\in C$ is completely compact if the functor $j_X: C\to \widehat{\mathrm{Set}}$ copresented by $X$ preserves small colimits.

Here $\widehat{\mathrm{Set}}$ is the (very large) category of sets in $V_{\kappa_1}$. Let us interpret what this definition means in the above systems.

  1. Here $C$ is any (possibly class-sized) category. Note that, especially in HTT, "locally small" is not a standard hypothesis, so this allows even morphisms between two objects to be proper sets. For this reason, the functor really has to go to $\widehat{\mathrm{Set}}$, and this is something we cannot talk about in this setting. So one would have to reformulate the condition to meet this objection; this should not be hard, but may be a bit nasty.

  2. I think it is implied in the definition that $C$ is any category that lies in $V_{\kappa_1}$. This is strictly capturing the setup of 1) in that if $C$ is small-definable as coming from 1), then any small colimit diagram in $C$ is automatically small-definable.

  3. Here we have two choices: Either the one from 1) or the one from 2), and they give different notions. In case of conflict the perspective from 1) is the correct one, so $C$ is small-definable, and one asks for commutation with colimits of small-definable diagrams. But while in 1) we had trouble formulating the condition, the universes at hand in 3) mean that the condition can now be formulated: we can ask that it takes small-definable colimits in $C$ to colimits in $\widehat{\mathrm{Set}}$. Here $\widehat{\mathrm{Set}}$ are the sets in $V_{\kappa_1}$.

So in this case, the upshot is that one has to be a bit careful in 3) about the interpretation, but guided by 1) one can give the right definition; and then the system actually helps.

Proposition 5.2.6.2: Let $C$ and $D$ be categories. Then the categories $\mathrm{Fun}^L(C,D)$ of left adjoint functors from $C$ to $D$, and $\mathrm{Fun}^R(D,C)$ of right adjoint functors from $D$ to $C$ are (canonically) equivalent to one another.

  1. In this perspective, this proposition only really makes sense when $C$ and $D$ are small, as otherwise $\mathrm{Fun}(C,D)$ is too large. (One wants to consider such functor categories when $C$ and $D$ are presentable (or accessible), but only when restricting to accessible functors. So that's a discussion that would appear later in Chapter 5.) Then the statement is clear enough, and the proof given applies.

  2. In this perspective, I think it's the same as in 1), except that one can also formulate the same result in a different universe.

  3. Same here.

Note however that as it stands, in 1) this proposition can not (yet) be applied in case $C$ and $D$ are presentable. In 2) and 3), (small-)presentable ones are special large categories, to which the result applies. Note however that the functor categories and their equivalence are all living in a larger universe, and we get no information of them lying in either $V_{\kappa_0+1}$ or $\mathrm{Def}(V_{\kappa_0})$.

The next proposition considers the presheaf category $\mathcal P(C)=\mathrm{Fun}(C^{\mathrm{op}},\mathrm{Set})$, and the proof is a typical argument involving passage to a larger universe to resolve homotopy-coherence issues.

Proposition 5.2.6.3: Let $f: C\to C'$ be a functor between small categories and let $G: \mathcal P(C')\to \mathcal P(C)$ be the induced functor of presheaf categories induced by composition with $f$. Then $G$ is right adjoint to $\mathcal P(f)$.

Here $\mathcal P(f)$ is defined to be the unique small colimit-preserving extension of $f$ (under the Yoneda embedding).

  1. Here, we have two class-sized categories and functors between them, all definable (as must be). The proposition would ask us to find (definable!) unit and counit transformations, making some diagrams commute. This seems not too hard. But in $\infty$-categories, it is famously hard to define functors by hand, so this is not actually how Lurie proceeds!

  2. Here $\mathcal P(C)$ and $\mathcal P(C')$ are special large categories. Lurie in fact applies the large Yoneda embedding in the proof. So this is really producing the unit and counit adjunctions only in some larger universe. As discussed above, I think this proof does not actually give what we wanted in 1)!

  3. We can argue as Lurie does to produce the data in some larger "universe". (Edit: Actually, as Tim Campion points out, one has to make a minimal detour in order to justify what's written. See comments to his answer.)

So when reading this proposition, in either system 2) or 3), one should make a mental marker that so far the statement proved is weaker than one might naively hope for. But this is corrected later, by observing that everything is determined by a small amount of data.

Upshot: While at first I thought there would be a substantial difference between 2) and 3), I actually think there is (almost) none. One difference is that $\mathrm{Def}(V_{\kappa_0})\subset V_{\kappa_0+1}$ is a proper inclusion, but in practice the way to guarantee containment in $V_{\kappa_0+1}$ seems to be to prove definability in $V_{\kappa_0}$ (for example, by proving that certain functors are accessible).

OK, now tell me why this doesn't work! :-)

Peter Scholze
  • 19,800
  • 3
  • 98
  • 118
  • This approach sounds promising to me. I've added a longer comment at the end of my answer, including some potential pitfalls in formalizing the proof of that last proposition. – Tim Campion Jan 29 '21 at 15:57
15

If I understand correctly, you're after a statement of the form :

"If something was proved in HTT using universes, it can be proved without them by restricting to some $V_\kappa$ for $\kappa$ large enough"

The rigourous answer to that, if we don't have more information about HTT, is that there can be no such statement if ZFC is consistent.

Indeed, it is possible that the existence of universes are inconsistent (in fact it's not possible to prove that it is consistent), and in that situation, anything can be proved using universes, and so such a statement would imply that anything can be proved, i.e. ZFC is inconsistent.

I'm being a bit sloppy about what's provable in what etc. but the main idea is there

Of course, we do know things about HTT, and if we read it carefully we can analyze where it uses universes, and see that they can in fact be replaced with transitive models of ZC+ replacement up to $\Sigma_{15}$-formulas, as Jacob points out. In that case, since there provably exist such nicely behaved models (of the form $V_\kappa$, for $\kappa$ well-chosen), this is not a problem; and HTT can be rewritten without universes - but this can not be proved without knowledge of what's in HTT.

The "moral" of this is that, in most mainstream category theoretic questions, universes are a time-saving device, and not an actual part of the mathematics.

Maxime Ramzi
  • 13,453
15

Answering this question depends strongly on exactly what you want from Higher Topos Theory, because expressing high logical strength is a different goal from expressing an aptly unified logical framework for algebraic geometry and number theory.  Unified strong foundations for general categorical mathematics are one fine goal, and seem to be the goal of many contributors here.  For that goal everything said in comments and answers to this question is relevant. But apt work in geometry and number theory does not call for vast logical strength.

While HTT is more intertwined with universes than SGA, neither HTT nor SGA makes real use of the (very strong) axiom scheme of replacement. Thus they can use "universes" radically weaker than Grothendieck's. As a typical and germane example, Grothendieck he made just one appeal to the axiom scheme of replacement.  That is in his quite crucial proof that every AB5 category with a generating set has enough injectives.   And this use of replacement turns out to be eliminable.  It worked, but Grothendieck actually did not need it to get his result.

To expand on Grothendieck's use of replacement: Reinhold Baer in the 1940s used transfinite induction (which requires the axiom scheme of replacement) to prove modules (over any given ring) have enough injectives.  He was consciously exploring new proof techniques and got a good result.  Grothendieck's Tohoku cast that proof in a form showing every AB5 category with a small set of generators has enough injectives--and a few years later Grothendieck found this was exactly the theorem he needed for topos cohomology.   Baer and Grothendieck both had practical goals, not tied to foundation concerns, but both wanted to get foundations right too.  And they did.  But it turns out they could have gotten those same theorems, correctly, without replacement, by nearly the same proofs, by specifying large enough function sets to begin with (using power set, but not replacement).  There are results that genuinely require the replacement axiom scheme.  But those results rarely occur outside of foundational research.

A lot of people coming from very different angles (some logicians, some disliking logic) since the 1960s have remarked that in the context of algebraic geometry and number theory, the high logical strength  of Grothendieck's universe axiom, is an actually unused by-product of Grothendieck's desire for a unified framework for cohomology.  That can now be made quite precise:  The entire Grothendieck apparatus including not just derived functor cohomology of toposes but the 2-category of toposes, and derived categories, can be formalized in almost exactly the same way as it was formalized by Grothendieck, but at logical strength far below Zermelo-Fraenkel or even Zermelo set theory.    The same is true for HTT.  You can get it without inaccessible universes or reflection as long as you do not need the vast (and rarely used) strength of replacement.  The proof has not actually been given for HTT.  It has been for Grothendieck's uses of universes. It seems clear the same will work for HTT.

The logical strength needed has been expressed indifferent ways:  Simple Type Theory (with arithmetic), Finite Order Arithmetic, the Elementary Theory of the Category of Sets, Bounded Quantifier Zermelo set theory.  Roughly put, you posit a set of natural numbers, and you posit that every set has a power set, but you do not posit unbounded iteration of power sets.  A fairly naive theory of universes can be given conservative over any one of these (the way Godel-Bernays set theory is conservative over ZFC) and adequate to all the large structure apparatus of the Grothendieck school.  

Colin McLarty
  • 10,817
  • I think what has come out in this discussion is that the use of universes is a bit more tightly woven into HTT than in SGA, even though they way they are used still doesn't really have anything to do with logical strength. For one thing, they appear (superficially) in the substructure of Quillen model categories used to get the theory going. For another, presentability theory / adjoint functor theorem - type arguments are used in a more essential way (constructing adjoints explicitly is less feasible when it entails higher coherence). – Tim Campion Feb 03 '21 at 17:07
  • My point is that reading or using HTT requires one to "think like a category theorist" even more than SGA does. So even if one is ultimately interested in theorems which will fit more naturally into weaker foundations, the advice (borne out in the story of FLT) that one should not worry about this slight increase in foundational requirements when applying the theory, trusting that the foundational requirements may be weakened for any particular application once it has been worked out, seems even more relevant. – Tim Campion Feb 03 '21 at 17:07
  • Of course, if one is building further layers of theory on top of HTT, and feels conceptually hindered by the extra foundational apparatus, that is another matter. – Tim Campion Feb 03 '21 at 17:07
  • (I should clarify that of course I don't mean to imply that any proof of FLT is some straightforward application of SGA. Just that the original proof, like so much mathematics, did in fact apply the results of SGA [in addition to layer upon layer of other insights, tools, theorems, etc].) – Tim Campion Feb 03 '21 at 17:40
  • @TimCampion Yes universes are more tightly involved in HTT than in SGA. But neither HTT nor SGA really uses the axiom scheme of replacement, and so both can have universes (in a weaker form than Grothendieck's, in principle, but giving all the same large scale apparatus) without needing inaccessible cardinals. – Colin McLarty Feb 03 '21 at 18:55
  • Thanks. I think maybe there's a subtlety involved in what exactly "all the same large scale apparatus" means. For instance, my understanding is that the principle "$X$ is a set of cardinality $<\kappa$ if and only if $X$ is bijective to a set in $V_\kappa$" is used all the time in category theory as soon as a universe gets involved (perhaps it's used only conceptually as a crutch and is never necessary, I don't know), but my understanding is that this statement is equivalent to $V_\kappa$ satisfying some (extremely weak -- at most $\Sigma_1$) form of replacement. – Tim Campion Feb 03 '21 at 19:20
  • If weakening the universes means I have to give up that principle, then it sounds like a conceptually nontrivial task to me. I suppose this objection applies already to SGA, so I'm betraying my ignorance here -- apparently this particular issue must have been already solved. – Tim Campion Feb 03 '21 at 19:21
  • 1
    @TimCampion The kind of statement you quote is used constantly in foundational explorations of strong foundations for category theory. But are used only rarely in SGA or HTT or contexts using them -- and the rare occurrences are (in every case I can find) eliminable just by noticing more explicitly what is being said. This is covered in detail in the reference https://doi.org/10.1017/S1755020319000340 which just appeared in 2020. – Colin McLarty Feb 03 '21 at 19:49
  • Thanks. I've linked to both "The large-scale structures..." and "What does it take..." several times here, and it's clear I should actually try to understand what's in them more carefully. The lesson is that in order to do category theory in such a setting, one must think carefully, and the necessary translation does not lend itself to comment boxes on the Internet. But that doesn't mean that one will no longer be "doing category theory" or that the possibility of the translation was ever in doubt. I will have to read your paper to understand what category theory looks like in this setting. – Tim Campion Feb 03 '21 at 19:55
  • 1
    I've finally registered from the closing paragraph: finite-order arithmetic is essentially equivalent to ETCS / BZC / simple type theory. So my above comment is misleading -- conceptually "The large-scale structure..." carries out exactly the sort of program one might describe as "eliminating universes within set theory". Previously I've put off reading it partly because I'm particularly ignorant of logic at the level of arithmetic. But this particular personal shortcoming should be no obstacle. – Tim Campion Feb 04 '21 at 16:16
13

Any theorem $T$ of $\mathsf{ZFC}$ follows from a finite subset of the axioms of $\mathsf{ZFC}$ or, to keep things simple, from $\mathsf{ZFC}$ where the axiom scheme of replacement is restricted to $\Sigma_n$ predicates¹, call this $\mathsf{ZFC}_n$. Now $\mathsf{ZFC}$, and more precisely $\mathsf{ZFC}_{n+1}$, proves the existence of arbitrarily large cardinals $\kappa$, strong limits of uncountable cofinality, such that $V_\kappa$ is a model of $\mathsf{ZFC}_n$, and, in particular, of the theorem $T$, and such that, moreover, that the truth value of any $\Sigma_n$ statement, with parameters in $V_\kappa$ is the same in $V_\kappa$ as in the (true) universe. We can call these $V_\kappa$ “limited universes” in that they are closed under most set-theoretic operation, like taking powersets, except that replacement needs to be either countable (included for convenience) or restricted to a $\Sigma_n$ predicate; and in particular, they are closed under whatever existence statements $T$ makes.

So the idea would be to apply the above to the conjunction $T$ of all the theorems which you consider to be part of Higher Topos Theory (and whatever other theories are used as prerequisisites) and find the appropriate $n$. (I actually suspect that $n=1$ should be sufficient: I would be very surprised to find an instance of replacement in ordinary mathematics which does not follow from $\Sigma_1$-replacement.) Then $\mathsf{ZFC}_n$ would prove $T$ (all the theorems of the theory) and $\mathsf{ZFC}_{n+1}$ would prove the existence of an endless supply of limited universes in which to use the theory.

Of course, to avoid an infinite loop, you cannot consider that theorem (the one asserting the existence of an endless supply of $V_\kappa$) to be part of the theory, or you need to move to a larger $n$.

To explain what might seem like a logical contradiction, here, it must be clarified that the statement that the existence of many models of $\mathsf{ZFC}_n$ can be proved in $\mathsf{ZFC}$ for every $n$, but not uniformly so (the proof gets longer and longer as $n$ grows), so $n$ must be a concrete natural number, the universally quantified (over $n$) statement is not provable in $\mathsf{ZFC}$. But this is not a problem so long as your theory is fixed and is formulated in $\mathsf{ZFC}$ (which demands that it does not, itself, contain such metatheorems as “for any concrete $n$ we can prove the following in $\mathsf{ZFC}$”). So it is up to you to ascertain that this is the case for HTT (and, if you're courageous enough, find the appropriate $n$).

(Just to give a sense of how the kinds of cardinals involved, the cardinals $\kappa$ such that $V_\kappa$ is a model of $\mathsf{ZFC}_1$ are the fixed points of the $\gamma \mapsto \beth_\gamma$ function. I don't think there's any hope for a reasonable description of the $\kappa$ such that $V_\kappa$ is a model of $\mathsf{ZFC}_n$ for any concrete $n\geq 2$. See also this question.)

  1. Meaning predicates with at most $n$ alternating sets of unbounded quantifiers, starting with existential quantifiers, followed by formula with bounded quantifiers (meaning of the form $\forall x\in y$ or $\exists x\in y$).
Gro-Tsen
  • 29,944
  • 4
  • 80
  • 335
  • Thanks, I found this very helpful! In fact, if I understand it right, then the cardinals of Lemma 4.1 here are precisely the $\Sigma_1$-correct cardinals of uncountable cofinality? – Peter Scholze Jan 28 '21 at 08:52
  • @PeterScholze That seems right, yes. And if you find (iii) useful, you can also do the same at higher levels: if I am not mistaken, $\mathsf{ZFC}{n+1}$ implies the existence not only of unboundedly-many $\Sigma_n$-correct cardinals of uncountable cofinality, but even of (unboundedly-many) $\Sigma_n$-correct cardinals $\kappa$ such that for all $\lambda<\kappa$ there is a $\Sigma_n$-correct $\kappa\lambda$ of cofinality $\geq\lambda$. – Gro-Tsen Jan 28 '21 at 10:56
  • (The key here is that if $C$ is a club (:= closed unbounded set of ordinals), and $(\alpha_\iota)$ indexes the elements of $C$ in increasing order, then the fixed points $C'$ of $\alpha$ themselves form a club, and if $\beta\in C'$ then for all regular cardinal $\lambda<\beta$ we can find $\gamma\in C$ with cofinality $\geq\lambda$, namely $\alpha_\lambda$.) – Gro-Tsen Jan 28 '21 at 11:04
  • Right! But I guess while this process describes the $\Sigma_1$-correct cardinals, it wouldn't inductively describe the $\Sigma_2$-correct cardinals, etc.? It's just a game you can play on top of $\Sigma_n$-correct cardinals. (I still need to figure out what $\Sigma_n$-correct means in practice, and whether it is just coincidental that I "rediscovered" $\Sigma_1$-correct cardinals.) – Peter Scholze Jan 28 '21 at 11:06
  • 1
    @PeterScholze Exactly, it's a small addition, far from the huge gap between $\Sigma_n$-correct and $\Sigma_{n+1}$-correct. Note that in the definition of that term there is some ambiguity as to whether it applies to $V_\kappa$ (the cumulative hierarchy) or $H(\kappa)$ (the sets hereditarily of cardinality $<\kappa$): the two coincide when $\kappa$ is a fixed point of $\beth$. – Gro-Tsen Jan 28 '21 at 11:56
10

A question that came up in the comments was about the motivation for asking the question. Let me try to address this here.

Foremost, it is about learning! As I mentioned in the original question, I had myself toyed around with some "stupid" cardinal bounds, and only later learned about the reflection principle, so I wanted to understand what it can do (and what it cannot do), and whether I can somehow automatically relegate any further complicated versions of such estimates into this machine. So it's the usual thing where you are just stumbling around in a dark room, and would very much like the room to be illuminated! So, thanks to you all for the illuminating answers!

Another reason is that recently I got a bit frustrated with the solution of Grothendieck universes to the problem at hand. Let me explain.

I very much want to talk about the category of all sets, or all groups, etc., and want to prove theorems about it. And, at least in the von Neumann-Bernays-Gödel (NBG) version of ZFC theory that allows classes, this is a perfectly valid notion. So I find it ontologically very pleasing to work in this setting, and would very much like the adjoint functor theorem to be a theorem about (presentable) categories in that sense.

Now presentable categories are determined by a small amount of data, so one could always work with this small amount of data and keep careful track of relative sizes. In fact, many proofs in HTT explicitly do keep track of such relative sizes, but there are still a few spots where it is nice to first take a "broader view" and look at these large categories as if they were small.

Indeed, the adjoint functor theorem is about functors between large categories, and it quickly becomes nasty to talk about this from within NBG/ZFC. Note that the statement of the adjoint functor theorem makes perfect sense -- it just asks that all the data of the adjunction is definable. But it's a bit nasty to try to speak of these things from "within". So it would definitely be nice to have some kind of meta-theory using which to argue about these large categories, and pretend that they are small. The subtle question of "definability from within" may be a priori lost in this meta-theory, but I regard this question of "definability from within" as central, because after all what I wanted was a theorem about all sets, so I'm fine with having to pay a little bit of attention to it -- and, to take the punchline away, it turns out that this is precisely the difference between working with Grothendieck universes and working with Feferman's "universes".

So this is the thing that Grothendieck universes are for: They always give you a larger universe for any universe you are currently working in. I find the existence of Grothendieck universes completely intuitive, and in fact positing their existence seems completely on par with positing an infinite set in the first place: You are just allowing to collect everything you already have into a bigger entity of its own.

But now suddenly what I used to think of as all sets are called the small sets, and there are also many larger sets. So even if I prove an adjoint functor theorem in this setting, it's not anymore a theorem about functors between categories of all sets/groups/..., but only one of functors between small sets/groups/.... So if you think about it, even in ZFC + Grothendieck universes, you will never prove that theorem you actually wanted, about the category of all sets. (Actually, until very recently, I assumed that the adjoint functor theorem (for $\infty$-categories) is a statement of ZFC that has been proved under "ZFC + Universes", but that's not quite right: The statement that has been proved can even only be formulated in ZFC + Universes.)

What has been proved is that it is consistent that the adjoint functor theorem holds. Namely, assuming consistency of ZFC + Universes, you now produced a model of ZFC -- that of small sets in your model of ZFC + Universes -- in which the theorem is true. So you could now work in the theory "ZFC + the adjoint functor theorem", in which the adjoint functor theorem can be applied to the category of all sets/groups/..., but that definitely feels like a cheat to me. You did not even prove that "ZFC + Universes + the adjoint functor theorem" is consistent! (You would get that if you start with the consistency of slightly more than ZFC + Universes, asking for $\kappa$ such that $V_\kappa$ satisfies ZFC + Universes. Again, that seems like a completely fair assumption to me -- just keep going.) But now you might get to see the danger that you inadvertently climb up the consistency ladder as you implicitly start to invoke more and more theorems proved for small sets also for all sets.

It would be much nicer if you would know that, in ZFC + Grothendieck universes, everything you proved about small sets is also a theorem about the whole ambient category of all sets. This is not automatic, but you can add this as an axiom schema. Mike Shulman in Section 12 of Set theory for category theory (arXiv:0810.1279) discusses this idea (that he denotes ZMC): I find it ontologically quite pleasing, it also seems to have a very simple axiomatization (even simpler than ZFC!), but

a) this additional axiom schema is not completely self-evident to me: Why should everything that is true in small sets also hold for all sets? (Especially if we had some trouble proving the desired result in the first place. Also, note that it definitely does not hold for any notion of small sets: Rather, the axiom schema guarantees that there is some notion of small sets for which this sort of reflection holds. Now this appears a bit dubious to me as in the first place I never wanted small sets, so now I'm positing them, and also ask that they still reflect the whole behaviour of all sets. Probably fine, but not self-evident to me.)

b) the consistency strength of this axiom schema is considerably higher: It is the same as the consistency of a Mahlo cardinal. This is still low-ish as large cardinals go, but it's much much higher than mere Grothendieck universes (which are really low at the bottom of the hierarchy).

Regarding a), the fact that we could prove the consistency of the adjoint functor theorem from the consistency of Grothendieck universes points in the right direction, but this does not in itself guarantee that the two together are consistent. I can imagine that I might convince myself that the axiom schema is reasonable, but I certainly think it needs much more justification than mere Grothendieck universes. (Side question: How large are the large cardinals one can justify using the idea of "allowing to collect together everything we already had"? I'm not sure if this is a completely well-defined question... but to me, a measurable cardinal is definitely not of that sort (but I'm happy to stand corrected), as it seems to posit the emergence of new combinatorial features.)

Another reason I recently got a bit unhappy with Grothendieck universes is that while in some sense we would like to use them to be able to ignore set-theoretic subtleties, in some ways they come back to bite you, as now you have to specify in which universe certain things live. Sometimes, you might then even have to specify several different universes for different types of objects (think of sheaves on profinite sets), and I find that it quickly gets quite ugly. I would much rather have all objects live together in one universe.

So, while thinking about sheaves on profinite sets, I came to find the solution with only one universe much more asthetically and ontologically pleasing, and this solution (condensed sets) can be formalized in ZFC without trouble.

OK, so I claim that Grothendieck universes did not really solve the problem they set out to solve, as

a) they still do not allow you to prove theorems about the category of all sets/groups/... (except as a consistency result, or under stronger large cardinal axioms)

b) working with them, you still have to worry about size issues -- your category of all sets now comes stratified into sets of all sorts of different sizes (i.e. in different universes).

Moreover, they also increase consistency strength.

Now, after this wonderful discussion here, I think Feferman's proposal is actually much better. However, as also Mike Shulman commented, I regard Feferman's axioms not as describing any ontologically correct world, but I regard the "universes" of Feferman's theory merely as conveniences, in order to talk about large categories as if they were small. In other words, Feferman's theory exactly gives you a meta-theory in which to argue about such large categories from the "outside". But it's a theory I would only ever use to give a proof of a theorem of ZFC. Comparing to Grothendieck universes, Feferman's theory

a) does allow you to prove theorems about the category of all sets/groups/..., because it explicitly includes an axiom schema that all theorems about small sets are also theorems about all sets.

b) Of course, within a proof of a theorem of ZFC that invokes some nontrivial size issues, it's very welcome that the theory enables you to talk about various sizes. Moreover, it does so in a way where you can still apply all axioms of ZFC to each of the "universes", and also is taking care "behind the scenes" of how to rewrite everything in terms of (potentially extremely subtle) cardinal bounds in ZFC itself. So it's like a high-level programming language for arguments involving difficult cardinal estimates in ZFC.

In addition, it does not increase consistency strength, and in fact any statements of ZFC proved in this language are theorems of ZFC. (As I recalled above, we could also have a) + b) with Grothendieck universes, but would then run up to the consistency of a Mahlo cardinal.)

So, the upshot is that I think Feferman's universes do a much better job at solving the problem of providing a meta-theory to "talk about large categories as if they were small" than Grothendieck universes do.

Let me add some final reasons for asking the question. I think higher-categorical techniques such as the ones laid out in HTT are of very central importance, not just in algebraic topology where they originated, but in all of mathematics. I can certainly attest to that as regards number theory and algebraic geometry. So their centrality is also an important reason to analyze their consistency strength.

Reading HTT is a very nontrivial matter -- it is long and complicated. Some number theory colleagues have however told that one of the primary reasons they could not read HTT is that it uses universes. Namely, they are so used to ZFC (and to checking with extreme care!) that they will automatically try to eliminate any use of universes in an argument. Now in SGA, at least if you were only interested in applications to etale cohomology of reasonable schemes, this was something you could by hand -- for example, just add some countability assumptions to make things small. However, in HTT I do not see any way that someone will be able to put cardinal bounds in as you read along -- the arguments are way too complex for this.

So now I hope I can tell them that they can check that everything works in ZFC, and they can still read HTT (essentially) as written, if they read it in Feferman's set theory. If they check carefully (which they will), they may still need to fill in a little lemma here and some small extra argument there -- but they would have to do so anyways, in any book of ~1000 pages, and I might imagine that less than half of these side remarks have to do with replacing Grothendieck universes by Feferman's "universes". Should anybody actually undertake that project, of course they deserve full credit if they succeed in this important work!

Let me end with a very brief note on what seems to be the key salient point in the translation to Feferman's theory. I came to appreciate the point that Tim Campion raised in his answer, and I now see that this was also mentioned in Jacob Lurie's second answer. Roughly, it is the following. If $C$ is a presentable category, then there is some small category $C_0$ such that $C=\mathrm{Ind}_\kappa(C_0)$

for some regular cardinal $\kappa$, adjoining freely all small $\kappa$-filtered colimits. This makes $C$ naturally a union of $C_\tau$'s, where $C_\tau$ only collects the $\tau$-small $\kappa$-filtered colimits. Here $\tau$ is a regular cardinal such that $\tau\gg \kappa$. This increasing structure of $C$ as a union of $C_\tau$'s is central in the theory of presentable categories, but the levels are really enumerated by (certain) regular cardinals $\tau$. If you increase your universe, then you also get a larger version $C'$ of $C$ itself, and in Grothendieck universes $C$ is now one of the nice layers $C'_\tau$ of $C$, where $\tau$ is the cutoff cardinal of the previous universe. But in Feferman's universes, this $\tau$ is not regular. This might make some arguments more subtle, but I would expect that one can usually solve this issue by simply embedding $C$ into some $C'_\tau$ with $\tau$ some regular cardinal larger than the cutoff cardinal of the smaller universe.

David Roberts
  • 33,851
Peter Scholze
  • 19,800
  • 3
  • 98
  • 118
  • This is interesting and clarifying! I'm shocked (and a bit alarmed!) to hear that the use of universes may actually be impeding the adoption of higher-categorical techniques by some number theorists, and I'd like to understand the issue better. Because of course, the whole point of introducing universes in the first place is precisely to make the lives of mathematicians like number theorists easier. Is there anywhere the concerns of number theorists regarding universes have been articulated in more detail? E.g. your discussion suggests there's an important historical dimension to the matter. – Tim Campion Jan 31 '21 at 17:15
  • 1
    My understanding was that because number theory imports techniques from so many other fields, number theorists become adept at "black-boxing" tools from other fields. So the picture of a number theorist going through SGA line-by-line with a particular eye toward set-theoretical concerns surprises me. I should avoid making sweeping generalizations about a large and diverse field I know little about. Going through HTT in similar detail just to apply the theory sounds very inefficient, risks missing the forest for the trees, and underscores the need for second-generation texts like Riehl-Verity – Tim Campion Jan 31 '21 at 17:41
  • 1
    Well, as you say, it's a diverse community. Some historical context is certainly given by the discussions around FLT. There have always been a few number theorists who pay great attention that they have good command over the whole infrastructure of a proof, and I think they play a critical role in the community. I believe Gabber, although working in areas and sometimes using methods very very close to HTT, has not read HTT (and consequently does not totally trust arguments relying on it); I can believe that one of the reasons is the cavalier use of universes, but am not sure in his case. – Peter Scholze Jan 31 '21 at 20:06
  • Note also that Gabber would only be satisfied with a treatment that actually gives all the details. He knows the pitfalls in the area very well, and he wants to see how the new generation can often magically avoid them using a new fancy language. He'd certainly be more impressed if it's all in ZFC. For another take, see here (fourth paragraph) -- I now that Buzzard is, to some degree, worried about universes. At least to the extent that he will keep track that if you're citing HTT, then by necessity you are using universes. – Peter Scholze Jan 31 '21 at 20:13
  • And now that I've thought about it more, I actually also dislike that the theorems formulated in HTT are not statements of ZFC proved under stronger assumptions, but are statements of ZFC + Universes. So if you want to apply them, you have to introduce universes. I actually don't want to introduce universes! They ought to be conveniences, not necessities! – Peter Scholze Jan 31 '21 at 20:18
  • 2
    I've previously fallen into the trap of dismissing many of these concerns, categorizing them variously as philosophical hangups, misplaced concern with logical details, or a misunderstanding of the "authority" of ZFC. One point you've brought out to me is a genuine conceptual concern: especially when going past just one universe to the second or third, "size" quickly begins to feel like a bookkeeping device run amok. Even though Universes doesn't technically add to the ontology, it feels like it does, and the "thing" it adds has no clear intuitive or "genuinely mathematical" meaning. – Tim Campion Jan 31 '21 at 20:24
  • 1
    (I am obviously writing everything here with the utmost respect to Lurie, and everyone else in the higher categorical community! If they wouldn't do such important work, I wouldn't be here to ramble about very minor implementation issues.) – Peter Scholze Jan 31 '21 at 20:25
  • 1
    On reflection, it's possible that such issues have impeded progress in higher category theory itself. For instance, one wants to systematically study 2-categories in such a way that Cat (= locally small categories) is an object of 2Cat. To do this, there seems nothing for it but to introduce a second universe. Immediately upon doing so, one feels that the universes are bizarre artifacts of the formal framework thrust gracelessly to the forefront of basic considerations like notions of completeness. It's hard to take seriously anything interesting one might say in such an artificial environment – Tim Campion Jan 31 '21 at 20:47
  • 3
    Re: "How large are the large cardinals one can justify using the idea of 'allowing to collect together everything we already had'?" This seems related to the often discussed matter in set theory of which large cardinals are justified by strong reflection principles. A good reference on this is Peter Koellner's "On Reflection Principles." – Elliot Glazer Feb 01 '21 at 05:16
  • Thanks, that was a nice read! – Peter Scholze Feb 01 '21 at 09:18
  • In particular, this seems to strongly suggest that I should be able to convince myself of Mahlo cardinals by simply reflecting enough on this principle of collecting things together. I'll try to see if that works. – Peter Scholze Feb 01 '21 at 10:51
  • 3
    Just a comment about Grothendieck universes. In SGA they are always careful to check that all constructions are compatible with enlarging the universe. So I think they avoid the criticism that one is not really proving a theorem about all groups, etc. But I don't think many later authors were so careful. – Dustin Clausen Feb 01 '21 at 10:52
  • 3
    @TimCampion My sense is that discomfort with universes among number theorists stems from a conviction that the theorems they care about are "concrete" and so "ought to be" theorems in a quasi-absolute sense; i.e., they should be provable in any halfway reasonable axiomatic system, and should not rely essentially on some arcane set-theoretic assumption. So it's fine to introduce universes as a façon de parler if it makes things easier, but only if the axiom is not essential to the proof. Another way to put it: if something is unprovable in ZFC then some may doubt if it's "really proved." – Timothy Chow Feb 01 '21 at 18:57
  • 1
    @TimothyChow Thanks. This is exactly the sort of thing that I tend to dismiss as a "mere philosophical hangup", getting in the way of doing one's job as a working mathematician. I will try to take such perspectives more seriously. – Tim Campion Feb 01 '21 at 19:01
  • 1
    @TimCampion I don't see why it gets in the way of doing one's job as a working mathematician. Isn't it standard practice, in all areas of mathematics, to examine the hypotheses of a theorem to see which ones are really needed? It would be remarkable, for example, if some concrete statement of number theory, such as the Riemann Hypothesis, were unprovable in ZFC but provable in ZFC + universes. It would also be embarrassing if we overlooked this remarkable result because we dismissed careful examination of the universe axiom as a "philosophical hangup." – Timothy Chow Feb 01 '21 at 19:07
  • @TimothyChow I agree it's important to carefully examine the universe axiom and hypotheses of theorems, and it's wrong to be dismissive of philosophical considerations. But I hope nobody starts reading HTT, sees the word "universe" in the footnotes, cries "Das ist keine Mathematik!" and throws the book away! It would be remarkable if RH were provable in ZFC but not PA, but if you allow ZFC proofs only as a façon de parler, you're doing logic, not number theory. The jump from PA to ZFC is much bigger than the jump from ZFC to ZFC+Universes -- why worry about the latter more than the former? – Tim Campion Feb 01 '21 at 20:18
  • 3
    I second what Timothy Chow said above. Actually, I believe the (very slight) uneasyness I have with the use of universes is a little similar to the use of noetherian hypothesis in many textbooks on algebraic geometry. The thing is that you actually acquire a much better understanding of the material, and prove better theorems, if you do not make these simplifying assumptions. I think for a long time the study of nonnoetherian phenomena was also dismissed as a "philosophical hangup". – Peter Scholze Feb 01 '21 at 20:27
  • @TimCampion I suspect that many, if not most, of those who worry about the jump from ZFC to ZFC+universes are also worried about the jump from PA to ZFC. But I agree with you that one should not reject universes, or infinite set theory, or the law of the excluded middle, or what have you, as "not mathematics." But is anyone rejecting HTT as "not mathematics"? I think they just want a clear understanding of how to eliminate universes if they want to. It's not that different from wanting to know (e.g.) if a theorem is ineffective or uses Choice. – Timothy Chow Feb 01 '21 at 21:54
  • 1
    @TimothyChow I see a big difference between (a) wanting to know whether a theorem uses choice and (b) deciding that because a theorem uses choice, one should put off learning about it. Perhaps the difference is not so big as I see it. I am starting to understand that it really is possible for a non-set-theorist to be so sensitive to the details of ZFC that even minor variations can really interrupt the flow of understanding arguments. This is a mathematical consideration I hadn't appreciated before. – Tim Campion Feb 01 '21 at 22:49
  • 1
    It's especially unfortunate if there are a just a few members of the community who feel this way, but the rest of the community, though not themselves particularly attuned to foundational matters, are accustomed to taking cues from those few who are. If the foundationally-sensitive number theorists are slow to adopt certain tools or results, their reluctance might be misunderstood by the rest of the community as stemming from doubt in the correctness those results, or perceived to mean that these tools are using strong foundational assumptions in untrustworthy ways. – Tim Campion Feb 02 '21 at 00:08
  • (Of course, I recognize that there are many reasons not to learn about or adopt new techniques or theories. My comments above should not be interpreted to mean "Everybody should drop what they're doing and learn higher category theory." I'm talking exclusively about the foundational reasoning.) – Tim Campion Feb 03 '21 at 16:45
  • If universes sometimes feel like a "bookkeeping device run amok", then won't the alternatives feel like an even more arduous bookkeeping device? With universes, the statements of your theorems (if you write them as actual theorems rather than metatheorems) will involve quantifying over universes, but the alternative of keeping track of multiple cardinals, their relationships, cofinality assumptions, etc. in the statement of each theorem seems even worse. – Cameron Zwarich Jan 13 '24 at 21:39
  • From my experience with both theorem proving assistants and large industrial software projects, the use of universes and universe polymorphism would probably be a minor inconvenience, whereas the hypothetical alternative of tracking all of the information to remain purely within ZFC seems like it would get in the way. In a sense, you've just replaced SGA universes with a more fine-grained and less composable notion of universe (for the benefit of not increasing consistency strength). – Cameron Zwarich Jan 13 '24 at 21:43
  • @CameronZwarich I claim that carefully thinking about universe issues may make it possible to eliminate them, but it requires one to think very carefully how to formulate things. (Again, staying in ZFC is not my actual concern, but a likely outcome of such a careful reflection.) I've recently thought more seriously about higher categories (in particular, categories of categories) and finally have a way of thinking about them that I like. Notably, this way does not involve a choice of universes (and instead features a form of $V\in V$ that happens to be literally true). – Peter Scholze Jan 14 '24 at 23:11
  • @PeterScholze That certainly sounds interesting. Are the details written up anywhere? Is it along the lines of existing non-well-founded set theories or something else? I’m curious how you avoid the Russell/Girard paradoxes while still being useful for mathematics, especially if the ambient logic is classical. – Cameron Zwarich Jan 15 '24 at 00:44
  • Unfortunately, my list of things to write is way too long. Short: As I wrote somewhere on MO, I like the notion of presentable $\infty$-category. In practice, all of them are $\omega_1$-compactly generated, and functors preserve $\omega_1$-compact objects. Such form an $\infty$-category $\mathrm{Pr}^{\omega_1}$ that happens itself to be $\omega_1$-compactly generated presentable, so $\mathrm{Pr}^{\omega_1}\in \mathrm{Pr}^{\omega_1}$. Presentable $(\infty,2)$-categories can then be defined as $\mathrm{Pr}^{\omega_1}$-modules in $\mathrm{Pr}^{\omega_1}$, etc. (I learned this from Ko Aoki.) – Peter Scholze Jan 15 '24 at 15:03
  • (One can replace $\omega_1$ here by any regular cardinal $\kappa$. One might object that I'm secretly introducing universes by this choice; I would argue against it. In particular, for any fixed $\kappa$ the above scheme can be used to inductively define $\kappa$-presentable $(\infty,n)$-categories for all $n$; no need for a chain of $n$ universes. Also, I have the feeling that it may actually be helpful to fix $\kappa=\omega_1$, similar to how in condensed math we recently switched to the light setting.) – Peter Scholze Jan 15 '24 at 15:10
  • (In any case, either take $\kappa=\omega_1$, or take a colimit over all $\kappa$ in the end; both ways lead to a theory of presentable $(\infty,n)$-categories that is independent of choices of universes.) – Peter Scholze Jan 15 '24 at 15:11
9

I would consider a conservative extension of ZFC obtained from ZFC by the adition of a constant $\alpha$ and the following axioms:

  1. $\alpha$ is an ordinal ($Ord(\alpha)$).

  2. The sentence $\phi\leftrightarrow\phi^{V_\alpha}$, for each sentence in the original language $\phi$ (axiom scheme).

$V_{\alpha}$ behaves as $V$ (for all sentences in the language of set theory). If two (or more) universes are needed, one can add another constant $\beta$ with the corresponding axioms, and the axiom $\alpha<\beta$.

The proof that the resulting theory is conservative over ZFC is easy.

Assume that $\phi$ is provable from the new axioms (axioms using $\alpha$), in which $\phi$ is in the original language. Since any proof is finite, there are finitely many sentences $\phi_1$, ..., $\phi_n$ such that

$Ord(\alpha)\wedge(\phi_1\leftrightarrow\phi_1^{V_{\alpha}})\wedge...\wedge(\phi_n\leftrightarrow\phi_n^{V_{\alpha}})\rightarrow \phi$

is provable without any new axioms. Therefore, one can think of $\alpha$ as a free variable and the above sentence is provable in ZFC (theorem on constants). Since $\alpha$ does not occur in $\phi$, the following implication is provable in ZFC ($\exists$-introduction):

$\exists\alpha(Ord(\alpha)\wedge(\phi_1\leftrightarrow\phi_1^{V_{\alpha}})\wedge...\wedge(\phi_n\leftrightarrow\phi_n^{V_{\alpha}}))\rightarrow \phi$

Now, the reflection principle for ZFC says that the antecedent is a ZFC-theorem. From modus ponens, ZFC proves $\phi$.

So you can work with the new axioms and $V_{\alpha}$ behaves as a universe, and everything that is proved that does not mention $\alpha$ can be proved already in ZFC.

  • Thanks for this proposal! I have to understand better what this means concretely. How is it avoided that $\alpha$ is strongly inaccessible? Is it again that when one takes unions of sets in $V_\alpha$, one has to make sure that everything is definable in $V_\alpha$? So $\alpha$ may have countable cofinality in $V$, because there are certain countable sequences of ordinals $<\alpha$ converging to $\alpha$, they are just not definable in $V_\alpha$? – Peter Scholze Jan 28 '21 at 15:09
  • Yes, $\alpha$ may have countable cofinality and your explanation is right. $V_{\alpha}$ satisfies each instance of (first-order) replacement only: For each definable function $f$ from $V_{\alpha}$ to $V_{\alpha}$, the direct image of an element of $V_{\alpha}$ is also an element of $V_{\alpha}$. If it were closed under direct images in general, then $\alpha$ would be strongly inaccessible. – Rodrigo Freire Jan 28 '21 at 15:23
  • OK, great! So basically this is a version of the $\Sigma_{15}$-universes mentioned above, but tweaked so that the precise value of $15$ has become irrelevant? – Peter Scholze Jan 28 '21 at 15:27
  • Yes, for each $n$ fixed in the metatheory, $V_{\alpha}$ is a $\Sigma_n$-universe. – Rodrigo Freire Jan 28 '21 at 15:39
  • 2
    Isn't this what Feferman does in the paper “Set-theoretical foundations of category theory" that I mentioned in a comment? – Mike Shulman Jan 28 '21 at 16:57
  • 4
    I think this strategy goes back to Levy, Definability in axiomatic set theory I, 1965, but Levy adds a constant M and axioms saying that M is transitive, countable and reflects every sentence. I have not seen your comment, sorry. – Rodrigo Freire Jan 28 '21 at 17:04
  • 2
    These conservative extensions do pose an ontological problem, though, since without large cardinal axioms they invariably mean that the universe has non-standard integers. Now this is not a problem per se, but it does mean that our "naive understanding of the integers" is incorrect as far as "the mathematical universe" is concerned. This is not necessarily a bad thing, of course, but it is something to contend when thinking about that. – Asaf Karagila Jan 29 '21 at 11:08
  • 1
    Rodrigo, @Mike: While I don't have the unfettered access to Azriel Levy that I had as his TA, I know that (1) he was a good friend of Feferman, and the two invariably influenced each other and shared ideas; and (2) he was extremely modest and hated the whole idea of getting credit for old ideas, when I'd ask him for who did what back in the early days of forcing he would usually say that it doesn't matter and who can remember anyway (although that's often a sign that it was him). My point is that Azriel would be okay with this being credited to Feferman, regardless to who came up with it. – Asaf Karagila Jan 29 '21 at 11:22
  • Thanks @AsafKaragila, I am big fan of Levy, it is good to hear this. I think of those conservative extensions as technical devices which are useful to prove ZFC theorems. These $\alpha$ and $\beta$ can be thought of as Hilbertian ideal objects (differently from inaccessibles) because the extension is conservative. It means that we may not commit ourselves with a strong notion of truth of these extensions. – Rodrigo Freire Jan 29 '21 at 11:41
  • @AsafKaragila Can you say exactly what you mean by "the universe" and "non-standard"? Under a naive reading I would think that "the universe has non-standard integers" would be a contradiction in terms, since the standard integers should by definition be those that exist in the universe. – Mike Shulman Jan 29 '21 at 14:50
  • @Mike: Yes, but the universe has its own meta-theory and meta-universe. It's turtles all the way down. – Asaf Karagila Jan 29 '21 at 15:03
  • @AsafKaragila So in this particular situation, which universe are you calling "the" universe and which are you calling the "meta-universe"? Are you saying that when you prove in ZFC the relative consistency of ZFC/S, you can also prove in ZFC that any model of ZFC/S will contain integers that are nonstandard relative to the ambient ZFC? – Mike Shulman Jan 29 '21 at 15:59
  • @Mike: If the integers are standard, then the model you have is a model of ZFC, internally, so that means that ZFC is consistent. Since the extension is conservative, it can't prove something like that. So if you didn't happen to already have this model in your ZFC universe to begin with, a universe of ZFC/S is going to be non-standard. – Asaf Karagila Jan 29 '21 at 18:01
  • 1
    @AsafKaragila Sorry, I'm lost. The proof of relative consistency doesn't go by constructing a model. I thought the whole point was that you can't prove internally in ZFC/S that $V_\alpha$ is a model of ZFC since it only satisfies all the axioms as a schema. What does that have to do with standardness of integers? (And can you please talk about "integers in model X" so that I don't get confused about which model you're calling "standard"?) – Mike Shulman Jan 29 '21 at 18:07
  • @Mike: In the ZFC/S universe, $V_\alpha$ is still a set, it still has a truth predicate, and it still provides us the consistency of whatever theory is true there. So to avoid having proved Con(ZFC) internally to that universe the only solution is that internal-ZFC now have new axioms, i.e. non-standard integers, relative to its meta-theory. – Asaf Karagila Jan 29 '21 at 18:19
  • @AsafKaragila Okay, maybe I almost get it. Say from a model $M$ of ZFC we build a model $M'$ of ZFC/S. If in $M'$ it were true that $V_\alpha \vDash$ ZFC, then we would have $M' \vDash$ Con(ZFC), and we can't expect to construct from any model of ZFC a model of ZFC+Con(ZFC). But otherwise, in $M'$ there must be an axiom of ZFC that $V_\alpha$ doesn't satisfy, which must be a non-standard axiom since $V_\alpha$ satisfies all the standard axioms. Right? – Mike Shulman Jan 29 '21 at 19:01
  • But it seems to me that there could still be some models of ZFC/S that do satisfy Con(ZFC) (at least, if Con(ZFC) is true). Were you claiming that all models of ZFC/S must contain nonstandard integers? – Mike Shulman Jan 29 '21 at 19:03
  • 1
    @Mike: That is correct. I didn't claim that all models of ZFC/S are non-standard, of course. But if there is a well-founded model, then it had to involve large cardinal axioms, which is what we are trying to avoid. (Note that Rodrigo's answer formulates the model as a $V_\alpha$, rather than any transitive model.) For example, assuming $0^#$ exists, let $\kappa=\omega_1$ in $V$, then $L_\kappa=V_\kappa^L$ is an elementary submodel of $L$, witnessing ZFC/S as a standard model. – Asaf Karagila Jan 29 '21 at 19:06
  • 1
    @AsafKaragila Okay, I think I finally get it; thanks for your patience. I think my attitude toward this would be the same as (what I took to be) Rodrigo's though: the goal is to prove theorems of ZFC, for which purpose we can just work syntactically with ZFC/S and then apply conservativity, rather than worry about any actual models of ZFC/S. – Mike Shulman Jan 29 '21 at 20:17
  • @Mike: I always thought that people outside of set theory (especially those espousing different foundations, e.g HTT) are not interested in proving "theorems in ZFC", but rather just proving theorems. To that end, I am a bit worried if there is a shift to ZFC/S that even more of those silly "set theory is a bad foundations because it proves junk theorems" arguments will be raised, and this time they will be "more well-founded" about the ill-foundedness of the universe (pun not intended, but a happy accident). – Asaf Karagila Jan 29 '21 at 22:24
  • 2
    I don't know about the HoTT community, but the number theory community (who is now using HTT) has a (bad?) habit of caring about ZFC. But they often appreciate cheats, so I'm sure they'd appreciate such a cheaty way of knowing that the main theorems of HTT are theorems of ZFC. – Peter Scholze Jan 29 '21 at 22:32
  • @AsafKaragila Well, someone who's just interested in "proving theorems" wouldn't be asking this question to begin with, would they? They'd just assume some Grothendieck universes and get on with it. – Mike Shulman Jan 30 '21 at 01:41
  • Anyway, the problem of ill-foundedness only arises if you (1) believe that the "real world" is ZFC and (2) want to use ZFC/S and (3) want to think about a model of ZFC/S rather than just its syntax. My guess would be that if there really were a shift to ZFC/S, then the average mathematician would just start to think of ZFC/S as the "real world" instead, in which case they wouldn't see any problem. – Mike Shulman Jan 30 '21 at 01:44
  • 1
    The main reason that I see for someone to be asking a question like this is not that they believe ZFC is the "real world" but don't believe that large cardinals are "real", but rather that they don't want to assume stronger hypotheses than necessary, i.e. that what they care about is not truth but consistency strength. For that purpose, the conservativity of ZFC/S over ZFC means we can feel free to use it without increasing the consistency strength, but also without worrying in practice about relating it back to ZFC at all or even talking about any models. – Mike Shulman Jan 30 '21 at 01:46
  • @MikeShulman I mostly agree with your last comment. To nitpick: there's a difference between caring about strength of hypotheses versus caring about consistency strength of hypotheses. For a silly example, I believe that $ZFC$ and $ZFC+\neg Con(ZFC)$ are equiconsistent, but you don't need to think that ZFC "is the real world" to prefer theorems of the former to the latter. In the present case, you might prefer ZFC theorems to ZFC+Universes theorems for portability: to interpret the former, you just need a model of ZFC, which can be constructed with "less overhead" than a ZFC+Universes model. – Tim Campion Jan 30 '21 at 16:03
  • @TimCampion Yes, that's absolutely true: generality is an important consideration too. – Mike Shulman Jan 30 '21 at 22:05
  • @AsafKaragila Some of your comments seem to equate "large cardinal assumption" with "anything that provides standard models of ZFC" or "anything that provides $V_\alpha$'s satisfying ZFC". One of my favorite metatheories is ZFC plus a satisfaction predicate $S$ for $\in$-formulas, subject to the usual clauses characterizing satisfaction plus replacement for formulas involving $S$. This gives a club of $\alpha$'s such that $V_\alpha\models$ ZFC, but I wouldn't call it a large cardinal axiom. – Andreas Blass Feb 03 '21 at 01:32
  • @Andreas: Yes, there's a lot of subtle points here. In general, though, I think that most people will agree that the existence of a worldly cardinal is a large cardinal axiom, which is $V_\alpha$ satisfying ZFC. Of course, in the case of a worldly cardinal we are talking about the internal version of ZFC. But if the internal and external versions agree... – Asaf Karagila Feb 03 '21 at 01:58
3

In response to the edit which nails things down to a formal system involving cardinals $\kappa_{-1} < \kappa_0 < \kappa_{1/2} < \kappa_1$:

I'm going to go out on a perhaps more ill-advised limb and predict that in order to fit Chapters 1-4 into this formal system, no real cardinal arithmetic will be necessary. Rather, for this part of the book, all you'll have to do is go through and add to various theorem statements hypotheses of the form "$X$ is $\kappa_{-1}$-small". After all, this part of the book really only deals with small objects, with the exception of a few particular large objects like the the category of small simplicial sets, the category of small simplicial categories, etc., and things like slice categories thereof. Various model structures are constructed, but I believe in each case one can get by with using the special case of the small object argument for generating cofibrations / acyclic cofibrations between finitely-presentable objects, so that no transfinite induction is needed. On the face of it, straightening / unstraightening have the look of constructions which might be using set theory in serious ways. But I'm going to go ahead and predict that they present no problems to the proposed formal system.

Chapter 5 becomes more annoying. I believe that one will have to make some careful choices about the core theorems of presentable ($\infty$)-categories. What makes presentable categories tick is that they package the adjoint functor theorem very cleanly, but as you say, the ordinary adjoint functor theorem comes with caveats in this setting. I might go so far as to say that the whole point of thinking about presentable categories in the first place is completely undone in this setting. You won't be able to prove basic things like "presentable categories are precisely the accessible localizations of presheaf categories". I predict that whatever choices are made about formulating weak versions of the core theorems of presentable categories in this setting, there will be some application or potential application which suffers.

Chapters 5 and 6 also contain some theorems about particular very large categories such as the $\infty$-category of presentable $\infty$-categories and the $\infty$-category of $\infty$-topoi [1]. The system seems to be such that this won't really be an issue per se, except that the issues encountered in basic presentability theory will now be compounded. You won't be able to prove that $Pr^L$ is dual to $Pr^R$. You won't be able to prove Giraud's theorem (well, the definitions are going to be in flux anyway, so I should clarify: you won't be able to prove that left exact accessible localizations of presheaf categories are the same as locally small categories satisfying a a list of completeness, generation, and exactness conditions). So any theorem about $\infty$-topoi whose proof proceeds by starting with the presheaf case and then localizing will have to be completely re-thought.

Maybe I'm off-base here, but I believe that significant extra work and genuinely new mathematical ideas would be required for Chapters 5 and 6, and the the result would be a theory which is substantially harder to use.

By contrast, I think if you're willing to restrict attention to large categories which are definable from small parameters, then although you'll be missing the beautiful ability to say "we proved this for small categories but now we can apply it to large ones", you'll end up with a much more useable theory of presentability, without leaving ZFC.

[1] Actually, in usual foundations these categories are (up to equivalence) only large and not very large (more precisely, they are have $\kappa_0$-many objects and $\kappa_0$-sized homs), but a modicum of work is required to show this. Will that still be the case in this formal system? I'm not sure.


EDIT: A long comment in response to Peter Scholze's answer.

  • One thing I just realized is that if $\kappa_0$ is not a $\beth$-fixed-point, then not every set in $V_{\kappa_0}$ has cardinality $<\kappa_0$, so that the notions of "smallness" are multiplied. Happily, I think your formal system proves that $V_{\kappa_0}$ has $\Sigma_1$-replacement, which implies that it's a $\beth$-fixed-point. Crisis averted!

  • Perhaps this approach of systematically using definability hypotheses within a "universe setting" will be workable -- combining the "best of both worlds". One nice thing is that even though you're explicitly using metamathematical hypotheses, it would appear that you'll still be able to state and prove these theorems as single theorems rather than schema.

  • I'm a bit confused about Proposition 5.2.6.3 (the last one you discuss, and a baby version of the adjoint functor theorem). I presume that the presheaf category $P(C)$ will be defined to comprise those functors $C^{op} \to Spaces$ which lie in $Def(V_{\kappa_0})$. When we pass to a larger universe, the transition is usually pretty seamless, because we expect $P(C)$ to have all colimits indexed by $\kappa_0$-small categories -- a perfectly natural property to work with in $V_{\kappa_1}$. Indeed, the first step of Lurie's proof of 5.2.6.3 is to show that a left adjoint exists using the fact that $P(C)$ has all small colimits [2]. However, in the present setting we can never assume that $\kappa_0$ is regular, and therefore we can never assume that $P(C)$ has all small colimits. The best we can say is that $V_{\kappa_0}$ thinks $P(C)$ has all small colimits. As long as we're working in $V_{\kappa_0}$, this property is "just as good" as actually having all small colimits. But when we move up to $V_{\kappa_1}$, suddenly we have to think of it for what it is -- a metamathematical property. Perhaps later I will sit down and try to see whether Lurie's proof of 5.2.6.3 can be made to work in this setting, but I think prima facie it's unclear.

[2] Only after verifying existence abstractly in this way does he show that the left adjoint must be the functor indicated. Of course, this maneuver is actually an additional complication which comes with the $\infty$-categorical setting -- in ordinary categories the formulas for the two functors can be directly verified to be adjoint, but in $\infty$-categories the formula for the left adjoint is not obviously functorial.

Tim Campion
  • 60,951
  • Sorry, I've misunderstood the proof of 5.2.6.3 a bit. The functor $P(f)$ was constructed earlier and is already known to be functorial. So the peculiar form of the proof is for some other reason which I don't fully understand. But at the very least, the proof does not work as written in the proposed formal system, even when the definitions involved are understood to have been modified. – Tim Campion Jan 29 '21 at 17:27
  • 1
    Thanks! Here's a cheap fix: Besides the "definable" version of $\mathcal P(C)$, one also has the version $\tilde{\mathcal P}(C)$ of functors $C^{op}\to \mathrm{Set}$ in $V_{\kappa_1}$. The proof works as written for $\tilde{\mathcal P}$, but both $\tilde{\mathcal P}(f)$ and $\tilde{G}$ preserve the subcategory $\mathcal P(C)\subset \tilde{\mathcal P}(C)$ (inducing $\mathcal P(f)$ and $G$), so one gets an induced adjunction on these full subcategories which is we one we want. – Peter Scholze Jan 29 '21 at 21:21
  • 1
    @PeterScholze Interesting. How do we know that $\tilde{ \mathcal P}(f)$ exists? Is there some sense in which we should expect $\tilde{\mathcal P}(C)$ to be a free cocompletion of $C$? After all, $Set$ doesn't have all small colimits ($V_{\kappa_0}$ just thinks it does), so I don't expect that $\tilde{\mathcal P}(C)$ has all small colimits. Of course, for these purposes having small colimits is less important than the freeness of those colimits, but it gives me pause. Or -- do we not need to know it exists, since the proof actually verifies the existence of a left adjoint abstractly? – Tim Campion Jan 29 '21 at 21:45
  • Or maybe I've misunderstood -- I assumed that $Set$ was the category of $\kappa_0$-small sets, but perhaps you're taking $\tilde P(C)$ to be category of $V_{\kappa_1}$-definable functors from $C^{op}$ to the category of $\kappa_1$-small sets. – Tim Campion Jan 29 '21 at 21:51
  • 1
    OK, I think this is a case where it's better to cut at a regular cardinal, so take say sets of size $\leq \kappa_0$; this has all colimits of size $\leq \kappa_0$, and has a corresponding freeness property. This is related to the fact that in the theory of presentable categories, all cuts happen at regular cardinals (and those cuts are basically independent of the "universe" cuts). – Peter Scholze Jan 29 '21 at 22:21
  • @PeterScholze I agree that should probably work. It's a bit annoying, because the logic of the book as written (as in most treatments of presentability) is (1) work out the properties of presheaf categories, then (2) apply this to properties of subcategories of presheaf categories. With this approach, it looks like we'll be weaving in some considerations from (2) to the treatment of (1). – Tim Campion Jan 30 '21 at 15:35
  • To start, we'll have to strengthen the statement of Thm 5.1.5.6, where the existence of $P(f)$ is deduced from the model category considerations of Lem 5.1.5.5. We might have to go back to earlier chapters and construct more model categories cut off at regular cardinals, using versions of the small object argument with various cardinal parameters inserted. At this point it feels to me at any rate that we're starting to need new mathematical ideas to push things through. So although I still agree that Grothendieck universes are really a convenience, it underscores just how convenient they are. – Tim Campion Jan 30 '21 at 15:35
  • It's difficult to reach a clear conclusion on an ambiguous question like "Does the book essentially work in a weaker formal system". Here's a slightly more precise version of that question: "To rewrite the book in a weaker formal system, would it be easier to keep the same theorem numbering and just make some line-by-line edits, or rather to invest time in rethinking the basic organization of the book, and write a new book which closely mirrors the original but does not attempt to be a direct translation?" I think the test cases we've seen suggest that the answer is the latter. – Tim Campion Jan 30 '21 at 18:15
  • Here's another precisification: If we were to rewrite HTT in a weaker formal system and then, before publishing our new book, upload its theorem statements onto wikipedia articles, would our edits be taken down for being "original research"? Again I think the the test cases we've seen suggest the answer is "yes". – Tim Campion Jan 30 '21 at 18:19
  • 2
    One more precisification: If a bright young professor were to rewrite HTT in a weaker formal system and then proudly present the results to the tenure committee as the centerpiece of their resume, should they get tenure? Of course, the question is still too ill-defined to answer as stated. Coming into this, I would have been somewhat sympathetic to the answer of "no", but I think after looking at these examples, I'd be more inclined to champion the cause of this hypothetical researcher. – Tim Campion Jan 30 '21 at 18:23
  • Well, I'm certainly rambling now, but I think we can at least say we've reached the conclusion that the question is MO-hard. – Tim Campion Jan 30 '21 at 18:37
  • 2
    OK, we seem to agree on "Probably yes, but you have to put some effort in". Your last argument convinced me: One should definitely get credit for a translation into ZFC. Later tonight, I'll try to write another longer answer to some questions that came up. – Peter Scholze Jan 30 '21 at 19:59
  • 2
    These all sound like arguments in favour of having Kerodon contain these more sophisticated treatments – David Roberts Jan 30 '21 at 22:27