Bacon on higher-order logic

In the dark old days of early logic, there was only syntax. People introduced formal languages and laid down axioms and inference rules, but there was nothing to justify these except a claim to "self-evidence". Of course, the languages were assumed to be meaningful, but there was no systematic theory of meaning, so the axioms and rules could not be justified by the meaning of the logical constants.

All this changed with the development of model theory. Now one could give a precise semantics for logical languages. The intuitive idea of entailment as necessary truth-preservation could be formalized. One could check that some proposed system of axioms and rules was sound, and one could confirm – what had been impossible before – that it was complete, so that any further, non-redundant axiom or rule would break the system's soundness.

That's how things played out in first-order logic. Systems of second-order logic turned out to be incomplete, and we found that there was no way to make them complete. Full second-order logic, it seems, can only be grasped semantically.

Or so I used to think. Andrew Bacon clearly disagrees. In his Philosophical Introduction to Higher-order Logics (Bacon 2023), he wants to take us back to the dark old days.

In Bacon's conception of logic, there is only syntax. The language of higher-order logic is assumed to be interpreted, but there is no systematic account of what an interpretation might look like; the standard set-theoretic interpretation is rejected. Axioms and inference rules can therefore only be justified by self-evidence. (There are a few chapters on model theory, starting on page 287, but Bacon makes clear that he doesn't think of model theory as elucidating or formalizing the meaning of higher-order sentences. Model theory is introduced as a tool to derive proof-theoretic results.)

What about the unformalizability ("incompleteness") of second-order logic? This, too, is dismissed as a mirage, as it is based on the supposedly faulty assumption that second-order entailment is defined in the standard model-theoretic fashion.

I can understand why a friend of higher-order logic might be reluctant to endorse the standard set-theoretic semantics, which effectively construes second-order quantifiers as first-order quantifiers over sets. Perhaps a truly faithful semantics for second-order logic should invoke second-order quantifiers in the meta-language. This is especially pressing when we think about second-order set theory: the standard semantics for second-order ZFC yields an interpretation that is surely unintended. But this much is true even for first-order ZFC: '\(\in\)' is not supposed to denote a set of ordered pairs.

In any case, I think it is important to think of models as models. In propositional logic, a model is a truth-value assignment to the sentence letters. Nobody seriously thinks sentence meanings can be identified with truth-values. The real meaning of a sentence should at least specify its truth-conditions. We could then determine whether one sentence entails another by checking if there's any conceivable scenario in which the first is true and the second false. To determine whether one sentence logically entails another, we would accordingly have to check if there is a conceivable scenario in which the first is true and the second false, on any assignment of truth-conditions to the sentence letters. Equivalently, we would have to check if every pair of a scenario and an assignment of truth-conditions to the sentence letters that makes the first sentence true also makes the second sentence true. If you think about this for a moment, you realize that it can be simplified: any pair of a scenario and an assignment of truth-conditions to the sentence letters determines an assignment of truth-values to the sentence letters. Conversely, any such truth-value assignment is determined by some pair of a scenario and an assignment of truth-conditions. This compressed information about a pair of a scenario and an interpretation is all we need, in propositional logic, to determine whether an arbitrary sentence is true in the scenario, under the interpretation. That's why we can model a pair of a scenario and an interpretation by an assignment of truth-values to the sentence letters. A model isn't the real thing. It simplifies the real thing down to whatever we need to determine whether any given sentence is true or false.

Return to second-order logic. Suppose the intended domain of discourse forms a set, as is usually the case. Whatever the true meaning of a predicate is, it is tempting to assume that for any subset of the domain, a predicate could apply to just the elements of that subset. If so, we can model predicate meanings by such subsets, and we can model the meaning of quantifiers by assuming that they range over all the subsets.

This leads to deep and fascinating questions. Once the domain is infinite, the notion of "all subsets of the domain" is far from innocent. If the domain consists of the real numbers, for example, it depends on the ambient set theory (the set theory in which the model theory is couched) whether there are any subsets of the domain whose cardinality is smaller than the cardinality of the domain, but larger than the cardinality of the naturals. Once again, this issue becomes more pressing if we think about set theory. Second-order ZFC is categorical: all its models are isomorphic. But whether it validates the Continuum Hypothesis depends on the ambient set theory!

I don't see how these deep and fascinating questions can be avoided by refusing to engage with semantics. If the higher-order quantifiers are plentitudinous, ranging over "properties" whose extension may be any subset of the domain, we face the question of exactly how plentitudinous they are. As Bacon points out in passing, there are pure second-order statements that decide the Continuum Hypothesis, one way or another. Is there a right way? Bacon seems to think so. But how could we tell which way is right? Surely neither statement is self-evident. And anyway, the determinacy assumption seems to go against a growing consensus in set theory according to which the Continuum Hypothesis resembles the Parallel Postulate in geometry: we can study structures in which it is true and structures in which it is false; neither is objectively more real than the other.

Bacon says very little about any of this. Most of the book is devoted to an entirely different issue.

That issue comes to light if we assume that our formal language doesn't just have second-order quantifiers, which are by their nature extensional, but also some further, non-extensional third-order predicates.

Bacon's running example is higher-order identity. He assumes that we have a predicate '=' that can be put between predicates and between sentences (which are zero-ary predicates). For Bacon, the central question of higher-order logic is how this predicate should behave – syntactically, of course: what are its introduction and elimination rules?

In the standard semantics of second-order logic, the denotation of a predicate is its extension. If we interpret '=' as identity of denotation, we get the result that 'F=G' is true whenever F and G are co-extensive. Things are even worse for zero-ary predicates: 'p=q' comes out true whenever p and q have the same truth-value.

This consequence is Bacon's main reason for rejecting the standard semantics. It would, he says, imply "that there are only two propositions", which, he says (on p.329), is "widely rejected in contemporary metaphysics".

Much of Bacon's book explores candidate rules for higher-order identity. Chapters 4-8, after some introductory chapters on lambdas, present a "classical" system in which '=' may be placed between two expressions iff they are provably equivalent. Chapters 9-13 discuss alternative systems that allow identity to fail even for provably equivalent expressions. According to Bacon, the choice tracks a deep question about the granularity of reality. This seems to be a common assumption in higher-order metaphysics. I don't get it.

Let's briefly take a different example first. Suppose we add to the standard language of propositional logic a box operator, whose intended interpretation is, say, 'it is known that' or 'it is provable in Peano Arithmetic that'. (As a sentence operator, the box is a higher-order predicate.) We clearly have to change our model theory. Since the box isn't truth-functional, our models must contain more information about the interpretation of sentence letters than their truth-values. That's why standard models in epistemic logic and provability logic have a lot more structure.

How fine-grained should the sentence meanings in the new models be? It depends. In my view, there are attractive intensional concepts that can claim to regiment the informal concept of knowledge. But there are also hyperintensional regimentations. For provability in Peano Arithmetic, we obviously need a very fine-grained, hyperintensional semantics.

Higher-order identity is special because it seems to rule out not only models that are too coarse-grained (like the box), but also models that are too fine-grained. If we have some pre-theoretic ideas about which sentences have the same meaning and which don't, the meanings in our models must have the same level of granularity.

But it would be entirely misguided, I think, to believe that sentences in a formal language have a fixed meaning, independent of any stipulations we might have made, so that there is an objective fact of the matter about whether or not two of them have the same meaning.

In provability logic, for example, the intended interpretation reads the sentence letters as standing for sentences of Peano Arithmetic (or their gödel numbers); \(\neg\neg p\) denotes a different sentence from \(p\). If we add a higher-order identity predicate to provability logic, we should adopt a "non-classical" system in which \(\neg\neg p = p\) isn't provable. In standard epistemic logic, by contrast, the sentence letters express possible ways the world might be. Here, \(\neg\neg p = p\) should be provable.

Note that I'm not reasoning from the granularity of the models. (This would be question-begging.) I'm not saying that the sentence letters in epistemic logic stand for subsets of some domain \(W\). We always start with an informal conception of meaning.

But informal or not, meanings don't fall from the sky. A language doesn't interpret itself. I certainly don't see why we would incur any controversial metaphysical commitments by adopting a version of provability logic in which \(\neg\neg p = p\) isn't provable.

Let's keep logic and metaphysics apart.

Bacon, Andrew. 2023. A Philosophical Introduction to Higher-Order Logics. Routledge.

Comments

# on 08 December 2025, 18:09

Hi Wolfgang!

Thanks for engaging with the book, I find your perspective very interesting!

There are a few points I want to make, I hope this isn't too rambly. I think John Burgess often makes a big deal of the distinction between a model theory and a semantics, where the former usually concerns a class of set-theoretic structures at most one of which could be a semantics (or maybe none, as we find in the case of set theory, higher-order logic, free logic, many-valued models of non-classical logics, and so on).

I think the question of the completeness of a HOL relative to this or that class of models is not a particularly important one. Provided the class contains the intended model, pretending there is one for a minute, the model theory will be sound (i.e. validate only truths). But as to how many truths you get, it depends on how many unintended models you throw into the class along with the intended one. If you have only the intended model of the logical constants in the class, then you get all the logical truths of that language, if you include all general models you get H, and you can get anything in between through some class of models. (If you're considering languages with non-logical constants, the *logical* truths should contain all models obtained by interpreting the non-logical constants in a single intended interpretation of logical constants.)

Note though that I don't dismiss the question of whether you can axiomitize the logical truths of a pure higher-order language. The above is a model theoretic way of putting it. In the book I consider the logical truths to be sentences that have a true universalization -- i.e. the result of replacing the non-logical constants with variables and universally generalizing them is a truth. Under a plausible axiom of infinity assumption, this means all the truths of arithmetic can be interpreted in the logical truths.

But if we're talking about the class of sentences valid in a class of models including unintended models, I'd question the significance of that class of sentences. There are myriad natural classes of models that contain the intended model, all with different logics.

# on 08 December 2025, 18:16

By the way, I'm not against giving a semantics for a higher-order language in a higher-order language itself. I do this in the paper "Radical Anti-disquotationalism", and there's a compressed version of that in the book. But it's not a model theory: you can describe a particular interpretation, the intended one, but can't quantify over different interpretations including unintended ones, as you do in model theory. (It has to do with the type strictures in HOL.)

# on 08 December 2025, 19:01

Oh, one more thing: on your point about whether "it is provable in Peano Arithmetic that" requires one to endorse certain higher-order hyperintensional theories of granularity. This depends on whether you think that the operator is an instance of the generalization:

For any operator O, if O(1=1), then O(every even number greater than 2 is the sum of two primes)

and also on your view about the truth value of that instance. Many people think it is not an instance. I.e. there is no self-standing meaningful operator expression that when applied to a sentence A means the same as "it is provable in Peano Arithmetic that A", because they will think that natural language is messy and the latter sentence has hidden quotation marks, and the only sensible logical form involves a metalinguistic predicate (something of type e->t, not t->t).

It's actually very hard to understand that expression as a propositional operator, since provability in PA is a syntactically defined thing, that is meaningful only in the context of the particular language and system you have defined it for. If you change the language at all, or replace sentences with propositions, even structured propositions, it's totally unclear what it means.

Also: I take it that whether there are two different planets, Hesperus and Phosphorus, or just one, can't be made true or false just by "choosing a different model". Of course, it is convenient for building models of propositional attitude operators to posit two different planets, but that would be revisionary astronomy! So what's the difference between saying that which identities between sentences are true depends on what we're trying to model, and saying that which identities between names are true depends on what we're trying to model. Why an asymmetry?

# on 10 December 2025, 13:03

Thanks Andrew! That's helpful.

I wonder if our main disagreement is simply about how to understand "logic". You seem to assume (correct me if I'm wrong) that there are metaphysically privileged domains of objects, properties, propositions, operations, etc., and that the expression types of a logical language must be interpreted as picking out elements of these domains (loosely speaking; you know what I mean). Personally, I doubt that reality neatly sorts itself into domains of these quasi-grammatical types. I suspect that reality has less structure. But even if reality does sort itself in the way you assume, I wouldn't make your suggested interpretation a constraint on what counts as logic. I think textbooks on first-order logic rightly don't say anything about the metaphysical status of the things picked out by individual constants. I would expect a textbook on higher-order logic to do the same.

The case of provability logic was meant to illustrate a use of logic that is clearly not meant to talk about the metaphysical types that your interpretation would enforce. In a higher-order version of provability logic, the 2nd-order quantifiers would effectively range over sentences, not over the metaphysically privileged domain of propositions. I wasn't thinking of provability logic as a fragment of English. The language is an artificial construct. It might be a version of the standard propositional modal language with propositional quantifiers. There can be no debate on whether the provability operator is a genuine sentence operator in that language: it is stipulatively defined to be one.

A smaller point: I disagree that a class of models is of interest only if it gives the logical constants their intended interpretation. Suppose the intended intepretation of '∧' is a function that maps two sets of worlds to their intersection. In the standard semantics of propositional logic, '∧' is instead interpreted as a function that maps pairs of truth-values to truth-values. But this doesn't mean that completeness wrt the class of truth-value assignments is uninteresting: if a proof system is incomplete in this way, it will also be incomplete wrt the class of intended models.

(I don't understand the point about Goldbach's conjecture. Are you assuming that it is provable in PA? Or disprovable? Or undecidable?)

# on 11 December 2025, 01:06

"I wonder if our main disagreement is simply about how to understand "logic"."

I don't really put much stock in the labels "metaphysics" or "logic"--feel free to use them how you like!

"You seem to assume (correct me if I'm wrong) that there are metaphysically privileged domains of objects, properties, propositions, operations, etc., and that the expression types of a logical language must be interpreted as picking out elements of these domains (loosely speaking; you know what I mean)."

I do not believe this. I'm pretty explicit in the textbook that you do not have to take this view either; in fact the interpretation of the higher-order formalism I prefer seems very far from that picture. Talk of "propositions", "properties", etc in English is really just a way of pronouncing sentences of higher-order logic, but sentences which might mean something else entirely.

"In a higher-order version of provability logic, the 2nd-order quantifiers would effectively range over sentences, not over the metaphysically privileged domain of propositions."

Most people using higher-order logic these days are roughly following Prior in "The Objects of Thought" (I'm certain many earlier logicians also implicitly understood it this way, but Prior was explicitly clear about this.) Higher-order generalizations don't need to be explained by defining what they "range over", or made meaningful by providing a "domain". They cannot be, for familiar reasons. A generalization, like "(F).John Fs if he Fs" can be understood up to logical equivalence by its logical connections to its instances, "John talks if John talks", "John walks if John walks", etc. (I.e. the introduction and elimination rules). So, on the interpretation of the higher-order quantifiers as expressing generalizations, in this stipulative sense, it really just amounts to whether you think "it's provable in PA that" is an instance of the schema I mentioned. There are other interpretations of higher-order quantifiers in which they can be said to "range over" a domain and they range over sentences, or propositions, or some other abstract objects. But those are just different interpreted languages that share a syntax with the language that expresses higher-order generalizations.

" (I don't understand the point about Goldbach's conjecture. Are you assuming that it is provable in PA? Or disprovable? Or undecidable?)"

Oh yes, my bad. I was thinking of Fermat's last theorem, and got mixed up. I meant a non-obvious truth that's identical to 1=1 on a course grained theory.

# on 11 December 2025, 09:16

I see. Right, I'm trying to reconstruct your view in semantic terms, and you refuse. Fair enough. (My talk about "domains" was meant to be understood loosely, as I said, but never mind.) I struggle to understand how an expression could get an interpretation by its connection to other sentences. So I was wrong to think that our disagreement is mainly about whether the label "higher-order logic" should be restricted to logics with a certain interpretation. It's also about how to think about interpretations. Unfortunately, that disagreement is a lot harder to resolve. Thanks again!

# on 13 December 2025, 01:43

Well, my way of turning model talk into stuff about the world is to assume the models are not purely mathematical structures, but interpreted models (like in the sciences; two interpreted models could have the same underlying mathematical structure, e.g. a single Kripke structure could represent two different peoples knowledge). Most models are only partially interpreted though: some aspects of the model are artefacts, like the origin in a model of space. I'm inclined to think things like the cardinality of an interpreted model of ZFC must be also be an artefact. That's also how I'm thinking about interpreted (set) models of higher-order logic: things like their cardinality, the fact that they are divided up into a hierarchy of domains, and so on, are uninterpreted features of the model. Nothing about using a typed language requires you to think that reality has a certain metaphysical structure. After all, even though the types don't go very high, first-order logic is still typed, and presumably doesn't require this. So if that's an OK way of reconstructing your loose talk of domains, then that's my position and I think that counts as denying it?

"I struggle to understand how an expression could get an interpretation by its connection to other sentences"

Well, it might be that you need to already understand some expressions that aren't sentences. It's like the way that you can come to understand a word by reading a dictionary, if you already understand the words used in the definition. You might think that if you understand a language without conjunction, and you add it and are told that the conjunction follows from the conjuncts and vice versa, you have been told enough to figure out what conjunction means, or at least to narrow its meaning down up to logical equivalence.

I don't think this is the only way to interpret a higher-order language: if you could somehow specify the interpreted models in a way that makes clear which bits are interpreted, then I'd be fine with it. But doing this always seems to lead people into taking too much of the model theory seriously, so it's no pedagogically helpful.

Add a comment

Please leave these fields blank (spam trap):

No HTML please.
You can edit this comment until 30 minutes after posting.