Why do we need a book about logic and information? Don’t we have a theory of information – Shannon’s? Well, yes, but a theory is good for what it’s designed for, and need not be good for other areas. Shannon’s theory is great if you are designing transmission codes for lossy communication channels. It’s not so good if you want to understand how people communicate with each other. But doesn’t good old-fashioned propositional logic cover that? Well, no. Consider the example sentence Devlin uses. Nancy says to John: “It’s cold in here”, with the intent to get John to close the window. Propositional logic doesn’t hack it. Devlin’s book uses situation theory, in order to formalise properties of information as it occurs in everyday discourse.
He starts off with some discussion of the difference between perception of information about the world, and cognition related to that perception. This cognition doesn’t have to be high level intelligence: even a thermostat can do it to some degree:
He starts building his theory from some “givens”: a preexisting ontology, and that the agents (that are transmitting and receiving information) are capable of “individuating” things from this ontology when they occur in the world.
I wonder if this is somewhat begging the question. If the theory is built on the given agent ontology, how can agents with different ontologies communicate? I am reminded of the William James quote “Other minds, other worlds from the same monotonous and inexpressive chaos!”. Still, maybe starting with a theory of matched ontologies helps, and we can wait for communication with alien ontologies until version 2.
This ontology includes situations (which are something akin to the context within which the information exists). These situations can be individuated by the agents and manipulated in the theory just like other things in the ontology.
Situations are a new kind of thing, different from the other kinds of things in the ontology. We might feel we understand individuals, locations, properties, relations, but just what are situations? Well, what they are is defined by how they behave. They are defined by the theory. Mathematically, this
Although he puts it in “scare quotes”, Devlin assumes that the reader knows what ‘extensional definition’ means. It means defining something by listing everything about it; the term is typically used of set definitions, where an extensional definition explicitly lists the members: {Tweety, (1, Rover), 42}. This contrasts with an ‘intensional definition’, given in terms of some membership property, for example, the set of “blue birds” (the members are all things that are blue and are birds) or “dangerous situations”. (Paulos, in his 1998 book Once Upon a Number, notes the dangerous things that can happen if these kinds of definitions get muddled in our minds. He also talks about situational logic.)
Since situations can’t in general be defined extensionally, it means there is always potentially something more to the situation: it is open. So agents necessarily have only partial (usually very partial) information about a situation.
I think the reason the term “roundabout means” is used here is that the intensional definition of the situation seems to be something like “all information that is about the situation”.
Devlin also has something to say about modelling, about being careful to distinguish properties of the model from those that are merely of the modelling techniques.
The [5] in the quotation refers to the 1983 book Situations and Attitudes. Barwise and Moss make similar points about modelling (in particular, just because you can model something as a set doesn’t mean that it is a set) in their 1996 book Vicious Circles, which is all about non-well-founded set theory.
Devlin then set up the machinery for defining information in situations. He has some unfortunate examples, though. Take:
Jon sold the house.This seems perfectly informational. Hearing me, and, let us suppose, being aware who Jon is, you learn something about this individual. … the fact that Jon had a house for which he had no further need, the fact that someone else wanted to buy this house, and so on, ….
It doesn’t seem perfectly informational to me, however. Reading this, I immediately ask “Is Jon the owner, or is he selling it as an estate agent [realtor]?” Indeed, is Jon the legitimate owner, or has he sold it fraudulently? (Compare “Jon sold the Brooklyn Bridge”.) These ambiguities can probably be resolved by the situational context. But here the sentence is used as a purely propositional example. [Or maybe the problem is just me reading everything with “Science Fiction reading protocols”.]
The theory uses relations between things (such as the relation “sells”, with slots for seller, purchaser, sold item, price, location, and time). One thing I found rather puzzling was the insistence on a fixed number of slots in a given relation. This seems overly restrictive, give the “open” nature of the situation within which the sale occurs. What if it becomes important to know something else about the sale? What is the “location” of the sale if it occurred on eBay? What is the “time” if the seller and buyer are in different time zones? etc, etc. However, since this fixed number of slots does not seem to become crucial in anything that follows, let’s move on.
An important part of the theory is that the agent need not be a powerful reasoner. There can be a logic that the theorist, peering at the agent and its discourse, uses to describe the system. And there can be a lesser theory used to describe the agent’s view of the system. And, harking back to the model not being the reality, there is also the process that the agent is actually using (as opposed to the theorist’s model of the agent’s logic).
This emphasis on the modelling approach potentially leaking through to the model of the system, made me wonder about:
a:Tfor a given object a, is determinate; that is to say, it is a fact of the world that either a is of type T or a is not of type T.
Consider the type T being “tall people”, with the corresponding defining property of “being tall”. Where is this sharp boundary between “being tall” and “not being tall”? This is a classic example of “fuzzy logic”: there are actually no crisp boundaries between some of our classifications. Our ontology is more subtle than the one considered here.
Having set up the infrastructure to capture “information” in a situated manner, Devlin moves on to an important feature of communicating agents: the information in their mental states. Some mental states are intentional: states that are “about” something in the world. These include beliefs (mental states about what you think the state of the world is), desires (mental state about how you would like the state of the world to be), and intents (mental states about what you are going to do to the state of the world). That these are different is shown by what happens if you discover your mental state and the state of the world do not correspond: if it is a belief, then your belief is mistaken and you change the belief to match the world; if it is a desire, you might (form an intent to) change the world to match the desire.
Note the words used here: “intentional” and “intent”, as well as the previous “intensional”. These similar-looking words can be confusing:
As well as these mental intentional states, agents have perceptual experiences. Things start to get interesting when formalising such experiences.
Devlin doesn’t go into much detail here, and I’m not sure if this “immediacy” is supposed to be an elliptical reference to something like qualia. I’ve always found the notion of qualia vacuous at best and incoherent at worst. But this might lead to some kind of coherent definition of qualia: something to do with self-reference.
The structure of an utterance (like “It’s cold in here”) involves many potentially different situations, all of which contribute to its meaning:
This made me wonder about the Amazonian people described in Don’t Sleep, There are Snakes – their language has a restricted grammar, and they seem to make use of very limited (only immediate) resource situations. Is there a connection, I wonder?
Discourse isn’t just about uttering propositions. Utterances can change the state of the world (particularly, the mental state of the listener): they have an impact.
Furthermore, the speaker has an intention to go with their utterance. Speakers don’t just convey propositional information (and indeed, they may lie or otherwise intentionally deceive). They can also also speak “indirectly”, by using a proposition to convey a request (saying “It’s cold in here”, with the intent to get the hearer to close the window). The intended communication is the request, not the proposition.
This covers most of the theory. There is a short section towards the end on the Liar paradox (“this sentence is false”, which in situational theory becomes “the proposition expressed by this utterance is false”). Situation theory provides a resolution to this paradox. It lies in difference between two negations of “situation s supports information t” /“t is a fact in situation s”; one is “situation s supports information not-t”/“not-t is a fact in situation s” (it is the case that the negation of t is a fact in s) and the other is “situation s does not supports information t”/“t is not a fact in situation s” (situation s may have nothing to say about t). A consequence of this is that the “world”, or what one might like to think of as the “universal situation”, is not in fact a well-formed situation at all. (This is gone into in a lot more detail in Barwise and Etchemendy’s book The Liar).
Finally Devlin winds up the book by discussing whether his work constitutes “mathematics”.
I think it just qualifies mathematics, in that it considers deeply subtle distinctions between concepts, and provides a framework for formalising them. It is very important to think deeply about such issues, and tease apart the relevant concepts. And the philosophical discussions throughout the book on how things should be formalised is worth the entrance price alone. What I would also like is the ability to do calculations. (This is a weaker requirement than the ability to prove theorems: theorem are generic, calculations are specific, but they both need a semantics.) Specifically, if I have certain information, what can I deduce? If I know another agent has certain information, what can I deduce about what they can deduce? This, I think, requires more machinery than Devlin has included so far, possibly including some consideration of the computational resource needed to perform the inference. And just how does the propositional utterance “It’s cold in here” succeed in getting the hearer to recognise that they have been asked to (form the intent to) close the window? I suspect this may require the addition of some formalisation of something like Gricean maxims or Relevance theory.
And, to wrap up my review, I quote the lovely little rant by Devlin against short-sighted research relevance policies, as he had to leave the UK to follow this line of research:
Indeed. And I have certainly found the work of use. I think there are two directions to go now. There is Devlin’s later book, Goodbye, Descartes, which appears to be a more “popular” account of some of the ideas here, and more. And there is Barwise and Seligman’s Information Flow, which appears to be a related push to further situation theory in a more computational setting. Hopefully someday I’ll get around to reading these, too.
This book collects together the succinct articles from Keith Devlin's wide-ranging Micromaths column in The Guardian. It's great for dipping into, but obviously nothing is done in much depth. However, it's good for whetting your appetite.
Topics include
Keith Devlin and Jonathan Borwein, two well-known mathematicians with expertise in different mathematical specialities but with a common interest in experimentation in mathematics, have joined forces to create this introduction to experimental mathematics. They cover a variety of topics and examples to give the reader a good sense of the current state of play in the rapidly growing new field of experimental mathematics.
The discipline of mathematics has many aspects. There is the process of exploration, “messing around”, spotting patterns, suggesting hypotheses and conjectures. Then there is the process of proof: going from the initial statement to the fully proven theorem, which in non-trivial cases will also involve significant elements of exploration. Finally, there is the theorem itself, a statement of mathematical fact, backed up by a rock-solid argument of the tidied-up proof. Often though, only that final proof, and none of the processes leading up to it, is made public. This can lead to an external view of mathematics as magical results of genius, rather than a process of creative hard graft.
In my discipline of computer science, there is some support for the process of proof, with a computer helping to make sure the steps really are rock solid, and not instead built on sand. Unfortunately, the computer is a super-pedant, even more so that the most nit-picking of mathematicians, and nothing but the most adamantine of rock will do for it. Valid short cuts such as “without loss of generality”, “by symmetry”, and “abusing the notation” are not allowed, making these tools somewhat exasperating to use at times (although they are improving all the time).
But what about that initial process of exploration? Experimental Mathematics augments that stage: using a computer to help “messing around”, taking away (some of) the tedium of exploratory calculations, providing more examples and help with pattern discovery.
This book provides an introduction to such Experimental Mathematics, with several examples, and suggested exercises if you feel inclined to pursue some of the ideas further. Many of the examples are of the form:
There are several examples of increasing complexity, covering sequences, series, integrals, and more, with discussion of the discoveries of the results and their proofs. There is discussion about the famous formula that allows the calculation of the nth binary digit of π, without needing to calculate all the preceding n–1 digits. I also learned of “spigot” algorithms to calculate numerical values, which produce the digits one by one, cutting down on memory requirements.
The book definitely gives a flavour of the overall process, with enough pointers to suitable software resources for the reader to take up some of the challenges. However, I was left feeling that the flavour was somewhat weak. Okay, now we can calculate π to a bazillion digits, or find its gazillionth (binary) digit without calculating the earlier ones, and we have proofs of lots of closed form solutions. And yet, I didn’t come away with a feeling of much mathematical depth in these results. At most, in some cases the process suggests links between seemingly unrelated expressions. But I was hoping for more significant results. For example, the book finishes with a short chapter on “discovery by visualisation”, an area where the computer surely reigns supreme, yet surprisingly little is made of this. Maybe I am simply unaware of a vast existing industry in mathematics that is about proving such seemingly-pedestrian but actually important results, which has now become much more automated and hence more productive?
Mathematical thinking is not the same as “doing math”—unless you are a professional mathematician. For most people, “doing math” means the application of procedures and symbolic manipulations. Mathematical thinking, in contrast, is what the name reflects, a way of thinking about things in the world that humans have developed over three thousand years. It does not have to be about mathematics at all, which means that many people can benefit from learning this powerful way of thinking, not just mathematicians and scientists.
You won’t learn much new mathematics from this book. But by acquiring the ability to think like a mathematician, you will be able to master new mathematics quickly, whenever you need it.