As it happens, most of the events upon which this book was based either happened in my recent lifetime, or are matters of well known historical events, such as Chamberlin meeting Hitler. Thus, I do have a perspective on these issues. All in all, this is a hard book to summarize. In part because it is basically clinical psychology, which is a soft science at the best of times.
The book wraps its arguments around several well known cases and/or events. It opens and closes with Sandra Bland, the young black woman who was pulled over in Texas (she drove down from Chicago) for "failing to signal", ended up being manhandled out of the car, cuffed and arrested, and then hung herself in the jail.
The first case described in detail was about a high ranking CIA official who turned out to be a Cuban spy. The second was about Chamberlin meeting Hitler. The basic message is that we suck at spotting liars. Ample evidence is provided that shows that a computer evaluating parole issues from raw data scores much better results than a judge who can look the perp in eye and judge their demeanor. This does not surprise me one bit. As a life long poker player, I have determined that I suck at it too, and I would be better off simply following the odds.
We rely way to much on our ability to tell when we are being lied to. And once we make the decision to trust, we are easier marks because we will defend the indefensible far too long. We are too trusting, and we "default to the truth". I think a philosopher would call this the Principle of Charity. It is, as the book acknowledges, both a very risky thing to do; and very necessary as it is the grease that keeps society alive.
In this vein, we tend to think of others as simple and ourselves as nuanced. E.g.: A cop might see a nuanced response to stimulus from you as proof you prevaricating. This is exactly what happened to the Amanda Knox (another case study in the book). She reacted in a way the prosecutors saw as irrational, and therefore she was guilty despite the fact the physical evidence pointed elsewhere. This went all the way to the Italian Supreme Court before it finally got tossed for the rubbish it was. What reaction set the cops off? Basically, she was cool, clam, collected and quite. If she had cried like a girl, or some such, she would have walked.
The book calls these encounters as two "mismatched" people. We might say "talking or acting at cross purposes".
Another case with which I was very familiar was Harry Markopolus and his take down of Bernie Madoff. The SEC "defaulted to truth" and believed Madoff. Markopolis is unusual in that he does not "default to truth", but he has paid price: paranoia.
"Transparency is a myth." This line is repeated and emphasized and I agree with it. Transparency is the idea that if we can see what is going on on the outside, we can tell what is going on on the inside. On TV, we see people say and do things, and their faces and their words match exactly. This is called emoting. In real life it is rare. We do not wear ourselves on our sleeves. We are not transparent. But we tend to think others are.
Another interesting point is that the facial expressions and other tells that we think we universal, are not. A smile is not just a smile. It depends on the culture you are in.
The book examines the Sandusky pedophile case; a well know fraternity sexual assault involving loads of alcohol; the KSM Guantanamo Bay torture story and several others. It also delves into policing a fair bit. The bottom line of the policing analysis is that cops are incredibly reluctant to give up their supposed god-like ability to see into men's souls and pick the good guys and the bad guys. When science gets involved, and it shows them incapable of doing what they think they can do, then and only then, do they-ever-so-slowly recognize the truth.
Which leads into the study of the opening story about Sandra Bland.
I read this book to the end because it did keep my interest. My overall feeling is that it is far to broad, and overly simplistic in its analysis. I am always skeptical when people are broken down into groups that too broad to useful. The book does not really give advice on what to do about it.
Chamberlain was played by Hitler, who was really good at it. For some people, lying is their first option, and they do so with remarkable ease. Hitler, Bernie Madoff and my ex wife have these traits in common. I "defaulted to truth" with her as I do with most people. But I am skeptic and would love to know how I can tell if me and my listener are "mismatched".
Other cases are also explainable without appealing the "mismatched" perspective of the book. Bernie's success turned on greed and he was an expert con man. And Sandra Bland can be explained by bad policing policy (pull over everyone you can, give out as many tickets as you can, dig for a reason to doubt "them", abuse your authority), and bigoted, poorly trained cops. IMHO, Bland was pulled over for being black. When she got uppity, the cop lost his sense of perspective, she died, and, thankfully, the asshole cop was fired.
The ideas in this book are useful, but I see very little in the way of practical advice other than the skeptics mantra: Doubt is the handmaiden of truth, and above all else, doubt yourself and your ability to detect lies.
Chamberlain had the facts… he was just a pussy who did not doubt his ability to read people.
George HW Bush made the same mistake with Putin (remember "I looked in his eyes?") and Donald Trump, the stupidest person to ever occupy the White House, is doing it again.
The pages of this book are small and the leading loose, so this book is fast read and I read it. Perhaps this is my disdain for clinical psychology (which has done a lot of damage over the years), but I cannot recommend this book unless you are a student of the subject.
When pondering on disaster relief, it is helpful to know what natural disasters we should prepare for.
War is hardly natural, and we seem to always be on the brink of it somewhere in the world, so I ignore it.
There are two major issues to deal with: How big might it be, and what are the chances of it happening. A crude break down of “how big” is Local vs Global.
Here is the list off the top of my head (Disaster; Size, Probability):
Only “Pandemic” stands out. It has the potential to go global; AND since it has happened in the past (Spanish Flu, Swine Flu, SARS), and, due largely to air travel, is more likely to happen in the future… the probability is high.
Covid 19 was not unlucky, it was inevitable. A moments thought, which is about as much as I put into this piece, will convince anyone that the smart money should go to prevention of pandemics.
Therefore, ergo, ipso facto, and thus, Donald Trump disbanded the pandemic teams, cut the budget for the CDC and tried to build a wall to keep out Mexicans. He has turned recklessness into an art form. Of course, he did not create C19, but he has, and is, making it much worse.
I always want to end on a happy thought. This can be fixed.
1) Reduce the number of people on the planet by 3 billion at least;
2) Reduce the number of Donald Trumps by one (ideally, remove the whole family tree); and
3) Stop treating every square inch of the planet as a tourist destination.
The following paper was written by my friend Dale Beyerstein for the journal Humanist Perspectives ( https://www.humanistperspectives.org/issue211/index.html ) .
People who debunk false or nonsensical claims sometimes tend to specialise – e.g., concentrating on pseudohistorical claims such as the 9/11 conspiracy, paranormal claims such as those involving extraterrestrial UFOs, alternative health claims such as the canard that MMR vaccine causes autism, or the claims of particular religions. This is understandable, because the one thing that unites modern day critics of these intellectual travesties is their commitment to thoroughness in researching and analysing claims before accepting them or debunking them. Of course, this is not to say that skeptics have always is lived up this ideal. But when one skeptic does not, fellow skeptics are generally just as critical of that lapse as they are towards any paranormalist.
But despite this tendency towards depth of knowledge, many skeptics also show a wide breadth. Many are quite well versed in what’s wrong with many different pseudosciences – from different versions of alternative or complementary medicine to astrology to UFOs. In addition, some are well versed in other areas as well, such as religion, politics, or anything you may wish to discuss around the watercooler. As well as being able to discuss these topics, they also can give a coherent account of scientific reasoning and critical thinking. Is this because these people are polymaths? Well, some are, but I think that the main reason has to with their curiosity.
This is not the way many people view skeptics. Many people see skeptics as, in the words of U.S. President Richard Nixon’s Vice President Spiro Agnew, “nattering nabobs of negativism.” This is because skeptics reject quite a few popularly held beliefs. But skeptics hold a number of beliefs, albeit provisionally. Sit down with a group of people that includes at least one skeptic, and you will find a genuine interest in ideas. But more important, you will hear a lot of “Why is that?”, “What’s the evidence for that?”, or “But what about…”. And they are curious, but they practise a special type of curiosity, which I shall call what I shall call guided curiosity.
What am I on about here? Well, everyone holds that curiosity is a good thing; but if you think about it, unbridled curiosity can actually inhibit understanding. Take conspiracy theorists, for example. Being curious about how every girder split from its mountings in the World Trade Center on 9/11 wouldn’t seem to be a bad thing. However, when you think about it, it’s obvious that such detailed knowledge about an event such as this just isn’t going to be available. Who would be wandering around the buildings as they were collapsing to gather it? And if anyone did, would they have survived long enough to tell us? And it begs the question about a conspiracy to assume that anyone was figuring these things out in advance in order to plan the attack. It’s obvious that flying a plane into an iconic building will cause chaos; the perpetrators didn’t need to know exactly how much or exactly how it would happen. So why should we be worried that we lack that information? But this lack of information is what starts some conspiracy theorists down the rabbit hole. Why don’t we have it? Who is hiding it? Why? Religious believers fall into the same trap. What happened before the Big Bang? Obviously, since we have no answer at present, we must conclude that there must have been a God to cause it. Asking questions that admit of no answers, or improperly formulating them in a way that prevents a sensible answer to be given, is a surefire method for generating false, and sometimes ridiculous beliefs. So there needs to be limits on our curiosity.
I’m not suggesting limiting the range of one’s curiosity to what is ‘practical’, or of immediate interest, or to easily answerable questions. Rather, the point of limiting curiosity is this: The wider one draws the curiosity net, the more information one receives, everything else being equal. But on the other hand, one will pick up more flotsam and jetsam as well. So the point is to maximise knowledge, or at least justified beliefs while minimising the amount of nonsense or outright falsehood.
The method which does the most to achieve this is best stated by the eighteenth-century philosopher David Hume: “A wise man, therefore, proportions his belief to the evidence” (Hume 1955). Let’s refer to this as Hume’s Rule. Sexism aside, the rule is stating that the stronger the evidence, the more confidence one should have in one’s belief, and the opposite holds as well. Less evidence should result in a weaker belief. And a corollary of this principle is that one should not have a belief at all until one has examined the evidence. Following Hume’s Rule is the basis of guided curiosity. In what follows I shall give some examples from religion and paranormal belief to show how guided curiosity keeps us from falling for nonsense.
The consequences for religious belief of Hume’s Rule are readily apparent. Very few religious believers, when pressed, will hold that the evidence for religious belief is very strong. This is where the argument typically takes a turn: to hold that there is, after all, an exception to Hume’s Rule, which applies only to religion. Religious belief is grounded on faith, not reason or evidence; and Hume’s Rule applies only to beliefs based on evidence. Faith is, of course, precisely belief in the absence of evidence, and is, according to the religionist, a virtue which not only elevates the religionist who possesses it, but shows the simple-mindedness and shallowness of thought and character of the atheist or agnostic who lacks it, and instead asks for evidence.
By the time the atheist has defended her character from the above charge, there probably won’t be much time left to return to the question why faith is only appropriate for religious claims. After all, faith has a companion in mundane affairs, gullibility – which is also belief in the absence of evidence – which is decidedly not considered a virtue in those who invested in Bernie Madoff’s Ponzi scheme. But if you do find the time to pursue this point with the religionist, it’s very unlikely that you will receive an informed answer. Instead you will probably be told that this comparison is insulting, and the debate will end there. Perhaps it is; but the fact that a person may be insulted by being told that he has a long nose doesn’t by itself prove that the claim is wrong.
Hume’s Rule has another important implication for religion, atheism and agnosticism. To see it, let’s introduce just a bit of probability theory. Since evidence is what makes a belief probable, it follows that evidence which establishes the probability of a claim to be less than .5 (or 50%) should be disbelieved. This is because the probability of belief p and the belief in the denial of p (Not-p) must add up to 1. The sentence “It will either rain or not rain on my house today” has a probability of 1, or, in other words, expresses a certainty (the Law of the Excluded Middle in logic). According to the app on my phone, the probability that it will rain here today is 40%, or .4. When a belief p has a probability of .4, its denial, Not-p has a probability of .6 (1 minus .4). So I should believe that it won’t rain today – but, applying Hume’s Rule, my belief shouldn’t be very strong, and I should be prepared to change it. Ditto my belief in a god, except that I assign the probability of there being a god to be much lower.
A claim with a probability of exactly .5 should be neither believed nor disbelieved. Thus, agnostics must be holding that the evidence for belief in a god is just as compelling as that for disbelief, or, in other words, a probability of .5 for each. A probability of less than .5 is grounds for atheism, for the reason just given. But most people who call themselves agnostics do not really believe that the probabilities are equal. They concede that the probability that there is a god is less than that of the belief that there isn’t one, but they stick to the claim that we cannot be sure that a god doesn’t exist. But this is just to miss the point of one of the basic axioms of probability theory given above. It is important to remember that the denial of the existence of something does not require evidence that the probability of its existence is 0. Most agnostics have no difficulty in dismissing the existence of Santa Claus or the Tooth Fairy, despite not having checked every claim of where Christmas presents or quarters under the pillow came from, and therefore not being in the position to say that the probability of their existence is 0.
It might be thought that the agnostic has an answer to this. Given that God is supposed to be transcendent, completely outside the realm of human experience, no evidence is possible for belief or disbelief, because there is no evidence at all. With no evidence leading us in either direction, suspension of belief (agnosticism) seems to be the only reasonable position. But this rebuttal isn’t conclusive.
The best reason for suspending belief is that we are awaiting further evidence that might require us to change our minds. Now let’s return to the agnostic’s strong point that we are considering the transcendent; for which no empirical evidence can be found. If this is so, then there would be no reason to suspend belief pending further evidence – the supposition is just that this won’t be any. Now, add this to another corollary of Hume’s Law: The onus of proof is always on the person who puts the idea forward. When the claim is presented without any evidence to support it, Hume’s wise person would disbelieve it. This is because there are always more ways of getting something wrong than of getting it right. Take for example, guessing the day of the week on which a total stranger was born. If you guess Thursday, you have one chance in seven of getting it right, and the smart money will be on you getting it wrong. So in the debate between the atheist and agnostic where both agree that there is no evidence available about the transcendent (literally the world for which no empirical evidence possible), the onus of proof is on the believer, and the believer has none. Therefore, the claim should be disbelieved, and the atheist wins by default. In addition, remember that the theist claims to believe in some god or another; one with certain properties (even if she admits that there is no empirical evidence for those properties. But now go to a second theist, who believes in another god, with somewhat different properties). Both of these theists will be implicitly recognising the onus of proof, since they apply it to each other: the first will deny the second’s god on the grounds that the evidence is insufficient, and vice versa. Repeat this a few thousands of times, and we have the point made so well by Richard Dawkins (2006): “We are all atheists about most of the gods that humanity has ever believed in. Some of us just go one god further.”
Pseudoscientific claims fudge the onus of proof too. In fact, they do so with such regularity that we might consider this error as one of the defining characteristics of pseudoscience. Take conspiracy theories for example. Why is all the evidence about how “they” killed Kennedy, or placed the explosives in the World Trade Center towers to supplement the work of the airplanes, completely hidden? That it is missing is just the evidence that skeptics are supposedly too dense to see; its absence shows how clever and powerful “they” are that “they” can hide it so well. You can diagnose the informal fallacy involved here as failing to respect the onus of proof; or you can equally well call it begging the question (appealing to the very fact you are attempting to prove as evidence for the very fact that you are attempting to prove). Or you can call it the argument from ignorance, which involves saying that because you cannot disprove my claim, I must be right – whether or not I have presented any evidence. But the main point is that they have managed the impossible feat of creating something out of nothing. This, by the way, is the thing about informal fallacies: They are like cockroaches, in that if you spot one you can be assured that there are a bunch more lurking where you can’t see them.
A corollary of Hume’s Rule is the requirement that we search for not only the evidence that supports our belief, but for that which goes against it. Looking only for the supporting evidence is confirmation bias. The religious, conspiracy theory and the paranormal believer on the other hand are notorious for this cherry picking. The Christian apologist, searching for miracles, concentrates on the one little baby that survives the plane crash, but ignores the 200 others who perish. The believer in her own telepathic powers zeroes in on the few times she ‘knows’ what her friend will say next, while remaining blindly oblivious to the many more times she guesses wrong. The 9/11 conspiracy theorist pays attention to any problem, no matter how insignificant, in the received account of how the towers collapsed, while not being troubled at all by the fact that there is no evidence whatsoever of anyone planting any explosives in the buildings beforehand. The believer in any sort of complementary or alternative medicine will keep track of every cold that cures itself after taking echinacea while ignoring the ones that cure themselves without taking it. The graphologist (handwriting analyst) keeps track of every case where a handwriting sample shows large loops on the descenders of fs, gs, js, ps, qs and zs and its writer has a better than average libido, and ignores those with large libidos but wih handwriting characteristics that cannot be described this way, as well as those whose handwriting has thin or small descenders but who are nevertheless quite libidinous. (If you think I’m making this up, check Paterson, (1980:11), and don’t ask how she measured libidinousness.)
Let us look at one more rule of guided curiosity. It’s not only important to pick up new information; it’s also essential to compare that new information with what you already have. Is the new information consistent with what you already believe? If not, you have work to do to reconcile these beliefs. Perhaps you will have to reject the new information; perhaps you will have to modify it to a certain degree. Or maybe you will have to do one or both of these things to your present beliefs to achieve a fit. This shuffling process will always be with us, as long as we are gaining new information. (For a better account of this process, see Quine and Ullian (1978).) If the new observation is consistent with our old beliefs, then you can ask how it enhances your understanding of what you already know. What implications are there from the new belief to other hitherto unsuspected claims? Do these implications suggest further claims which can be tested? An example will show what I mean. Therapeutic Touch (TT) is a healing modality which involves practitioners manipulating the “energy field” which surrounds our body without actually touching the body itself. (This has led me to describe TT as “neither therapy nor touch.”} One shape of the field is healthy, others are making us sick. The crucial thing to note about TT theory is that practitioners work on just this energy field, not on the body underneath it. So, let’s take these claims at face value. If the healers can manipulate the field, they must be able to discern its presence somehow without simply inferring it from the presence of a body. But if we can perceive its presence through sight, smell, touch or hearing (probably taste isn’t an option here), then everyone should be aware of it. But we are not. Only TT practitioners are. Well, that must be because there’s another sensory mechanism which not all of us have – only those who would be good therapeutic touch practitioners have it. The inference from the claim that the energy field can be worked with to the claim that practitioners must be able to recognise its presence is not one that is often made by TT believers. But it was made by a nine-year-old from Loveland, Colorado, Emily Rosa. And though no TT practitioner had thought of doing this, the young Ms Rosa thought about how to test this claim. For her science project she set up a solid barrier dividing a table in half. The barrier left enough room for a TT practitioner to pass her hand underneath, just high enough for the experimenter to have her hand underneath it, or not. Whether it was or was not was determined by a randomizer. So, if the TT practitioner could do better than chance, on a test designed to rule out the other sensory modalities, this would be evidence of the energy field that some gifted people could detect. Needless to say, Ms Rosa’s experiment did not confirm this hypothesis, but it did lead to her being the youngest person ever to be published in a top rank medical journal (Rosa , et al, 1998).
There is another important implication of Hume’s Rule. He tells us that belief should be based on the preponderance of evidence, or on the probability that evidence confers on the belief. But the evidence for or against the belief is continually shifting as more of it becomes available. Along with this, the probabilities will fluctuate. Remembering this is how the skeptic following guided curiosity avoids dogmatism even when she has a fairly strongly held belief. She is always ready to modify her beliefs, and in some cases switch from belief to disbelief or vice versa as the new evidence comes in. And it is also why a skeptic should not only state her beliefs, but state them along with the degree of confidence – her estimate of the likelihood that more evidence will require her to revise or abandon those beliefs. Or better yet, always be prepared to state the belief along with the evidence for it. With these qualifications, there is no harm in provisionally stating a belief with a probability not much higher than .5, or disbelief even when the probability is a bit less than .5. Taking a belief seriously confers the benefit that, once a belief is stated along with the evidence for it, it can be examined, and implications drawn from it, which in turn can lead to new understanding. However, dismissing it because the probability is not much above .5 forgoes this possibility. The important thing to remember about dogmatism is that what is wrong with it is not the forceful stating of the belief, but concentrating on the belief rather than the evidence for it, or the unwillingness to budge from it when new evidence comes to light.
Some skeptics will be disappointed that I have gone all this way without mentioning one of the cardinal principles of skepticism, that there are times – quite a lot of them, actually – when you shouldn’t express a belief at all; you should straightforwardly admit that you do not know. There are two advantages to this admission. The first is that it serves as a stimulus to curiosity: having admitted that you don’t already know gives you a good reason to try to find out. Second, it prevents you from misleading others (and yourself). When they think you know and you don’t, they might follow you when they shouldn’t.
There are two situations where one should say that one doesn’t know. The first is when this expression is simply a substitute for “I don’t care.” This expression is tantamount to admitting that you have very little evidence and you are not prepared to gather any more. For example, just by reading the headlines and deciding that I have no interest in the articles they head, I couldn’t help finding out that Prince William, the Duke of Cambridge, and his wife, Kate Middleton, recently had a son, whom they named “Archie”. But his birthweight? I don’t know; meaning …. After all, one cannot expect to have time to look into everything; one must prioritise.
The second situation is where the evidence you have at present is about equally compelling for belief and disbelief, and it is possible to get more. In this situation it makes sense to suspend belief and wait for further evidence. But there’s an important exception here: sometimes waiting isn’t a viable option; circumstances require an immediate response. Fortunately, these special cases are quite rare; so withholding judgement while awaiting more evidence is an available option, and a good one.
Otto von Neurath (1921) compared our belief system to a leaky ship at sea. We are continually replacing rotten planks with fresh ones, but never are we able to replace the whole bottom at once, given that we wish to remain afloat. Thus we will never have a perfect set of beliefs; there will always be some false ones in there that we haven’t found yet. The best we can hope for is a gradual improvement. To continue with his metaphor, when we find a particularly strong plank on the boat which doesn’t fit very well with the old ones already in place, going to the all the work to make it fit may result in a much less leaky boat overall. Similarly, encountering a new belief that is inconsistent with some old ones, but with a lot of evidence backing it up may require the modification of several of the old beliefs at once. But the result may be a more coherent belief system overall. But not a perfect one. Non-skeptics may find this disconcerting. They are like the sailors who aren’t in it for the pleasure, but just want to get somewhere – anywhere -- and who just want to go along for the ride. But the true skeptic enjoys the sailing for its own sake.
Dawkins, Richard 2006: The God Delusion.Boston, Houghton Mifflin Company.
Hume, David: “Of Miracles”, in Hume 1955, Indianapolis, The Bobbs-Merill Company Inc., An Inquiry Concerning Human Understanding.
Paterson, Jane 1980: Know Yourself Through Your Handwriting. Montreal, Readers Digest Assn.
Quine, W.v.O. Quine and J.S. Ullian, 1978: The Web of Belief. New York, Random House.
Rosa, L., E. Rosa, L. Sarner, S. Barrett 1998: "A Close Look at Therapeutic Touch". Journal of the American Medical Association, Vol 279(13):1005–1010.
Von Neurath, Otto 1921: Anti-Spengler. Munich, G.D.W. Callwey
Dyer is a Canadian war researcher who wrote this book in 1985. From what I can gather, it is well respected. This edition has been revised and updated.
The book tours war historically, examining its roots and its evolution over the millennia.
The opening comments discuss some of the big picture history. For example, major battles prior to the 20th century would generate casualty rates as high as 50-60%, with an average of 20. In modern battles, this figure rarely exceeds 1 percent. This was due largely to changes in technology (weapons got better). This in turn changed the nature of soldiers and their mission. In the old days, soldiers did not get combat fatigue (shell shock, or today: PTSD) because they would die before it ever happened. Today, all armies recognize that troops can only take so many combat days (in WWII, it was 240) before they fall apart.
Dyer has always impressed me especially with regard to basic training, which he discusses in a chapter called "Any one's Son Will Do". Get them when they are young. Break them down to the same level, making them all equals, and then build up a small group dynamic where each man relies on the others to help him stay alive. Drill Sergeants are masters of psychological manipulation, and they know it.
In his opening chapter, Dyer makes this startling revelation: In WWII in the ETO, only 15% of combat infantry riflemen ever fired their weapon in anger! Most soldiers would never have reveal such a fact, until they were told they were not at all alone. More than four fifths of combat soldiers got through the war with killing anyone, and without firing a shot! More research has backed this up. Gettysburg was a hugely costly battle for Americans because it was Americans on both sides. More than half of the recovered muskets from the battle were loaded with more than one round, and only 5% were ready to be fired when they were abandoned or dropped. Six thousand had as many as 10 rounds in the barrel. In other words, most of the fighters were spending their time reloading a loaded gun, and, one assumes, ducking. No one is suggesting these men were cowards. Many were simply principled and did not want to kill.
The upshot of this is startling: if you could get the malingerers to enter the fray, the other guys didn't stand a chance. Which they did. New training got the numbers up to 50% in Korea and 80% in Vietnam. The US had figured out how to get men to kill automatically. Dehumanizing the enemy was a big part of it.
War goes back a long way. Chimps war on each other and so do we. We did not invent war, we inherited it. Dyer examines hunter gatherer groups and their interaction; ritualized warfare in groups like New Guinea bushmen; and other evidence of our more primitive past. Some suggest that ritualized warfare is not the "real deal". It does generate casualties at a low rate, which meant that they could do it a lot, which in turn meant high casualty rates over time. Hobbs, Rousseau and Darwin are invoked and examined for their viewpoints.
The birth of war was driven by human life style choices. Hunting and gathering can keep you alive. So can farming. But a new type of living was now being made: pastoralism (i.e.: nomads). They lived by herding domesticated animals. Inevitably, these groups clashed. The nomads would win fights because they were mobile, which meant they could concentrate their forces when needed. To combat this, walls around towns were built. When horses were domesticated, things accelerated. It is worth noting that Egypt was largely spared from nomadic attacks by its geography.
The Sumerians hit upon the idea that religion could be a better way of settling disputes. It worked for a while. Priests liked it because it gave them power. But as we know, religion is not a cure for war, but more often an excuse or a direct cause.
Aside: Women were equal partners in life until civilization and agriculture came along, driving the need for a power structure which eschewed women because it could.
Sargon was the worlds first emperor over a militarized society (circa 2300 BC). Around then, there was a major technological innovation: the compound bow. The bow and arrow has killed more people than any other weapon, ever. Unlike the English longbow (yet to be invented), the composite bow was short, powerful and could be fired from horseback.
Other inventions over time include the chariot (fast, hit and run, archery platform); the pike and phalanx; war galleys; improved, harder metals; bigger horses; the saddle and stirrup (700 AD); gun powder (1300 AD); and organized armies (which gave the powers-that-be the willies).
Aside: Japan had a warrior based culture. When the musket was introduced, they were appalled. A samurai could be killed by a commoner! Unacceptable… so they just stopped making them. That obviously changed later in history.
Armies got bigger, the percentage of soldiers dying went up, battles were huge but infrequent; and the average citizen was left alone.
Armies had to be trained and fed. You could not just conscript someone and throw them into battle. Standing armies were a part of every major European power from 1700 on.
What happened next was the introduction over time of the concept of "total war".
New weapons were being developed at a quick pace. It was soon discovered that a few men behind cover with guns could stop a large number of advancing troops. Tactics changed. WWI introduced the concept of the continuous front. The tank was used for the first time. It gave professional soldiers the hope that it might end the wars of attrition. It did, but it caused the continuous front to become mobile. This and aircraft took a terrible toll on civilians. The tank eventually spawned "blitzkrieg", or mobile "lightning" war. The world wars were the first where civilian deaths outnumber the deaths of combatants.
The next chapter in the book documents the history of nukes. There have been a lot of ideas over the years on this subject, but the prevailing idea is that you only need enough to destroy the other guy. This was known as MAD (Mutually Assured Destruction). And so far, this math has kept the nuclear peace for 75 years. The advent of nukes gave rise to the phrase "conventional war".
The birth of the professional army came in 1803 Prussia with the creation of the Kriegsakademie. This approach of using well trained professional soldiers was soon adopted by others. One hundred Germans in WWII were a match for 125 British or 250 Russian troops. Why? Because they had 10x as many good generals who knew how to get the most from men and equipment. The old game is now protected turf and generals everywhere want to keep it that way (the start of the military-industrial complex?). Technology has changed the equation a great deal. One example: a Spitfire in WWII cost 5,000 pounds to make. The supersonic Tornado, its 3rd generation successor, cost 17 million, 172 times more after adjusting for inflation.
The last chapter speculates about the future of war:
Baboons are nasty creatures. The males are obsessed with status and fight all the time. But in one troop something unusual happened. The aggressive males all died off at once from eating infected meat. Overnight, the troop settled down to a much more egalitarian society. And it stayed that way even after the demographics of the tribe returned to normal! The ritualized war fare on New Guinea bushmen killed a large percentage of combatants, one at a time, year after year. The government stepped in and told them this had to stop. They all agreed enthusiastically and never looked back! It seems both were caught in a local stability point that they could not get out of without a nudge. Perhaps there is hope in both those stories.
While not as eye-opening a book as Guns Germs and Steal, this book is a must read for anyone who wishes to grapple with these thorny issues. War has been with us for all of our history. But perhaps we can first ritualize it, and then dump it as a bad idea. I doubt it. The New Guinea bushmen were offered an alternative to their wars… essentially third party arbitration. This is something the top dogs will never agree to.
One final tidbit: Have you ever wondered if you were safer as an officer in combat, or as a soldier? In WWII, the answer was "soldier". In Vietnam, it did not matter.
Dyer himself has served, so he has seen some of all this from the inside.
Very Stable Genius, A: Donald J. Trump's Testing of America; Philip Rucker & Carol Leoning; 2020; Penguin Press; 416 pgs; notes, index
This was a much awaited book when it showed up. It covers the Trump history from getting elected (when the first thing he did was lie and claim he won the popular vote) to just after the infamous July 25th phone call to the president of Ukraine. Like books from other reporters like Bob Woodward, this book is heavy on basic facts. It is clearly written and as the title of the book implies, the authors know about whom they are writing. Unlike Woodward's books, which I found dry and dull, I rather enjoyed reviewing the events of the last few years in one compact reference.
Trump fired Comey -- who he knew was on the west coast -- by sending a letter a letter through one of his henchmen. He also torpedoed him on Twitter. When told that he had screwed up in sending the letter as he did, he replied "I know, fucking incompetence. Drives me crazy!" (referring to his staff). Trump never errs. Period.
The material in the book is largely familiar to anyone who has followed the election of the Mango Mussolini. If you had negative opinions of him, and who doesn't, this book will be satisfying and scary at the same time. If sheds some light on the Mueller report and why it fell so flat. It also illuminates the mind of Trump.
He is a petty, pompous, pugnacious, pinhead (at that is just the "P"s) in charge of a country he does not understand. Every person who has been in contact with him for any length of time has walked away from him covered in bullshit and fleas. Some still serve, but most have been arrested or driven from office.
In years to come, this book will become a reference for the times. The only unsatisfying aspect of the book is that it ends before the story is over.
As a skeptic, I have read a lot about cults. We had at least two ex-cult members lecture the BC Skeptics . I spoke with them in person. The difference between a cult and a religion is often subtle. In this case, we are talking about a political figure who demands utmost fealty and believes he can do no wrong. The is nothing subtle about Trump , and the parallels between the Trump movement and the rise of Nazi Germany continue to trouble me.
I recently read Prisoner's of Isolation which focused on solitary confinement in the Canadian penal system. Canada's practices were, as I described, barbaric. There is some light at the end of the tunnel for Canada.
This book is a much broader indictment of the US penal system, and specifically Attica. Canada was bad. The US was and is far, far worse.
Attica is a name familiar to anyone of my generation. I was a teen when it happened, and I must admit I did not pay much attention. Attica is a town and a prison in upstate New Your. On Sept 9, 1971, after months of simmering issues, the shit hit the fan. The prisoners took over, controlled the yard, and had many CO (correction officers) hostages. In the process, a CO was killed. After that, the cons took responsibility for the hostages and protected them. Rumors fostered by the state spread: castrations, torture, slit throats and such. The COs and the SPs (State Police) smelled blood. A few days later, the state (lead by the warden and the state head of corrections) let to dogs loose. What followed was chaos, torture and barbarity. Many cons and several hostage COs were killed, all by other COs. No attempts were made to match bullets to guns, guns to COs, or do any of that police professionalism stuff.
Bigotry was a major issue. Most of the prison population was non-white. Guard training was an issue: they had none. General cowboy attitudes, tough on crime types, and the general view that COs walked on water and cons were animals did the rest. These were jail towns and they hired any moron who could tote a gun from the local populace to do a "professional" job.
The top names in law were embroiled in a 40 year long cover up of crimes. The state was utterly heartless to everyone involved, including the families of injured or killed COs. The state refused to pay the salaries of the hostage COs because they were "off the clock". The state lost evidence, or denied it existed, or claimed privilege, or said it existed but it was unimportant throughout it all. The state even sent "support cheques" to widows, knowing that if they were cashed, they would, under NY state law, constitute an acceptance of compensation… i.e.: take this pittance which you desperately need, but if you do, you can never, ever sue. IOW: the state conned its own citizens.
You would think things would have gotten better, but not in the US of A. Some precedents were set, some rules changed, and some people made whole (ish). But mostly, the US doubled down, prison populations soared, , conditions got worse, for profit prisons made it worse still, and it is worse than ever now.
This is a long but gripping read. There are real heroes, real bad guys (most of whom died before the law could ever catch up to them), and lots of victims. The worst thing about it, though, is just how trivial it is for politicians to completely derail the justice system whenever they want and use it as a cudgel.
Universe From Nothing, A; Lawrence Krauss, Afterword by Richard Dawkins; 2012; Simon and Schuster; 191 pgs; index
I have known about this book for some time and finally got around to reading it.
Nothing is big topic. What is nothing? Just empty space? It still has time and volume, and the potential to have other stuff too, and the potential is something. Or: nothing is not nothing if it has the potential to be something.
One hundred years ago, the fastest form of commercial travel was the train and ship, and Einstein's special theory was still being debated . The General Theory of Relativity took longer to accept. One hundred years ago, we believed we lived in an island universe called The Milky Way. But now we know that the universe is far, far more vast. At this point, the author turns to Douglas Adams ("the universe is so big, you wouldn't believe it" etc).
To get a handle on the universe, we need to know how much stuff is in it. I.e.: what does it weigh?
Standard candles (Cepheid variable stars and Type 1A supernovae are the two biggies) have opened our eyes not just to the size of the universe, but its age and its destiny. The universe is expanding and will continue to do so forever. Here is an interesting idea: in 2 thousand billion years (2 trillion years), stars and galaxies will still exist. But future astronomers will literally have no way of knowing the deep history that gave rise to their universe. The rest of the universe would have disappeared beyond the horizon (or, if you prefer, they will be so far red-shifted that they will be moving away at over the speed of light and thus become physically undetectable). We are privileged and can see our history, but this will not be true forever.
The book gives a nice summary of what we know today and how we know it. The reasoning is fascinating. The bottom line is that dark energy is driving the universe to expand.
Another fun fact: if you are old enough to remember TV in the days of yore, you will remember that TV was broadcast and picked up with antennae. In the dead of night, the broadcasters stopped broadcasting and the channel in question would appear as "snow"… just a lot of static. About 1% of the snow you see is the after-glow of the big bang! This microwave background radiation can be sued to calculate the age of the universe: 13.72 billion years.
The fundamental base of the book is one of book keeping. We got a universe from nothing. But how. Energy (i.e.: matter and other stuff) is a zero sum game and so, suggests Krauss is the universe. If there are equal amounts of positive energy and negative energy (yes, there is such a thing), then the books still balance and sum to zero! And we get a flat universe…. Which is nice because curvy universes are more complex.
Krauss quotes Chris Hitchens often when it comes to the implications of all this. One thing seems clear, for Krauss there is no need of a creator. And the universe cares not what you might think of it. Stephen Weinberg is known for saying that science does not make it impossible to believe in god, it makes it possible to not believe in god.
This is a quick read. But the contents are deep and I may come back and read it again in a few years.
One analogy I liked for how it came about is this:
Take a ball and throw it in the air. It will fall to Earth. Thrown it harder, and it will still fall back to Earth. But throw it hard enough and it will never come done. This is a kind of symmetry breaking. One can imagine the potential for space doing this for eternities stacked on eternities, until one day, the ball did not come back down, and the universe was born.
One final quip from the book: One answer to "Why is there something rather than nothing?" is "There won't be for long!"
Miracle at Midway; Prange, Gordon W.; 1982; Open Road Media; 414 pgs; notes, index, Order of Battle, bibliography, chronology;
Dec 7, 41: the US enters the war after Pearl Harbor
Jun 4, 42: Midway
Dec, 42: Stalingrad
Jun 6, 44: D-Day
May 8: The Germans surrender
Sep 2: After six years (and one day)… the war ends with VJ Day.
When it comes to battles during WWII, the two that stand out the most are the Battle of the Bulge and Midway. There are others, of course: the Battles of the Atlantic/Britain/Stalingrad/Kursk; D-Day, Alamein, Market Garden and many more… and that is just the EOT. But these two stand out.
The Battle of the Bulge was the most costly, especially for the Americans. But it lasted for a month, and its ultimate outcome was inevitable. The decisive action at Midway was over in under seven minutes; the final outcome was by no means clear; and ownership of the Pacific was at stake.
Prior to Midway was the Battle of the Coral Sea. This was the first carrier vs. carrier battle ever fought. No ships saw the enemy. The Japanese had two of its Pearl Harbor veteran carriers damaged. The US also had two carriers damaged. Due to a non-battle related incident, the Lexington was sunk (actually, scuttled). Both sides towed their wounded carriers back to their home bases. The Coral Sea was a draw. But it set the stage for what happened next.
The Japanese were already planning an attempt to take and occupy Midway, a small atoll north and west of Hawaii, consisting to two small islands, Sand and Eastern, where the US had an airbase and refueling port. The American code breakers lead by Rocheport had broken much of the Japanese JN25 code and knew that AF, as the Japanese code worded it, aka Midway, was the target.
The Japanese were told that their two damaged Pearl Harbor veteran carriers would take three months to repair. The Midway invasion fleet set sail with the other four carriers that hit Midway: Akagi, Horyu, Soryu and Kagi. The Japanese also wanted to hit the Aleutians as a feint. Following behind the two attack fleets was the rest of the Japanese navy, including the largest battleship ever built, the Yamato, which carried Yamamoto, the fleet admiral. Nagumo, who lead the Pearl Harbor attack, lead the Midway attack force.
The Yorktown was towed back to Pearl and there, the shore crews also said three months were required to fix her up. Nimitz gave them two days. In a minor miracle, Yorktown sailed two days later to join her sister carriers Enterprise and Hornet. Fletcher and Spruance were to lead the American side.
As history records, the two fleets met north and west of Midway. A couple of squadrons of bombers found the Japanese carriers without fighter cover and hurriedly attempting to re-arm their planes to fight the American fleet (rather than to attack Midway itself). To this point, the US had had zero luck. But this time, in under seven minutes, three Japanese carriers were burning and subsequent action sunk the fourth. The cream of Japan's naval fliers and ground crews were wiped out.
Later, the Yorktown survived two airborne attacks, only to be sunk by a submarine's torpedo.
One difference between the Americans and the Japanese: When the Yorktown was sinking, the captain confirmed that all crew were off the ship, and then grabbed his kit and left. When Akagi went down, they Japanese spent most of the time trying to figure out who should or should not commit Hara Kiri.
I knew a lot about Midway, but this was still a stimulating read. The main reason the Japanese lost the battle was arrogance… what they called "victory disease". On paper, they were way ahead. There were other reasons too: Failure to deploy their battle ships; Inferior damage control technology (a consequence of the Bushido code); Poor use of scouts; Biting off more than you can chew; etc.
War is hell. The US launched around 30 torpedo bombers… old slow and carrying shitty US torpedoes. Not one hit its target, or if it did, it failed to explode. The crews knew their likely fate and only one or two of them escaped it.
If the Japanese had succeeded, the Pacific would have been theirs for some time. As it was, from June 6, 1942 (two years before D-Day), every subsequent move of the Japanese took them closer to home.
If the Japanese had been succeeded, I suspect that Tokyo would have been nuked. In any case, many more people, mostly Japanese, would have died.
The book is well researched with loads of detail.
Impeach: The Case Against Donald Trump; Neal Katyal, Sam Koppleman; 2019 (Dec); Mariner Books; 156 pgs, notes, exhibits
This book was banged out in two weeks and published just before Trump was actually impeached. In this case, I think the word "impeach" really means "impeach and convict (i.e.: remove from office)". It is remarkably current considering the pace of events.
Katyal is a regular on MSNBC and worked in the Obama Whitehouse. He is a Constitutional scholar and Professor of Constitutional Law. The book is a quick read. The writing is concise and clear.
He outlines the history of impeachment first. He then examines the evidence, and the case against DJT. The legal issues are well explained . He finds himself teaching the American people law in the usual way: establish a set of hypothetical circumstances, and then explore the legal consequences.
Except this time, the circumstances are real, and USA is dealing with a bizzaro hypothetical come to life. Finally, he offers up some legislative patches that might make it harder for the next out-of-control President to follow in DJT's footprints.
The Whistler-Blower's statement is in it in full. It was compelling reading. The WB really did his homework well.
Katyal lays out the case why DJT must be removed from office very convincingly. I could go through the long list of things that Trump has done, but they are well known and there is no point in my summarizing a summary.
The book is not expensive and is out in paperback. Read it for yourself.
I really liked the film series of the same name that documented the experiences of Easy Company of the 101st Airborne. The book 's chapters each correspond to a 1 hour piece of the film series. I re-watched the series as I read the book.
Easy was trained by a much hated Captain Sobel. But Sobel's strict emphasis on fitness made Easy the most fit unit in the ETO (European Theater of Operations). Easy landed at Normandy, fought through Holland and Belgium, participated in Operation Market Garden, and famously played a key role in defending Bastogne from the German onslaught during The Battle of the Bulge.
The book is an easy read. It is about as close to fiction as I read these days, but of course, it is not fiction. The attention to military detail was fantastic in the film. I spotted things that were referred to in the book, but never pointed out in the film. So unless you were there, or were a veteran of other battles, the film had aspects that you would never notice without reading the book. Having said that, the film viewing time is longer than the book reading time. So if you have seen the film, you have not missed much. In fact, for most people, you can read the book faster than you can watch the show.
What these men had to endure is beyond anything I will ever experience. I stand in awe of what they accomplished… beyond simple survival.
The film ends with interviews of some of the soldiers… now quite old. The paradox of war is in the title. When you arrive as a recruit, you and your mates are equals, regardless of background. And if you survive the whole affair, they are your best friends to whom you owe your life, and vice versa. All of the survivors are glad for the experience, but only with the after the fact knowledge that they make it. Regardless, the emotional ties are strong and very touching.
The cliché "blood is thicker than water" does not refer to kinship ties, but to spilt blood between soldiers.
Lee Moller is a life-long skeptic and atheist and the author of The God Con.