Operations is the middle layer of military analysis, below strategy and above tactics. Operations concerns the movement of forces (often over multiple lines of advance to fully utilize the transportation network available) and their logistical support. Fundamentally, operations are about getting forces to the objectives specified in your strategy with sufficient supply to sustain themselves, so that once there they can employ your tactics to achieve victory. The specific task of crafting operations which will achieve a set of strategic objectives is called “operational art” in US doctrine. Operational failures typically manifest as logistics and maneuver failures – particularly operational plans with unreasonable timetables – both of which have been particularly in evidence in the initial Russian invasion [of Ukraine in 2022].
[…]
Strategy is the upper layer of military analysis. Fundamentally strategy concerns the identification of final objectives, the way those objectives can be achieved and the resources to be used to achieve those objectives; these three components of strategy in US doctrine are termed “Ends, Ways, and Means” respectively. Strategy is thus the “big picture” thinking behind an action, including the decisions to both commence hostilities and end them.
[…]
Tactics are the lowest layer of military analysis. Tactics concern the methods to be used to win battles. Things like flanking, suppressive fire, ambushes, etc. are tactics. A military’s tactical system is often spelled out in doctrine. In theory, operations is designed to deliver forces to battles in such a way (positioning, comparative force, etc.) that their tactics can win those battles, while strategy should aim to ensure that winning those particular battles will achieve the desired political end (whatever concessions are desired). It is important to distinguish actions which are strategy (designed to directly produce a desired end to the conflict) from those which are merely tactical (designed to achieve a local success or advantage in a given engagement). It is important when assessing failures in war to distinguish between strategic failures (typically a failure to come up with realistic goals and the means to reach them), operational failures (e.g. logistics failures or unreasonable maneuver timetables) and tactical failures (e.g. failure to use combined arms effectively).
Bret Devereaux, “Miscellanea: A Very Short Glossary of Military Terminology”, A Collection of Unmitigated Pedantry, 2022-03-25.
March 6, 2026
QotD: Operations, strategy, and tactics
February 28, 2026
QotD: The “Balance of Terror” in the missile age
The advance of missile and rocket technology in the late 1950s started to change the strategic picture; the significance of Sputnik (launched in 1957) was always that if the USSR could orbit a small satellite around the Earth, they could do the same with a nuclear weapon. By 1959, both the USA and the USSR had mounted nuclear warheads on intercontinental ballistic missiles (ICBMs), fulfilling Brodie’s prophecy that nuclear weapons would accelerate the development of longer-range and harder to intercept platforms: now the platforms had effectively infinite range and were effectively impossible to intercept.
This also meant that a devastating nuclear “first strike” could now be delivered before an opponent would know it was coming, or at least on extremely short notice. A nuclear power could no longer count on having enough warning to get its nuclear weapons off before the enemy’s nuclear strike had arrived. Bernard Brodie grappled with these problems in Strategy in the Missile Age (1959) but let’s focus on a different theorist, Albert Wohlstetter, also with the RAND Corporation, who wrote The Delicate Balance of Terror (1958) the year prior.
Wohlstetter argued that deterrence was not assured, but was in fact fragile: any development which allowed one party to break the other’s nuclear strike capability (e.g. the ability to deliver your strike so powerfully that the enemy’s retaliation was impossible) would encourage that power to strike in the window of vulnerability. Wohstetter, writing in the post-Sputnik shock, saw the likelihood that the USSR’s momentary advantage in missile technology would create such a moment of vulnerability for the United States.
Like Brodie, Wohlstetter concluded that the only way to avoid being the victim of a nuclear first strike (that having the enemy hit you with their nukes) was being able to credibly deliver a second strike. This is an important distinction that is often misunderstood; there is a tendency to read these theorists (Dr. Strangelove does this to a degree and influences public perception on this point) as planning for a “winnable” nuclear war (and some did, just not these fellows here), but indeed the point is quite the opposite: they assume nuclear war is fundamentally unwinnable and to be avoided, but that the only way to avoid it successfully is through deterrence and deterrence can only be maintained if the second strike (that is, your retaliation after your opponent’s nuclear weapons have already gone off) can be assured. Consequently, planning for nuclear war is the only way to avoid nuclear war – a point we’ll come back to.
Wohlstetter identifies six hurdles that must be overcome in order to provide a durable, credible second strike system – and remember, it is the perception of the system, not its reality that matters (though reality may be the best way to create perception). Such systems need to be stable in peacetime (and Wohlstetter notes that stability is both in the sense of being able to work in the event after a period of peace, but also such that they do not cause unintended escalation; he thus warns against, for instance, just keeping lots of nuclear-armed bombers in the air all of the time), they must be able to survive the enemy’s initial nuclear strikes, it must be possible to decide to retaliate and communicate that to the units with the nuclear weapons, then they must be able to reach enemy territory, then they have to penetrate enemy defenses, and finally they have to be powerful enough to guarantee that whatever fraction do penetrate those defenses are powerful enough to inflict irrecoverable damage.
You can think of these hurdles as a series of filters. You start a conflict with a certain number of systems and then each hurdle filters some of them out. Some may not work in the event, some may be destroyed by the enemy attack, some may be out of communication, some may be intercepted by enemy defenses. You need enough at the end to do so much damage that it would never be worth it to sustain such damage.
This is the logic behind the otherwise preposterously large nuclear arsenals of the United States and the Russian Federation (inherited from the USSR). In order to sustain your nuclear deterrent, you need more weapons than you would need in the event because you are planning for scenarios in which some large number of weapons are lost in the enemy’s first strike. At the same time, as you overbuild nuclear weapons to counter this, you both look more like you are planning a first strike and your opponent has to estimate that a larger portion of their nuclear arsenal may be destroyed in that (theoretical) first strike, which means they too need more missiles.
What I want to note about this logic is that it neatly explains why nuclear disarmament is so hard: nuclear weapons are, in a deterrence scenario, both necessary and useless. Necessary, because your nuclear arsenal is the only thing which can deter an enemy with nuclear weapons, but that very deterrence renders the weapons useless in the sense that you are trying to avoid any scenario in which you use them. If one side unilaterally disarmed, nuclear weapons would suddenly become useful – if only one side has them, well, they are the “absolute” weapon, able to make up for essentially any deficiency in conventional strength – and once useful, they would be used. Humanity has never once developed a useful weapon they would not use in extremis; and war is the land of in extremis.
Thus the absurd-sounding conclusion to fairly solid chain of logic: to avoid the use of nuclear weapons, you have to build so many nuclear weapons that it is impossible for a nuclear-armed opponent to destroy them all in a first strike, ensuring your second-strike lands. You build extra missiles for the purpose of not having to fire them.
(I should note here that these concerns were not the only things driving the US and USSR’s buildup of nuclear weapons. Often politics and a lack of clear information contributed as well. In the 1960s, US fears of a “missile gap” – which were unfounded and which many of the politicians pushing them knew were unfounded – were used to push for more investment in the US’s nuclear arsenal despite the United States already having at that time a stronger position in terms of nuclear weapons. In the 1970s and 1980s, the push for the development of precision guidance systems – partly driven by inter-agency rivalry in the USA and not designed to make a first strike possible – played a role in the massive Soviet nuclear buildup in that period; the USSR feared that precision systems might be designed for a “counter-force” first strike (that is a first strike targeting Soviet nuclear weapons themselves) and so built up to try to have enough missiles to ensure survivable second strike capability. This buildup, driven by concerns beyond even deterrence did lead to absurdities: when the SIOP (“Single Integrated Operational Plan”) for a nuclear war was assessed by General George Lee Butler in 1991, he declared it, “the single most absurd and irresponsible document I had ever reviewed in my life”. Having more warheads than targets had lead to the assignment of absurd amounts of nuclear firepower on increasingly trivial targets.)
All of this theory eventually filtered into American policy making in the form of “mutually assured destruction” (initially phrased as “assured destruction” by Secretary of Defense Robert McNamara in 1964). The idea here was, as we have laid out, that US nuclear forces would be designed to withstand a first nuclear strike still able to launch a retaliatory second strike of such scale that the attacker would be utterly destroyed; by doing so it was hoped that one would avoid nuclear war in general. Because different kinds of systems would have different survivability capabilities, it also led to procurement focused on a nuclear “triad” with nuclear systems split between land-based ICBMs in hardened silos, forward-deployed long-range bombers operating from bases in Europe and nuclear-armed missiles launched from submarines which could lurk off an enemy coast undetected. The idea here is that with a triad it would be impossible for an enemy to assure themselves that they could neutralize all of these systems, which assures the second strike, which assures the destruction, which deters the nuclear war you don’t want to have in the first place.
It is worth noting that while the United States and the USSR both developed such a nuclear triad, other nuclear powers have often seen this sort of secure, absolute second-strike capability as not being essential to create deterrence. The People’s Republic of China, for instance, has generally focused their resources on a fewer number of systems, confident that even with a smaller number of bombs, the risk of any of them striking an enemy city (typically an American city) would be enough to deter an enemy. As I’ve heard it phrased informally by one western observer, a strategy of, “one bomb and we’ll be sure to get it to L.A.” though of course that requires more than one bomb and one doubts the PRC phrases their doctrine so glibly (note that China is, in theory committed to developing a triad, they just haven’t bothered to actually really do so).
Bret Devereaux, “Collections: Nuclear Deterrence 101”, A Collection of Unmitigated Pedantry, 2022-03-11.
February 22, 2026
QotD: The shift from “motte-and-bailey” construction to stone castles
As we move to stone construction and especially full stone construction (which we’ll define as the point when at least one complete curtain wall – don’t worry, we’ll define that in a second – is in stone) in the 12th century, we’re beginning to contemplate a different kind of defense. The wooden motte and bailey, as we’ve seen, mostly served to resist both raids and “hasty” assaults, thus forcing less coordinated or numerous attackers to set in to starve the castle out or go home. But stone walls are a much larger investment in time and resources; they also require a fair bit more careful design in order to be structurally sound. For all of that expense, the builder wants quite a bit of a security, and in the design of stone castles it is hard not to notice increasing attention towards resisting a deliberate assault; stone castles of the 12th century and beyond are increasingly being designed to stand up to the best that the “small army” playbook can throw at them. Of course it is no accident that this is coming at the same time that medieval European population and wealth is beginning to increase more rapidly, leaving political authorities (read: the high nobility) with both the resources for impressive new castles (although generally the number of castles falls during this period – fewer, stronger castles) and at the same time with more resources to invest in the expertise of siegecraft (meaning that an attacker is more likely to have fancy tools like towers, catapults and better coordination to use them).
To talk about how these designs work, we need to clear some terminology. The (typically thin) wall that runs the circuit of the castle and encloses the bailey is called a “curtain wall“. In stone castles, there may be multiple curtain walls, arranged concentrically (a design that seems to emerge in the Near East and makes its way to Europe in the 13th century via the crusades); the outermost complete circuit (the primary wall, as it were) is called the enceinte. Increasingly, the keep in stone castles is moved into the bailey (that is, it sits at the center of the castle rather than off to one side), although of course stone versions of motte and bailey designs exist. In some castle design systems, with stone the keep itself drops away, since the stone walls and towers often provided themselves enough space to house the necessary peacetime functions; in Germany there often was no keep (that is, no core structure that contained the core of the fortified house), but there often was a bergfriede, a smaller but still tall “fighting tower” to serve the tactical role of the keep (an elevated, core position of last-resort in a defense-in-depth arrangement) without the peacetime role.
While the wooden palisade curtain walls of earlier motte and bailey castles often lacked many defensive features (though sometimes you’d have towers and gatehouses to provide fighting positions around the gates), stone castles tend to have lots of projecting towers which stick out from the curtain wall. The value of projecting towers is that soldiers up on those towers have clear lines of fire running down the walls, allowing them to target enemies at the base of the curtain wall (the term for this sort of fire is “enfilade” fire – when you are being hit in the side). Clearly what is being envisaged here is the ability to engage enemies doing things like undermining the base of walls or setting up ladders or other scaling devices.
The curtain walls themselves also become fighting positions. Whether on a tower or on the wall itself, the term for the fighting position at the top is a “battlement”. Battlements often have a jagged “tooth” pattern of gaps to provide firing positions; the term for the overall system is crenellation; the areas which have stone are merlons, while the gaps to fire through are crenals. The walkway behind both atop the wall is the chemin de ronde, allure or “wall-walk”. One problem with using the walls themselves as fighting positions is that it is very hard to engage enemies directly beneath the wall or along it without leaning out beyond the protection of the wall and exposing yourself to enemy fire. The older solution to this were wooden, shed-like projections from the wall called “hoarding”; these were temporary, built when a siege was expected. During the crusades, European armies encountered Near Eastern fortification design which instead used stone overhangs (with the merlons on the outside) with gaps through which one might fire (or just drop things) directly down at the base of the wall; these are called machicolations and were swiftly adopted to replace hoardings, since machicolations were safer from both literal fire (wood burns, stone does not) and catapult fire, and also permanent. All of this work on the walls and the towers is designed to allow a small number of defenders to exchange fire effectively with a large number of attackers, and in so doing to keep those attackers from being able to “set up shop” beneath the walls.
[I]t is worth noting something about the amount of fire being developed by these projecting towers: the goal is to prevent the enemy operating safely at the wall’s base, not to prohibit approaches to the wall. These defenses simply aren’t designed to support that much fire, which makes sense: castle garrisons were generally quite small, often dozens or a few hundred men. While Hollywood loves sieges where all of the walls of the castle are lined with soldiers multiple ranks deep, more often the problem for the defender was having enough soldiers just to watch the whole perimeter around the clock (recall the example at Antioch: Bohemond only needs one traitor to access Antioch because one of its defensive towers was regularly defended by only one guy at night). It is actually not hard to see that merely by looking at the battlements: notice in the images here so far often how spaced out the merlons of the crenellation are. The idea here isn’t maximizing fire for a given length of wall but protecting a relatively small number of combatants on the wall. As we’ll see, that is a significant design choice: castle design assumes the enemy will reach the walls and aims to prevent escalade once they are there; later in this series we’ll see defenses designed to prohibit effective approach itself.
As with the simpler motte and bailey, stone castles often employ a system of defense in depth to raise the cost of an attack. At minimum, generally, that system consists of a moat (either wet or dry), the main curtain walls (with their towers and gatehouses) and then a central keep. Larger castles, especially in the 13th century and beyond, adopting cues from castle design in the Levant (via the crusades) employed multiple concentric rings of walls. Generally these were set up so that the central ring was taller, either by dint of terrain (as with a castle set on a hill) or by building taller walls, than the outer ring. The idea here seems not to be stacking fire on approaching enemies, but ensuring that the inner ring could dominate the outer ring if the latter fell to attackers; defenders could fire down on attackers who would lack cover (since the merlons of the outer ring would face the other way). As an aside, the concern to be firing down is less about the energy imparted by a falling arrow (though this is more meaningful with javelins or thrown rocks) and more about a firing position that denies enemies cover by shooting down at them (think about attackers, for instance, crossing a dry moat – if your wall is the right height and the edges of the moat are carefully angled, you can set up a situation where the ditch never actually offers the attackers any usable cover, but you need to be high up to do it!).
Speaking of the moat, this is a common defensive element (essentially just a big ditch!) which often gets left out of pop culture depictions of castles and siege warfare, but it accomplishes so many things at such a low cost premium. Even assuming the moat is “dry”! For attackers on foot (say, with ladders) looking to approach the wall, the moat is an obstacle that slows them down without potentially providing any additional cover (it is also likely to disorder an attack). For sappers (attackers looking to tunnel under the walls and then collapse the tunnel to generate a breach), the depth of the ditch forces them to dig deeper, which in turn raises the demands in both labor and engineering to dig their tunnel. For any attack with siege engines (towers, rams, or covered protective housings made so that the wall can be approached safely), the moat is an obstruction that has to be filled in before those engines can move forward – a task which in turn broadcasts the intended route well in advance, giving the defenders a lot of time to prepare.
Well-built stone castles of this sort were stunningly resistant to assault, even with relatively small garrisons (dozens or a few hundred, not thousands). That said, building them was very expensive; maintaining them wasn’t cheap either. For both castles and fortified cities, one ubiquitous element in warfare of the period (and in the ancient period too, by the by) was the rush when war was in the offing to repair castle and town walls, dig out the moat and to clear buildings that during peace had been built int he firing lines of the castle or city walls.
Bret Devereaux, “Collections: Fortification, Part III: Castling”, A Collection of Unmitigated Pedantry, 2021-12-10.
February 16, 2026
QotD: How nuclear weapons were viewed right after WW2
In that context [clear Soviet superiority of conventional forces in Europe], the fact that it had been the United States which had been the first to successfully develop nuclear weapons (and use them in anger, a decision which remains hotly debated to this day) must have seemed like an act of divine providence, as it enabled the western allies to retain a form of military parity with the USSR (and thus deterrence) while still demobilizing. US airbases in Europe put much of the Soviet Union in range of American bombers which could carry nuclear weapons, which served to “balance” the conventional disparity. It’s important to keep in mind also that nuclear weapons emerged in the context where “strategic” urban bombing had been extensively normalized during the Second World War; the idea that the next major war would include the destruction of cities from the air wasn’t quite as shocking to them as it was to us – indeed, it was assumed. Consequently, planners in the US military went about planning how they would use nuclear weapons on the battlefield (and beyond it) should a war with a non-nuclear Soviet Union occur.
At the same time, US strategists (particularly associated with the RAND Corporation) were beginning to puzzle out the long term strategic implications of nuclear weapons. In 1946, Bernard Brodie published The Absolute Weapon which set out the basic outlines of deterrence theory; he did this, to be clear, three years before the USSR successfully tested its first nuclear weapon in 1949 (far earlier than anyone expected because the USSR had spies in the Manhattan Project). Brodie is thus predicting what the strategic situation will be like when the USSR developed nuclear weapons; his predictions proved startlingly accurate, in the event.
Brodie’s argument proceeds as a series of propositions (paraphrased):
- The power of a nuclear bomb is such that any city can be destroyed by less than ten bombs.
- No adequate defense against the bomb exists and the possibilities of such are very unlikely.
- Nuclear weapons will motivate the development of newer, longer range and harder to stop delivery systems.
- Superiority in the air is not going to be enough to stop sufficient nuclear weapons getting through.
- Superiority in nuclear arms also cannot guarantee meaningful strategic superiority. It does not matter that you had more bombs if all of your cities are rubble.
- Within five to ten years (of 1946), other powers will have nuclear weapons. [This happened in just three years.]
All of which, in the following years were shown to be true. Consequently, Brodie notes that while nuclear weapons are “the apotheosis of aggressive instruments”, any attacker who used them would fear retaliation with their enemy’s nuclear weapons which would in turn also be so destructive such that “no victory, even if guaranteed in advance – which it never is – would be worth the price”. Crucially, it is not the fact of retaliation, but the fear of it, which matters and “the threat of retaliation does not have to be 100 per cent certain; it is sufficient if there is a good chance of it, or if there is a belief that there is a good chance of it. The prediction is more important than the fact.” [emphasis mine]
This does not “make war impossible” by any means, but rather turns strategy towards focusing on making sure that nuclear weapons are not used, by making it clear to any potential aggressor that nuclear weapons would be used against them. And that leads to Brodie’s final, key conclusion:
Thus, the first and most vital step in any American security program for the age of atomic bombs is to take measures to guarantee to ourselves in case of attack the possibility of retaliation in kind. The writer in making that statement is not for the moment concerned about who will win the next war in which atomic bombs are used. Thus far the chief purpose of our military establishment has been to win wars. From now on its chief purpose must be to avert them. It can have almost no other useful purpose.
To sum that up, because both the United States and its key enemies will have nuclear weapons and because their destructive power is effectively absolute (so high as to make any “victory” meaningless) and because there is no effective defense against such weapons, consequently the only rational response is to avoid the use of nuclear weapons and the only way to do that is to be able to credibly threaten to retaliate with nuclear weapons in the event of war (since if you cannot so retaliate, your opponent could use their nuclear weapons without fear).
That thinking actually took a while to take hold in actual American policy and instead during the 1940s and 1950s, the United States focused resources on bomber fleets with the assumption that they would match Soviet superiority in conventional arms in Europe with American nuclear superiority, striking military and industrial targets (“precision attacks with an area weapon”, a notion that is as preposterous as it feels) to immediately cripple the USSR in the event of war, or else aim to “win” a “limited” nuclear exchange.
Bret Devereaux, “Collections: Nuclear Deterrence 101”, A Collection of Unmitigated Pedantry, 2022-03-11.
February 10, 2026
QotD: The (historical) walls of Jericho
These strategic (and operational) considerations dictate some of the tactical realities of most sieges. The attacker’s army is generally going to be larger and stronger, typically a lot larger and stronger, because if the two sides were anywhere near parity with each other the defender would risk a battle rather than submit to a siege. Thus the main problem the attacker faces is access: if the attacker can get into the settlement, that will typically be sufficient to ensure victory.
The problem standing between that attacking army and access was, of course, walls (though as we will see, walls rarely stand alone as part of a defensive system). Even very early Neolithic settlements often show concerns for defense and signs of fortification. The oldest set of city walls belong to one of the oldest excavated cities (which should tell us how short the interval between the development of large population centers and the need to fortify those population centers was), Jericho in the West Bank. The site was inhabited beginning around 10,000 BC and the initial phase of construction on what appears to be a city wall reinforced with a defensive tower was c. 8000 BC. It is striking just how substantial the fortifications are, given how early they were constructed: initially the wall was a 3.6m stone perimeter wall, supported by a 8.5m tall tower, all in stone. That setup was eventually reinforced with a defensive ditch dug 2.7m deep and 8.2m wide cutting through the bedrock (that is a ditch even Roel Konijnendijk could be proud of!), by which point the main wall was enhanced to be some 1.5-2m thick and anywhere from 3.7-5.2m high. That is a serious wall and unlikely the first defensive system protecting the site; chances are there were older fortifications, perhaps in perishable materials, which do not survive. Simply put, no one starts by building a 4m by 2m stone wall reinforced by a massive stone tower and a huge ditch through the bedrock; clearly city walls [were] something people had already been thinking about for some time.
I want to stress just how deep into the past a site like Jericho is. At 8000 BC, Jericho’s wall and tower pre-date the earliest writing anywhere (the Kish tablet, c. 3200 BC) by c. 4,800 years. The tower of Jericho was more ancient to the Great Pyramid of Giza (c. 2600 BC), than the Great Pyramid is to us. In short, the problem of walled cities – and taking walled cities – was a very old problem, one which predated writing by thousands of years. By the time the arrival of writing allows us to see even a little more clearly, Egypt, Mesopotamia and the Levant are already filled with walled cities, often with stunningly impressive stone or brick walls. Gilgamesh (r. 2900-2700 BC) brags about the walls of Uruk in the Epic of Gilgamesh (composed c. 2100) as enclosing more than three square miles and being made of superior baked bricks (rather than inferior mudbrick); there is evidence to suggest, by the by, that the historical Gilgamesh (or Bilgames) did build Uruk’s walls and that they would have lived up to the poem’s billing. Meanwhile, in Egypt, we have artwork like the Towns Palette, which appears to commemorate the successful sieges of a number of walled towns
So a would-be agrarian conqueror in Egypt, Mesopotamia or the Levant, from well before the Bronze Age would have already had to contest with the problem of how to seize fortified towns. Of course depictions like these make it difficult to reconstruct siege tactics (the animals on the Towns Palette likely represent armies, rather than a strategy of “use a giant bird as a siege weapon”), so we’re going to jump ahead to the (Neo)Assyrian Empire (911-609 BC; note that we are jumping ahead thousands of years).
Bret Devereaux, “Collections: Fortification, Part I: The Besieger’s Playbook”, A Collection of Unmitigated Pedantry, 2021-10-29.
February 4, 2026
QotD: The impact of quasi-official monotheism on the Roman Empire
This trend towards calcification [into the relatively rigid categorizations of honestiores and humiliores (“respectable” and “humble” people, but in practice, “wealthy” and “commoners”)] had been matched by the loss of Rome’s (admittedly opportunistic and unevenly applied) religious tolerance. This is often attributed to Christianity itself, but is perhaps better understood in light of the increasing demands of emperors during and after the Crisis of the Third Century to insist on unity through uniformity. The first empire-wide systemic persecution of Christians, the Decian Persecution (250 AD) was exactly this – an effort to have all Romans everywhere sacrifice for the safety of the emperor as an act of unity to strengthen his reign which rather backfired because it seems not to have occurred to Decius that Christians (of whom, by 250, there were many) would be unable to participate. Diocletian likewise launched the Great Persecution in 303 as part of a program to stress unity in worship and try to bind the fractured Roman Empire together, particularly by emphasizing the cults of Jupiter and Hercules. From that perspective, Christians were a threat to the enforced, homogeneous unity Diocletian wanted to foster and thus had to be brought back or removed, though of course in the event Christianity’s roots were by 303 far too deep for it to be uprooted.
That is part of the context where we should understand Constantine (r. 306-337). Constantine is famous for declaring the toleration of Christianity in the empire and being the first emperor to convert to Christianity (only on on his death-bed). What is less well known is that, having selected Christianity as his favored religion, Constantine – seeking unity again – promptly set out to unify his new favored religion, by force if necessary. A schism had arose as a consequence of Diocletian’s persecution and – now that Christianity was in the good graces of the emperor – both sides sought Constantine’s aid in suppressing the other in what became known as the Donatism controversy, as the side which was eventually branded heretical supported a Christian bishop named Donatus. Constantine, after failing to get the two groups to agree settled on persecuting one of them (the Donatists) out of existence (which didn’t work either).
It is in that context that later Christian emperors’ efforts to unify the empire behind Christianity (leading to the Edict of Thessalonica in 380) ought to be understood – as the culmination of, by that point, more than a century of on-again, off-again efforts by emperors to try to strengthen the empire by enforcing religious unity. By the end of the fourth century, the Christian empire was persecuting pagans and Jews, not even a full century after it had been persecuting Christians.
These efforts to violently enforce unity through homogeneity had the exact opposite effect. Efforts to persecute Arian Christians (who rejected the Nicene Creed) created further divisions in the empire; they also made it even more difficult to incorporate the newly arriving Germanic peoples, who had mostly converted to the “wrong” (Arian) Christianity. Meanwhile, in the fifth century, the church in the East splintered further, leading to the “Nestorian” (the term is contested) churches of Syria and the Coptic Church in Egypt on the “outs” with the official (Eastern) Roman Church and thus also facing persecution after the Council of Ephesus in 431. The resentment created by the policy of persecution in the East seems to have played a fairly significant role in limiting the amount of local popular resistance faced by the Muslim armies of the Rashidun Caliphate during the conquests of Syria, the Levant and Egypt in the 630s, since in many cases Christian communities viewed as “heretical” by Constantinople could actually expect potentially better treatment under Muslim rule. Needless to say, this both made the Muslim conquests of those regions easier but also go some distance to explaining why Roman/Byzantine reconquest was such a non-starter. Efforts to enforce unity in the empire had, perhaps paradoxically, made it more fragile rather than more resilient.
Bret Devereaux, “Collections: The Queen’s Latin or Who Were the Romans, Part V: Saving and Losing and Empire”, A Collection of Unmitigated Pedantry, 2021-07-30.
January 29, 2026
QotD: Nitpicking the Roman army in Gladiator (2000)
We pick up in an improbably mud-soaked clearing with a title card informing us that we’re in “Germania”, which is correct in a very broad sense that this is the Second Marcomannic War and the enemies here are the Marcomanni and Quadi, who are Germani (Germanic-language speakers), but the army here isn’t operating out of the Roman provinces of Germania (superior and inferior) which are on the Rhine, but rather on the Danube, from the provinces of Noricum and Pannonia (Superior). But in the sense that we’re in Germania magna, the greater zone of Germanic cultural influence, sure, fine.
In the process of Maximus riding up, the failure of negotiations and Maximus riding to join his cavalry, we get something of an overview of the Roman army and its position and both are wrong. Let’s start with the soldiers: we see a very clear distinction between two kinds of soldiers, the mail-clad auxilia, all archers, and the legionaries wearing the lorica segmentata and there appear to be about the same number of both groups. And here is where we first see the clear influence of the Column of Trajan (and to an unfortunately lesser degree, the far more appropriate Column of Marcus Aurelius) on the depiction, because this use of armor to distinctly signal the Roman citizen legionaries and non-citizen auxilia is straight from the Column of Trajan, completed probably around 113 and commemorating Trajan’s two Dacian Wars (101-102, 105-106).
What this sequence gets correct is that the Roman army was divided into those two groups, they were roughly equal in number (by this period, the auxilia probably modestly outnumber the legions in total manpower)1 and Trajan’s Column does use that visual signifier to distinguish them. This component is the crux of the verisimilitude that leads people to trust the rest of this sequence.
The problems start almost immediately from there. Roman auxilia were far more varied than what we see here in terms of equipment and tactics and only very few of them were archers. So let’s break down Roman auxiliary contingents. With all due caveats about the limits of our evidence, infantry auxilia outnumber cavalry by about 2:1 in attested auxilia units (auxilia were grouped into cavalry alae and infantry cohortes, generally of 480 men (sometimes around 800), but unlike for legionary citizen-infantry, these cohorts were not grouped into larger legions).2 So we ought to expect about a third of our auxilia to be cavalry, which is important because the cavalry detachments of Roman legions were very small (and mostly for scouting and messenger duties). Auxilia cavalry ranged in equipment and could include horse archers and even ultra-heavy cataphract cavalry, but most were mailed shock cavalry, equipped quite a lot like how Gallic or Germanic warrior-aristocrats or Roman legionary cavalry would be.
Of the remainder, the most common kind of infantry auxilia by far seem to have been heavy infantry, fighting in fairly heavy armor. These fellows get depicted in Roman artwork generally in mail armor, with flat oval shields (as opposed to the curved, rectangular imperial-period Roman scutum), spears and swords. These fellows, totally absent in this sequence are all over the Column of Trajan, with their flat oval shields being frequently seen (although one must distinguish them from Dacians who carry the same shield; the auxilia stick out for their mail and helmets). A bit less than 10% of auxilia units are attested as cohortes sagittariorum (“cohort of archers”). We also know the Romans used slingers within the auxilia, but as far as we can tell, not in specialized units; they may have been brigaded in with other auxilia cohorts. In either case, they appear in fairly small numbers. Finally, we also see on things like the Column of Trajan Roman allied or auxiliary units that are substantially lighter infantry: on the Column of Trajan, these are local troops shown wielding large clubs and stripped to the waist, presumably representing troops local to the Danube region, fighting in local (unarmored, with heavy two-handed weapons) style.
So whereas the army we see is a nearly even split between legionary heavy infantry and auxilia archers (with a small amount of legionary cavalry waiting for Maximus to show up to lead them), in practice a typical Roman field army would have far fewer archers, indeed around ten times fewer: not almost 50% of the force, but in fact probably a bit less than 5% of the force (since they’re less than 10% of the auxilia who would make up around half of a Roman field army). Meanwhile we’re simply missing the – by far – two most common sorts of auxilia cohorts, those of heavy infantry or heavy cavalry. This mangling of the structure of a Roman army is going to have implications when we get to Maximus’ overall plan for the battle as well.
Meanwhile, the legionary infantry are also much too uniform, literally. This is easily the most pardonable error, because what has happened here is that director Ridley Scott has copied the Column of Trajan but far too uncritically. After all, the Column of Trajan is not a photograph and thus has space for the artists producing it to take liberties, particularly in the name of imperial ideology and propaganda. In this case, showing large numbers of identically equipped soldiers, often moving in unified formation, serves the same rhetorical purpose in antiquity as it does today, suggesting an impressive, inhumanly uniformed and disciplined source. Moreover, the segmented Roman body armor, which we call the lorica segmentata (we don’t know what the Romans called it), was very distinctive to the legions, as it was the one armor that it seems like the auxilia probably (the evidence here can be tricky) didn’t share. And keeping the legions distinct from the auxilia also matters, as the legionary soldiers are higher status citizens who thus get “higher billing” in the imagery, as it were, than the auxilia. So showing all of the legions equipped neatly with this armor makes them seem distinct, impressive and uniform.
In short, it served Trajan’s image (and thus the artists aim) to suggest that all of his legions wore this armor.
Archaeology tells us quite clearly it was not so. Indeed, the lorica segmentata, so iconic because of its use in this way on the Column of Trajan, was probably the least common of the three major types of Roman legionary body armor in this period. The most common armor of the Roman legions was almost certainly still – as it had been in the Late Republic – mail, exactly the same as we see the auxilia wearing. We find fragments of Roman mail in legionary sites in all corners of the Empire and it remained common everywhere. To head off a standard question: no, it does not seem that the Romans ever got the idea to layer other defenses over mail, so when it was worn, it was the “primary” armor (worn over a padded textile defense called a subarmalis, but not under any other armor).3 We also see mail represented in Roman artwork, including on very high status soldiers, like senior centurions.
The next most common armor was probably scale armor, which we find very frequently in the East (that is, on the frontier with the Parthians/Sassanids) and often enough (if less frequently) in the West (that is, the Rhine/Danube frontier). We also know that some auxilia units wore this armor too and we see quite a bit of scale armor – wholly absent in this sequence – on the Column of – wait for it – Marcus Aurelius (completed c. 193). That’s the column that commemorates this war. Contemporary with this fictional battle. But it is less famous and somewhat less well-preserved than 70-years-earlier Column of Trajan, which they pretty evidently used quite a bit more of.
The lorica segmentata shows up the least often and – to my knowledge – effectively exclusively in the west on the Rhine/Danube frontier, where it is still probably not the most common (although it may have been more common than scale on that frontier). So what we ought to see in this army are legionaries who are marked out by their large scuta (the big Roman shield, by this period distinctly rectangular and also (as in the republic) curved), but in a range of mail, scale and lorica segmentata (with mail and segmentata being the most common, because we are on the Danube frontier, but scale hardly rare), along with auxilia divided into specialist cohorts (480 man units) each with different sets of armor and weapons: a few missile cohorts (archers, slingers), a lot more heavy infantry cohorts with spears and long shields, some lighter troops, and so on. The auxilia ought to be wearing basically every armor under the sun except for the lorica segmentata (which to my knowledge we’ve only ever found in sites associated with the legions).
Finally, these units are backed up by a whole load of catapults. We see two kinds, dual-arm arrow-throwing machines (which most folks would casually call ballistae) and single-armed pot-throwing machines (which most folks would casually call catapults), all of them in stationary mounts. Now on the one hand, “the Romans use lots of torsion-based catapults as artillery” is a true statement about the Roman army of this period, but on the other hand once again beyond that basic idea, most of this is wrong. Once again there’s an issue of verisimilitude here: the appearance of strange catapults and the true fact that the Romans used a lot of unusual catapults is likely to lead the viewer to assume some research has been done here and thus that these are the right catapults. For the most part, they are not.
We can start with the easy one, the larger single-armed pot-throwers. These are onagers, a late-Roman simplified single-arm torsion catapult, named for their fearsome “kick” (like an ass, an onager). These are popular favorites for Roman artillery, for instance showing up in both Rome: Total War and Total War: Rome II (both of which have main campaigns set during the Late Republic). There’s only one problem, which is that Gladiator (much less the even earlier Total War games) is set substantially too early for an onager to appear. Our first attestation of the onager is in Ammianus Marcellinus, writing in the last two decades of the 300s AD about the events of 353-378 (his work was broader than this, but only the back end survives). Vegetius, writing roughly contemporary with Ammianus also mentions them. But before the late fourth century, we don’t have any evidence for this design and it doesn’t show up on the Columns of Trajan or Marcus. So this isn’t just a little bit too early for these catapults but, given the evidence we have, around 150 years too early, the equivalent of having a line of M1 Abrams show up in a film about the Battle of Gettysburg.
What we do have are a number of twin-armed bolt or arrow-throwing machines and the Romans certainly had those, though what we see doesn’t match up well with what the Romans used. What we see is a single size of fairly large arrow-throwing engines, aimed upward to fire in fairly high arcs and built with large metal cases containing the torsion springs (generally made of hair or sinew, tightly coiled up; it is the coiling of these springs which stores the energy of the machine).
These two-armed torsion catapults came in a wide range of sizes and could be designed to throw either arrows/bolts or stones (the latter carved into spheres of rather precise caliber for specific machines). And we ought to see a pretty wide range of sizes here, from massive one-talent engines, which threw a 1 talent (26kg) stone and stood about three times the height of a man, to much smaller anti-personnel weapons (scorpiones) that were more like a “crew served” weapon than a large artillery piece. By Trajan’s time, the Romans had even taken to mounting these smaller crew-served engines on mule-drawn carts (called carroballistae) to allow them to be rapidly repositioned, something like early modern “horse artillery” (they were not meant to fire on the move; when we see them on Trajan’s Column, at least one of the operators is usually standing on the ground outside of the cart to winch the machine). These smaller machines, which would have made up the bulk of those deployed in a field battle, seem mostly absent in the sequence.
The result of all of this is that the Roman army presented in the opening moments of Gladiator manages to strike a remarkably unhappy balance: having just enough of the appearance of accuracy to decisively influence two decades of subsequent depictions of the Roman army without actually being particularly correct about anything beyond a very surface level. But subsequent pop-culture (again, I think Rome: Total War played a significant role here) would codify this vision of the Roman army – fire-throwing onagers, lots of auxilia archers, legionary rather than auxiliary cavalry, uniform use of the lorica segmentata – as the dominant model for quite some time.
But the army isn’t the only thing that’s wrong.
Bret Devereaux, “Collections: Nitpicking Gladiator’s Iconic Opening Battle, Part I”, A Collection of Unmitigated Pedantry, 2025-06-06.
- See figures in P. Holder, Auxiliary Deployment in the Reign of Hadrian (2003).
- There’s some complexity here because some infantry auxilia cohorts had small attached cavalry contingents too.
- I suppose I should note that is an odd exception for a type of very fine armor sometimes called lorica plumata (“feathered armor”) by modern writers where metal scales were mounted on mail armor (typically with extremely fine, small rings), rather than on a textile backing. This armor type seems to have been rare and must have been very expensive.
January 23, 2026
QotD: The peasant – historically, the overwhelming majority of humanity
Prior to the industrial revolution, peasant farmers of varying types made up the overwhelming majority of people in settled societies (the sort with cities and writing). And when I say overwhelming, I mean overwhelming: we generally estimate these societies to have consisted of upwards of 80% peasant farmers, often as high as 90 or even 95%. Yet when we talk about these periods, we are often focused on aristocrats, priests, knights, warriors, kings and literate bureaucrats, the sort of folks who write to us or on smiths, masons and artists, the sort of folk whose work sometimes survives for us to see. But this series is going to be about what life was like for the great majority of people who lived in small farming households.
We’re actually doing two things in this series. First, of course, we’ll be discussing what we know about the patterns of life for peasant households. But we’re also laying out a method. The tricky thing with discussing peasants, after all, is that they generally do note write to us (not being literate) and the writers we do have from the past are generally uninterested in them. This is a mix of snobbery – aristocrats rarely actually care very much how the “other half” (again, the other 95%) live – but also a product of familiarity: it was simply unnecessary to describe what life for the peasantry was like because everyone could see it and most people were living it. But that can actually make investigating the lives of these farming folks quite hard, because their lives are almost never described to us as such. Functionally no one in antiquity or the middle ages is writing a biography of a small peasant farmer who remained a peasant farmer their whole life.1 But the result is that I generally cannot tell you the story of a specific ancient or medieval small peasant farmer.
What we can do, however is uncover the lives of these peasant households through modelling. Because we mostly do have enough scattered evidence to chart the basic contours, as very simply mathematical models, of what it was like to live in these households: when one married, the work one did, the household size, and so on. So while I cannot pick a poor small farmer from antiquity and tell you their story, I can, in a sense, tell you the story of every small farmer in the aggregate, modelling our best guess at what a typical small farming household would look like.
So that’s what we’re going to do here. This week we’re going to introduce our basic building blocks, households and villages, and talk about their shape and particularly their size. Then next week (hopefully), we’ll get into marriage, birth and mortality patterns to talk about why they are the size they are. Then, ideally, the week after that, we’ll talk about labor and survival for these households: how they produce enough to survive, generation to generation and what “survival” means. And throughout, we’ll get a sense of both what a “typical” peasant household might look and work like, and also the tools historians use to answer those questions.
But first, a necessary caveat: I am a specialist on the Roman economy and so my “default” is to use estimates and data from the Roman Republic and Roman Empire (mostly the latter). I have some grounding in modelling other ancient and medieval economies in the broader Mediterranean, where the staple crops are wheat and barley (which matters). So the models we’re going to set up are going to be most applicable in that space: towards the end of antiquity in the Mediterranean. They’ll also be pretty applicable to the European/Mediterranean Middle Ages and some parts – particularly mortality patterns – are going to apply universally to all pre-modern agrarian societies. I’ll try to be clear as we move what elements of the model are which are more broadly universal and which are very context sensitive (meaning they differ place-to-place or period-to-period) and to the degree I can say, how they vary. But our “anchor point” is going to be the Romans, operating in the (broadly defined) iron age, at the tail end of antiquity.
Bret Devereaux, “Collections: Life, Work, Death and the Peasant, Part I: Households”, A Collection of Unmitigated Pedantry, 2025-07-11.
- And, as we’ll see, these societies generally have almost no social mobility, so extremely few – functionally none – of the sort of people who write to us will have ever been peasant farmers.
January 17, 2026
QotD: The introduction of tanks on the western front did not break the trench stalemate
Where the Germans tried tactics, the British tried tools. If the problems were trenches, what was needed was a trench removal machine: the tank.
In theory, a good tank ought to be effectively immune to machine-gun fire, able to cross trenches without slowing and physically protect the infantry (who could advance huddled behind the mass of it), all while bringing its own firepower to the battle. Tracked armored vehicles had been an idea considered casually by a number of the pre-war powers but not seriously attempted. The British put the first serious effort into tank development with the Landship Committee, formed in February of 1915; the first real tanks, 49 British Mark I tanks, made their first battlefield appearance during the Battle of the Somme in 1916. Reliability proved to be a problem: of the 49 tanks that stepped off on the attack on September 15th, only 3 were operational on the 16th, mostly due to mechanical failures and breakdowns.
Nevertheless there was promise in the idea that was clearly recognized and a major effort to show what tanks could do what attempted at Cambrai in November of 1917; this time hundreds of tanks were deployed and they had a real impact, breaking through the barbed wire and scattering the initial German defenses. But then came the inevitable German counter-attacks and most of the ground taken was lost. It was obvious that tanks had great potential; the French had by 1917 already developed their own, the light Renault FT tank, which would end up being the most successful tank of the war despite its small size (it is the first tank to have its main armament in a rotating turret and so in some sense the first “real” tank). This was hardly an under-invested in technology. So did tanks break the trench stalemate?
No.
It’s understandable that many people have the impression that they did. Interwar armored doctrine, particularly German Maneuver Warfare (bewegungskrieg) and Soviet Deep Battle both aimed to use the mobility and striking power of tanks in concentrated actions to break the trench stalemate in future wars (the two doctrines are not identical, mind you, but in this they share an objective). But these were doctrines constructed around the performance capabilities of interwar tanks, particularly by two countries (Germany and the USSR) who were not saddled with large numbers of WWI era tanks (and so could premise their doctrine entirely on more advanced models). The Panzer II, with a 24.5mph top speed and an operational range of around 100 miles, depending on conditions, was actually in a position to race the train and win; the same of course true of the Soviet interwar T-26 light tank (19.3mph on roads, 81-150 mile operational range). Such tanks could have radios for coordination and communication on the move (something not done with WWI tanks or even French tanks in WWII).
By contrast, that Renault FT had a top speed of 4.3mph and an operational range of just 37 miles. The British Mark V tank, introduced in 1918, moved at only 5mph and had just 45 miles of range. Such tanks struggled to keep up with the infantry; they certainly were not going to win any race the infantry could not. It is little surprise that the French, posed with the doctrinal problem of having to make use of the many thousands of WWI tanks they had, settled on a doctrine whereby most tanks would simply be the armored gauntlet stretched over the infantry’s fist: it was all those tanks could do! The sort of tank that could do more than just dent the trench-lines (the same way a good infiltration assault with infantry could) were a decade or more away when the war ended.
Moreover, of course, the doctrine – briefly the systems of thinking and patterns of training, habit and action – to actually pull off what tanks would do in 1939 and 1940 were also years away. It seems absurd to fault World War I era commanders for not coming up with a novel tactical and operational system in 1918 for using vehicles that wouldn’t exist for another 15 years and yet more so assuming that they would get it right (since there were quite a number of different ideas post-war about how tanks ought to be used and while many of them seemed plausible, not all of them were practical or effective in the field). It is hard to see how any amount of support into R&D or doctrine was going to make tanks capable of breakthroughs even in the late 1920s or early 1930s (honestly, look at the “best” tanks of the early 1930s; they’re still not up to the task in most cases) much less by 1918.
Bret Devereaux, “Collections: No Man’s Land, Part II: Breaking the Stalemate”, A Collection of Unmitigated Pedantry, 2021-09-24.
January 11, 2026
QotD: The limits of foreign policy realism
Longtime readers will remember that we’ve actually already talked about “realism” as a school of international relations study before, in the context of our discussion of Europa Universalis. But let’s briefly start out with what we mean when we say IR realism (properly “neo-realism” in its modern form): this is not simply being “realistic” about international politics. “Realism” is amazing branding, but “realists” are not simply claiming that they are observing reality – they have a broader claim about how reality works.
Instead realism is the view that international politics is fundamentally structured by the fact that states seek to maximize their power, act more or less rationally to do so, and are unrestrained by customs or international law. Thus the classic Thucydidean formulation in its most simple terms, “the strong do what they will, the weak suffer what they must”,1 with the additional proviso that, this being the case, all states seek to be as strong as possible.
If you accept those premises, you can chart a fairly consistent analytical vision of interstate activity basically from first principles, describing all sorts of behavior – balancing, coercion, hegemony and so on – that ought to occur in such systems and which does occur in the real world. Naturally, theory being what it is, neo-realist theory (which is what we call the modern post-1979 version of this thinking) is split into its own sub-schools based on exactly how they imagine this all works out, with defensive realism (“states aim to survive”) and offensive realism (“states aim to maximize power”), but we needn’t get into the details.
So when someone says they are a “foreign policy realist”, assuming they know what they’re talking about, they’re not saying they have a realistic vision of international politics, but that they instead believe that the actions of states are governed mostly by the pursuit of power and security, which they pursue mostly rationally, without moral, customary or legal constraint. This is, I must stress, not the only theory of the case (and we’ll get into some limits in a second).
The first problem with IR Realists is that they run into a contradiction between realism as an analytical tool and realism as a set of normative behaviors. Put another way, IR realism runs the risk of conflating “states generally act this way”, with “states should generally act this way”. You can see that specific contradiction manifested grotesquely in John Mearsheimer’s career as of late, where his principle argument is that because a realist perspective suggests that Russia would attack Ukraine that Russia was right to do so and therefore, somehow, the United States should not contest this (despite it being in the United States’ power-maximizing interest to do so). Note the jump from the analytical statement (“Russia was always likely to do this”) to the normative statement (“Russia carries no guilt, this is NATO’s fault, we should not stop this”). The former, of course, can always be true without the latter being necessary.
I should note, this sort of “normative smuggling” in realism is not remotely new: it is exactly how the very first instances of realist political thought are framed. The first expressions of IR realism are in Thucydides, where the Athenians – first at Corinth and then at Melos – make realist arguments expressly to get other states to do something, namely to acquiesce to Athenian Empire. The arguments in both cases are explicitly normative, that Athens did not act “contrary to the common practice of mankind” (expressed in realist dog-eat-dog terms) and so in the first case shouldn’t be punished with war by Sparta and in the latter case, that the Melians should submit to Athenian rule. In both cases, the Athenians are smuggling in a normative statement about what a state should do (in the former case, seemingly against interest!) into a description of what states supposedly always do.
I should note that one of my persistent complaints against international relations study in political science in general is that political scientists often read Thucydides very shallowly, dipping in for the theory and out for the rest. But Thucydides’ reader would not have missed that it is always the Athenians who make the realist arguments and they lost both the arguments [AND] the war. When Thucydides has the Melians caution that the Athenians’ “realist” ruthlessness would mean “your fall would be a signal for the heaviest vengeance and an example for the world to meditate upon”2 the ancient Greek reader knows they are right, in a way that it often seems to me political science students seem to miss.
And there’s a logical contradiction inherent in this sort of normative smuggling, which is that the smuggling is even necessary at all. After all, if states are mostly rational and largely pursue their own interests, loudly insisting that they should do so seems a bit pointless, doesn’t it? Using realism as a way to describe the world or to predict the actions of other states is consistent with the logical system, but using it to persuade other states – or your own state – seems to defeat the purpose. If you believe realism is true, your state and every other is going to act to maximize its power, regardless of what you do or say. If they can do otherwise than there must be some significant space for institutions, customs, morals, norms or simple mistakes and suddenly the air-tight logical framework of realism begins to break down.
That latter vision gives rise to constructivism (“international relations are shaped by ideology and culture”) and IR liberalism (“international relations are also shaped by institutions, which can bend the system away from the endless conflict realism anticipates”). The great irony of realism is that to think that having more realists in power would cause a country to behave in a more realist way is inconsistent with neo-Realism which would suggest countries ought to behave in realist ways even in the absence of realist theory or thinkers.
In practice – and this is the punchline – in my experience most “realists”, intentionally or not, use realism as a cover for strong ideological convictions, typically convictions which are uncomfortable to utter in the highly educated spaces that foreign policy chatter tends to happen. Sometimes those convictions are fairly benign – it is not an accident that there’s a vocal subset of IR-realists with ties to the CATO Institute, for instance. They’re libertarians who think the foreign policy adventures that often flew under the banner of constructivist or liberal internationalist label – that’s where you’d find “spreading democracy will make the world more peaceful” – were really expensive and they really dislike taxes. But “we should just spend a lot less on foreign policy” is a tough sell in the foreign policy space; realism can provide a more intellectually sophisticated gloss to the idea. Sometimes those convictions are less benign; one can’t help but notice the realist pretensions of some figures in the orbit of the current administration have a whiff of authoritarianism or ethnocentrism in them, since a realist framework can be used to drain imperial exploitation and butchery of its moral component, rendering it “just states maximizing their power – and better to be exploiter than exploited”.
One question I find useful to ask of any foreign policy framework, but especially of self-claimed realist frameworks is, “what compromise, what tradeoff does this demand of you?” Strategy, after all, is the art of priorities and that means accepting some things you want are lower priority; in the case of realism which holds that states seek to maximize power, it may mean assigning a high priority to things you do not want the state to do at all but which maximize its power. A realism deserving of the name, in applied practice would be endlessly caveated: “I hate, this but …” “I don’t like this, but …” “I would want to do this, but …” If a neo-realist analysis leads only to comfortable conclusions that someone and their priorities were right everywhere all along, it is simply ideology, wearing realism as a mask. And that is, to be frank, the most common form, as far as I can tell.
That isn’t to say there is nothing to neo-realism or foreign policy realists. I think as an analytical and predict tool, realism is quite valuable. States very often do behave in the way realist theory would suggest they ought, they just don’t always do so and it turns out norms and expectations matter a lot. Not the least of which because, as we’ve noted before, the economic model on which realist and neo-realist thinking was predicted basically no longer exists. To return to the current Ukraine War: is Putin really behaving rationally in a power-maximizing mode by putting his army to the torch capturing burned out Ukrainian farmland one centimeter at a time and no faster? It sure seems like Russian power has been reduced rather than enhanced by this move, even though realists will insist that Russia’s effort to dominate states near it is rational power-maximizing under offensive realism.
Bret Devereaux, “Fireside Friday, June 27, 2025 (On the Limits of Realism)”, A Collection of Unmitigated Pedantry, 2025-06-27.
- Thuc. 5.89.
- Thuc. 5.90.
January 5, 2026
QotD: Nitpicking the opening battle in Gladiator (2000)
This week, we’re going to take a close look at arguably the most famous and recognizable Roman battle sequence in film: the iconic opening battle from Gladiator (2000).1 Despite being a relatively short sequence (about ten minutes), there’s actually enough to talk about here that we’re going to split it over two weeks, talking about the setup – the battlefield, army composition, equipment and battle plan – this week and then the actual conduct of the battle next week.
The iconic opening battle, set in the Marcomannic Wars (166-180) during the reign of Marcus Aurelius (r. 161-180) dominates the pop-cultural reference points for the Roman army in battle and you can see its heavy influence in things like how the Total War series presents Roman armies (particularly in trailers and other promotional material). Students and enthusiasts alike will often cite this sequence as the thing which sparked their interest in the Roman army. It is hard to overstate how pervasive its influence is in the public imagination of what the Roman army, particularly of the imperial period, was like, especially as its style is imitated by later pop culture works.
Which is why it is so unfortunate that it is such a deceptive historical mess. This sequence in particular is a banner example of what I’ve termed elsewhere the “perils of historical verisimilitude“, the habit of historically based popular-culture works including what we might think of as fake signifiers of research, things that seem historically grounded rather than being historically grounded, as a way to cheaply cash in on the cachet that an actually grounded representation gets.
Gladiator actually provides a perfect metaphor for this: its main character’s name. Russell Crowe proudly informs us he is, “Maximus Decimus Meridius”, a name that certainly sounds suitably Roman, picking up the three-part name with that standard second declension -us ending. It sounds like it could be a real name – if you didn’t know Latin you would probably assume that it could be a real Roman name. But, as we’ve noted, it isn’t a Roman name and in fact gets nearly all of the Roman naming conventions wrong: Roman names are ordered as praenomen, nomen and cognomen, with the nomen indicating one’s gens (“clan” more or less) and the praenomen selected from just a couple dozen common personal names. Decimus is one of those two-dozen common praenomina (which also means it is never going to show up as the name of a gens), so it ought to go first as it is actually his personal name. Meanwhile Maximus (“the greatest”) is very much not one of those roughly two-dozen praenomina, instead being always cognomen (essentially a nickname). Finally Meridius isn’t a Latin word at all (so it can’t be a praenomen personal name nor a cognomen nickname),2 meaning it has to be the nomen (referencing a fictive gens Meridia). Every part of his name is wrong and it should read Decimus Meridius Maximus.
It sounds just right enough to fool your average viewer, while being entirely wrong. It is “truthy” rather than true – verisimilitudinous (like truth), rather than veristic (realistic, true).
In the case of Gladiator‘s opening battle scene, the attention is on creating verisimilitude (without fidelity, as we’ll see) in the visual elements of the sequence and only the visual elements. The visual representation of a Roman army – the equipment in particular – is heavily based on the Column of Trajan (including replicating the Column’s own deceptions) and since that is the one thing a viewer can easily check, that verisimilitude leads a lot of viewers to conclude that the entire sequence is much more historically grounded than it is. They take their cues from the one thing they can judge – “do these fellows wear that strange armor I saw on that picture of a Roman column?” – and assume everything is about as well researched, when in fact none of it is.
Instead, apart from the equipment – which has its own deep flaws – this is a sequence that bears almost no resemblance to the way Roman armies fought and expected to win their battles. The Roman army in this sequence has the wrong composition, is deployed incorrectly, uses the wrong tactics, has the wrong theory of victory and employs the wrong weapons and then employs them incorrectly. Perhaps most importantly the sequence suggests an oddly cavalry-and-archer focused Roman army which is simply not how the Romans in this period expected to win their battles.
Now I want to be clear here that this isn’t a review of the film Gladiator (2000) or my opinion in general on the film. To be honest, unlike the recent sequel, I enjoy Gladiator even though it is historical gibberish. So I am not telling you that you aren’t “allowed” to like Gladiator, but rather simply that, despite appearances, it is historical gibberish, particularly this opening scene, which I often find folks who are aware the rest of the film is historical gibberish nevertheless assume this opening scene is at least somewhat grounded. It is not.
Bret Devereaux, “Collections: Nitpicking Gladiator’s Iconic Opening Battle, Part I”, A Collection of Unmitigated Pedantry, 2025-06-06.
- I’d think its only real rival for prominence would be Spartacus (1960).
- If you are wondering, “but then were does our word “meridian” come from, the answer is from Latin meridies, meaning “midday”.
December 31, 2025
QotD: The cloth trade in the ancient and medieval world
Fabric as a finished product is somewhat different from the other products (grain and iron) we’ve discussed in this series so far. Bulk grain is a commodity – which means that one kilogram of grain is pretty much like the next. When grain gets traded in bulk it is because of differences in supply, not generally because grain in one region or other is particularly tasty. Consequently, you only get bulk grain trading when one area is producing a surplus and another not producing enough. The same is more loosely true of iron; Iron wasn’t a perfect commodity in the pre-modern world since some ores produced better quality metal than others and some trade in high quality ores or metal (like wootz) happened. But for every-day use, local iron was generally good enough and typically available. Iron could be – and often was – treated as a commodity too.
Fabric is not like that. While the lower classes will often have had to make do with whatever sorts of cloth is produced cheaply and locally, for people who could afford to have some choices the very nature of textile production produces lots of regional differences which incentivized trade. Almost everything about textile production is subject to regional variations:
- What fibers are being used. Major wool and linen producing regions tended to be separate, with wool (and sheep) in colder climates and uplands while (as noted) flax tended to be grown in warmer river-valleys with rich alluvial soil. Meanwhile, the cotton-and-silk producing regions generally (with some notable exceptions, mind) did not overlap with the wool and linen producing regions.
- The nature of the local fibers. We’ll get to some specific examples in a moment but ancient and medieval writers were well aware that different growing conditions, breeds of sheep or flax, climate and so on produced fibers of subtly different qualities. Consequently, even with the exact same processes, cloth produced in one region might be different from cloth produced in another region. But the processes were almost never the same because …
- Local variation in production processes. Again, we’ll have some examples in a moment, but the variance here could be considerable. As we’ve seen, the various tasks in cloth production are all pretty involved and give a lot of room for skill and thus for location variations in methods and patterns. One might see different weaving patterns, different spinning techniques, different chemical treatments, different growing or shearing methods and so on producing fabrics of different qualities which might thus be ideal for different purposes.
- Dye methods and availability. And as we discussed last time, available dyestuffs (and the craft knowledge about how to use them) was also often very local, leading to certain colors or patterns of color being associated with different regions and creating a demand for those. While it was often possible to ship dyestuffs (although not all dyestuffs responded well to long-distance shipping), it was often more economical to shift dyed fabric.
Added on top of this, fabric is a great trade-good. It is relatively low bulk when wrapped around in a roll (a bolt of fabric might hold fabric anywhere from 35-91m long and 100-150cm wide. Standard English broadcloth was 24 yards x 1.75 yards; that’s a lot of fabric in both cases!) and could be very high value, especially for high quality or foreign fabrics (or fabrics dyed in rare or difficult colors). Moreover, fabric isn’t perishable, temperature sensitive (short of an actual fire) or particularly fragile, meaning that as long as it is kept reasonably dry it will keep over long distances and adverse conditions. And everyone needs it; fabrics are almost perfect stock trade goods.
Consequently, we have ample evidence to the trade of both raw fibers (that is, wool or flax rovings) and finished fabrics from some of the earliest periods of written records (spotting textile trade earlier than that is hard, since fabric is so rarely preserved in the archaeological record). Records from Presargonic Mesopotamia (c. 2400-2300) record wool trading both between Mesopotamian cities but wool being used as a trade good for merchants heading through the Persian Gulf, to Elam and appears to have been one of, if not the primary export good for Sumerian cities (W. Sallaberger in Breniquet and Michel, op. cit.). More evidence comes later, for instance, palace letters and records from the Old Babylonian Empire (1894-1595) reporting the commercialization of wool produced under the auspices of the palace or the temple (Mesopotamian economies being centralized in this way in what is sometimes termed a “redistribution economy” though this term and the model it implies is increasingly contested as it becomes clearer from our evidence that economic activity outside of the “palace economy” also existed) and being traded with other cities like Sippar (on this, note K. De Graef and C. Michel’s chapters in Breniquet and Michel, op. cit.).
Pliny the Elder, in detailing wool and linen producing regions provides some clues for the outline of the cloth trade in the Roman world. Pliny notes that the region between the Po and Ticino river (in Northern Italy) produced linen that was never bleached while linen from Faventia (modern Faenza) was always bleached and renowned for its whiteness (Pliny, NH 19.9). Linen from Spain around Tarragona was thought by Pliny to be exceptionally fine while linens from Zeola (modern Oiartzun, Spain) was particularly durable and good for making nets (Pliny, NH 19.10). Meanwhile Egyptian flax he notes is the least strong but the most fine and thus the most expensive (Pliny NH 19.14). Meanwhile on wool, Pliny notes that natural wool color varied by region; the best white wool he thought came from the Po River valley, the best black wool from the Alps, the best red wool from Asia Minor, the best brown wool from Canusium (in Apulia) and so on (Pliny, Natural History 8.188-191). He also notes different local manufacture processes producing different results, noting Gallic embroidery and felting (NH 8.192). And of course, being Pliny, he must rank them all, with wool from Tarentum and Canusium (Taranto and Canosa di Puglia) being the best, followed more generally by Italian wools and in third place wools from Miletus in Asia Minor (Pliny NH 8.190). Agreement was not quite universal, Columella gives the best wools as those from Miletus, Calabria, and Apulia, with Tarantine wool being the best, but demoting the rest of the Italian wools out of the list entirely (Col. De Re Rust. 7.2).
In medieval Europe, wool merchants were a common feature of economic activity in towns, with the Low Countries and Northern Italy in particular being hubs of trade in wool and other fabrics (Italian ports also being one of the major routes by which cottons and silks from India and China might find their way, via the Mediterranean, to European aristocrats). Wool produced in Britain (which was a major production center) would be shipped either as rovings or as undyed “broadcloths” (called “whites”) to the Low Countries for dyeing and sale abroad (though there was also quite a lot of cloth dyeing happening in Britain as well).
Bret Devereaux, “Collections: Clothing, How Did They Make It? Part IVb: Cloth Money”, A Collection of Unmitigated Pedantry, 2021-04-09.
December 16, 2025
QotD: Arms and the (pre-modern) man (at arms)
… how much value might a heavily armored fighter or warrior be carrying around on their backs in the real world? Because I think the answer here is informative.
Here we do have some significant price data, but of course its tricky to be able to correlate a given value for arms and armor with something concrete like wages in every period, because of course prices are not stable. But here are some of the data points I’ve encountered:
We don’t have good Roman price data from the Republic or early/high Empire, unfortunately (and indeed, the reason I have been collecting late antique and medieval comparanda is to use it to understand the structure of earlier Roman costs). Hugh Elton1 notes that a law of Valens (r. 364-378) assessed the cost of clothing, equipment and such for a new infantry recruit to be 6 solidi and for a cavalryman, 13 solidi (the extra 7 being for the horse). The solidus was a 4.5g gold coin at the time (roughly equal to the earlier aureus) so that is a substantial expense to kit out an individual soldier. For comparison, the annual rations for soldiers in the same period seem to have been 4-5 solidi, so we might suggest a Roman soldier is wearing something like a year’s worth of living expenses.2
We don’t see a huge change in the Early Middle Ages either. The seventh century Lex Ripuaria,3 quotes the following prices for military equipment: 12 solidi for a coat of mail, 6 solidi for a metal helmet, 7 for a sword with its scabbard, 6 for mail leggings, 2 solidi for a lance and shield for a rider (wood is cheap!); a warhorse was 12 solidi, whereas a whole damn cow was just 3 solidi. On the one hand, the armor for this rider has gotten somewhat more extensive – mail leggings (chausses) were a new thing (the Romans didn’t have them) – but clearly the price of metal equipment here is higher: equipping a mailed infantryman would have some to something like 25ish solidi compared to 12 for the warhorse (so 2x the cost of the horse) compared to the near 1-to-1 armor-to-horse price from Valens. I should note, however, warhorses even compared to other goods, show high volatility in the medieval price data.
As we get further one, we get more and more price data. Verbruggen (op. cit. 170-1) also notes prices for the equipment of the heavy infantry militia of Bruges in 1304; the average price of the heavy infantry equipment was a staggering £21, with the priciest item by far being the required body armor (still a coat of mail) coming in between £10 and £15. Now you will recall the continental livre by this point is hardly the Carolingian unit (or the English one), but the £21 here would have represented something around two-thirds of a year’s wages for a skilled artisan.
Almost contemporary in English, we have some data from Yorkshire.4 Villages had to supply a certain number of infantrymen for military service and around 1300, the cost to equip them was 5 shillings per man, as unarmored light infantry. When Edward II (r. 1307-1327) demanded quite minimally armored men (a metal helmet and a textile padded jack or gambeson), the cost jumped four-fold to £1, which ended up causing the experiment in recruiting heavier infantry this way to fail. And I should note, a gambeson and a helmet is hardly very heavy infantry!
For comparison, in the same period an English longbowman out on campaign was paid just 2d per day, so that £1 of kit would have represented 120 days wages. By contrast, the average cost of a good quality longbow in the same period was just 1s, 6d, which the longbowman could earn back in just over a week.5 Once again: wood is cheap, metal is expensive.
Finally, we have the prices from our ever-handy Medieval Price List and its sources. We see quite a range in this price data, both in that we see truly elite pieces of armor (gilt armor for a prince at £340, a full set of Milanese 15th century plate at more than £8, etc) and its tricky to use these figures too without taking careful note of the year and checking the source citation to figure out which region’s currency we’re using. One other thing to note here that comes out clearly: plate cuirasses are often quite a bit cheaper than the mail armor (or mail voiders) they’re worn over, though hardly cheap. Still, full sets of armor ranging from single to low-double digit livres and pounds seem standard and we already know from last week’s exercise that a single livre or pound is likely reflecting a pretty big chunk of money, potentially close to a year’s wage for a regular worker.
So while your heavily armored knight or man-at-arms or Roman legionary was, of course, not walking around with the Great Pyramid’s worth of labor-value on his back, even the “standard” equipment for a heavy infantryman or heavy cavalryman – not counting the horse! – might represent a year or even years of a regular workers’ wages. On the flipside, for societies that could afford it, heavy infantry was worth it: putting heavy, armored infantry in contact with light infantry in pre-gunpowder warfare generally produces horrific one-sided slaughters. But relatively few societies could afford it: the Romans are very unusual for either ancient or medieval European societies in that they deploy large numbers of armored heavy infantry (predominately in mail in any period, although in the empire we also see scale and the famed lorica segmentata), a topic that forms a pretty substantial part of my upcoming book, Of Arms and Men, which I will miss no opportunity to plug over the next however long it takes to come out.6 Obviously armored heavy cavalry is even harder to get and generally restricted to simply putting a society’s aristocracy on the battlefield, since the Big Men can afford both the horses and the armor.
But the other thing I want to note here is the social gap this sort of difference in value creates. As noted above with the bowman’s wages, it would take a year or even years of wages for a regular light soldier (or civilian laborers of his class) to put together enough money to purchase the sort of equipment required to serve as a soldier of higher status (who also gets higher pay). Of course it isn’t as simple as, “work as a bowman for a year and then buy some armor”, because nearly all of that pay the longbowman is getting is being absorbed by food and living expenses. The result is that the high cost of equipment means that for many of these men, the social gap between them and either an unmounted man-at-arms or the mounted knight is economically unbridgeable.7
Bret Devereaux, “Fireside Friday, January 10, 2025”, A Collection of Unmitigated Pedantry, 2025-01-10.
- Warfare in Roman Europe AD 350-425 (1996), 122.
- If you are wondering why I’m not comparing to wages, the answer is that by this point, Roman military wages are super irregular, consisting mostly of donatives – special disbursements at the accession of a new emperor or a successful military campaign – rather than regular pay, making it really hard to do a direct comparison.
- Here my citation is not from the text directly, but Verbruggen, The Art of War in Western Europe during the Middle Ages (1997), 23.
- From Prestwich, Armies and Warfare in the Middle Ages (1996).
- These prices via Hardy, Longbow: A Social and Military History (1992).
- Expect it no earlier than late this year; as I write this, the core text is actually done (but needs revising), but that’s hardly the end of the publication process.
- Man-at-arms is one of those annoyingly plastic terms which is used to mean “a man of non-knightly status, equipped in the sort of kit a knight would have”, which sometimes implies heavy armored non-noble infantry and sometimes implies non-knightly heavy cavalry retainers of knights and other nobles.
December 9, 2025
QotD: The development of army discipline and drill in pre-modern armies
The usual solution to this difficulty [maneuvring units on the battlefield] often goes by the terms “drill” or “discipline” though we should be clear here exactly what we mean. Discipline in particular has a number of meanings: it can mean the personal restraint of an individual, a system of rewards and punishments (and the effects of that system; the punishments are typically corporal) and what we are actually interested in: the ability of a large body of humans to move and act effectively in concert (all of these meanings are present to some degree in the root Latin word disciplina). For clarity’s sake then I am going to borrow a term (as is my habit) from W. Lee, Waging War (2016), synchronized discipline to describe the “humans moving an acting in concert” component of discipline that we’re most interested in here. That said, it is worth noting that those three components: personal restraint, corporal punishments and the synchronized component of discipline are frequently (but not universally) associated for reasons we’ll get to, not merely in the Roman concept of disciplina, but note also for instance their close association in Sun Tzu’s Art of War in the first chapter (section 13).
The reason we cannot just call this “drill” is because while drill is the most common way agrarian societies produce this result, it is not the only way to this end. For instance as we’ve discussed before, steppe nomads could achieve a very high degree of coordination and synchronization without the same formal systems of drill because the training that produced that coordination was embedded in their culture (particularly in hunting methods) and so young steppe nomad males were acculturated into the synchronicity that way. That said for the rest of this we’re going to place those systems aside and mostly focus on synchronized discipline as a result of drill because for most armies that developed a great deal of synchronized discipline, that’s how they did it.
Fundamentally the principle behind using drill to build synchronized discipline is that the way to get a whole lot of humans to act effectively in concert together is to force them to practice doing exactly the things they’ll be asked to do on the battlefield a lot until the motions are practically second nature. Indeed, the ideal in developing this kind of drill was often to ingrain the actions the soldiers were to perform so deeply that in the midst of the terror of battle when they couldn’t even really think straight those soldiers would fall back on simply mechanically performing the actions they were trained to perform. That in turn creates an important element of predictability: an individual soldier does not need to be checking their action or position against the others around them as much because they’ve done this very maneuver with these very fellows and so already know where everyone is going to be.
The context that drill tends to emerge in (this is an idea invented more than once) tends to give it a highly regimented, fairly brutal character. For instance in early modern Europe, the structure of drill for gunpowder armies was conditioned by elite snobbery: European officer-aristocrats (in many cases the direct continuation of the medieval aristocracy) had an extremely poor view of their common soldiers (drawn from the peasantry). Assuming they lacked any natural valor, harsh drill was settled upon as a solution to make the actions of battle merely mechanical, to reduce the man to a machine. Roman commanders seemed to have thought somewhat better of their soldiers’ bravery, but assumed that harsh discipline was necessary to control, restrain and direct the native fiery virtus (“strength/bravery/valor”) of the common soldier who, unlike the aristocrat, could not be expected to control himself (again, in the snobbish view of the aristocrats).
In short, drill tends to appear in highly stratified agrarian societies, the very nature of which tends to mean that drill is instituted by a class of aristocrats who have at best a dim view of their common soldiers. Consequently, while the core of drill is to simply practice the actions of battle over and over again until they become natural, drill tends to also be encrusted with lots of corporal punishments and intense regulation as a product of those elite attitudes. And though it falls outside of our topic today it seems worth noting that our systems of drill to produce synchronized discipline have the same roots (deriving from early modern musket drill).
Bret Devereaux, “Collections: Total Generalship: Commanding Pre-Modern Armies, Part IIIa”, A Collection of Unmitigated Pedantry, 2022-06-17.




