Tag Trevor N. Dupuy

Human Factors in Warfare: Measuring Morale

Figure One: Second Army, Weekly Admissions per 1,000 to General Hospitals and Casualty Clearing Stations for Sickness, Battle Exhaustion and SIW, 11 June 1944 to 5 May 1945. Morale scale equivalents are presented on the right hand Y-axis. (Dr. Jonathan Fennell)
Figure One: Second Army, Weekly Admissions per 1,000 to General Hospitals and Casualty Clearing Stations for Sickness, Battle Exhaustion and SIW, 11 June 1944 to 5 May 1945. Morale scale equivalents are presented on the right hand Y-axis. (Dr. Jonathan Fennell)

One of Trevor Dupuy’s more important and controversial contributions to a theory of combat was the assertion that outcomes were dictated in part by behavioral factors, i.e. the human element. Among the influential human factors he identified were leadership, training, experience, and manpower quality. He also recognized the importance of morale.

Morale is an ephemeral quality of military forces and is certainly intangible. Yet even though it may not be easily defined and can probably never be quantified, troop morale is very real and can be very important as a contributor to victory or defeat. The significance of morale is probably inversely proportional to the quality of troops. A well-trained, well-led, cohesive force of veterans will fight well and effectively even if morale is low… Yet for ordinary armies, poor morale can contribute to defeat.[1]

Dr. Jonathan Fennell of the Defence Studies Department at King’s College London recently set out to determine if there were ways of measuring morale by looking at the combat experiences of the British Army in World War II. Fennell proposed

that the concept of morale has no place in a critical analysis of the past unless it is clearly differentiated from definitions associated solely or primarily with mood or cohesion and the group. Instead, for morale to have explanatory value, particularly in a combat environment, a functional conceptualisation is proposed, which, while not excluding the role of mood or group cohesion, focuses its meaning and relevance on motivation and the willingness to act in a manner required by an authority or institution.

Fennell constructed a multi-dimensional model of morale

By drawing on studies made across the social sciences and on primary archival evidence from the British and Commonwealth Army’s experiences in North Africa in the Second World War… It suggests that morale can best be understood as emerging from the subtle interdependencies and interrelationships of the many factors known to affect military means.

Fennell tested his methodology by developing a weekly morale score using bi-weekly censorship summaries of letters and correspondence from members of the British Second Army in the Northwest Europe Campaign in 1944-45.

These summaries proved a useful source to describe and ‘quantify’ levels of morale (through the use of a numerical morale scale). Where morale was described as ‘excellent’, it was awarded a score of 3. ‘High’ morale was given a score of 2 and ‘good’ morale was scored 1. ‘Satisfactory’ morale was given a score of 0 (neither positive or negative). Morale described as ‘severely tried’ was scored -1, while ‘low’ and ‘very low’ morale were scored -2 and -3 respectively.

He then correlated these scores with weekly statistics compiled by the British Second Army and 21st Army Group on rates of sickness, battle exhaustion, desertion, absence without leave (AWOL) and self-inflicted wounds (SIW).

The results of the correlation analysis showed that the tabulated rates (the combined rate of sickness, battle exhaustion, desertion, AWOL and SIW) had an extremely strong negative correlation with morale (-0.949, P<0.001), i.e. when morale was high, sickness rates etc. were low, and when morale was low, sickness rates etc. were high. This is a remarkably strong relationship and shows that these factors when taken together can be used as a quantitative method to assess levels of morale, at the very least for the Army and campaign under discussion.

The results are shown on the graph above. According to Fennell,

This analysis of morale supports the conclusions of much of the recent historiography on the British Army in Northwest Europe; morale was a necessary component of combat effectiveness (morale in Second Army was broadly speaking high throughout the victorious campaign); however, morale was not a sufficient explanation for Second Army’s successes and failures on the battlefield. For example, morale would appear to have been at its highest before and during Operation ‘Market Garden’. But ‘Market Garden’ was a failure. It is likely, as John Buckley has argued, that ‘Market Garden’ was a conceptual failure rather than a morale one. Morale would also appear to have been mostly high during operations in the Low Countries and Germany, but these operations were beset with setbacks and delays.

Fennell further explored the relationship between morale and combat performance, and combat performance and strategy, in his contribution to Anthony King, ed., Frontline: Combat and Cohesion in the Twenty-First Century (Oxford: Oxford University Press, 2015). An earlier version of his chapter can be found here.

NOTES

[1] Trevor N. Dupuy, Understanding Defeat: How To Recover From Loss In Battle To Gain Victory In War (New York: Paragon House, 1990), p. 67

Assessing the TNDA 1990-91 Gulf War Forecast

Map of ground operations of Operation Desert Storm starting invasion February 24-28th 1991. Shows allied and Iraqi forces. Special arrows indicate the American 101st Airborne division moved by air and where the French 6st light division and American 3rd Armored Cavalry Regiment provided security. Image created by Jeff Dahl and reposted under the terms of the GNU Free Documentation License, Version 1.2.
Map of ground operations of Operation Desert Storm starting invasion February 24-28th 1991. Shows allied and Iraqi forces. Special arrows indicate the American 101st Airborne division moved by air and where the French 6st light division and American 3rd Armored Cavalry Regiment provided security. Image created by Jeff Dahl and reposted under the terms of the GNU Free Documentation License, Version 1.2.

[NOTE: This post has been edited to more accurately characterize Trevor Dupuy’s public comments on TNDA’s estimates.]

Operation DESERT STORM began on 17 January 1991 with an extended aerial campaign that lasted 38 days. Ground combat operations were initiated on 24 February and concluded after four days and four hours, with U.S. and Coalition forces having routed the Iraqi Army in Kuwait and in position to annihilate surviving elements rapidly retreating northward. According to official accounting, U.S. forces suffered 148 killed in action and 467 wounded in action, for a total of 614 combat casualties. An additional 235 were killed in non-hostile circumstances.[1]

In retrospect, TNDA’s casualty forecasts turned out to be high, with the actual number of casualties falling below the lowest +/- 50% range of estimates. Forecasts, of course, are sensitive to the initial assumptions they are based upon. In public comments made after the air campaign had started but before the ground phase began, Trevor Dupuy forthrightly stated that TNDA’s estimates were likely to be too high.[2]

In a post-mortem on the forecast in March 1991, Dupuy identified three factors which TNDA’s estimates miscalculated:

  • an underestimation of the effects of the air campaign on Iraqi ground forces;
  • the apparent surprise of Iraqi forces; and
  • an underestimation of the combat effectiveness superiority of U.S. and Coalition forces.[3]

There were also factors that influenced the outcome that TNDA could not have known beforehand. Its estimates were based on an Iraqi Army force of 480,000, a figure derived from open source reports available at the time. However, the U.S. Air Force’s 1993 Gulf War Air Power Survey, using intelligence collected from U.S. government sources, calculated that there were only 336,000 Iraqi Army troops in and near Kuwait in January 1991 (out of a nominal 540,000) due to unit undermanning and troops on leave. The extended air campaign led a further 25-30% to desert and inflicted about 10% casualties, leaving only 200,000-220,000 depleted and demoralized Iraqi troops to face the U.S. and Coalition ground offensive.[4].

TNDA also underestimated the number of U.S. and Coalition ground troops, crediting them with a total of 435,000, when the actual number was approximately 540,000.[5] Instead of the Iraqi Army slightly outnumbering its opponents in Kuwait as TNDA approximated (480,000 to 435,000), U.S. and Coalition forces probably possessed a manpower advantage approaching 2 to 1 or more at the outset of the ground campaign.

There were some aspects of TNDA’s estimate that were remarkably accurate. Although no one foresaw the 38-day air campaign or the four-day ground battle, TNDA did come quite close to anticipating the overall duration of 42 days.

DESERT STORM as planned and executed also bore a striking resemblance to TNDA’s recommended course of action. The opening air campaign, followed by the “left hook” into the western desert by armored and airmobile forces, coupled with holding attacks and penetration of the Iraqi lines on the Kuwaiti-Saudi border were much like a combination of TNDA’s “Colorado Springs,” “Leavenworth,” and “Siege” scenarios. The only substantive difference was the absence of border raids and the use of U.S. airborne/airmobile forces to extend the depth of the “left hook” rather than seal off Kuwait from Iraq. The extended air campaign served much the same intent as TNDA’s “Siege” concept. TNDA even anticipated the potential benefit of the unprecedented effectiveness of the DESERT STORM aerial attack.

How effective “Colorado Springs” will be in damaging and destroying the military effectiveness of the Iraqi ground forces is debatable….On the other hand, the circumstances of this operation are different from past efforts of air forces to “go it alone.” The terrain and vegetation (or lack thereof) favor air attacks to an exceptional degree. And the air forces will be operating with weapons of hitherto unsuspected accuracy and effectiveness against fortified targets. Given these new circumstances, and considering recent historical examples in the 1967 and 1973 Arab-Israeli Wars, the possibility that airpower alone can cause such devastation, destruction, and demoralization as to destroy totally the effectiveness of the Iraqi ground forces cannot be ignored. [6]

In actuality, the U.S. Central Command air planners specifically targeted Saddam’s government in the belief that air power alone might force regime change, which would lead the Iraqi Army to withdraw from Kuwait. Another objective of the air campaign was to reduce the effectiveness of the Iraqi Army by 50% before initiating the ground offensive.[7]

Dupuy and his TNDA colleagues did anticipate that a combination of extended siege-like assault on Iraqi forces in Kuwait could enable the execution of a quick ground attack coup de grace with minimized losses.

The potential of success for such an operation, in the wake of both air and ground efforts made to reduce the Iraqi capacity for offensive along the lines of either Operation “Leavenworth’…or the more elaborate and somewhat riskier “RazzleDazzle”…would produce significant results within a short time. In such a case, losses for these follow-on ground operations would almost certainly be lower than if they had been launched shortly after the war began.[8]

Unfortunately, TNDA did not hazard a casualty estimate for a potential “Colorado Springs/ Siege/Leavenworth/RazzleDazzle” combination scenario, a forecast for which might very well have come closer to the actual outcome.

Dupuy took quite a risk in making such a prominently public forecast, opening his theories and methodology to criticism and judgement. In my next post, I will examine how it stacked up with other predictions and estimates made at the time.

NOTES

[1] Nese F. DeBruyne and Anne Leland, “American War and Military Operations Casualties: Lists and Statistics,” (Washington, D.C.: Congressional Research Service, 2 January 2015), pp. 3, 11

[2] Christopher A. Lawrence, America’s Modern Wars: Understanding Iraq, Afghanistan and Vietnam (Philadelphia, PA: Casemate, 2015) p. 52

[3] Trevor N. Dupuy, “Report on Pre-War Forecasting For Information and Comment: Accuracy of Pre-Kuwait War Forecasts by T.N. Dupuy and HERO-TNDA,” 18 March, 1991. This was published in the April 1991 edition of the online wargaming “fanzine” Simulations Online. The post-mortem also included a revised TNDM casualty calculation for U.S. forces in the ground war phase, using the revised assumptions, of 70 killed and 417 wounded, for a total of 496 casualties. The details used in this revised calculation were not provided in the post-mortem report, so its veracity cannot be currently assessed.

[4] Thomas A. Keaney and Eliot A. Cohen, Gulf War Airpower Survey Summary Report (Washington, D.C.: U.S. Department of the Air Force, 1993), pp. 7, 9-10, 107

[5] Keaney and Cohen, Gulf War Airpower Survey Summary Report, p. 7

[6] Trevor N. Dupuy, Curt Johnson, David L. Bongard, Arnold C. Dupuy, How To Defeat Saddam Hussein: Scenarios and Strategies for the Gulf War (New York: Warner Books, 1991), p. 58

[7] Gulf War Airpower Survey, Vol. I: Planning and Command and Control, Pt. 1 (Washington, D.C.: U.S. Department of the Air Force, 1993), pp. 157, 162-165

[8] Dupuy, et al, How To Defeat Saddam Hussein, p. 114

Forecasting the 1990-1991 Gulf War

DoD photo by Regina Ali
DoD photo by Regina Ali

In my last post on the subject of prediction in security studies, I mentioned that TDI has a public forecasting track record. The first of these, and possibly the most well know, involves the 1990-1991 Gulf War.

On 13 December 1990, Trevor N. Dupuy, President of Trevor N. Dupuy & Associates (TNDA), testified before the House Armed Services Committee on the topic of the looming military confrontation between the military forces of the United States and United Nations Coalition allies and those of Iraq.[1] He offered TNDA’s assessment of the potential character of the forthcoming conflict, as well as estimates of the likely casualties that both sides would suffer. Dupuy published a refined and expanded version of TNDA’s analysis in January 1991.[2]

Based on a methodology derived from Dupuy’s combat models and synthesized data on historical personnel and material combat attrition, TNDA forecast a successful U.S. and Coalition air/ground offensive campaign into Kuwait.[3] Using publicly available sources, TNDA calculated that Iraqi forces in Iraq numbered 480,000, U.S. forces at 310,000, and Coalition allies at 125,000.

The estimated number of casualties varied based on a campaign anticipated to last from 10 to 40 days depending on five projected alternate operational scenarios:

  • Operation “Colorado Springs.” A 10-day air campaign aimed at achieving air superiority and attacking Iraq’s ground forces and war-making infrastructure. While TNDA believed an air campaign would proceed any ground offensive option, Dupuy suggested that it could potentially force an Iraqi surrender without the need for a land attack.
  • Operation “Bulldozer.” A frontal assault on Iraqi forces in Kuwait, lasting 10-20 days.
  • Operation “Leavenworth.” A double envelopment of Iraqi forces in Kuwait using an armored turning force in the west and a U.S. Marine amphibious landing in the east.
  • Operation “RazzleDazzle.” Similar to “Leavenworth,” but combined with an assault along the Iraq-Kuwait border by airborne/airmobile forces for a triple envelopment to encircle all Iraqi forces in Kuwait.
  • Operation “Siege.” A combination of an extended Operation “Colorado Springs” and ground force raids on all of Iraq’s borders. After 10-20 days, one of the three ground attack options (“Bulldozer,” “Leavenworth,” or “RazzleDazzle”) would be initiated to conclude the war.

Based on these assumptions, TNDA produced a range of casualty predictions for U.S. forces that TNDA asserted would probably be accurate to within +/- 50%. These ranged from a low of 380 for a 10-day “Colorado Springs” air-only campaign, to a top-end calculation of 16,645 for a 10-day “Colorado Springs” followed by a 20-day “Bulldozer” frontal assault.

TNDA’s Projection of Likely U.S. Casualties

Scenario Duration

Killed

Wounded

Total

+/-50%

Colorado Springs

10-40 days

190-315

190-315

380-630

Bulldozer*

10-20 days

1,858-2,068

8,332-9,222

10,190-11,290

5,335-16,645

Leavenworth*

10-20 days

1,454-1,664

6,309-7,199

7,763-8,863

4,122-12,995

RazzleDazzle*

10-20 days

1,319-1,529

5,534-6,524

6,853-8,053

3,717-11,790

Siege*

10-30 days

564-1,339

1,858-5,470

2,422-6,809

1,451-10,479

* Figures include air casualties

Based on these calculations, TNDA recommended the following course of action:

If the above figures are close to accurate (and history tells us they should should be), then the proper solution is to begin the war with the air campaign of Operation “Colorado Springs.” If this should result in an Iraqi surrender, so much the better. If not, then after about ten days of “Colorado Springs,“ to continue the air campaign for about ten more days while initiating Operation “Siege.” If this does not bring about an Iraqi surrender, the ground campaign should be concluded with Operation “RazzleDazzle.” If this has not brought about an Iraqi surrender, then an advance should be made through the desert to destroy any resisting Iraqi forces and to occupy Baghdad if necessary.[4]

In my next post, I will assess the accuracy of TNDA’s forecast and how it stacked up against others made at the time.

Notes

[1] Armed Services Committee, U.S. House of Representatives, Testimony of Col. T. N. Dupuy, USA, Ret. (Washington D.C.: 13 December 1990)

[2] Trevor N. Dupuy, Curt Johnson, David L. Bongard, Arnold C. Dupuy, If War Comes, How To Defeat Saddam Hussein (McLean, VA.: HERO Books, 1991); subsequently republished as How To Defeat Saddam Hussein: Scenarios and Strategies for the Gulf War (New York: Warner Books, 1991).

[3] These are the Quantified Judgement Model (QJM) and Tactical Numerical Deterministic Model (TNDM). Dupuy’s methodological approach and his first cut on a Gulf War estimate are described in Chapter 7 of Trevor N. Dupuy, Attrition: Forecasting Battle Casualties and Equipment Losses in Modern War (McLean, VA.: HERO Books, 1990).

[4] Dupuy, et al, How To Defeat Saddam Hussein, 126

Will This Weapon Change Infantry Warfare Forever? Maybe, But Probably Not

XM25 Counter Defilade Target Engagement (CDTE) System

The weapon pictured above is the XM25 Counter Defilade Target Engagement (CDTE) precision-guided grenade launcher. According to its manufacturer, Orbital ATK,

The XM25 is a next-generation, semi-automatic weapon designed for effectiveness against enemies protected by walls, dug into foxholes or hidden in hard-to-reach places.

The XM25 provides the soldier with a 300 percent to 500 percent increase in hit probability to defeat point, area and defilade targets out to 500 meters. The weapon features revolutionary high-explosive, airburst ammunition programmed by the weapon’s target acquisition/fire control system.

Following field testing in Afghanistan that reportedly produced mixed results, the U.S. Army is seeking funding the Fiscal Year 2017 defense budget to acquire 105 of the weapons for issue to specifically-trained personnel at the tactical unit level.

The purported capabilities of the weapon have certainly raised expectations for its utility. A program manager in the Army’s Program Executive Office declared “The introduction of the XM25 is akin to other revolutionary systems such as the machine gun, the airplane and the tank, all of which changed battlefield tactics.” An industry observer concurred, claiming that “The weapon’s potential revolutionary impact on infantry tactics is undeniable.”

Well…maybe. There is little doubt that the availability of precision-guided standoff weapons at the squad or platoon level will afford significant tactical advantages. Whatever technical problems that currently exist will be addressed and there will surely be improvements and upgrades.

It seems unlikely, however, that the XM25 will bring revolutionary change to the battlefield. In his 1980 study The Evolution of Weapons and Warfare, Trevor N. Dupuy explored the ongoing historical relationship between technological change and adaptation on the battlefield. The introduction of increasingly lethal weapons has led to corresponding changes in the ways armies fight.

Assimilation of a significant increase in [weapon] lethality has generally been marked (a) by dispersion, thus reducing the number of people exposed to the new weapon in the enemy’s hands; (b) by giving greater freedom of maneuver; and (c) by improving cooperation among the different arms and services. [p. 337]

As the chart below illustrates (click for a larger version), as weapons have become more lethal over time, combat forces have adjusted by dispersing in greater frontage and depth on the battlefield (as reflected by the red line).

[pp. 288-289]

Dupuy noted that there is a lag between the introduction of a new weapon and its full integration into an army’s tactics and force structure.

In modern times — and to some extent in earlier eras — there has been an interval of approximately twenty years between introduction and assimilation of new weapons…it is significant that, despite the rising tempo of invention, this time lag remained relatively constant. [p. 338]

Moreover, Dupuy observed that true military revolutions are historically rare, and require more than technological change to occur.

Save for the recent significant exception of strategic nuclear weapons, there have been no historical instances in which new and more lethal weapons have, of themselves, altered the conduct of war or the balance of power until they have been incorporated into a new tactical system exploiting their lethality and permitting their coordination with other weapons. [p. 340]

Looking at the trends over time suggests that any resulting changes will be evolutionary rather than revolutionary. The ways armies historically have adapted to new weapons — dispersion, tactical flexibility, and combined arms —- are hallmarks of the fire and movement concept that is at the heart of modern combat tactics, which evolved in the early years of the 20th century, particularly during the First World War. However effective the XM25 may prove to be, it’s impact is unlikely to alter the basic elements of fire and movement tactics. Enemy combatants will likely adapt through even greater dispersion (the modern “empty battlefield“), tactical innovation, and combinations of countering weapons. It is also likely that it will take time, trial and error, and effective organizational leadership in order to take full advantage of the XM25’s capabilities.

[Edited]

Screw Theory! We Need More Prediction in Security Studies!

Johnny Carson as Carnac the Magnificent; taken from the January 24, 2005 broadcast of The Tonight Show.
Johnny Carson as Carnac the Magnificent; taken from the January 24, 2005 broadcast of The Tonight Show.

My previous post touched on the apparent analytical bankruptcy underlying the U.S. government’s approach to counterterrorism policy. While many fingers were pointed at the government for this state of affairs, at least one scholar admitted that “the leading [academic] terrorism research was mostly just political theory and anecdotes” which has “left policy makers to design counterterrorism strategies without the benefit of facts.”

So what can be done about this? Well, Michael D. Ward, a Professor of Political Science at Duke University, has suggested a radical solution: test the theories to see if they can accurately predict real world outcomes. Ward recently published an article in the Journal of Global Security Studies (read it now before it goes behind the paywall) arguing in favor of less theory and more prediction in the fields of international relations and security studies.

[W]e need less theory because most theory is an attempt to rescue or adapt extant theory. We need more predictions in order to keep track of how well we understand the world around us. They will tell us how good our theories are and where we need better explanations.

As Ward explained,

[P]rediction is deeply embedded in the philosophy of science… The argument is that if you can develop models that provide an understanding—without a teleology of why things happen—you should be able to generate predictions that will not only be accurate, but may also be useful in a larger societal context.

Ward argued that “until very recently, most of this thread of work in security studies had been lost, or if not lost, at least abandoned.” The reason for this was the existence of a longstanding epistemological disagreement: “Many social scientists see a sharp distinction between explanation on the one hand and prediction on the other. Indeed, this distinction is often sharp enough that it is argued that doing one of these things cuts you out of doing the other.”

For the most part, Ward asserted, the theorists have won out over the empiricists.

[M]any scholars (but few others) will tell you that we need more theory. Doubtless they are right. Few of them really mean “theory” in the sense that I reserve for the term. Few of them mean “theory” in the sense of analytical narratives. Many of them mean “detailed, plausible stories” about how stuff occurs.

In light of the uncomfortable conclusion that more detailed, plausible stories about how stuff occurs does not actually yield more insight, Ward has adopted a decidedly contrarian stance.

I am here to suggest that less is more. Thus, let me be the first to call for less theory in security studies. We should winnow the many, many such “theories” that occupy the world of security studies.

Instead, we need more predictions.

He went on to detail his argument.

We need these predictions for four reasons. First, we need these predictions to help us make relevant statements about the world around us. We also need these predictions to help us throw out the bad “theories” that continue to flourish. These predictions will help drive our research into new areas, away from moribund approaches that have been followed for many decades. Finally, and perhaps most important, predictions will force us to keep on track.

But making predictions is only part of the process. Tracking them and accounting for their accuracy is the vital corollary to improving both accuracy and theory. As Ward pointed out, “One reason that many hate predictions is that talking heads make many predictions in the media, but few of them ever keep track of how well they are doing.” Most, in fact, are wrong; few are held accountable for it.

Of course, the use of empirical methods to predict the outcomes of future events animated much of Trevor N. Dupuy’s approach to historical analysis and is at the heart of what The Dupuy Institute carries on doing today. Both have made well-documented predictions that have also been remarkably accurate. More about those in the next post.

Are They Channeling Trevor Dupuy?

TrevorCoverShot

Continuing the RAND description of their hex boardgame:

Ground unit combat strengths were based on a systematic scoring of individual weapons, from tanks and artillery down to light machine guns, which were then aggregated according to the tables of organization and equipment for the various classes of NATO and Russian units. Overall unit scores were adjusted to account for differences in training, sustainment, and other factors not otherwise captured. Air unit combat strengths were derived from the results of offline engagement, mission, and campaign-level modeling.

This looks like some kind of firepower or combat power score, or perhaps Trevor Dupuy’s OLIs (Operational Lethality Indexes). As they say “systematic scoring” one wonders what system they used. Know of only one scoring system that is systematic (meaning the OLIs, which are based upon formulae). The subject is probably best summarized in Dr. James Taylor’s article on “Consistent Scoring of Weapons and Aggregation of Forces:” http://www.dupuyinstitute.org/pdf/v2n2.pdf. This is the same James Taylor who wrote the definitive two-volume work on Lanchester equations.

I do note with interest the adjustment for “differences in training, sustainment, and other factors.” That is always good to see.

Also noted:

Full documentation of the gaming platform will be forthcoming in a subsequent report.

Look forward to reading it.