Tag strategic studies

Measuring the Effects of Combat in Cities, Phase II – part 2

There was actually supposed to be a part 2 to this Phase II contract, which was analysis of urban combat at the army-level based upon 50 operations, of which a half-dozen would include significant urban terrain. This effort was not funded.

On the other hand, the quantitative analysis of battles of Kharkov only took up the first 41 pages of the report. A significant part of the rest of the report was a more detailed analysis and case study of the three fights over Kharkov in February, March and August of 1943. Kharkov was a large city, according to the January 1939 census, it has a population of 1,344,200, although a Soviet-era encyclopedia gives the pre-war population as 840,000. We never were able to figure out why there was a discrepancy. The whole area was populated with many villages. The January 1939 gives Kharkov Oblast (region) a population of 1,209,496. This is in addition to the city, so the region had a total population of 2,552,686. Soviet-era sources state that when the city was liberated in August 1943, the remaining population was only 190,000. Kharkov was a much larger city than any of the others ones covered in Phase I effort (except for Paris, but the liberation of that city was hardly a major urban battle).

The report then does a day-by-day review of the urban fighting in Kharkov. Doing a book or two on the battles of Kharkov is on my short list of books to write, as I have already done a lot of the research. We do have daily logistical expenditures of the SS Panzer Corps for February and March (tons of ammo fired, gasoline used and diesel used). In March when the SS Panzer Corps re-took Kharkov, we noted that the daily average for the four days of urban combat from 12 to 15 March was 97.25 tons of ammunition, 92 cubic meters of gasoline and 10 cubic meters of diesel. For the previous five days (7-11 March) the daily average was 93.20 tons of ammunition, 145 cubic meters of gasoline and 9 cubic meters of diesel. Thus it does not produce a lot of support for the idea that–as has sometimes been expressed (for example in RAND’s earlier reports on the subject)–that ammunition and other supplies will be consumed at a higher rate in urban operations.

We do observe from the three battles of Kharkov that (page 95):

There is no question that the most important lesson found in the three battles of Kharkov is that one should just bypass cities rather than attack them. The Phase I study also points out that the attacker is usually aware that faster progress can be made outside the urban terrain, and that the tendency is to weight one or both flanks and not bother to attack the city until it is enveloped. This is indeed what happened in two of the three cases at Kharkov and was also the order given by the Fourth Panzer Army that was violated by the SS Panzer Corps in March.

One must also note that since this study began the United States invaded Iraq and conducted operations in some major urban areas, albeit against somewhat desultory and ineffective opposition. In the southern part of Iraq the two major port cities Umm Qasar and Basra were first enveloped before any forces were sent in to clear them. In the case of Baghdad, it could have been enveloped if sufficient forces were available. As it was, it was not seriously defended. The recent operations in Iraq again confirmed that observations made in the two phases of this study.

P.S. The picture is of Kharkov in 1942, when it was under German occupation.

Measuring the Effects of Combat in Cities, Phase II – part 1

Our first urban warfare report that we did had a big impact. It clearly showed that the intensity of urban warfare was not what some of the “experts” out there were claiming. In particular, it called into question some of the claims being made by RAND. But, the report was based upon Aachen, Cherbourg, and a collection of mop-up operations along the Channel Coast. Although this was a good starting point because of the ease of research and availability of data, we did not feel that this was a fully representative collection of cases. We also did not feel that it was based upon enough cases, although we had already assembled more cases than most “experts” were using. We therefore convinced CAA (Center for Army Analysis) to fund a similar effort for the Eastern Front in World War II.

For this second phase, we again assembled a collection of Eastern Front urban warfare engagements in our DLEDB (Division-level Engagement Data Base) and compared it to Eastern Front non-urban engagements. We had, of course, a considerable collection of non-urban engagements already assembled from the Battle of Kursk in July 1943. We therefore needed a good urban engagement nearby. Kharkov is the nearest major city to where these non-urban engagements occurred and it was fought over three times in 1943. It was taken by the Red Army in February, it was retaken by the German Army in March, and it was taken again by the Red Army in August. Many of the units involved were the same units involved in the Battle of Kursk. This was a good close match. It has the additional advantage that both sides were at times on the offense.

Furthermore, Kharkov was a big city. At the time it was the fourth biggest city in the Soviet Union, being bigger than Stalingrad (as measured by pre-war population). A picture of its Red Square in March 1943, after the Germans retook it, is above.

We did have good German records for 1943 and we were able to get access to Soviet division-level records from February, March and August from the Soviet military archives in Podolsk. Therefore, we were able to assembled all the engagements based upon the unit records of both sides. No secondary sources were used, and those that were available were incomplete, usually one-sided, sometimes biased and often riddled with factual errors.

So, we ended up with 51 urban and conurban engagements from the fighting around Kharkov, along with 65 non-urban engagements from Kursk (we have more now).

The Phase II effort was completed on 30 June 2003. The conclusions of Phase II (pages 40-41) were similar to Phase I:

.Phase II Conclusions:

  1. Mission Accomplishment: This [Phase I] conclusion was further supported. The data does show a tendency for urban engagements not to generate penetrations.
  2. Casualty Rates: This [Phase I] conclusion was further supported. If urban combat influenced the casualty rate, it appears that it resulted in a reduction of the attacker casualty rate and a more favorable casualty exchange ratio compared to nonurban warfare. There still appears to be no basis to the claim that urban combat is significantly more intense with regards to casualties than is nonurban warfare.
  3. Advance Rates: There is no strong evidence of a reduction in the advance rates in urban terrain in the Eastern Front data. TDI still stands by its original conclusion that the average advance rate in urban combat should be one-half to one-third that of nonurban combat.
  4. Linear Density: Again, there is little evidence that the presence of urban terrain results in a higher linear density of troops, but unlike the ETO data, the data did not show a tendency to trend in that direction.
  5. Armor Losses: This conclusion was further supported (Phase I conclusion was: Overall, it appears that the loss of armor in urban terrain is the same as or less than that found in nonurban terrain, and in some cases is significantly lower.)
  6. Force Ratios: The conclusion was further supported (Phase I conclusion was: Urban combat did not significantly influence the Force Ratio required to achieve success or effectively conduct combat operations).
  7. Duration of Combat: Nothing could be determined from an analysis of the data regarding the Duration of Combat (Time) in urban versus nonurban terrain.

There is a part 2 to this effort that I will pick up in a later post.

Measuring The Effects Of Combat In Cities, Phase I

“Catalina Kid,” a M4 medium tank of Company C, 745th Tank Battalion, U.S. Army, drives through the entrance of the Aachen-Rothe Erde railroad station during the fighting around the city viaduct on Oct. 20, 1944. [Courtesy of First Division Museum/Daily Herald]

In 2002, TDI submitted a report to the U.S. Army Center for Army Analysis (CAA) on the first phase of a study examining the effects of combat in cities, or what was then called “military operations on urbanized terrain,” or MOUT. This first phase of a series of studies on urban warfare focused on the impact of urban terrain on division-level engagements and army-level operations, based on data drawn from TDI’s DuWar database suite.

This included engagements in France during 1944 including the Channel and Brittany port cities of Brest, Boulogne, Le Havre, Calais, and Cherbourg, as well as Paris, and the extended series of battles in and around Aachen in 1944. These were then compared to data on fighting in contrasting non-urban terrain in Western Europe in 1944-45.

The conclusions of Phase I of that study (pp. 85-86) were as follows:

The Effect of Urban Terrain on Outcome

The data appears to support a null hypothesis, that is, that the urban terrain had no significantly measurable influence on the outcome of battle.

The Effect of Urban Terrain on Casualties

Overall, any way the data is sectioned, the attacker casualties in the urban engagements are less than in the non-urban engagements and the casualty exchange ratio favors the attacker as well. Because of the selection of the data, there is some question whether these observations can be extended beyond this data, but it does not provide much support to the notion that urban combat is a more intense environment than non-urban combat.

The Effect of Urban Terrain on Advance Rates

It would appear that one of the primary effects of urban terrain is that it slows opposed advance rates. One can conclude that the average advance rate in urban combat should be one-half to one-third that of non-urban combat.

The Effect of Urban Terrain on Force Density

Overall, there is little evidence that combat operations in urban terrain result in a higher linear density of troops, although the data does seem to trend in that direction.

The Effect of Urban Terrain on Armor

Overall, it appears that armor losses in urban terrain are the same as, or lower than armor losses in non-urban terrain. And in some cases it appears that armor losses are significantly lower in urban than non-urban terrain.

The Effect of Urban Terrain on Force Ratios

Urban terrain did not significantly influence the force ratio required to achieve success or effectively conduct combat operations.

The Effect of Urban Terrain on Stress in Combat

Overall, it appears that urban terrain was no more stressful a combat environment during actual combat operations than was non-urban terrain.

The Effect of Urban Terrain on Logistics

Overall, the evidence appears to be that the expenditure of artillery ammunition in urban operations was not greater than that in non-urban operations. In the two cases where exact comparisons could be made, the average expenditure rates were about one-third to one-quarter the average expenditure rates expected for an attack posture in the European Theater of Operations as a whole.

The evidence regarding the expenditure of other types of ammunition is less conclusive, but again does not appear to be significantly greater than the expenditures in non-urban terrain. Expenditures of specialized ordnance may have been higher, but the total weight expended was a minor fraction of that for all of the ammunition expended.

There is no evidence that the expenditure of other consumable items (rations, water or POL) was significantly different in urban as opposed to non-urban combat.

The Effect of Urban Combat on Time Requirements

It was impossible to draw significant conclusions from the data set as a whole. However, in the five significant urban operations that were carefully studied, the maximum length of time required to secure the urban area was twelve days in the case of Aachen, followed by six days in the case of Brest. But the other operations all required little more than a day to complete (Cherbourg, Boulogne and Calais).

However, since it was found that advance rates in urban combat were significantly reduced, then it is obvious that these two effects (advance rates and time) are interrelated. It does appear that the primary impact of urban combat is to slow the tempo of operations.

This in turn leads to a hypothetical construct, where the reduced tempo of urban operations (reduced casualties, reduced opposed advance rates and increased time) compared to non-urban operations, results in two possible scenarios.

The first is if the urban area is bounded by non-urban terrain. In this case the urban area will tend to be enveloped during combat, since the pace of battle in the non-urban terrain is quicker. Thus, the urban battle becomes more a mopping-up operation, as it historically has usually been, rather than a full-fledged battle.

The alternate scenario is that created by an urban area that cannot be enveloped and must therefore be directly attacked. This may be caused by geography, as in a city on an island or peninsula, by operational requirements, as in the case of Cherbourg, Brest and the Channel Ports, or by political requirements, as in the case of Stalingrad, Suez City and Grozny.

Of course these last three cases are also those usually included as examples of combat in urban terrain that resulted in high casualty rates. However, all three of them had significant political requirements that influenced the nature, tempo and even the simple necessity of conducting the operation. And, in the case of Stalingrad and Suez City, significant geographical limitations effected the operations as well. These may well be better used to quantify the impact of political agendas on casualties, rather than to quantify the effects of urban terrain on casualties.

The effects of urban terrain at the operational level, and the effect of urban terrain on the tempo of operations, will be further addressed in Phase II of this study.

Air Combat And Technology

Any model of air combat needs to address the effect of weapons on the opposing forces.  In the Dupuy Air Combat Model (DACM), this was rifled bullets fired from machine guns, as well as small caliber cannon in the 20-30 millimeter (mm) class.  Such was the state of air combat in World War II.  This page is an excellent, in-depth analysis of the fighter guns and cannon.  Of course, technology has effects beyond firepower.  One of the most notable technologies to go into active use during World War II was radar, contributing to the effectiveness of the Royal Air Force (RAF), successfully holding off the Wehrmacht’s Luftwaffe in the Battle of Britain.

Since that time, driven by “great power competition”, technology continues to advance the art of warfare in the air.  This happened in several notable stages during the Cold War, and was on display in subsequent contemporary conflicts when client or proxy states fought on behalf of the great powers.  Examples include well-known conflicts, such as the Korean and Vietnam conflicts, but also the conflicts between the Arabs and Israelis.  In the Korean War, archives now illustrate than Russian pilots secretly flew alongside North Korean and Chinese pilots against the allied forces.

Stages in technology are often characterized by generation.  Many of the features that are associated with the generations are driven by the Cold War arms race, and the back and forth development cycles and innovation cycles by the aircraft designers.  This was evident in comments by Aviation Week’s Bill Sweetman, remarking that the Jas-39 Grippen is actually a sixth generation fighter, based upon the alternative focus on maintainability, operability from short runways / austere airbases (or roadways!), the focus on cost reduction, but most importantly, software: “The reason that the JAS 39E may earn a Gen 6 tag is that it has been designed with these issues in mind. Software comes first: The new hardware runs Mission System 21 software, the latest roughly biennial release in the series that started with the JAS 39A/B.”

Upon close inspection of the DACM parameters, we can observe a few important data elements and metadata definitions: avionics (aka software & hardware), and sensor performance.  Those two are about data and information.  A concise method to assign values to these parameters is needed.  The U.S. Air Force (USAF) Air Combat Command (ACC) has used the generation of fighters as a proxy for this in the past, at least at a notional level:

[Source: 5th Generation Fighters, Lt Gen Hawk Carlisle, USAF ACC]

The Fleet Series game that has been reviewed in previous posts has a different method.  The Air-to-Air Combat Resolution Table does not seem to resonate well, as the damage effects are imposed against either one side or the other.  This does not jive with the stated concerns of the USAF, which has been worried about an exchange in which both Red and Blue forces are destroyed or eliminated in a mutual fashion, with a more or less one-for-one exchange ratio.

The Beyond Visual Range (BVR) version, named Long Range Air-to-Air (LRAA) combat in Asian Fleet, is a better model of this, in which each side rolls a die to determine the effect of long range missiles, and each side may take losses on non-stealthy units, as the stealthy units are immune to damage at BVR.

One important factor that the Fleet Series combat process does resolve is a solid determination of which side “holds” the airspace, and this is capable of using other support aircraft, such as AWACS, tankers, reconnaissance, etc.  Part of this determination is the relative morale of the opposing forces.  These effects have been clearly evident in air campaigns such as the strategic bombing campaign on Germany and Japan in the latter portion of World War II.

Dealing with this conundrum, I decided to relax by watching some dogfight videos on YouTube, Dogfights Greatest Air Battles, and this was rather entertaining, it included a series of engagements in aerial combat, taken from the exploits of American aces over the course of major wars:

  1. Eddie Rickenbacker, flying a Spad 13 in World War I,
  2. Clarence Emil “Bud” Anderson, flying a P-51B “Old Crow” in European skies during World War II, flying 67 missions in P-51Ds, 35 missions in F-80s and 121 missions in F-86s. He wrote “No Guts, No Glory,” a how to manual with lots of graphics of named maneuvers like the “Scissors.”
  3. Frederick Corbin “Boots” Blesse, flying a F-86 Sabre in “MiG Alley” in North Korea close to the Chinese border,
  4. Several engagements and interviews of aces from the Vietnam War:
    1. Steve Ritchie, who said “Surprise is a key element.” Previously discussed.
    2. Robin Olds – a triple ace in both WWII (P-38 and P-51) and Vietnam (F-4), and the mastermind of Operation Bolo, a fantastic application of deception.
    3. Randy “Duke” Cunningham and William P “Irish” Discol, flying an F-4 Phantom, “Showtime 100”, and up against North Vietnamese MiG-17s.

An interesting paraphrase by Cunningham of Manfred von Richthofen, the Red Baron’s statement: “When he sees the enemy, he attacks and kills, everything else is rubbish.”  What Richthofen said (according to skygod.com), was “The duty of the fighter pilot is to patrol his area of the sky, and shoot down any enemy fighters in that area. Anything else is rubbish.” Richtofen would not let members of his Staffel strafe troops in the trenches.

The list above is a great reference, and it got me to consider an alternative form of generation, including the earlier wars, and the experiences gained in those wars.  Indeed, we can press on in time to include the combat performance of the US and Allied militaries in the first Gulf War, 1990, as previously discussed.

There was a reference to the principles of aerial combat, such as the Dicta Boelcke:

  1. Secure the benefits of aerial combat (speed, altitude, numerical superiority, position) before attacking. Always attack from the sun.
  2. If you start the attack, bring it to an end.
  3. Fire the machine gun up close and only if you are sure to target your opponent.
  4. Do not lose sight of the enemy.
  5. In any form of attack, an approach to the opponent from behind is required.
  6. If the enemy attacks you in a dive, do not try to dodge the attack, but turn to the attacker.
  7. If you are above the enemy lines, always keep your own retreat in mind.
  8. For squadrons: In principle attack only in groups of four to six. If the fight breaks up in noisy single battles, make sure that not many comrades pounce on an opponent.

Appendix A – my own attempt to classify the generations of jet aircraft, in an attempt to rationalize the numerous schemes … until I decided that it was a fool’s errand:

  • Generation Zero:
    • World War II, 1948 Arab Israeli conflict
    • Blue: Spitfire, P-51 Mustang,
    • Red: Bf-109, FW-190, Mitsubishi Zero/George
    • Propeller engines, machine guns & cannons
  • First Generation:
    • Korean War, China & Taiwan conflicts
    • Blue: F-86 Sabre,
    • Red: MiG-15, Me-262?
    • Jet engines, swept wings, machine guns & cannons, early air-to-air missiles
  • Second Generation –
    • 1967 and Cuban Missile Crisis
    • Blue: F-100, F-102, F-104, F-5, F-8
    • Grey: Mirage III, Mirage F1
    • Red: MiG-19, MiG-21
    • Multi-mach speeds, improved air-to-air missiles, but largely within-visual range (WVR), early radar warning receivers (RWR), early countermeasures.
  • Third Generation:
    • 1973 Arab Israeli Wars, Vietnam War
    • Blue: F-4 Phantom, F-111 Ardvark, F-106?
    • Grey: Mirage III
    • Red: MiG-23, MiG-25, Su-15
    • Look-down/Shoot-down capability, radar-guided missiles, Beyond Visual Range (BVR), Identification Friend or Foe (IFF), all-aspect infrared missiles.
  • Fourth Generation:
    • 1980’s Cold War, 1990 Gulf War, 1982 Lebanon, 1980-88 Iran-Iraq War
    • Blue: F-15 Eagle, F-16 Viper, F-14 Tomcat, F/A-18 Hornet
    • Grey: Mirage 2000
    • Red: MiG-29, MiG-31, Su-27/30
  • Fourth Plus Generation:
    • 2003 Gulf War, 2011 Libiya
    • Blue: F/A-18E/F Super Hornet, F-15 improved (F-15E, F-15I, F-15SG, F-15SK…)
    • Grey: Eurofighter Typhoon, Rafale
    • Red: Su-35S
  • Fifth Generation:
    • Marketing term used by aircraft producers
    • Blue: Adanced Tactical Fighter (ATF) = F-22 Raptor, Joint Strike Fighter (JSF) = F-35 Lightening II
    • Grey: Grippen?
    • Red: PAK-FA Su-57, J-20
  • Sixth Generation – the current frontier
    • Blue: Next Generation Air Dominance (NGAD) program, UAS ?
    • Red: ?
    • Grey: Two seat, Twin tail “drone-herder”?

Scoring Weapons And Aggregation In Trevor Dupuy’s Combat Models

[The article below is reprinted from the October 1997 edition of The International TNDM Newsletter.]

Consistent Scoring of Weapons and Aggregation of Forces:
The Cornerstone of Dupuy’s Quantitative Analysis of Historical Land Battles
by
James G. Taylor, PhD,
Dept. of Operations Research, Naval Postgraduate School

Introduction

Col. Trevor N. Dupuy was an American original, especially as regards the quantitative study of warfare. As with many prophets, he was not entirely appreciated in his own land, particularly its Military Operations Research (OR) community. However, after becoming rather familiar with the details of his mathematical modeling of ground combat based on historical data, I became aware of the basic scientific soundness of his approach. Unfortunately, his documentation of methodology was not always accepted by others, many of whom appeared to confuse lack of mathematical sophistication in his documentation with lack of scientific validity of his basic methodology.

The purpose of this brief paper is to review the salient points of Dupuy’s methodology from a system’s perspective, i.e., to view his methodology as a system, functioning as an organic whole to capture the essence of past combat experience (with an eye towards extrapolation into the future). The advantage of this perspective is that it immediately leads one to the conclusion that if one wants to use some functional relationship derived from Dupuy’s work, then one should use his methodologies for scoring weapons, aggregating forces, and adjusting for operational circumstances; since this consistency is the only guarantee of being able to reproduce historical results and to project them into the future.

Implications (of this system’s perspective on Dupuy’s work) for current DOD models will be discussed. In particular, the Military OR community has developed quantitative methods for imputing values to weapon systems based on their attrition capability against opposing forces and force interactions.[1] One such approach is the so-called antipotential-potential method[2] used in TACWAR[3] to score weapons. However, one should not expect such scores to provide valid casualty estimates when combined with historically derived functional relationships such as the so-called ATLAS casualty-rate curves[4] used in TACWAR, because a different “yard-stick” (i.e. measuring system for estimating the relative combat potential of opposing forces) was used to develop such a curve.

Overview of Dupuy’s Approach

This section briefly outlines the salient features of Dupuy’s approach to the quantitative analysis and modeling of ground combat as embodied in his Tactical Numerical Deterministic Model (TNDM) and its predecessor the Quantified Judgment Model (QJM). The interested reader can find details in Dupuy [1979] (see also Dupuy [1985][5], [1987], [1990]). Here we will view Dupuy’s methodology from a system approach, which seeks to discern its various components and their interactions and to view these components as an organic whole. Essentially Dupuy’s approach involves the development of functional relationships from historical combat data (see Fig. 1) and then using these functional relationships to model future combat (see Fig, 2).

At the heart of Dupuy’s method is the investigation of historical battles and comparing the relationship of inputs (as quantified by relative combat power, denoted as Pa/Pd for that of the attacker relative to that of the defender in Fig. l)(e.g. see Dupuy [1979, pp. 59-64]) to outputs (as quantified by extent of mission accomplishment, casualty effectiveness, and territorial effectiveness; see Fig. 2) (e.g. see Dupuy [1979, pp. 47-50]), The salient point is that within this scheme, the main input[6] (i.e. relative combat power) to a historical battle is a derived quantity. It is computed from formulas that involve three essential aspects: (1) the scoring of weapons (e.g, see Dupuy [1979, Chapter 2 and also Appendix A]), (2) aggregation methodology for a force (e.g. see Dupuy [1979, pp. 43-46 and 202-203]), and (3) situational-adjustment methodology for determining the relative combat power of opposing forces (e.g. see Dupuy [1979, pp. 46-47 and 203-204]). In the force-aggregation step the effects on weapons of Dupuy’s environmental variables and one operational variable (air superiority) are considered[7], while in the situation-adjustment step the effects on forces of his behavioral variables[8] (aggregated into a single factor called the relative combat effectiveness value (CEV)) and also the other operational variables are considered (Dupuy [1987, pp. 86-89])

Figure 1.

Moreover, any functional relationships developed by Dupuy depend (unless shown otherwise) on his computational system for derived quantities, namely OLls, force strengths, and relative combat power. Thus, Dupuy’s results depend in an essential manner on his overall computational system described immediately above. Consequently, any such functional relationship (e.g. casualty-rate curve) directly or indirectly derivative from Dupuy‘s work should still use his computational methodology for determination of independent-variable values.

Fig l also reveals another important aspect of Dupuy’s work, the development of reliable data on historical battles, Military judgment plays an essential role in this development of such historical data for a variety of reasons. Dupuy was essentially the only source of new secondary historical data developed from primary sources (see McQuie [1970] for further details). These primary sources are well known to be both incomplete and inconsistent, so that military judgment must be used to fill in the many gaps and reconcile observed inconsistencies. Moreover, military judgment also generates the working hypotheses for model development (e.g. identification of significant variables).

At the heart of Dupuy’s quantitative investigation of historical battles and subsequent model development is his own weapons-scoring methodology, which slowly evolved out of study efforts by the Historical Evaluation Research Organization (HERO) and its successor organizations (cf. HERO [1967] and compare with Dupuy [1979]). Early HERO [1967, pp. 7-8] work revealed that what one would today call weapons scores developed by other organizations were so poorly documented that HERO had to create its own methodology for developing the relative lethality of weapons, which eventually evolved into Dupuy’s Operational Lethality Indices (OLIs). Dupuy realized that his method was arbitrary (as indeed is its counterpart, called the operational definition, in formal scientific work), but felt that this would be ameliorated if the weapons-scoring methodology be consistently applied to historical battles. Unfortunately, this point is not clearly stated in Dupuy’s formal writings, although it was clearly (and compellingly) made by him in numerous briefings that this author heard over the years.

Figure 2.

In other words, from a system’s perspective, the functional relationships developed by Colonel Dupuy are part of his analysis system that includes this weapons-scoring methodology consistently applied (see Fig. l again). The derived functional relationships do not stand alone (unless further empirical analysis shows them to hold for any weapons-scoring methodology), but function in concert with computational procedures. Another essential part of this system is Dupuy‘s aggregation methodology, which combines numbers, environmental circumstances, and weapons scores to compute the strength (S) of a military force. A key innovation by Colonel Dupuy [1979, pp. 202- 203] was to use a nonlinear (more precisely, a piecewise-linear) model for certain elements of force strength. This innovation precluded the occurrence of military absurdities such as air firepower being fully substitutable for ground firepower, antitank weapons being fully effective when armor targets are lacking, etc‘ The final part of this computational system is Dupuy’s situational-adjustment methodology, which combines the effects of operational circumstances with force strengths to determine relative combat power, e.g. Pa/Pd.

To recapitulate, the determination of an Operational Lethality Index (OLI) for a weapon involves the combination of weapon lethality, quantified in terms of a Theoretical Lethality Index (TLI) (e.g. see Dupuy [1987, p. 84]), and troop dispersion[9] (e.g. see Dupuy [1987, pp. 84- 85]). Weapons scores (i.e. the OLIs) are then combined with numbers (own side and enemy) and combat- environment factors to yield force strength. Six[10] different categories of weapons are aggregated, with nonlinear (i.e. piecewise-linear) models being used for the following three categories of weapons: antitank, air defense, and air firepower (i.e. c1ose—air support). Operational, e.g. mobility, posture, surprise, etc. (Dupuy [1987, p. 87]), and behavioral variables (quantified as a relative combat effectiveness value (CEV)) are then applied to force strength to determine a side’s combat-power potential.

Requirement for Consistent Scoring of Weapons, Force Aggregation, and Situational Adjustment for Operational Circumstances

The salient point to be gleaned from Fig.1 and 2 is that the same (or at least consistent) weapons—scoring, aggregation, and situational—adjustment methodologies be used for both developing functional relationships and then playing them to model future combat. The corresponding computational methods function as a system (organic whole) for determining relative combat power, e.g. Pa/Pd. For the development of functional relationships from historical data, a force ratio (relative combat power of the two opposing sides, e.g. attacker’s combat power divided by that of the defender, Pa/Pd is computed (i.e. it is a derived quantity) as the independent variable, with observed combat outcome being the dependent variable. Thus, as discussed above, this force ratio depends on the methodologies for scoring weapons, aggregating force strengths, and adjusting a force’s combat power for the operational circumstances of the engagement. It is a priori not clear that different scoring, aggregation, and situational-adjustment methodologies will lead to similar derived values. If such different computational procedures were to be used, these derived values should be recomputed and the corresponding functional relationships rederived and replotted.

However, users of the Tactical Numerical Deterministic Model (TNDM) (or for that matter, its predecessor, the Quantified Judgment Model (QJM)) need not worry about this point because it was apparently meticulously observed by Colonel Dupuy in all his work. However, portions of his work have found their way into a surprisingly large number of DOD models (usually not explicitly acknowledged), but the context and range of validity of historical results have been largely ignored by others. The need for recalibration of the historical data and corresponding functional relationships has not been considered in applying Dupuy’s results for some important current DOD models.

Implications for Current DOD Models

A number of important current DOD models (namely, TACWAR and JICM discussed below) make use of some of Dupuy’s historical results without recalibrating functional relationships such as loss rates and rates of advance as a function of some force ratio (e.g. Pa/Pd). As discussed above, it is not clear that such a procedure will capture the essence of past combat experience. Moreover, in calculating losses, Dupuy first determines personnel losses (expressed as a percent loss of personnel strength, i.e., number of combatants on a side) and then calculates equipment losses as a function of this casualty rate (e.g., see Dupuy [1971, pp. 219-223], also [1990, Chapters 5 through 7][11]). These latter functional relationships are apparently not observed in the models discussed below. In fact, only Dupuy (going back to Dupuy [1979][12] takes personnel losses to depend on a force ratio and other pertinent variables, with materiel losses being taken as derivative from this casualty rate.

For example, TACWAR determines personnel losses[13] by computing a force ratio and then consulting an appropriate casualty-rate curve (referred to as empirical data), much in the same fashion as ATLAS did[14]. However, such a force ratio is computed using a linear model with weapon values determined by the so-called antipotential-potential method[15]. Unfortunately, this procedure may not be consistent with how the empirical data (i.e. the casualty-rate curves) was developed. Further research is required to demonstrate that valid casualty estimates are obtained when different weapon scoring, aggregation, and situational-adjustment methodologies are used to develop casualty-rate curves from historical data and to use them to assess losses in aggregated combat models. Furthermore, TACWAR does not use Dupuy’s model for equipment losses (see above), although it does purport, as just noted above, to use “historical data” (e.g., see Kerlin et al. [1975, p. 22]) to compute personnel losses as a function (among other things) of a force ratio (given by a linear relationship), involving close air support values in a way never used by Dupuy. Although their force-ratio determination methodology does have logical and mathematical merit, it is not the way that the historical data was developed.

Moreover, RAND (Allen [1992]) has more recently developed what is called the situational force scoring (SFS) methodology for calculating force ratios in large-scale, aggregated-force combat situations to determine loss and movement rates. Here, SFS refers essentially to a force- aggregation and situation-adjustment methodology, which has many conceptual elements in common with Dupuy‘s methodology (except, most notably, extensive testing against historical data, especially documentation of such efforts). This SFS was originally developed for RSAS[16] and is today used in JICM[17]. It also apparently uses a weapon-scoring system developed at RAND[18]. It purports (no documentation given [citation of unpublished work]) to be consistent with historical data (including the ATLAS casualty-rate curves) (Allen [1992, p.41]), but again no consideration is given to recalibration of historical results for different weapon scoring, force-aggregation, and situational-adjustment methodologies. SFS emphasizes adjusting force strengths according to operational circumstances (the “situation”) of the engagement (including surprise), with many innovative ideas (but in some major ways has little connection with previous work of others[19]). The resulting model contains many more details than historical combat data would support. It also is methodology that differs in many essential ways from that used previously by any investigator. In particular, it is doubtful that it develops force ratios in a manner consistent with Dupuy’s work.

Final Comments

Use of (sophisticated) mathematics for modeling past historical combat (and extrapolating it into the future for planning purposes) is no reason for ignoring Dupuy’s work. One would think that the current Military OR community would try to understand Dupuy’s work before trying to improve and extend it. In particular, Colonel Dupuy’s various computational procedures (including constants) must be considered as an organic whole (i.e. a system) supporting the development of functional relationships. If one ignores this computational system and simply tries to use some isolated aspect, the result may be interesting and even logically sound, but it probably lacks any scientific validity.

REFERENCES

P. Allen, “Situational Force Scoring: Accounting for Combined Arms Effects in Aggregate Combat Models,” N-3423-NA, The RAND Corporation, Santa Monica, CA, 1992.

L. B. Anderson, “A Briefing on Anti-Potential Potential (The Eigen-value Method for Computing Weapon Values), WP-2, Project 23-31, Institute for Defense Analyses, Arlington, VA, March 1974.

B. W. Bennett, et al, “RSAS 4.6 Summary,” N-3534-NA, The RAND Corporation, Santa Monica, CA, 1992.

B. W. Bennett, A. M. Bullock, D. B. Fox, C. M. Jones, J. Schrader, R. Weissler, and B. A. Wilson, “JICM 1.0 Summary,” MR-383-NA, The RAND Corporation, Santa Monica, CA, 1994.

P. K. Davis and J. A. Winnefeld, “The RAND Strategic Assessment Center: An Overview and Interim Conclusions About Utility and Development Options,” R-2945-DNA, The RAND Corporation, Santa Monica, CA, March 1983.

T.N, Dupuy, Numbers. Predictions and War: Using History to Evaluate Combat Factors and Predict the Outcome of Battles, The Bobbs-Merrill Company, Indianapolis/New York, 1979,

T.N. Dupuy, Numbers Predictions and War, Revised Edition, HERO Books, Fairfax, VA 1985.

T.N. Dupuy, Understanding War: History and Theory of Combat, Paragon House Publishers, New York, 1987.

T.N. Dupuy, Attrition: Forecasting Battle Casualties and Equipment Losses in Modem War, HERO Books, Fairfax, VA, 1990.

General Research Corporation (GRC), “A Hierarchy of Combat Analysis Models,” McLean, VA, January 1973.

Historical Evaluation and Research Organization (HERO), “Average Casualty Rates for War Games, Based on Historical Data,” 3 Volumes in 1, Dunn Loring, VA, February 1967.

E. P. Kerlin and R. H. Cole, “ATLAS: A Tactical, Logistical, and Air Simulation: Documentation and User’s Guide,” RAC-TP-338, Research Analysis Corporation, McLean, VA, April 1969 (AD 850 355).

E.P. Kerlin, L.A. Schmidt, A.J. Rolfe, M.J. Hutzler, and D,L. Moody, “The IDA Tactical Warfare Model: A Theater-Level Model of Conventional, Nuclear, and Chemical Warfare, Volume II- Detailed Description” R-21 1, Institute for Defense Analyses, Arlington, VA, October 1975 (AD B009 692L).

R. McQuie, “Military History and Mathematical Analysis,” Military Review 50, No, 5, 8-17 (1970).

S.M. Robinson, “Shadow Prices for Measures of Effectiveness, I: Linear Model,” Operations Research 41, 518-535 (1993).

J.G. Taylor, Lanchester Models of Warfare. Vols, I & II. Operations Research Society of America, Alexandria, VA, 1983. (a)

J.G. Taylor, “A Lanchester-Type Aggregated-Force Model of Conventional Ground Combat,” Naval Research Logistics Quarterly 30, 237-260 (1983). (b)

NOTES

[1] For example, see Taylor [1983a, Section 7.18], which contains a number of examples. The basic references given there may be more accessible through Robinson [I993].

[2] This term was apparently coined by L.B. Anderson [I974] (see also Kerlin et al. [1975, Chapter I, Section D.3]).

[3] The Tactical Warfare (TACWAR) model is a theater-level, joint-warfare, computer-based combat model that is currently used for decision support by the Joint Staff and essentially all CINC staffs. It was originally developed by the Institute for Defense Analyses in the mid-1970s (see Kerlin et al. [1975]), originally referred to as TACNUC, which has been continually upgraded until (and including) the present day.

[4] For example, see Kerlin and Cole [1969], GRC [1973, Fig. 6-6], or Taylor [1983b, Fig. 5] (also Taylor [1983a, Section 7.13]).

[5] The only apparent difference between Dupuy [1979] and Dupuy [1985] is the addition of an appendix (Appendix C “Modified Quantified Judgment Analysis of the Bekaa Valley Battle”) to the end of the latter (pp. 241-251). Hence, the page content is apparently the same for these two books for pp. 1-239.

[6] Technically speaking, one also has the engagement type and possibly several other descriptors (denoted in Fig. 1 as reduced list of operational circumstances) as other inputs to a historical battle.

[7] In Dupuy [1979, e.g. pp. 43-46] only environmental variables are mentioned, although basically the same formulas underlie both Dupuy [1979] and Dupuy [1987]. For simplicity, Fig. 1 and 2 follow this usage and employ the term “environmental circumstances.”

[8] In Dupuy [1979, e.g. pp. 46-47] only operational variables are mentioned, although basically the same formulas underlie both Dupuy [1979] and Dupuy [1987]. For simplicity, Fig. 1 and 2 follow this usage and employ the term “operational circumstances.”

[9] Chris Lawrence has kindly brought to my attention that since the same value for troop dispersion from an historical period (e.g. see Dupuy [1987, p. 84]) is used for both the attacker and also the defender, troop dispersion does not actually affect the determination of relative combat power PM/Pd.

[10] Eight different weapon types are considered, with three being classified as infantry weapons (e.g. see Dupuy [1979, pp, 43-44], [1981 pp. 85-86]).

[11] Chris Lawrence has kindly informed me that Dupuy‘s work on relating equipment losses to personnel losses goes back to the early 1970s and even earlier (e.g. see HERO [1966]). Moreover, Dupuy‘s [1992] book Future Wars gives some additional empirical evidence concerning the dependence of equipment losses on casualty rates.

[12] But actually going back much earlier as pointed out in the previous footnote.

[13] See Kerlin et al. [1975, Chapter I, Section D.l].

[14] See Footnote 4 above.

[15] See Kerlin et al. [1975, Chapter I, Section D.3]; see also Footnotes 1 and 2 above.

[16] The RAND Strategy Assessment System (RSAS) is a multi-theater aggregated combat model developed at RAND in the early l980s (for further details see Davis and Winnefeld [1983] and Bennett et al. [1992]). It evolved into the Joint Integrated Contingency Model (JICM), which is a post-Cold War redesign of the RSAS (starting in FY92).

[17] The Joint Integrated Contingency Model (JICM) is a game-structured computer-based combat model of major regional contingencies and higher-level conflicts, covering strategic mobility, regional conventional and nuclear warfare in multiple theaters, naval warfare, and strategic nuclear warfare (for further details, see Bennett et al. [1994]).

[18] RAND apparently replaced one weapon-scoring system by another (e.g. see Allen [1992, pp. 9, l5, and 87-89]) without making any other changes in their SFS System.

[19] For example, both Dupuy’s early HERO work (e.g. see Dupuy [1967]), reworks of these results by the Research Analysis Corporation (RAC) (e.g. see RAC [1973, Fig. 6-6]), and Dupuy’s later work (e.g. see Dupuy [1979]) considered daily fractional casualties for the attacker and also for the defender as basic casualty-outcome descriptors (see also Taylor [1983b]). However, RAND does not do this, but considers the defender’s loss rate and a casualty exchange ratio as being the basic casualty-production descriptors (Allen [1992, pp. 41-42]). The great value of using the former set of descriptors (i.e. attacker and defender fractional loss rates) is that not only is casualty assessment more straight forward (especially development of functional relationships from historical data) but also qualitative model behavior is readily deduced (see Taylor [1983b] for further details).

The Lanchester Equations and Historical Warfare

Allied force dispositions at the Battle of Anzio, on 1 February 1944. [U.S. Army/Wikipedia]

[The article below is reprinted from History, Numbers And War: A HERO Journal, Vol. 1, No. 1, Spring 1977, pp. 34-52]

The Lanchester Equations and Historical Warfare: An Analysis of Sixty World War II Land Engagements

By Janice B. Fain

Background and Objectives

The method by which combat losses are computed is one of the most critical parts of any combat model. The Lanchester equations, which state that a unit’s combat losses depend on the size of its opponent, are widely used for this purpose.

In addition to their use in complex dynamic simulations of warfare, the Lanchester equations have also sewed as simple mathematical models. In fact, during the last decade or so there has been an explosion of theoretical developments based on them. By now their variations and modifications are numerous, and “Lanchester theory” has become almost a separate branch of applied mathematics. However, compared with the effort devoted to theoretical developments, there has been relatively little empirical testing of the basic thesis that combat losses are related to force sizes.

One of the first empirical studies of the Lanchester equations was Engel’s classic work on the Iwo Jima campaign in which he found a reasonable fit between computed and actual U.S. casualties (Note 1). Later studies were somewhat less supportive (Notes 2 and 3), but an investigation of Korean war battles showed that, when the simulated combat units were constrained to follow the tactics of their historical counterparts, casualties during combat could be predicted to within 1 to 13 percent (Note 4).

Taken together, these various studies suggest that, while the Lanchester equations may be poor descriptors of large battles extending over periods during which the forces were not constantly in combat, they may be adequate for predicting losses while the forces are actually engaged in fighting. The purpose of the work reported here is to investigate 60 carefully selected World War II engagements. Since the durations of these battles were short (typically two to three days), it was expected that the Lanchester equations would show a closer fit than was found in studies of larger battles. In particular, one of the objectives was to repeat, in part, Willard’s work on battles of the historical past (Note 3).

The Data Base

Probably the most nearly complete and accurate collection of combat data is the data on World War II compiled by the Historical Evaluation and Research Organization (HERO). From their data HERO analysts selected, for quantitative analysis, the following 60 engagements from four major Italian campaigns:

Salerno, 9-18 Sep 1943, 9 engagements

Volturno, 12 Oct-8 Dec 1943, 20 engagements

Anzio, 22 Jan-29 Feb 1944, 11 engagements

Rome, 14 May-4 June 1944, 20 engagements

The complete data base is described in a HERO report (Note 5). The work described here is not the first analysis of these data. Statistical analyses of weapon effectiveness and the testing of a combat model (the Quantified Judgment Method, QJM) have been carried out (Note 6). The work discussed here examines these engagements from the viewpoint of the Lanchester equations to consider the question: “Are casualties during combat related to the numbers of men in the opposing forces?”

The variables chosen for this analysis are shown in Table 1. The “winners” of the engagements were specified by HERO on the basis of casualties suffered, distance advanced, and subjective estimates of the percentage of the commander’s objective achieved. Variable 12, the Combat Power Ratio, is based on the Operational Lethality Indices (OLI) of the units (Note 7).

The general characteristics of the engagements are briefly described. Of the 60, there were 19 attacks by British forces, 28 by U.S. forces, and 13 by German forces. The attacker was successful in 34 cases; the defender, in 23; and the outcomes of 3 were ambiguous. With respect to terrain, 19 engagements occurred in flat terrain; 24 in rolling, or intermediate, terrain; and 17 in rugged, or difficult, terrain. Clear weather prevailed in 40 cases; 13 engagements were fought in light or intermittent rain; and 7 in medium or heavy rain. There were 28 spring and summer engagements and 32 fall and winter engagements.

Comparison of World War II Engagements With Historical Battles

Since one purpose of this work is to repeat, in part, Willard’s analysis, comparison of these World War II engagements with the historical battles (1618-1905) studied by him will be useful. Table 2 shows a comparison of the distribution of battles by type. Willard’s cases were divided into two categories: I. meeting engagements, and II. sieges, attacks on forts, and similar operations. HERO’s World War II engagements were divided into four types based on the posture of the defender: 1. delay, 2. hasty defense, 3. prepared position, and 4. fortified position. If postures 1 and 2 are considered very roughly equivalent to Willard’s category I, then in both data sets the division into the two gross categories is approximately even.

The distribution of engagements across force ratios, given in Table 3, indicated some differences. Willard’s engagements tend to cluster at the lower end of the scale (1-2) and at the higher end (4 and above), while the majority of the World War II engagements were found in mid-range (1.5 – 4) (Note 8). The frequency with which the numerically inferior force achieved victory is shown in Table 4. It is seen that in neither data set are force ratios good predictors of success in battle (Note 9).

Table 3.

Results of the Analysis Willard’s Correlation Analysis

There are two forms of the Lanchester equations. One represents the case in which firing units on both sides know the locations of their opponents and can shift their fire to a new target when a “kill” is achieved. This leads to the “square” law where the loss rate is proportional to the opponent’s size. The second form represents that situation in which only the general location of the opponent is known. This leads to the “linear” law in which the loss rate is proportional to the product of both force sizes.

As Willard points out, large battles are made up of many smaller fights. Some of these obey one law while others obey the other, so that the overall result should be a combination of the two. Starting with a general formulation of Lanchester’s equations, where g is the exponent of the target unit’s size (that is, g is 0 for the square law and 1 for the linear law), he derives the following linear equation:

log (nc/mc) = log E + g log (mo/no) (1)

where nc and mc are the casualties, E is related to the exchange ratio, and mo and no are the initial force sizes. Linear regression produces a value for g. However, instead of lying between 0 and 1, as expected, the) g‘s range from -.27 to -.87, with the majority lying around -.5. (Willard obtains several values for g by dividing his data base in various ways—by force ratio, by casualty ratio, by historical period, and so forth.) A negative g value is unpleasant. As Willard notes:

Military theorists should be disconcerted to find g < 0, for in this range the results seem to imply that if the Lanchester formulation is valid, the casualty-producing power of troops increases as they suffer casualties (Note 3).

From his results, Willard concludes that his analysis does not justify the use of Lanchester equations in large-scale situations (Note 10).

Analysis of the World War II Engagements

Willard’s computations were repeated for the HERO data set. For these engagements, regression produced a value of -.594 for g (Note 11), in striking agreement with Willard’s results. Following his reasoning would lead to the conclusion that either the Lanchester equations do not represent these engagements, or that the casualty producing power of forces increases as their size decreases.

However, since the Lanchester equations are so convenient analytically and their use is so widespread, it appeared worthwhile to reconsider this conclusion. In deriving equation (1), Willard used binomial expansions in which he retained only the leading terms. It seemed possible that the poor results might he due, in part, to this approximation. If the first two terms of these expansions are retained, the following equation results:

log (nc/mc) = log E + log (Mo-mc)/(no-nc) (2)

Repeating this regression on the basis of this equation leads to g = -.413 (Note 12), hardly an improvement over the initial results.

A second attempt was made to salvage this approach. Starting with raw OLI scores (Note 7), HERO analysts have computed “combat potentials” for both sides in these engagements, taking into account the operational factors of posture, vulnerability, and mobility; environmental factors like weather, season, and terrain; and (when the record warrants) psychological factors like troop training, morale, and the quality of leadership. Replacing the factor (mo/no) in Equation (1) by the combat power ratio produces the result) g = .466 (Note 13).

While this is an apparent improvement in the value of g, it is achieved at the expense of somewhat distorting the Lanchester concept. It does preserve the functional form of the equations, but it requires a somewhat strange definition of “killing rates.”

Analysis Based on the Differential Lanchester Equations

Analysis of the type carried out by Willard appears to produce very poor results for these World War II engagements. Part of the reason for this is apparent from Figure 1, which shows the scatterplot of the dependent variable, log (nc/mc), against the independent variable, log (mo/no). It is clear that no straight line will fit these data very well, and one with a positive slope would not be much worse than the “best” line found by regression. To expect the exponent to account for the wide variation in these data seems unreasonable.

Here, a simpler approach will be taken. Rather than use the data to attempt to discriminate directly between the square and the linear laws, they will be used to estimate linear coefficients under each assumption in turn, starting with the differential formulation rather than the integrated equations used by Willard.

In their simplest differential form, the Lanchester equations may be written;

Square Law; dA/dt = -kdD and dD/dt = kaA (3)

Linear law: dA/dt = -k’dAD and dD/dt = k’aAD (4)

where

A(D) is the size of the attacker (defender)

dA/dt (dD/dt) is the attacker’s (defender’s) loss rate,

ka, k’a (kd, k’d) are the attacker’s (defender’s) killing rates

For this analysis, the day is taken as the basic time unit, and the loss rate per day is approximated by the casualties per day. Results of the linear regressions are given in Table 5. No conclusions should be drawn from the fact that the correlation coefficients are higher in the linear law case since this is expected for purely technical reasons (Note 14). A better picture of the relationships is again provided by the scatterplots in Figure 2. It is clear from these plots that, as in the case of the logarithmic forms, a single straight line will not fit the entire set of 60 engagements for either of the dependent variables.

To investigate ways in which the data set might profitably be subdivided for analysis, T-tests of the means of the dependent variable were made for several partitionings of the data set. The results, shown in Table 6, suggest that dividing the engagements by defense posture might prove worthwhile.

Results of the linear regressions by defense posture are shown in Table 7. For each posture, the equation that seemed to give a better fit to the data is underlined (Note 15). From this table, the following very tentative conclusions might be drawn:

  • In an attack on a fortified position, the attacker suffers casualties by the square law; the defender suffers casualties by the linear law. That is, the defender is aware of the attacker’s position, while the attacker knows only the general location of the defender. (This is similar to Deitchman’s guerrilla model. Note 16).
  • This situation is apparently reversed in the cases of attacks on prepared positions and hasty defenses.
  • Delaying situations seem to be treated better by the square law for both attacker and defender.

Table 8 summarizes the killing rates by defense posture. The defender has a much higher killing rate than the attacker (almost 3 to 1) in a fortified position. In a prepared position and hasty defense, the attacker appears to have the advantage. However, in a delaying action, the defender’s killing rate is again greater than the attacker’s (Note 17).

Figure 3 shows the scatterplots for these cases. Examination of these plots suggests that a tentative answer to the study question posed above might be: “Yes, casualties do appear to be related to the force sizes, but the relationship may not be a simple linear one.”

In several of these plots it appears that two or more functional forms may be involved. Consider, for example, the defender‘s casualties as a function of the attacker’s initial strength in the case of a hasty defense. This plot is repeated in Figure 4, where the points appear to fit the curves sketched there. It would appear that there are at least two, possibly three, separate relationships. Also on that plot, the individual engagements have been identified, and it is interesting to note that on the curve marked (1), five of the seven attacks were made by Germans—four of them from the Salerno campaign. It would appear from this that German attacks are associated with higher than average defender casualties for the attacking force size. Since there are so few data points, this cannot be more than a hint or interesting suggestion.

Future Research

This work suggests two conclusions that might have an impact on future lines of research on combat dynamics:

  • Tactics appear to be an important determinant of combat results. This conclusion, in itself, does not appear startling, at least not to the military. However, it does not always seem to have been the case that tactical questions have been considered seriously by analysts in their studies of the effects of varying force levels and force mixes.
  • Historical data of this type offer rich opportunities for studying the effects of tactics. For example, consideration of the narrative accounts of these battles might permit re-coding the engagements into a larger, more sensitive set of engagement categories. (It would, of course, then be highly desirable to add more engagements to the data set.)

While predictions of the future are always dangerous, I would nevertheless like to suggest what appears to be a possible trend. While military analysis of the past two decades has focused almost exclusively on the hardware of weapons systems, at least part of our future analysis will be devoted to the more behavioral aspects of combat.

Janice Bloom Fain, a Senior Associate of CACI, lnc., is a physicist whose special interests are in the applications of computer simulation techniques to industrial and military operations; she is the author of numerous reports and articles in this field. This paper was presented by Dr. Fain at the Military Operations Research Symposium at Fort Eustis, Virginia.

NOTES

[1.] J. H. Engel, “A Verification of Lanchester’s Law,” Operations Research 2, 163-171 (1954).

[2.] For example, see R. L. Helmbold, “Some Observations on the Use of Lanchester’s Theory for Prediction,” Operations Research 12, 778-781 (1964); H. K. Weiss, “Lanchester-Type Models of Warfare,” Proceedings of the First International Conference on Operational Research, 82-98, ORSA (1957); H. K. Weiss, “Combat Models and Historical Data; The U.S. Civil War,” Operations Research 14, 750-790 (1966).

[3.] D. Willard, “Lanchester as a Force in History: An Analysis of Land Battles of the Years 1618-1905,” RAC-TD-74, Research Analysis Corporation (1962). what appears to be a possible trend. While military analysis of the past two decades has focused almost exclusively on the hardware of weapons systems, at least part of our future analysis will be devoted to the more behavioral aspects of combat.

[4.] The method of computing the killing rates forced a fit at the beginning and end of the battles. See W. Fain, J. B. Fain, L. Feldman, and S. Simon, “Validation of Combat Models Against Historical Data,” Professional Paper No. 27, Center for Naval Analyses, Arlington, Virginia (1970).

[5.] HERO, “A Study of the Relationship of Tactical Air Support Operations to Land Combat, Appendix B, Historical Data Base.” Historical Evaluation and Research Organization, report prepared for the Defense Operational Analysis Establishment, U.K.T.S.D., Contract D-4052 (1971).

[6.] T. N. Dupuy, The Quantified Judgment Method of Analysis of Historical Combat Data, HERO Monograph, (January 1973); HERO, “Statistical Inference in Analysis in Combat,” Annex F, Historical Data Research on Tactical Air Operations, prepared for Headquarters USAF, Assistant Chief of Staff for Studies and Analysis, Contract No. F-44620-70-C-0058 (1972).

[7.] The Operational Lethality Index (OLI) is a measure of weapon effectiveness developed by HERO.

[8.] Since Willard’s data did not indicate which side was the attacker, his force ratio is defined to be (larger force/smaller force). The HERO force ratio is (attacker/defender).

[9.] Since the criteria for success may have been rather different for the two sets of battles, this comparison may not be very meaningful.

[10.] This work includes more complex analysis in which the possibility that the two forces may be engaging in different types of combat is considered, leading to the use of two exponents rather than the single one, Stochastic combat processes are also treated.

[11.] Correlation coefficient = -.262;

Intercept = .00115; slope = -.594.

[12.] Correlation coefficient = -.184;

Intercept = .0539; slope = -,413.

[13.] Correlation coefficient = .303;

Intercept = -.638; slope = .466.

[14.] Correlation coefficients for the linear law are inflated with respect to the square law since the independent variable is a product of force sizes and, thus, has a higher variance than the single force size unit in the square law case.

[15.] This is a subjective judgment based on the following considerations Since the correlation coefficient is inflated for the linear law, when it is lower the square law case is chosen. When the linear law correlation coefficient is higher, the case with the intercept closer to 0 is chosen.

[16.] S. J. Deitchman, “A Lanchester Model of Guerrilla Warfare,” Operations Research 10, 818-812 (1962).

[17.] As pointed out by Mr. Alan Washburn, who prepared a critique on this paper, when comparing numerical values of the square law and linear law killing rates, the differences in units must be considered. (See footnotes to Table 7).

What Is A Breakpoint?

French retreat from Russia in 1812 by Illarion Mikhailovich Pryanishnikov (1812) [Wikipedia]

After discussing with Chris the series of recent posts on the subject of breakpoints, it seemed appropriate to provide a better definition of exactly what a breakpoint is.

Dorothy Kneeland Clark was the first to define the notion of a breakpoint in her study, Casualties as a Measure of the Loss of Combat Effectiveness of an Infantry Battalion (Operations Research Office, The Johns Hopkins University: Baltimore, 1954). She found it was not quite as clear-cut as it seemed and the working definition she arrived at was based on discussions and the specific combat outcomes she found in her data set [pp 9-12].

DETERMINATION OF BREAKPOINT

The following definitions were developed out of many discussions. A unit is considered to have lost its combat effectiveness when it is unable to carry out its mission. The onset of this inability constitutes a breakpoint. A unit’s mission is the objective assigned in the current operations order or any other instructional directive, written or verbal. The objective may be, for example, to attack in order to take certain positions, or to defend certain positions. 

How does one determine when a unit is unable to carry out its mission? The obvious indication is a change in operational directive: the unit is ordered to stop short of its original goal, to hold instead of attack, to withdraw instead of hold. But one or more extraneous elements may cause the issue of such orders: 

(1) Some other unit taking part in the operation may have lost its combat effectiveness, and its predicament may force changes, in the tactical plan. For example the inability of one infantry battalion to take a hill may require that the two adjoining battalions be stopped to prevent exposing their flanks by advancing beyond it. 

(2) A unit may have been assigned an objective on the basis of a G-2 estimate of enemy weakness which, as the action proceeds, proves to have been over-optimistic. The operations plan may, therefore, be revised before the unit has carried out its orders to the point of losing combat effectiveness. 

(3) The commanding officer, for reasons quite apart from the tactical attrition, may change his operations plan. For instance, General Ridgway in May 1951 was obliged to cancel his plans for a major offensive north of the 38th parallel in Korea in obedience to top level orders dictated by political considerations. 

(4) Even if the supposed combat effectiveness of the unit is the determining factor in the issuance of a revised operations order, a serious difficulty in evaluating the situation remains. The commanding officer’s decision is necessarily made on the basis of information available to him plus his estimate of his unit’s capacities. Either or both of these bases may be faulty. The order may belatedly recognize a collapse which has in factor occurred hours earlier, or a commanding officer may withdraw a unit which could hold for a much longer time. 

It was usually not hard to discover when changes in orders resulted from conditions such as the first three listed above, but it proved extremely difficult to distinguish between revised orders based on a correct appraisal of the unit’s combat effectiveness and those issued in error. It was concluded that the formal order for a change in mission cannot be taken as a definitive indication of the breakpoint of a unit. It seemed necessary to go one step farther and search the records to learn what a given battalion did regardless of provisions in formal orders… 

CATEGORIES OF BREAKPOINTS SELECTED 

In the engagements studied the following categories of breakpoint were finally selected: 

Category of Breakpoint 

No. Analyzed 

I. Attack [Symbol] rapid reorganization [Symbol] attack 

9 

II. Attack [Symbol] defense (no longer able to attack without a few days of recuperation and reinforcement 

21 

III. Defense [Symbol] withdrawal by order to a secondary line 

13 

IV. Defense [Symbol] collapse 

5 

Disorganization and panic were taken as unquestionable evidence of loss of combat effectiveness. It appeared, however, that there were distinct degrees of magnitude in these experiences. In addition to the expected breakpoints at attack [Symbol] defense and defense [Symbol] collapse, a further category, I, seemed to be indicated to include situations in which an attacking battalion was ‘pinned down” or forced to withdraw in partial disorder but was able to reorganize in 4 to 24 hours and continue attacking successfully. 

Category II includes (a) situations in which an attacking battalion was ordered into the defensive after severe fighting or temporary panic; (b) situations in which a battalion, after attacking successfully, failed to gain ground although still attempting to advance and was finally ordered into defense, the breakpoint being taken as occurring at the end of successful advance. In other words, the evident inability of the unit to fulfill its mission was used as the criterion for the breakpoint whether orders did or did not recognize its inability. Battalions after experiencing such a breakpoint might be able to recuperate in a few days to the point of renewing successful attack or might be able to continue for some time in defense. 

The sample of breakpoints coming under category IV, defense [Symbol] collapse, proved to be very small (5) and unduly weighted in that four of the examples came from the same engagement. It was, therefore, discarded as probably not representative of the universe of category IV breakpoints,* and another category (III) was added: situations in which battalions on the defense were ordered withdrawn to a quieter sector. Because only those instances were included in which the withdrawal orders appeared to have been dictated by the condition of the unit itself, it is believed that casualty levels for this category can be regarded as but slightly lower than those associated with defense [Symbol] collapse. 

In both categories II and III, “‘defense” represents an active situation in which the enemy is attacking aggressively. 

* It had been expected that breakpoints in this category would be associated with very high losses. Such did not prove to be the case. In whatever way the data were approached, most of the casualty averages were only slightly higher than those associated with category II (attack [Symbol] defense), although the spread in data was wider. It is believed that factors other than casualties, such as bad weather, difficult terrain, and heavy enemy artillery fire undoubtedly played major roles in bringing about the collapse in the four units taking part in the same engagement. Furthermore, the casualty figures for the four units themselves is in question because, as the situation deteriorated, many of the men developed severe cases of trench foot and combat exhaustion, but were not evacuated, as they would have been in a less desperate situation, and did not appear in the casualty records until they had made their way to the rear after their units had collapsed.

In 1987-1988, Trevor Dupuy and colleagues at Data Memory Systems, Inc. (DMSi), Janice Fain, Rich Anderson, Gay Hammerman, and Chuck Hawkins sought to create a broader, more generally applicable definition for breakpoints for the study, Forced Changes of Combat Posture (DMSi, Fairfax, VA, 1988) [pp. I-2-3]

The combat posture of a military force is the immediate intention of its commander and troops toward the opposing enemy force, together with the preparations and deployment to carry out that intention. The chief combat postures are attack, defend, delay, and withdraw.

A change in combat posture (or posture change) is a shift from one posture to another, as, for example, from defend to attack or defend to withdraw. A posture change can be either voluntary or forced. 

A forced posture change (FPC) is a change in combat posture by a military unit that is brought about, directly or indirectly, by enemy action. Forced posture changes are characteristically and almost always changes to a less aggressive posture. The most usual FPCs are from attack to defend and from defend to withdraw (or retrograde movement). A change from withdraw to combat ineffectiveness is also possible. 

Breakpoint is a term sometimes used as synonymous with forced posture change, and sometimes used to mean the collapse of a unit into ineffectiveness or rout. The latter meaning is probably more common in general usage, while forced posture change is the more precise term for the subject of this study. However, for brevity and convenience, and because this study has been known informally since its inception as the “Breakpoints” study, the term breakpoint is sometimes used in this report. When it is used, it is synonymous with forced posture change.

Hopefully this will help clarify the previous discussions of breakpoints on the blog.

Human Factors In Combat: Syrian Strike Edition

Missile fire lit up the Damascus sky last week as the U.S. and allies launched an attack on chemical weapons sites. [Hassan Ammar, AP/USA Today]

Even as pundits and wonks debate the political and strategic impact of the 14 April combined U.S., British, and French cruise missile strike on Assad regime chemical warfare targets in Syria, it has become clear that effort was a notable tactical success.

Despite ample warning that the strike was coming, the Syrian regime’s Russian-made S-200 surface-to-air missile defense system failed to shoot down a single incoming missile. The U.S. Defense Department claimed that all 105 cruise missiles fired struck their targets. It also reported that the Syrians fired 40 interceptor missiles but nearly all launched after the incoming cruise missiles had already struck their targets.

Although cruise missiles are difficult to track and engage even with fully modernized air defense systems, the dismal performance of the Syrian network was a surprise to many analysts given the wary respect paid to it by U.S. military leaders in the recent past. Although the S-200 dates from the 1960s, many surmise an erosion in the combat effectiveness of the personnel manning the system is the real culprit.

[A] lack of training, command and control and other human factors are probably responsible for the failure, analysts said.

“It’s not just about the physical capability of the air defense system,” said David Deptula, a retired, three-star Air Force general. “It’s about the people who are operating the system.”

The Syrian regime has become dependent upon assistance from Russia and Iran to train, equip, and maintain its military forces. Russian forces in Syria have deployed the more sophisticated S-400 air defense system to protect their air and naval bases, which reportedly tracked but did not engage the cruise missile strike. The Assad regime is also believed to field the Russian-made Pantsir missile and air-defense artillery system, but it likely was not deployed near enough to the targeted facilities to help.

Despite the pervasive role technology plays in modern warfare, the human element remains the most important factor in determining combat effectiveness.

U.S. Army Invests In Revitalizing Long Range Precision Fires Capabilities

U.S. Marines from the The 11th MEU fire their M777 Lightweight 155mm Howitzer during Exercise Alligator Dagger, Dec. 18, 2016. (U.S. Marine Corps/Lance Cpl. Zachery C. Laning/Military.com)

In 2016, Michael Jacobson and Robert H. Scales amplified a warning that after years of neglect during the counterinsurgency war in Iraq and Afghanistan, the U.S. was falling behind potential adversaries in artillery and long range precision fires capabilities. The U.S. Army had already taken note of the performance of Russian artillery in Ukraine, particularly the strike at Zelenopillya in 2014.

Since then, the U.S. Army and Marine Corps have started working on a new Multi-Domain Battle concept aimed at countering the anti-access/area denial (A2/AD) capabilities of potential foes. In 2017, U.S. Army Chief of Staff General Mark Milley made rapid improvement in long range precision fires capabilities the top priority for the service’s modernization effort. It currently aims to field new field artillery, rocket, and missile weapons capable of striking at distances from 70 to 500 kilometers – double the existing ranges – within five years.

The value of ground-based long-range precision fires has been demonstrated recently by the effectiveness of U.S. artillery support, particularly U.S. Army and Marine Corps 155mm howitzers, for Iraqi security forces in retaking Mosul, Syrian Democratic Forces assaulting Raqaa, and in protection of Syrian Kurds being attacked by Russian mercenaries and Syrian regime forces.

According to Army historian Luke O’Brian, the Fiscal Year 2019 Defense budget includes funds to buy 28,737 XM1156 Precision Guided Kit (PGK) 155mm howitzer munitions, which includes replacements for the 6,269 rounds expended during Operation INHERENT RESOLVE. O’Brian also notes that the Army will also buy 2,162 M982 Excalibur 155mm rounds in 2019 and several hundred each in following years.

In addition, in an effort to reduce the dependence on potentially vulnerable Global Positioning System (GPS) satellite networks for precision fires capabilities, the Army has awarded a contract to BAE Systems to develop Precision Guided Kit-Modernization (PGK-M) rounds with internal navigational capacity.

While the numbers appear large at first glance, data on U.S. artillery expenditures in Operation DESERT STORM and IRAQI FREEDOM (also via Luke O’Brian) shows just how much the volume of long-range fires has changed just since 1991. For the U.S. at least, precision fires have indeed replaced mass fires on the battlefield.

Breakpoints in U.S. Army Doctrine

U.S. Army prisoners of war captured by German forces during the Battle of the Bulge in 1944. [Wikipedia]

One of the least studied aspects of combat is battle termination. Why do units in combat stop attacking or defending? Shifts in combat posture (attack, defend, delay, withdrawal) are usually voluntary, directed by a commander, but they can also be involuntary, as a result of direct or indirect enemy action. Why do involuntary changes in combat posture, known as breakpoints, occur?

As Chris pointed out in a previous post, the topic of breakpoints has only been addressed by two known studies since 1954. Most existing military combat models and wargames address breakpoints in at least a cursory way, usually through some calculation based on personnel casualties. Both of the breakpoints studies suggest that involuntary changes in posture are seldom related to casualties alone, however.

Current U.S. Army doctrine addresses changes in combat posture through discussions of culmination points in the attack, and transitions from attack to defense, defense to counterattack, and defense to retrograde. But these all pertain to voluntary changes, not breakpoints.

Army doctrinal literature has little to say about breakpoints, either in the context of friendly forces or potential enemy combatants. The little it does say relates to the effects of fire on enemy forces and is based on personnel and material attrition.

According to ADRP 1-02 Terms and Military Symbols, an enemy combat unit is considered suppressed after suffering 3% personnel casualties or material losses, neutralized by 10% losses, and destroyed upon sustaining 30% losses. The sources and methodology for deriving these figures is unknown, although these specific terms and numbers have been a part of Army doctrine for decades.

The joint U.S. Army and U.S. Marine Corps vision of future land combat foresees battlefields that are highly lethal and demanding on human endurance. How will such a future operational environment affect combat performance? Past experience undoubtedly offers useful insights but there seems to be little interest in seeking out such knowledge.

Trevor Dupuy criticized the U.S. military in the 1980s for its lack of understanding of the phenomenon of suppression and other effects of fire on the battlefield, and its seeming disinterest in studying it. Not much appears to have changed since then.