Tag Multi-Domain Battle

TDI Friday Read: U.S. Airpower

[Image by Geopol Intelligence]

This weekend’s edition of TDI’s Friday Read is a collection of posts on the current state of U.S. airpower by guest contributor Geoffery Clark. The same factors changing the character of land warfare are changing the way conflict will be waged in the air. Clark’s posts highlight some of the way these changes are influencing current and future U.S. airpower plans and concepts.

F-22 vs. F-35: Thoughts On Fifth Generation Fighters

The F-35 Is Not A Fighter

U.S. Armed Forces Vision For Future Air Warfare

The U.S. Navy and U.S. Air Force Debate Future Air Superiority

U.S. Marine Corps Concepts of Operation with the F-35B

The State of U.S. Air Force Air Power

Fifth Generation Deterrence

 

The Effects Of Dispersion On Combat

[The article below is reprinted from the December 1996 edition of The International TNDM Newsletter. A revised version appears in Christopher A. Lawrence, War by Numbers: Understanding Conventional Combat (Potomac Books, 2017), Chapter 13.]

The Effects of Dispersion on Combat
by Christopher A. Lawrence

The TNDM[1] does not play dispersion. But it is clear that dispersion has continued to increase over time, and this must have some effect on combat. This effect was identified by Trevor N. Dupuy in his various writings, starting with the Evolution of Weapons and Warfare. His graph in Understanding War of the battle casualties trends over time is presented here as Figure 1. As dispersion changes over time (dramatically), one would expect the casualties would change over time. I therefore went back to the Land Warfare Database (the 605 engagement version[2]) and proceeded to look at casualties over time and dispersion from every angle that l could.

l eventually realized that l was going to need some better definition of the time periods l was measuring to, as measuring by years scattered the data, measuring by century assembled the data in too gross a manner, and measuring by war left a confusing picture due to the number of small wars with only two or three battles in them in the Land Warfare Database. I eventually defined the wars into 14 categories, so I could fit them onto one readable graph:

To give some idea of how representative the battles listed in the LWDB were for covering the period, I have included a count of the number of battles listed in Michael Clodfelter’s two-volume book Warfare and Armed Conflict, 1618-1991. In the case of WWI, WWII and later, battles tend to be defined as a divisional-level engagement, and there were literally tens of thousands of those.

I then tested my data again looking at the 14 wars that I defined:

  • Average Strength by War (Figure 2)
  • Average Losses by War (Figure 3)
  • Percent Losses Per Day By War (Figure 4)a
  • Average People Per Kilometer By War (Figure 5)
  • Losses per Kilometer of Front by War (Figure 6)
  • Strength and Losses Per Kilometer of Front By War (Figure 7)
  • Ratio of Strength and Losses per Kilometer of Front by War (Figure 8)
  • Ratio of Strength and Loses per Kilometer of Front by Century (Figure 9)

A review of average strengths over time by century and by war showed no surprises (see Figure 2). Up through around 1900, battles were easy to define: they were one- to three-day affairs between clearly defined forces at a locale. The forces had a clear left flank and right flank that was not bounded by other friendly forces. After 1900 (and in a few cases before), warfare was fought on continuous fronts

with a ‘battle’ often being a large multi-corps operation. It is no longer clearly understood what is meant by a battle, as the forces, area covered, and duration can vary widely. For the LWDB, each battle was defined as the analyst wished. ln the case of WWI, there are a lot of very large battles which drive the average battle size up. ln the cases of the WWII, there are a lot of division-level battles, which bring the average down. In the case of the Arab-Israeli Wars, there are nothing but division and brigade-level battles, which bring the average down.

The interesting point to notice is that the average attacker strength in the 16th and 17th century is lower than the average defender strength. Later it is higher. This may be due to anomalies in our data selection.

Average loses by war (see Figure 3) suffers from the same battle definition problem.

Percent losses per day (see Figure 4) is a useful comparison through the end of the 19th Century. After that, the battles get longer and the definition of a duration of the battle is up to the analyst. Note the very dear and definite downward pattern of percent loses per day from the Napoleonic Wars through the Arab-Israeli Wars. Here is a very clear indication of the effects of dispersion. It would appear that from the 1600s to the 1800s the pattern was effectively constant and level, then declines in a very systematic pattern. This partially contradicts Trevor Dupuy’s writing and graphs (see Figure 1). It does appear that after this period of decline that the percent losses per day are being set at a new, much lower plateau. Percent losses per day by war is attached.

Looking at the actual subject of the dispersion of people (measured in people per kilometer of front) remained relatively constant from 1600 through the American Civil War (see Figure 5). Trevor Dupuy defined dispersion as the number of people in a box-like area. Unfortunately, l do not know how to measure that. lean clearly identify the left and right of a unit, but it is more difficult to tell how deep it is Furthermore, density of occupation of this box is far from uniform, with a very forward bias By the same token, fire delivered into this box is also not uniform, with a very forward bias. Therefore, l am quite comfortable measuring dispersion based upon unit frontage, more so than front multiplied by depth.

Note, when comparing the Napoleonic Wars to the American Civil War that the dispersion remains about the same. Yet, if you look at the average casualties (Figure 3) and the average percent casualties per day (Figure 4), it is clear that the rate of casualty accumulation is lower in the American Civil War (this again partially contradicts Dupuy‘s writings). There is no question that with the advent of the Minié ball, allowing for rapid-fire rifled muskets, the ability to deliver accurate firepower increased.

As you will also note, the average people per linear kilometer between WWI and WWII differs by a factor of a little over 1.5 to 1. Yet the actual difference in casualties (see Figure 4) is much greater. While one can just postulate that the difference is the change in dispersion squared (basically Dupuy‘s approach), this does not seem to explain the complete difference, especially the difference between the Napoleonic Wars and the Civil War.

lnstead of discussing dispersion, we should be discussing “casualty reduction efforts.” This basically consists of three elements:

  • Dispersion (D)
  • Increased engagement ranges (R)
  • More individual use of cover and concealment (C&C).

These three factors together result in the reduced chance to hit. They are also partially interrelated, as one cannot make more individual use of cover and concealment unless one is allowed to disperse. So, therefore. The need for cover and concealment increases the desire to disperse and the process of dispersing allows one to use more cover and concealment.

Command and control are integrated into this construct as being something that allows dispersion, and dispersion creates the need for better command control. Therefore, improved command and control in this construct does not operate as a force modifier, but enables a force to disperse.

Intelligence becomes more necessary as the opposing forces use cover and concealment and the ranges of engagement increase. By the same token, improved intelligence allows you to increase the range of engagement and forces the enemy to use better concealment.

This whole construct could be represented by the diagram at the top of the next page.

Now, I may have said the obvious here, but this construct is probably provable in each individual element, and the overall outcome is measurable. Each individual connection between these boxes may also be measurable.

Therefore, to measure the effects of reduced chance to hit, one would need to measure the following formula (assuming these formulae are close to being correct):

(K * ΔD) + (K * ΔC&C) + (K * ΔR) = H

(K * ΔC2) = ΔD

(K * ΔD) = ΔC&C

(K * ΔW) + (K * ΔI) = ΔR

K = a constant
Δ = the change in….. (alias “Delta”)
D = Dispersion
C&C = Cover & Concealment
R = Engagement Range
W = Weapon’s Characteristics
H = the chance to hit
C2 = Command and control
I = Intelligence or ability to observe

Also, certain actions lead to a desire for certain technological and system improvements. This includes the effect of increased dispersion leading to a need for better C2 and increased range leading to a need for better intelligence. I am not sure these are measurable.

I have also shown in the diagram how the enemy impacts upon this. There is also an interrelated mirror image of this construct for the other side.

I am focusing on this because l really want to come up with some means of measuring the effects of a “revolution in warfare.” The last 400 years of human history have given us more revolutionary inventions impacting war than we can reasonably expect to see in the next 100 years. In particular, I would like to measure the impact of increased weapon accuracy, improved intelligence, and improved C2 on combat.

For the purposes of the TNDM, I would very specifically like to work out an attrition multiplier for battles before WWII (and theoretically after WWII) based upon reduced chance to be hit (“dispersion”). For example, Dave Bongard is currently using an attrition multiplier of 4 for his WWI engagements that he is running for the battalion-level validation data base.[3] No one can point to a piece of paper saying this is the value that should be used. Dave picked this value based upon experience and familiarity with the period.

I have also attached Average Loses per Kilometer of Front by War (see Figure 6 above), and a summary chart showing the two on the same chart (see figure 7 above).

The values from these charts are:

The TNDM sets WWII dispersion factor at 3,000 (which l gather translates into 30,000 men per square kilometer). The above data shows a linear dispersion per kilometer of 2,992 men, so this number parallels Dupuy‘s figures.

The final chart I have included is the Ratio of Strength and Losses per Kilometer of Front by War (Figure 8). Each line on the bar graph measures the average ratio of strength over casualties for either the attacker or defender. Being a ratio, unusual outcomes resulted in some really unusually high ratios. I took the liberty of taking out six

data points because they appeared unusually lop-sided. Three of these points are from the English Civil War and were way out of line with everything else. These were the three Scottish battles where you had a small group of mostly sword-armed troops defeating a “modem” army. Also, Walcourt (1689), Front Royal (1862), and Calbritto (1943) were removed. L also have included the same chart, except by century (Figure 9).
Again, one sees a consistency in results in over 300+ years of war, in this case going all the way through WWI, then sees an entirely different pattern with WWII and the Arab-Israeli Wars

A very tentative set of conclusions from all this is:

  1. Dispersion has been relatively constant and driven by factors other than firepower from 1600-1815.
  2. Since the Napoleonic Wars, units have increasingly dispersed (found ways to reduce their chance to be hit) in response to increased lethality of weapons.
  3. As a result of this increased dispersion, casualties in a given space have declined.
  4. The ratio of this decline in casualties over area have been roughly proportional to the strength over an area from 1600 through WWI. Starting with WWII, it appears that people have dispersed faster than weapons lethality, and this trend has continued.
  5. In effect, people dispersed in direct relation to increased firepower from 1815 through 1920, and then after that time dispersed faster than the increase in lethality.
  6. It appears that since WWII, people have gone back to dispersing (reducing their chance to be hit) at the same rate that firepower is increasing.
  7. Effectively, there are four patterns of casualties in modem war:

Period 1 (1600 – 1815): Period of Stability

  • Short battles
  • Short frontages
  • High attrition per day
  • Constant dispersion
  • Dispersion decreasing slightly after late 1700s
  • Attrition decreasing slightly after mid-1700s.

Period 2 (1816 – 1905): Period of Adjustment

  • Longer battles
  • Longer frontages
  • Lower attrition per day
  • Increasing dispersion
  • Dispersion increasing slightly faster than lethality

Period 3 (1912 – 1920): Period of Transition

  • Long Battles
  • Continuous Frontages
  • Lower attrition per day
  • Increasing dispersion
  • Relative lethality per kilometer similar to past, but lower
  • Dispersion increasing slightly faster than lethality

Period 4 (1937 – present): Modern Warfare

  • Long Battles
  • Continuous Frontages
  • Low Attrition per day
  • High dispersion (perhaps constant?)
  • Relatively lethality per kilometer much lower than the past
  • Dispersion increased much faster than lethality going into the period.
  • Dispersion increased at the same rate as lethality within the period.

So the question is whether warfare of the next 50 years will see a new “period of adjustment,” where the rate of dispersion (and other factors) adjusts in direct proportion to increased lethality, or will there be a significant change in the nature of war?

Note that when l use the word “dispersion” above, l often mean “reduced chance to be hit,” which consists of dispersion, increased engagement ranges, and use of cover & concealment.

One of the reasons l wandered into this subject was to see if the TNDM can be used for predicting combat before WWII. l then spent the next few days attempting to find some correlation between dispersion and casualties. Using the data on historical dispersion provided above, l created a mathematical formulation and tested that against the actual historical data points, and could not get any type of fit.

I then locked at the length of battles over time, at one-day battles, and attempted to find a pattern. I could find none. I also looked at other permutations, but did not keep a record of my attempts. I then looked through the work done by Dean Hartley (Oakridge) with the LWDB and called Paul Davis (RAND) to see if there was anyone who had found any correlation between dispersion and casualties, and they had not noted any.

It became clear to me that if there is any such correlation, it is buried so deep in the data that it cannot be found by any casual search. I suspect that I can find a mathematical correlation between weapon lethality, reduced chance to hit (including dispersion), and casualties. This would require some improvement to the data, some systematic measure of weapons lethality, and some serious regression analysis. I unfortunately cannot pursue this at this time.

Finally, for reference, l have attached two charts showing the duration of the battles in the LWDB in days (Figure 10, Duration of Battles Over Time and Figure 11, A Count of the Duration of Battles by War).

NOTES

[1] The Tactical Numerical Deterministic Model, a combat model developed by Trevor Dupuy in 1990-1991 as the follow-up to his Quantified Judgement Model. Dr. James G. Taylor and Jose Perez also contributed to the TNDM’s development.

[2] TDI’s Land Warfare Database (LWDB) was a revised version of a database created by the Historical Evaluation Research Organization (HERO) for the then-U.S. Army Concepts and Analysis Agency (now known as the U.S. Army Center for Army Analysis (CAA)) in 1984. Since the original publication of this article, TDI expanded and revised the data into a suite of databases.

[3] This matter is discussed in Christopher A. Lawrence, “The Second Test of the TNDM Battalion-Level Validations: Predicting Casualties,” The International TNDM Newsletter, April 1997, pp. 40-50.

U.S. Army Updates Draft Multi-Domain Battle Operating Concept

The U.S. Army Training and Doctrine Command has released a revised draft version of its Multi-Domain Battle operating concept, titled “Multi-Domain Battle: Evolution of Combined Arms for the 21st Century, 2025-2040.” Clearly a work in progress, the document is listed as version 1.0, dated October 2017, and as a draft and not for implementation. Sydney J. Freeberg, Jr. has an excellent run-down on the revision at Breaking Defense.

The update is the result of the initial round of work between the U.S. Army and U.S. Air Force to redefine the scope of the multi-domain battlespace for the Joint Force. More work will be needed to refine the concept, but it shows remarkable cooperation in forging a common warfighting perspective between services long-noted for their independent thinking.

On a related note, Albert Palazzo, an Australian defense thinker and one of the early contributors to the Multi-Domain Battle concept, has published the first of a series of articles at The Strategy Bridge offering constructive criticism of the U.S. military’s approach to defining the concept. Palazzo warns that the U.S. may be over-emphasizing countering potential Russian and Chinese capabilities in its efforts and not enough on the broad general implications of long-range fires with global reach.

What difference can it make if those designing Multi-Domain Battle are acting on possibly the wrong threat diagnosis? Designing a solution for a misdiagnosed problem can result in the inculcation of a way of war unsuited for the wars of the future. One is reminded of the French Army during the interwar period. No one can accuse the French of not thinking seriously about war during these years, but, in the doctrine of the methodical battle, they got it wrong and misread the opportunities presented by mechanisation. There were many factors contributing to France’s defeat, but at their core was a misinterpretation of the art of the possible and a singular focus on a particular way of war. Shaping Multi-Domain Battle for the wrong problem may see the United States similarly sow the seeds for a military disaster that is avoidable.

He suggests that it would be wise for U.S. doctrine writers to take a more considered look at potential implications before venturing too far ahead with specific solutions.

Combat Readiness And The U.S. Army’s “Identity Crisis”

Servicemen of the U.S. Army’s 173rd Airborne Brigade Combat Team (standing) train Ukrainian National Guard members during a joint military exercise called “Fearless Guardian 2015,” at the International Peacekeeping and Security Center near the western village of Starychy, Ukraine, on May 7, 2015. [Newsweek]

Last week, Wesley Morgan reported in POLITICO about an internal readiness study recently conducted by the U.S. Army 173rd Airborne Infantry Brigade Combat Team. As U.S. European Command’s only airborne unit, the 173rd Airborne Brigade has been participating in exercises in the Baltic States and the Ukraine since 2014 to demonstrate the North Atlantic Treaty Organization’s (NATO) resolve to counter potential Russian aggression in Eastern Europe.

The experience the brigade gained working with Baltic and particularly Ukrainian military units that had engaged with Russian and Russian-backed Ukrainian Separatist forces has been sobering. Colonel Gregory Anderson, the 173rd Airborne Brigade commander, commissioned the study as a result. “The lessons we learned from our Ukrainian partners were substantial. It was a real eye-opener on the absolute need to look at ourselves critically,” he told POLITICO.

The study candidly assessed that the 173rd Airborne Brigade currently lacked “essential capabilities needed to accomplish its mission effectively and with decisive speed” against near-peer adversaries or sophisticated non-state actors. Among the capability gaps the study cited were

  • The lack of air defense and electronic warfare units and over-reliance on satellite communications and Global Positioning Systems (GPS) navigation systems;
  • simple countermeasures such as camouflage nets to hide vehicles from enemy helicopters or drones are “hard-to-find luxuries for tactical units”;
  • the urgent need to replace up-armored Humvees with the forthcoming Ground Mobility Vehicle, a much lighter-weight, more mobile truck; and
  • the likewise urgent need to field the projected Mobile Protected Firepower armored vehicle companies the U.S. Army is planning to add to each infantry brigade combat team.

The report also stressed the vulnerability of the brigade to demonstrated Russian electronic warfare capabilities, which would likely deprive it of GPS navigation and targeting and satellite communications in combat. While the brigade has been purchasing electronic warfare gear of its own from over-the-counter suppliers, it would need additional specialized personnel to use the equipment.

As analyst Adrian Bonenberger commented, “The report is framed as being about the 173rd, but it’s really about more than the 173rd. It’s about what the Army needs to do… If Russia uses electronic warfare to jam the brigade’s artillery, and its anti-tank weapons can’t penetrate any of the Russian armor, and they’re able to confuse and disrupt and quickly overwhelm those paratroopers, we could be in for a long war.”

While the report is a wake-up call with regard to the combat readiness in the short-term, it also pointedly demonstrates the complexity of the strategic “identity crisis” that faces the U.S. Army in general. Many of the 173rd Airborne Brigade’s current challenges can be traced directly to the previous decade and a half of deployments conducting wide area security missions during counterinsurgency operations in Iraq and Afghanistan. The brigade’s perceived shortcomings for combined arms maneuver missions are either logical adaptations to the demands of counterinsurgency warfare or capabilities that atrophied through disuse.

The Army’s specific lack of readiness to wage combined arms maneuver warfare against potential peer or near-peer opponents in Europe can be remedied given time and resourcing in the short-term. This will not solve the long-term strategic conundrum the Army faces in needing to be prepared to fight conventional and irregular conflicts at the same time, however. Unless the U.S. is willing to 1) increase defense spending to balance force structure to the demands of foreign and military policy objectives, or 2) realign foreign and military policy goals with the available force structure, it will have to resort to patching up short-term readiness issues as best as possible and continue to muddle through. Given the current state of U.S. domestic politics, muddling through will likely be the default option unless or until the consequences of doing so force a change.

Tanks With Frickin’ Laser Beams On Their Heads

Portent Of The Future: This Mobile High-Energy Laser-equipped Stryker was evaluated during the 2017 Maneuver Fires Integrated Experiment at Fort Sill, Oklahoma. The MEHEL can shoot a drone out of the sky using a 5kW laser. (Photo Credit: C. Todd Lopez)

As the U.S. Army ponders its Multi-Domain Battle concept for future warfare, it is also considering what types of weapons it will need to conduct it. Among these is a replacement for the venerable M1 Abrams Main Battle Tank (MBT), which is now 40 years old. Recent trends in combat are leading some to postulate a next-generation MBT that is lighter and more maneuverable, but equipped with a variety of new defensive capabilities to make them more survivable against modern anti-tank weapons. These include electronic jamming and anti-missile missiles, collectively referred to as Active Protection Systems, as well as unmanned turrets. Manned vehicles will be augmented with unmanned ground vehicles.The Army is also exploring new advanced composite armor and nanotechnology.

Also under consideration are replacements for the traditional MBT long gun, including high-power lasers and railguns. Some of these could be powered by hydrogen power cells and biofuels.

As the U.S. looks toward lighter armored vehicles, some countries appear to going in the other direction. Both Russia and Israel are developing beefed-up versions of existing vehicles designed specifically for fighting in urban environments.

The strategic demands on U.S. ground combat forces don’t allow for the luxury of fielding larger combat vehicles that complicate the challenge of rapid deployment to face global threats. Even as the historical trend toward increasing lethality and greater dispersion on the battlefield continues, the U.S. may have little choice other than to rely on technological innovation to balance the evolving capabilities of potential adversaries.

Trevor Dupuy on Military Innovation

In an article published by the Association of the U.S. Army last November that I missed on the first go around, U.S. Army Colonel Eric E. Aslakson and Lieutenant Colonel Richard T. Brown, (ret.) make the argument that “Staff colonels are the Army’s innovation center of gravity.”

The U.S. defense community has settled upon innovation as one of the key methods for overcoming the challenges posed by new technologies and strategies adapted by potential adversaries, as articulated in the Third Offset Strategy developed by the late Obama administration. It is becoming clear however, that a desire to innovate is not the same as actual innovation. Aslakson and Brown make the point that innovation is not simply technological development and identify what they believe is a crucial institutional component of military innovation in the U.S. Army.

Innovation is differentiated from other forms of change such as improvisation and adaptation by the scale, scope and impact of that value creation. Innovation is not about a new widget or process, but the decisive value created and the competitive advantage gained when that new widget or process is applied throughout the Army or joint force…

However, none of these inventions or activities can rise to the level of innovation unless there are skilled professionals within the Army who can convert these ideas into competitive advantage across the enterprise. That is the role of a colonel serving in a major command staff leadership assignment…

These leaders do not typically create the change. But they have the necessary institutional and operational expertise and experience, contacts, resources and risk tolerance to manage processes across the entire framework of doctrine, organization, training, materiel, leadership and education, personnel and facilities, converting invention into competitive advantage.

In his seminal book, The Evolution of Weapons and Warfare (Indianapolis, IN: The Bobbs-Merrill Company, Inc., 1980), Trevor Dupuy noted a pattern in the historical relationship between development of weapons of increasing lethality and their incorporation in warfare. He too noted that the crucial factor was not the technology itself, but the organizational approach to using it.

When a radically new weapon appears and is first adopted, it is inherently incongruous with existing weapons and doctrine. This is reflected in a number of ways; uncertainty and hesitation in coordination of the new weapon with earlier ones; inability to use it consistently, effectively, and flexibly in offensive action, which often leads to tactical stalemate; vulnerability of the weapon and of its users to hostile countermeasures; heavy losses incident to the employment of the new weapon, or in attempting to oppose it in combat. From this it is possible to establish the following criteria of assimilation:

  1. Confident employment of the weapon in accordance with a doctrine that assures its coordination with other weapons in a manner compatible with the characteristics of each.
  2. Consistently effective, flexible use of the weapon in offensive warfare, permitting full employment of the advantages of superior leadership and/or superior resources.
  3. Capability of dealing effectively with anticipated and unanticipated countermeasures.
  4. Sharp decline in casualties for those employing the weapon, often combined with a capability for inflicting disproportionately heavy losses on the enemy.

Based on his assessment of this historical pattern, Dupuy derived a set of preconditions necessary for a successful assimilation of new technology into warfare.

  1. An imaginative, knowledgeable leadership focused on military affairs, supported by extensive knowledge of, and competence in, the nature and background of the existing military system.
  2. Effective coordination of the nation’s economic, technological-scientific, and military resources.
    1. There must exist industrial or developmental research institutions, basic research institutions, military staffs and their supporting institutions, together with administrative arrangements for linking these with one another and with top decision-making echelons of government.
    2. These bodies must conduct their research, developmental, and testing activities according to mutually familiar methods so that their personnel can communicate, can be mutually supporting, and can evaluate each other’s results.
    3. The efforts of these institutions—in related matters—must be directed toward a common goal.
  3. Opportunity for battlefield experimentation as a basis for evaluation and analysis.

Does the U.S. defense establishment’s organizational and institutional approach to innovation meet these preconditions? Good question.

The U.S. Army’s Stryker Conundrum

Soldiers, of the 1st Stryker Brigade Combat Team, 4th Infantry Division, tactically move a Stryker over the Mojave Desert during Decisive Action Rotation 15-10 at the National Training Center on Fort Irwin, Calif., Sept. 24, 2015. The Stryker and other ground combat vehicles are undergoing a number of upgrades, according to officials. (Photo Credit: Sgt. William Howard)

As part of an initiative to modernize the U.S. Army in the late 1990s, then-Chief of Staff General Eric Shinseki articulated a need for combat units that were more mobile and rapidly deployable than the existing heavy armor brigades. This was the genesis of the Army’s medium-weight Stryker combat vehicle and the Stryker Brigade Combat Teams (SBCTs).

Since the Stryker’s introduction in 2002, SBCTs have participated successfully in U.S. expeditionary operations in Iraq and Afghanistan, validating for many the usefulness of Shinseki’s original medium-weight armor concept. However, changes in the strategic landscape and advances in technology and operational doctrine by potential adversaries are calling the medium armor concept back into question.

Medium armor faces the same conundrum that currently confronts the U.S. Army in general: should it optimize to conduct wide area security operations (which is the most likely type of future conflict) or combined arms maneuver (the most dangerous potential future conflict), or should it continue to hedge against strategic uncertainty by fielding a balanced, general purpose force which does a tolerable job of both, as it does now?

The Problem

In the current edition of Military Review, U.S. Army Captain Matthew D. Allgeyer presents an interesting critique of the Army’s existing medium-weight armor concept. He asserts that it is “is suffering from a lack of direction and focus…” Several improvements for the Stryker have been proposed based on perceptions of evolving Russian military capabilities, namely “a modern heavy-force threat supported by aviation assets.” The problem, according to Allgeyer, is that

The SBCT community wants all the positive aspects of a light force: lower cost, a small tooth-to-tail ratio, greater operational-level speed, etc. But, it also wants the ability to confront a heavy-armored force on its own terms without having to adopt the cost, support, and deployment time required by an armored force. Since these two ideas are mutuality exclusive, we have been forced to adopt a piecemeal response to shortcomings identified during training and training center rotations.

Even if the currently proposed improvements are adopted however, Allgeyer argues that updated Strykers would only provide the U.S. with a medium weight armor capability analogous to the 1960’s era Soviet motor-rifle regiment, a doctrinal step backward.

Allgeyer identifies the SBCT’s biggest weaknesses as a lack of firepower capable of successfully engaging enemy heavy armor and the absence of an organic air defense capability. Neither of these is a problem in wide area security missions such as peacekeeping or counterinsurgency, where deployability and mobility are priority considerations. However, both shortcomings are critical disadvantages in combined arms maneuver scenarios, particularly against Russian or Russian-equipped opposing forces.

Potential Solutions

These observations are not new. A 2001 TDI study of the historical effectiveness of lighter-weight armor pointed out its vulnerability to heavy armored forces, but also its utility in stability and contingency operations. The Russians long ago addressed these issues with their Bronetransporter (BTR)-equipped motor-rifle regiments by adding organic tank battalions to them, incorporating air defense platoons in each battalion, and upgunning the BTRs with 30mm cannons and anti-tank guided missiles (ATGMs).

The U.S. Army has similar solutions available. It has already sought to add 30mm cannons and TOW-2 ATGMs to the Styker. The Mobile Protected Firepower program that will provide a company of light-weight armored vehicles with high-caliber cannons to each infantry brigade combat team could easily be expanded to add a company or battalion of such vehicles to the SBCT. No proposals exist for improving air defense capabilities, but this too could be addressed.

Allgeyer agrees with the need for these improvements, but he is dissatisfied with the Army “simply reinventing on its own the wheel Russia made a long time ago.”  His proposed “solution is a radical restructuring of thought around the Stryker concept.” He advocates ditching the term “Stryker” in favor of the more generic “medium armor” to encourage doctrinal thinking about the concept instead of the platform. Medium armor advocates should accept the need for a combined arms solution to engaging adversary heavy forces and incorporate more joint training into their mission-essential task lists. The Army should also do a better job of analyzing foreign medium armor platforms and doctrine to see what may be appropriate for U.S. adoption.

Allgeyer’s proposals are certainly worthy, but they may not add up to the radical restructuring he seeks. Even if adopted, they are not likely to change the fundamental characteristics of medium armor that make it more suitable to the wide area security mission than to combined arms maneuver. Optimizing it for one mission will invariably make it less useful for the other. Whether or not this is a wise choice is also the same question the Army must ponder with regard to its overall strategic mission.

More On The U.S. Army’s ‘Identity Crisis’

The new edition of the U.S. Army War College’s quarterly journal Parameters contains additional commentary on the question of whether the Army should be optimizing to wage combined arms maneuver warfare or wide-area security/Security Force Assistance.

Conrad Crane, the chief of historical services at the U.S. Army Heritage and Education Center offers some comments and criticism of an article by Gates Brown, “The Army’s Identity Crisis” in the Winter 2016–17 issue of Parameters. Brown then responds to Crane’s comments.

Attrition In Future Land Combat

Soldiers with Battery C, 1st Battalion, 82nd Field Artillery Regiment, 1st Brigade Combat Team, 1st Cavalry Division maneuver their Paladins through Hohenfels Training Area, Oct. 26. Photo Credit: Capt. John Farmer, 1st Brigade Combat Team, 1st Cav

Last autumn, U.S. Army Chief of Staff General Mark Milley asserted that “we are on the cusp of a fundamental change in the character of warfare, and specifically ground warfare. It will be highly lethal, very highly lethal, unlike anything our Army has experienced, at least since World War II.” He made these comments while describing the Army’s evolving Multi-Domain Battle concept for waging future combat against peer or near-peer adversaries.

How lethal will combat on future battlefields be? Forecasting the future is, of course, an undertaking fraught with uncertainties. Milley’s comments undoubtedly reflect the Army’s best guesses about the likely impact of new weapons systems of greater lethality and accuracy, as well as improved capabilities for acquiring targets. Many observers have been closely watching the use of such weapons on the battlefield in the Ukraine. The spectacular success of the Zelenopillya rocket strike in 2014 was a convincing display of the lethality of long-range precision strike capabilities.

It is possible that ground combat attrition in the future between peer or near-peer combatants may be comparable to the U.S. experience in World War II (although there were considerable differences between the experiences of the various belligerents). Combat losses could be heavier. It certainly seems likely that they would be higher than those experienced by U.S. forces in recent counterinsurgency operations.

Unfortunately, the U.S. Defense Department has demonstrated a tenuous understanding of the phenomenon of combat attrition. Despite wildly inaccurate estimates for combat losses in the 1991 Gulf War, only modest effort has been made since then to improve understanding of the relationship between combat and casualties. The U.S. Army currently does not have either an approved tool or a formal methodology for casualty estimation.

Historical Trends in Combat Attrition

Trevor Dupuy did a great deal of historical research on attrition in combat. He found several trends that had strong enough empirical backing that he deemed them to be verities. He detailed his conclusions in Understanding War: History and Theory of Combat (1987) and Attrition: Forecasting Battle Casualties and Equipment Losses in Modern War (1995).

Dupuy documented a clear relationship over time between increasing weapon lethality, greater battlefield dispersion, and declining casualty rates in conventional combat. Even as weapons became more lethal, greater dispersal in frontage and depth among ground forces led daily personnel loss rates in battle to decrease.

The average daily battle casualty rate in combat has been declining since 1600 as a consequence. Since battlefield weapons continue to increase in lethality and troops continue to disperse in response, it seems logical to presume the trend in loss rates continues to decline, although this may not necessarily be the case. There were two instances in the 19th century where daily battle casualty rates increased—during the Napoleonic Wars and the American Civil War—before declining again. Dupuy noted that combat casualty rates in the 1973 Arab-Israeli War remained roughly the same as those in World War II (1939-45), almost thirty years earlier. Further research is needed to determine if average daily personnel loss rates have indeed continued to decrease into the 21st century.

Dupuy also discovered that, as with battle outcomes, casualty rates are influenced by the circumstantial variables of combat. Posture, weather, terrain, season, time of day, surprise, fatigue, level of fortification, and “all out” efforts affect loss rates. (The combat loss rates of armored vehicles, artillery, and other other weapons systems are directly related to personnel loss rates, and are affected by many of the same factors.) Consequently, yet counterintuitively, he could find no direct relationship between numerical force ratios and combat casualty rates. Combat power ratios which take into account the circumstances of combat do affect casualty rates; forces with greater combat power inflict higher rates of casualties than less powerful forces do.

Winning forces suffer lower rates of combat losses than losing forces do, whether attacking or defending. (It should be noted that there is a difference between combat loss rates and numbers of losses. Depending on the circumstances, Dupuy found that the numerical losses of the winning and losing forces may often be similar, even if the winner’s casualty rate is lower.)

Dupuy’s research confirmed the fact that the combat loss rates of smaller forces is higher than that of larger forces. This is in part due to the fact that smaller forces have a larger proportion of their troops exposed to enemy weapons; combat casualties tend to concentrated in the forward-deployed combat and combat support elements. Dupuy also surmised that Prussian military theorist Carl von Clausewitz’s concept of friction plays a role in this. The complexity of interactions between increasing numbers of troops and weapons simply diminishes the lethal effects of weapons systems on real world battlefields.

Somewhat unsurprisingly, higher quality forces (that better manage the ambient effects of friction in combat) inflict casualties at higher rates than those with less effectiveness. This can be seen clearly in the disparities in casualties between German and Soviet forces during World War II, Israeli and Arab combatants in 1973, and U.S. and coalition forces and the Iraqis in 1991 and 2003.

Combat Loss Rates on Future Battlefields

What do Dupuy’s combat attrition verities imply about casualties in future battles? As a baseline, he found that the average daily combat casualty rate in Western Europe during World War II for divisional-level engagements was 1-2% for winning forces and 2-3% for losing ones. For a divisional slice of 15,000 personnel, this meant daily combat losses of 150-450 troops, concentrated in the maneuver battalions (The ratio of wounded to killed in modern combat has been found to be consistently about 4:1. 20% are killed in action; the other 80% include mortally wounded/wounded in action, missing, and captured).

It seems reasonable to conclude that future battlefields will be less densely occupied. Brigades, battalions, and companies will be fighting in spaces formerly filled with armies, corps, and divisions. Fewer troops mean fewer overall casualties, but the daily casualty rates of individual smaller units may well exceed those of WWII divisions. Smaller forces experience significant variation in daily casualties, but Dupuy established average daily rates for them as shown below.

For example, based on Dupuy’s methodology, the average daily loss rate unmodified by combat variables for brigade combat teams would be 1.8% per day, battalions would be 8% per day, and companies 21% per day. For a brigade of 4,500, that would result in 81 battle casualties per day, a battalion of 800 would suffer 64 casualties, and a company of 120 would lose 27 troops. These rates would then be modified by the circumstances of each particular engagement.

Several factors could push daily casualty rates down. Milley envisions that U.S. units engaged in an anti-access/area denial environment will be constantly moving. A low density, highly mobile battlefield with fluid lines would be expected to reduce casualty rates for all sides. High mobility might also limit opportunities for infantry assaults and close quarters combat. The high operational tempo will be exhausting, according to Milley. This could also lower loss rates, as the casualty inflicting capabilities of combat units decline with each successive day in battle.

It is not immediately clear how cyberwarfare and information operations might influence casualty rates. One combat variable they might directly impact would be surprise. Dupuy identified surprise as one of the most potent combat power multipliers. A surprised force suffers a higher casualty rate and surprisers enjoy lower loss rates. Russian combat doctrine emphasizes using cyber and information operations to achieve it and forces with degraded situational awareness are highly susceptible to it. As Zelenopillya demonstrated, surprise attacks with modern weapons can be devastating.

Some factors could push combat loss rates up. Long-range precision weapons could expose greater numbers of troops to enemy fires, which would drive casualties up among combat support and combat service support elements. Casualty rates historically drop during night time hours, although modern night-vision technology and persistent drone reconnaissance might will likely enable continuous night and day battle, which could result in higher losses.

Drawing solid conclusions is difficult but the question of future battlefield attrition is far too important not to be studied with greater urgency. Current policy debates over whether or not the draft should be reinstated and the proper size and distribution of manpower in active and reserve components of the Army hinge on getting this right. The trend away from mass on the battlefield means that there may not be a large margin of error should future combat forces suffer higher combat casualties than expected.

On Domains And Doctrine

Sorry. I could not resist.

There is a lot of smart writing being published over at The Strategy Bridge. If you don’t follow it, you should. Among the recent offerings is a very good piece by Erik Heftye, a retired Air Force officer and senior military analyst in the U.S. Army’s Mission Command Battle Laboratory at Fort Leavenworth. His article “Multi-Domain Confusion: All Domains Are Not Created Equal,” takes a look at an au courant topic central to the new concept of Multi-Domain Battle (MDB).

Defense grognards have been grumbling that MDB is just a new term for an old concept. This is true, insofar as it goes. I am not sure this is in dispute. After all, the subtitle of the U.S. Army/U.S. Marine Corps MDB white paper is “Combined Arms For The 21st Century.” However, such comments do raise the issue of whether new terms for old concepts are serving to clarify or cloud current understanding.

This is Heftye’s concern regarding the use of the term domain: “An ambiguous categorization of separate operating domains in warfare could actually pose an intractable conceptual threat to an integrated joint force, which is ironically the stated purpose of multi-domain battle.” Noting the vagueness of the concept, Heftye traces how the term entered into U.S. military doctrine in the 1990s and evolved into the current four physical (land, sea, air, and space) and one virtual (cyberspace) warfighting realms. He then discusses the etymology of the word and how its meanings imply that all of the domains are equivalent in nature and importance. While this makes sense for air, sea, and land, the physical aspects of those domains do not translate to space or cyberspace. He argues that treating them all analogously will inevitably lead to conceptual confusion.

Heftye recommends a solution: “In order to minimize the problem of domain equivalence, a revised construct should distinguish different types of domains in relation to relevant and advantageous warfighting effects. Focusing on effects rather than domains would allow for the elimination of unnecessary bureaucratic seams, gaps, and turf wars.” He offers up a simplified variation of the domain construct that had been published in the superseded 2005 edition of the Joint Chiefs of Staff’s Capstone Concept for Joint Operations, which defined a domain as “any potential operating ‘space’ through which the target system can be influenced—not only the domains of land, sea, air, and space, but also the virtual (information and cyber) and human (cognitive, moral, and social) domains.”

This version not only simplifies by cutting the five existing categories to three, but it also groups like with like. “Land, sea, air, and space are physical domains involving material reality. Cyberspace and information, as well as the electromagnetic spectrum are virtual domains involving sensing and perception. The construct also included a human category involving value judgements.” Heftye acknowledges that looking at domains in terms of effects runs contrary to then-Joint Forces Commander General (and current Defense Secretary) James Mattis’s ban on the use of the Effects Based Operations (EBO) concept by the U.S. military 2008. He also points out that the concept of domains will not be going away anytime soon, either.

Of course, conceptual confusion is not a unique problem in U.S. military doctrine argues Steve Leonard in “Broken and Unreadable: Our Unbearable Aversion to Doctrine,” over at the Modern War Institute at West Point (which is also publishing great material these days). Leonard (aka Doctrine Man!!), a former U.S. Army senior strategist, ruminates about dissatisfaction and discontent with the American “rules of the game.” He offers up a personal anecdote about a popular military staff pastime: defining doctrinal terms.

We also tend to over-define our terminology. Words in common usage since the days of Noah Webster can find new life—and new meaning—in Army doctrine. Several years ago, I endured an hours-long argument among a group of doctrine writers trying to define the term “asymmetric.” The suggestion—after three full hours of debate—that the group consult a dictionary was not well-received.

I have no doubt Erik Heftye feels your pain, Doctrine Man.