SlideShare una empresa de Scribd logo
1 de 6
Descargar para leer sin conexión
42 • February 2015	 www.usni.org
Trouble
I
f the United States is serious about its “pivot” to the
Pacific, the Department of Defense should invest in
a long-range stealthy autonomous combat air vehicle
that can operate deep behind enemy lines and engage
adversaries without human intervention. Autonomy might
be the only effective way for a long-range unmanned air-
craft to operate in a theater where the United States faces
By Dave Majumdar
The legal and moral questions that come with
unleashing weapons that can seemingly think for
themselves pose a dilemma for the U.S. military.
High-Tech
The
with
off against a near-peer potential adversary—but the prospect
of such a system raises legal and moral questions.
In the Western Pacific, China is building up its anti-
access/area-denial (A2/AD) capabilities—including com-
munications jamming, cyber-warfare, and antisatellite
weapons. In the event of a conflict, Chinese forces are
likely to attack those vital communication links that enable
www.usni.org	 PROCEEDINGS • 43
U.S. NAVY (KELLY SCHINDLER)
An MQ-4C Triton unmanned aircraft system prepares to land at Naval
Air Station Patuxent River, Maryland, after completing a flight from
Northrop Grumman’s California facility. The Triton and the RQ-4B
Global Hawk “have far more autonomy but are not armed,” the author
notes. “Nonetheless, a number of autonomous weapon systems are
already either in service or have reached the prototype stage.”
stealth aircraft may not have the range or the survivability
to operate in those theaters.2
‘The Best Option’
In that case, the best option for the Navy might be to
develop a long-range unmanned strike aircraft with wide-
band all-aspect stealth technology that would be able to
persist inside even the densest of enemy air defenses. By
necessity, given that such an advanced adversary would
be able to deny or degrade communications significantly,
such an aircraft would have to be fully autonomous. In
other words, the unmanned aircraft would have to be able
to operate independently of prolonged communications
with its human masters, and it would also need to be able
to make the decision to release weapons without phoning
home for a human operator’s consent.
Moreover, the U.S. Air Force also faces basing chal-
lenges in the Western Pacific, as existing air bases such
as Kadena and Misawa in Japan and Andersen Air Force
Base in Guam are vulnerable to concerted air and mis-
sile attacks.3
A very stealthy long-range autonomous un-
manned strike aircraft could be used to complement the
service’s prospective long-range strike bomber—going
into places that are far too dangerous for a manned aircraft
or to perform missions like stand-in jamming from inside
hostile territory.
While the initial cost of developing such an autono-
mous unmanned aircraft might be high, there could be
significant savings longer-term. An autonomous unmanned
aircraft only needs to be flown occasionally during peace-
time to keep up the proficiency of maintainers. Further,
an autonomous aircraft has no need to fly training sorties
or to practice—a computer can simply be programmed to
do what needs to be done.
Such an autonomous unmanned aircraft also would not
need downtime between deployments—just the occasional
depot-level maintenance overhaul. That means that the
Navy—or the Air Force if it bought some—would need
only as many aircraft as required to fill the number of
deployed carriers and account for attrition reserves and
planes laid up in depot maintenance. There could also be
significant personnel cost savings because a fully autono-
mous aircraft would not require pilots, and the smaller
fleet would require fewer maintainers.
We Have the Technology
The technology to develop and build such an aircraft
mostly already exists. Most current unmanned aircraft
like the General Atomics Aeronautical Systems MQ-1
Predator and MQ-9 Reaper are remotely controlled by a
U.S. forces to operate cohesively. In those communica-
tions-degraded/-denied environments, unless a system is
manned, autonomy might be the only way to go.
For the U.S. Navy, there is an added dimension—as
was postulated by Jan van Tol, Mark Gunzinger, Andrew
Krepinevich, and Jim Thomas at the Center for Strategic
and Budgetary Assessments. The service’s aircraft carriers
no longer have a safe haven in coastal waters 200 nautical
miles off shore.1
With the rising threats to the carrier in the
form of antiship cruise and ballistic missiles, those ships
may be forced to stand off a significant distance—more
than 1,000 nautical miles—from the enemy shoreline.
In addition, with the proliferation of advanced inte-
grated air-defense networks and low-frequency radars
that can detect and track low-observable targets, existing
44 • February 2015	 www.usni.org
human operator. Others—such as the Northrop Grumman
MQ-4C Triton or RQ-4B Global Hawk—have far more
autonomy but are not armed. Nonetheless, a number of
autonomous weapon systems are already either in service
or have reached the prototype stage that can engage hostile
targets without human intervention.
Perhaps the two most obvious examples are cruise mis-
siles and intercontinental ballistic missiles. Once those
weapons are launched, they proceed autonomously to their
preprogrammed targets—without human intervention.
If one were to imagine a U.S. Navy destroyer launching
a Tomahawk cruise missile at a fixed target somewhere
in the Western Pacific, a sequence of events would be
followed. The crew of the destroyer would receive orders
to attack a particular target. The crew would then pro-
gram that information into the missile. Once launched,
the Tomahawk navigates its way to the target in a man-
ner similar to a manned aircraft—but completely without
human intervention.
Against a fixed target—a bunker or factory, for exam-
ple—a fully autonomous unmanned aerial vehicle (UAV)
would be very similar to a cruise missile. Like a Toma-
hawk, the UAV would receive a particular target location
and instructions for how to engage that target with the cor-
rect weapons. Also like the Tomahawk, the UAV would be
able to navigate to that target completely autonomously. If
the UAV were then to engage that fixed target with a joint
direct-attack munition (JDAM) or some other weapon,
there is no real difference in practical terms between an
unmanned aircraft and a cruise missile. The effect would
be identical. The only change would be that the UAV
could make a second pass, fly on to another target, or fly
home to be rearmed. And it could be argued that with its
jet engine and wings, a Tomahawk is really just a small
UAV on a one-way trip.
The more challenging scenario comes when there is an
unexpected “pop-up” threat such as an S-400 surface-to-
air missile battery that might be encountered by an au-
tonomous unmanned combat air vehicle (UCAV) during
a wartime sortie. Human pilots are assumed to inherently
have the judgment to decide whether or not to engage
such a threat. But those human pilots are making their
decisions based on sensor information being processed by
the aircraft’s computer. In fact, the pilot is often entirely
dependent on the aircraft’s sensors and the avionics to
perform a combat identification of a contact.
The Lockheed Martin F-22 Raptor and F-35 Joint
Strike Fighter epitomize this concept—especially in the
realm of beyond-visual-range air-to-air combat. Both the
F-22 and the F-35 fuse data correlated from the aircraft’s
radar, electronic support measures, and other sensors
into a track file that the computer identifies as hostile,
friendly, or unknown. The pilot is entirely reliant on the
computer to determine a proper combat identification.
It would be a very small technological step for the sys-
tem to engage targets autonomously—without human
intervention.
The air-to-ground arena is somewhat more challeng-
ing because of target-location errors that are inherent in
sensors and navigation systems (and also environmental
effects and enemy camouflage). But with a combination of
electro-optical/infrared cameras, synthetic aperture radar
(SAR), ground moving-target indication (GMTI) radar or
even hyperspectral sensors, a computer can ascertain a
positive combat identification of ground targets—assum-
ing that the data being gathered are geo-registered. Once
the computer can determine a positive identification, either
a manned or unmanned aircraft can engage a target. But
ultimately, the computer is still making the determination
that a contact is hostile.
‘Zero in for the Kill’
In fact, autonomous systems capable of identifying and
attacking targets at their own discretion have existed in the
Civil engineers fold a fiberglass mat at Kadena Air Base, Japan. “The U.S. Air Force faces basing challenges in the Western Pacific, as existing
air bases such as Kadena and Misawa in Japan and Andersen Air Force Base in Guam are vulnerable to concerted air and missile attacks.” One
way to rectify the basing shortfall could be “a very stealthy long-range autonomous unmanned strike aircraft.”
U.S.AIRFORCE(JARVIEZ.WALLACE)
www.usni.org	 PROCEEDINGS • 45
past. One example is the Northrop AGM-136 Tacit Rain-
bow antiradiation cruise missile that was canceled in 1991.
The weapon was designed to be preprogrammed for a
designated target area over which it would loiter. It would
remain in that designated box until it detected emissions
from a hostile radar. Once the Tacit Rainbow detected and
identified an enemy emitter, the missile would zero in for
the kill—all without human intervention.
A later example is the Lockheed Martin low-cost au-
tonomous attack system (LOCASS). The now-defunct
miniature loitering cruise missile demonstrator was guided
by GPS/INS to a target box. It would then use a laser
radar to illuminate targets and match them with preloaded
signatures. The weapon would then go after the highest-
priority target while at the same time selecting the ap-
propriate mode for the warhead to best engage the target
autonomously—without human intervention.
Other prominent examples include the Aegis combat
system, which in its full automatic mode can engage mul-
tiple aircraft or missiles simultaneously—without human
intervention. Similarly, the shipboard close-in weapon sys-
tem (CIWS, or Phalanx) has an autonomous-engagement
capability.
What all of that means is that fully autonomous combat
identification and engagement are technically feasible for
unmanned aircraft—given sophisticated avionics and smart
precision-guided weapons. But while feasible, what of the
moral and legal implications?
Moral and Legal
The Pentagon already preemptively issued policy guid-
ance on the development and operational use of autono-
mous and semi-autonomous weapons in November 2012.
DOD directive 3000.09 states: “Autonomous and semi-
autonomous weapon systems shall be designed to allow
commanders and operators to exercise appropriate lev-
els of human judgment over the use of force.”4
But the
policy does not expressly forbid the development of fully
autonomous lethal weapon systems, it merely states that
senior DOD leadership would
closely supervise any such de-
velopment.5
To prevent what the DOD
calls an “unintended engage-
ment,” those who authorize or
direct the operation of autono-
mous and semi-autonomous
weapon systems are required
to use “appropriate care and in
accordance with the law of war,
applicable treaties, weapon sys-
tem safety rules, and applicable
rules of engagement (ROE).”6
Thus it would seem that the
U.S. government views the use
of autonomous weapon systems
as legal under the laws of war—
provided certain conditions are
met. A number of lawyers spe-
cializing in national-security
law have suggested that fully
autonomous weapons are law-
ful. The responsibility for the
use of such a weapon would
ultimately fall to the person
who authorized its employ-
ment—which is similar to any
other manned weapon.
But there are those who
are adamantly opposed to any
fully autonomous weapon systems—organizations such
as Human Rights Watch (HRW). HRW has called for
an international treaty that would preemptively ban all
autonomous weapons in a November 2012 report titled
“Losing Humanity: The Case against Killer Robots.”7
In
fact, it seems likely that the DOD policy guidance on
the development of autonomous weapons stems from the
conclusions of the HRW report.
The report makes three recommendations.8
The first is
to “Prohibit the development, production, and use of fully
autonomous weapons through an international legally bind-
ing instrument.” The second is to “Adopt national laws and
policies to prohibit the development, production, and use of
The guided-missile destroyer USS Arleigh Burke (DDG-51) launches Tomahawk cruise missiles from the Red
Sea in strikes against the Islamic State of Iraq and the Levant, or ISIL, on 23 September 2014. “Against a
fixed target, a fully autonomous unmanned aerial vehicle (UAV) would be very similar to a cruise missile. . . .
With its jet engine and wings, a Tomahawk is really just a small UAV on a one-way trip.”
U.S.NAVY(CARLOSM.VAZQUEZ)
46 • February 2015	 www.usni.org
fully autonomous weapons.” And the third is to “Commence
reviews of technologies and components that could lead
to fully autonomous weapons. These reviews should take
place at the very beginning of the development process and
continue throughout the development and testing phases.”9
HRW asserts that autonomous systems are unable to
meet the standards set forth under international humani-
tarian law. “The rules of distinction, proportionality, and
military necessity are especially important tools for pro-
tecting civilians from the effects of war, and fully autono-
mous weapons would not be able to abide by those rules,”
the report states.10
Machines Are Better than Humans
But critics such as legal scholar Benjamin Wittes at the
Brookings Institution have challenged such statements.
Wittes has written that there are situations where machines
can “distinguish military targets far better and more ac-
curately than humans can.” Those familiar with unmanned
technology, sensor hardware, and software can attest that
is indeed the case.11
If a computer is given a certain set of parameters—for
example, a series of rules of engagement—it will follow
those instructions precisely. If the autonomous weapon
is designed and built to operate within the laws of war,
then there should be no objection to their use.12
Under
Article 36 of the 1977 Additional Protocol to the Geneva
Conventions, weapons cannot be inherently indiscriminate
and are prohibited from causing unnecessary suffering or
superfluous injury. “The fact that an autonomous weapon
system selects the target or undertakes the attack does not
violate the rule,” write Hoover Institution legal scholars
Kenneth Anderson and Matthew Waxman in a paper titled
“Law and Ethics For Autonomous Weapon Systems.”13
Technology is continuously moving forward, and while
autonomous systems may not be able to operate under
all circumstances, it may only be a matter of time before
engineers find a technical solution. Under many circum-
stances—with the right sensors and algorithms—autono-
mous systems would have the ability to distinguish lawful
targets from unlawful ones, but that is not currently the case
under all circumstances. Thus there are some limitations
inherent to autonomous weapon systems for the time being.
A Matter of Proportion
However, those will not always be there, as technol-
ogy continues to advance and engineers make progress.
As Wittes points out correctly, “To call for a per se ban
on autonomous weapons is to insist as a matter of IHL
[international humanitarian law] on preserving a minimum
level of human error in targeting.”14
Machines are gener-
ally far more precise than human beings.
Along with being able to distinguish between targets,
the law requires that combatants weigh the proportionality
of their actions.15
“Any use of a weapon must also involve
evaluation that sets the anticipated military advantage to
be gained against the anticipated civilian harm,” Anderson
and Waxman write. “The harm to civilians must not be
excessive relative to the expected military gain.”
Though technically challenging, a completely autono-
mous weapon system would have to be required to address
proportionality as well as distinction. But the difficulty is
entirely dependent on the specific operational scenario. For
example, while an unmanned aircraft could identify and
The technology is here. Air Force F-22 Raptors line up on the runway at Hill Air Force Base in Utah. The Raptor and the F-35 Joint Strike Fighter
are specially equipped for “beyond-visual-range air-to-air combat. Both the F-22 and the F-35 fuse data correlated from the aircraft’s radar, elec-
tonic support measures, and other sensors into a track file that the computer identifies as hostile, friendly, or unknown.”
U.S.AIRFORCE
www.usni.org	 PROCEEDINGS • 47
attack a hostile surface-to-air missile system deep behind
enemy lines or an enemy warship at sea—where there is
little chance of encountering civilians—targets inside a
highly populated area are more difficult to prosecute.
Some of the most problematic scenarios—which would
not necessarily be factors in a high-end campaign against
an A2/AD threat—would be challenging for a human
pilot, let alone a machine. For ex-
ample, during a counter-insurgency
campaign, if two school buses were
driving side-by-side in a built-up area,
but one of the vehicles was carrying
nuns and the other carrying heavily
armed terrorists, it would be very dif-
ficult for a human pilot to determine
which bus is the proper target until
one of them commits a hostile act.
The same would be true for an autonomous system, but
in the near term it could be a technological challenge.
The human pilot would also have to determine what
kind of weapon to use—judging the proportionality. Does
he or she select a 2,000-pound JDAM, a smaller 250-
pound small-diameter bomb, a 20-mm cannon, or do
nothing since the risk of civilian casualties is too high?
Likewise, an autonomous weapon system would need to
be programmed to select an appropriate low-collateral-
damage munition or to disengage if the danger of civilian
casualties is too great once the target has been positively
identified. But it would take time and investment before
such an autonomous system could become a reality.
Playing Fair?
Thus, for the near future, autonomous weapons would
have to be developed incrementally, starting with systems
that could engage fixed targets and “obviously” military
targets like surface-to-air missile sites or tank columns on
the open battlefield during a conventional war. Likewise,
in the maritime environment, where there are few civil-
ians to speak of, autonomous systems could offer huge
advantages with little in the way of drawbacks.
For the time being, autonomous weapons should not
be used in complex scenarios such as counterinsurgency,
where there is significant possibility that it could cause
inadvertent civilian casualties or unintended collateral
damage. It may also be unwise to use a fully autono-
mous UCAV for missions like close-air support—particu-
larly during “danger close”–type situations where friendly
troops are in close contact with the enemy—until the
technology has been proven operationally in other roles.
Human pilots have a hard enough time with those types of
missions. Oftentimes, even human pilots cannot determine
friend from foe—depending on the scenario—as the U.S.
Central Command is discovering over Iraq and Syria with
the ongoing fight against the Islamic State terrorist group.
While at present some technological limitations do exist,
those are not likely to remain roadblocks forever. Autono-
mous technology is advancing rapidly and could one day
be precise and reliable enough to not only distinguish cor-
rect targets but also make proportionality judgments in
complex scenarios based on parameters programmed into
the machine. Those parameters would not be unlike rules
of engagement given to human operators. Already, cam-
eras and software exist that can identify individual human
faces. Once a target is precisely identified, it would not be
a huge leap then for an autonomous
system to use a low-collateral-damage
weapon to eliminate hostile targets
while minimizing harm to civilians.
Much of the objection to fully au-
tonomous weapons seems to stem
from a sense of “fair play” rather
than any real legal issues—most of
which are likely to be overcome. But
any time new weapon technology
emerges, there is opposition from those who believe that
the technology fundamentally unbalances war.16
Objec-
tions have been raised throughout history to new technolo-
gies ranging from crossbows and longbows to machine
guns and submarines because the use of such weapons was
considered to be “unfair” or “unsporting.” But ultimately,
the use of such weapons became a fact of life. War is not
a game, and as Anderson and Waxman write: “The law, to
be sure, makes no requirement that sides limit themselves
to the weapons available to the other side; weapons supe-
riority is perfectly lawful and indeed assumed as part of
military necessity.”17
Indeed, there is no legal requirement for war to be fair—
in fact throughout history war has been anything but.
1. Jan van Tol, Mark Gunzinger, Andrew Krepinevich, and Jim Thomas, “AirSea Battle:
A Point-of-Departure Operational Concept,” Center for Strategic and Budgetary As-
sessments, www.csbaonline.org/publications/2010/05/airsea-battle-concept/.
2. Ibid.
3. Ibid.
4. DOD directive 3000.09, www.dtic.mil/whs/directives/corres/pdf/300009p.pdf.
5. Ibid.
6. Ibid.
7. Human Rights Watch, “Losing Humanity: The Case against Killer Robots,” www.
hrw.org/reports/2012/11/19/losing-humanity-0.
8. Ibid.
9. Ibid.
10. Ibid.
11. Benjamin Wittes, Brookings Institution, www.lawfareblog.com/2012/12/does-
human-rights-watch-prefer-disproportionate-and-indiscriminate-humans-to-discrimi-
nating-and-proportionate-robots/#.UunS7GRdXB8
12. Kenneth Anderson and Matthew Waxman, “Law and Ethics For Autono-
mous Weapon Systems,” Hoover Institution http://papers.ssrn.com/sol3/papers.
cfm?abstract_id=2250126.
13. Ibid.
14. Wittes, Brookings Institution.
15. Anderson and Waxman, Hoover Institution.
16. COL Lawrence Spinetta, USAF, commander of the 69th Reconnaissance Group,
interview with author.
17. Anderson and Waxman, Hoover Institution.
Mr. Majumdar has been covering defense since 2004. In addition
to Proceedings, he writes for USNI News, Aviation Week & Space
Technology, and The Daily Beast, among others. He previously covered
national-security issues at Flight International, Defense News, and
C4ISR Journal.
Human Rights Watch
has called for an
international treaty
that would ban all
autonomous weapons.

Más contenido relacionado

La actualidad más candente

The Future Of Unmanned Air Systems (2 May2011)
The Future Of Unmanned Air Systems (2 May2011)The Future Of Unmanned Air Systems (2 May2011)
The Future Of Unmanned Air Systems (2 May2011)Mark Tempestilli
 
UAV (Unmanned Aerial Vehicle)
UAV (Unmanned Aerial Vehicle)UAV (Unmanned Aerial Vehicle)
UAV (Unmanned Aerial Vehicle)UDIT PATEL
 
Joint Attack Pathfinder Teams v3.0
Joint Attack Pathfinder Teams v3.0Joint Attack Pathfinder Teams v3.0
Joint Attack Pathfinder Teams v3.01st_TSG_Airborne
 
The Automation and Proliferation of Military Drones and the Protection of Civ...
The Automation and Proliferation of Military Drones and the Protection of Civ...The Automation and Proliferation of Military Drones and the Protection of Civ...
The Automation and Proliferation of Military Drones and the Protection of Civ...Angelo State University
 
ACO9 - Aviation Firefighting for Structual Trucks
 ACO9 - Aviation Firefighting for Structual Trucks  ACO9 - Aviation Firefighting for Structual Trucks
ACO9 - Aviation Firefighting for Structual Trucks Brock Jester
 
ACO-2 Aircraft Familiarization
ACO-2 Aircraft Familiarization ACO-2 Aircraft Familiarization
ACO-2 Aircraft Familiarization Brock Jester
 
Crenshaw - Licence Plates for Drones
Crenshaw - Licence Plates for DronesCrenshaw - Licence Plates for Drones
Crenshaw - Licence Plates for DronesMichael Crenshaw
 
Joint Attack Pathfinder Teams (Light) v8.0
Joint Attack Pathfinder Teams (Light) v8.0Joint Attack Pathfinder Teams (Light) v8.0
Joint Attack Pathfinder Teams (Light) v8.01st_TSG_Airborne
 
Mohawk Maneuver Air Support (MAS) v4.0
Mohawk Maneuver Air Support (MAS) v4.0Mohawk Maneuver Air Support (MAS) v4.0
Mohawk Maneuver Air Support (MAS) v4.01st_TSG_Airborne
 
Defeat Surveillance-Strike-Complexes v1.0
Defeat Surveillance-Strike-Complexes v1.0Defeat Surveillance-Strike-Complexes v1.0
Defeat Surveillance-Strike-Complexes v1.01st_TSG_Airborne
 
Naval Aircraft & Missiles Web
Naval Aircraft & Missiles WebNaval Aircraft & Missiles Web
Naval Aircraft & Missiles WebLynn Seckinger
 
Joint Attack Pathfinder Teams v1.0
Joint Attack Pathfinder Teams v1.0Joint Attack Pathfinder Teams v1.0
Joint Attack Pathfinder Teams v1.01st_TSG_Airborne
 
Helicopters at Warfare
Helicopters at WarfareHelicopters at Warfare
Helicopters at WarfareBradfordRay7
 
Drone-Unmanned Aerial Vehicle
Drone-Unmanned Aerial VehicleDrone-Unmanned Aerial Vehicle
Drone-Unmanned Aerial Vehicleshivu1234
 
Media Object File Flt Ops Gnd Hand Seq01
Media Object File Flt Ops Gnd Hand Seq01Media Object File Flt Ops Gnd Hand Seq01
Media Object File Flt Ops Gnd Hand Seq01syed viquar
 

La actualidad más candente (19)

The Future Of Unmanned Air Systems (2 May2011)
The Future Of Unmanned Air Systems (2 May2011)The Future Of Unmanned Air Systems (2 May2011)
The Future Of Unmanned Air Systems (2 May2011)
 
final pppt
final ppptfinal pppt
final pppt
 
UAV (Unmanned Aerial Vehicle)
UAV (Unmanned Aerial Vehicle)UAV (Unmanned Aerial Vehicle)
UAV (Unmanned Aerial Vehicle)
 
Joint Attack Pathfinder Teams v3.0
Joint Attack Pathfinder Teams v3.0Joint Attack Pathfinder Teams v3.0
Joint Attack Pathfinder Teams v3.0
 
The Automation and Proliferation of Military Drones and the Protection of Civ...
The Automation and Proliferation of Military Drones and the Protection of Civ...The Automation and Proliferation of Military Drones and the Protection of Civ...
The Automation and Proliferation of Military Drones and the Protection of Civ...
 
ACO9 - Aviation Firefighting for Structual Trucks
 ACO9 - Aviation Firefighting for Structual Trucks  ACO9 - Aviation Firefighting for Structual Trucks
ACO9 - Aviation Firefighting for Structual Trucks
 
V4n3 fog of_war
V4n3 fog of_warV4n3 fog of_war
V4n3 fog of_war
 
Research assignment
Research assignmentResearch assignment
Research assignment
 
ACO-2 Aircraft Familiarization
ACO-2 Aircraft Familiarization ACO-2 Aircraft Familiarization
ACO-2 Aircraft Familiarization
 
Group1
Group1Group1
Group1
 
Crenshaw - Licence Plates for Drones
Crenshaw - Licence Plates for DronesCrenshaw - Licence Plates for Drones
Crenshaw - Licence Plates for Drones
 
Joint Attack Pathfinder Teams (Light) v8.0
Joint Attack Pathfinder Teams (Light) v8.0Joint Attack Pathfinder Teams (Light) v8.0
Joint Attack Pathfinder Teams (Light) v8.0
 
Mohawk Maneuver Air Support (MAS) v4.0
Mohawk Maneuver Air Support (MAS) v4.0Mohawk Maneuver Air Support (MAS) v4.0
Mohawk Maneuver Air Support (MAS) v4.0
 
Defeat Surveillance-Strike-Complexes v1.0
Defeat Surveillance-Strike-Complexes v1.0Defeat Surveillance-Strike-Complexes v1.0
Defeat Surveillance-Strike-Complexes v1.0
 
Naval Aircraft & Missiles Web
Naval Aircraft & Missiles WebNaval Aircraft & Missiles Web
Naval Aircraft & Missiles Web
 
Joint Attack Pathfinder Teams v1.0
Joint Attack Pathfinder Teams v1.0Joint Attack Pathfinder Teams v1.0
Joint Attack Pathfinder Teams v1.0
 
Helicopters at Warfare
Helicopters at WarfareHelicopters at Warfare
Helicopters at Warfare
 
Drone-Unmanned Aerial Vehicle
Drone-Unmanned Aerial VehicleDrone-Unmanned Aerial Vehicle
Drone-Unmanned Aerial Vehicle
 
Media Object File Flt Ops Gnd Hand Seq01
Media Object File Flt Ops Gnd Hand Seq01Media Object File Flt Ops Gnd Hand Seq01
Media Object File Flt Ops Gnd Hand Seq01
 

Destacado

Dispozita normative i_
Dispozita normative i_Dispozita normative i_
Dispozita normative i_hasanmarjan
 
Stylish tunics collection
Stylish tunics collectionStylish tunics collection
Stylish tunics collectionSD Desi Store
 
The mahabharata( shri rajagopalachari)
The mahabharata( shri rajagopalachari)The mahabharata( shri rajagopalachari)
The mahabharata( shri rajagopalachari)Nikhil Vooda
 
Designer vikram-phadnis-suits
Designer vikram-phadnis-suitsDesigner vikram-phadnis-suits
Designer vikram-phadnis-suitsSD Desi Store
 
Charcter Casting and Why
Charcter Casting and WhyCharcter Casting and Why
Charcter Casting and WhyOMGWTFITSJAMES
 
Andrea Rooney CV 2015
Andrea Rooney CV 2015Andrea Rooney CV 2015
Andrea Rooney CV 2015Andrea Rooney
 
Collection of Beautiful Designer Sarees
Collection of Beautiful Designer SareesCollection of Beautiful Designer Sarees
Collection of Beautiful Designer SareesSD Desi Store
 
Mem consultora habilitada-eia_dgaam
Mem   consultora habilitada-eia_dgaamMem   consultora habilitada-eia_dgaam
Mem consultora habilitada-eia_dgaamJCesar Nap
 
Beautiful Collection of Indo Western Dresses
Beautiful Collection of Indo Western DressesBeautiful Collection of Indo Western Dresses
Beautiful Collection of Indo Western DressesSD Desi Store
 

Destacado (11)

Dispozita normative i_
Dispozita normative i_Dispozita normative i_
Dispozita normative i_
 
Stylish tunics collection
Stylish tunics collectionStylish tunics collection
Stylish tunics collection
 
The mahabharata( shri rajagopalachari)
The mahabharata( shri rajagopalachari)The mahabharata( shri rajagopalachari)
The mahabharata( shri rajagopalachari)
 
Designer vikram-phadnis-suits
Designer vikram-phadnis-suitsDesigner vikram-phadnis-suits
Designer vikram-phadnis-suits
 
Charcter Casting and Why
Charcter Casting and WhyCharcter Casting and Why
Charcter Casting and Why
 
Andrea Rooney CV 2015
Andrea Rooney CV 2015Andrea Rooney CV 2015
Andrea Rooney CV 2015
 
Collection of Beautiful Designer Sarees
Collection of Beautiful Designer SareesCollection of Beautiful Designer Sarees
Collection of Beautiful Designer Sarees
 
Mem consultora habilitada-eia_dgaam
Mem   consultora habilitada-eia_dgaamMem   consultora habilitada-eia_dgaam
Mem consultora habilitada-eia_dgaam
 
Beautiful Collection of Indo Western Dresses
Beautiful Collection of Indo Western DressesBeautiful Collection of Indo Western Dresses
Beautiful Collection of Indo Western Dresses
 
Binder1
Binder1Binder1
Binder1
 
Sigform Catalog New
Sigform Catalog NewSigform Catalog New
Sigform Catalog New
 

Similar a Majumdar, Dave - 2015, 141-2-1344

UAS Threat to Aviation - January 2019
UAS Threat to Aviation - January 2019UAS Threat to Aviation - January 2019
UAS Threat to Aviation - January 2019ChadCogan
 
Island Air Defence: Challenges, Novel Surveillance Concepts and Advanced Rada...
Island Air Defence: Challenges, Novel Surveillance Concepts and Advanced Rada...Island Air Defence: Challenges, Novel Surveillance Concepts and Advanced Rada...
Island Air Defence: Challenges, Novel Surveillance Concepts and Advanced Rada...Ashwin Samales
 
UNMANNED AERIAL VEHICLE
UNMANNED AERIAL VEHICLEUNMANNED AERIAL VEHICLE
UNMANNED AERIAL VEHICLEThirumal Aero
 
Prop For Std UAV in CO EM [03102016]
Prop For Std UAV in CO EM [03102016]Prop For Std UAV in CO EM [03102016]
Prop For Std UAV in CO EM [03102016]Francis Song
 
Operational Maneuver from the Air_Schenck
Operational Maneuver from the Air_SchenckOperational Maneuver from the Air_Schenck
Operational Maneuver from the Air_SchenckJohn Schenck, PMP
 
weg-2016-vol-3-naval-systems
 weg-2016-vol-3-naval-systems weg-2016-vol-3-naval-systems
weg-2016-vol-3-naval-systemsTriplo Sof
 
Responding to the Emerging Threat of Chinese DF-21D (CSS-5 MOD 4) Anti-Ship B...
Responding to the Emerging Threat of Chinese DF-21D (CSS-5 MOD 4) Anti-Ship B...Responding to the Emerging Threat of Chinese DF-21D (CSS-5 MOD 4) Anti-Ship B...
Responding to the Emerging Threat of Chinese DF-21D (CSS-5 MOD 4) Anti-Ship B...Georgia Department of Defense
 
UnderseaWarfar_WritingSample
UnderseaWarfar_WritingSampleUnderseaWarfar_WritingSample
UnderseaWarfar_WritingSampleRoyce Morales
 
Cruise Missile Technology Seminar Presentation
Cruise Missile Technology Seminar PresentationCruise Missile Technology Seminar Presentation
Cruise Missile Technology Seminar PresentationVaibhav R. R. Chichmalkar
 
Air Defence Systems And Weapons
Air Defence Systems And WeaponsAir Defence Systems And Weapons
Air Defence Systems And WeaponsAndrew Parish
 
vdocument.in_stealth-technology-report.pdf
vdocument.in_stealth-technology-report.pdfvdocument.in_stealth-technology-report.pdf
vdocument.in_stealth-technology-report.pdfzishansingh
 
Types of US Military Aircraft
Types of US Military AircraftTypes of US Military Aircraft
Types of US Military Aircrafttoothsomeauger257
 
The Sharp Curve of Technology.pdf
The Sharp Curve of Technology.pdfThe Sharp Curve of Technology.pdf
The Sharp Curve of Technology.pdfMandeep Singh
 

Similar a Majumdar, Dave - 2015, 141-2-1344 (20)

0822b
0822b0822b
0822b
 
UAS Threat to Aviation - January 2019
UAS Threat to Aviation - January 2019UAS Threat to Aviation - January 2019
UAS Threat to Aviation - January 2019
 
Island Air Defence: Challenges, Novel Surveillance Concepts and Advanced Rada...
Island Air Defence: Challenges, Novel Surveillance Concepts and Advanced Rada...Island Air Defence: Challenges, Novel Surveillance Concepts and Advanced Rada...
Island Air Defence: Challenges, Novel Surveillance Concepts and Advanced Rada...
 
UNMANNED AERIAL VEHICLE
UNMANNED AERIAL VEHICLEUNMANNED AERIAL VEHICLE
UNMANNED AERIAL VEHICLE
 
Prop For Std UAV in CO EM [03102016]
Prop For Std UAV in CO EM [03102016]Prop For Std UAV in CO EM [03102016]
Prop For Std UAV in CO EM [03102016]
 
Operational Maneuver from the Air_Schenck
Operational Maneuver from the Air_SchenckOperational Maneuver from the Air_Schenck
Operational Maneuver from the Air_Schenck
 
weg-2016-vol-3-naval-systems
 weg-2016-vol-3-naval-systems weg-2016-vol-3-naval-systems
weg-2016-vol-3-naval-systems
 
UAV
UAVUAV
UAV
 
Responding to the Emerging Threat of Chinese DF-21D (CSS-5 MOD 4) Anti-Ship B...
Responding to the Emerging Threat of Chinese DF-21D (CSS-5 MOD 4) Anti-Ship B...Responding to the Emerging Threat of Chinese DF-21D (CSS-5 MOD 4) Anti-Ship B...
Responding to the Emerging Threat of Chinese DF-21D (CSS-5 MOD 4) Anti-Ship B...
 
UnderseaWarfar_WritingSample
UnderseaWarfar_WritingSampleUnderseaWarfar_WritingSample
UnderseaWarfar_WritingSample
 
Proceedings Article
Proceedings ArticleProceedings Article
Proceedings Article
 
Cruise Missile Technology Seminar Presentation
Cruise Missile Technology Seminar PresentationCruise Missile Technology Seminar Presentation
Cruise Missile Technology Seminar Presentation
 
Missile
MissileMissile
Missile
 
8717796.ppt
8717796.ppt8717796.ppt
8717796.ppt
 
Air Defence Systems And Weapons
Air Defence Systems And WeaponsAir Defence Systems And Weapons
Air Defence Systems And Weapons
 
Intersec January 2008
Intersec January 2008Intersec January 2008
Intersec January 2008
 
403 8
403 8403 8
403 8
 
vdocument.in_stealth-technology-report.pdf
vdocument.in_stealth-technology-report.pdfvdocument.in_stealth-technology-report.pdf
vdocument.in_stealth-technology-report.pdf
 
Types of US Military Aircraft
Types of US Military AircraftTypes of US Military Aircraft
Types of US Military Aircraft
 
The Sharp Curve of Technology.pdf
The Sharp Curve of Technology.pdfThe Sharp Curve of Technology.pdf
The Sharp Curve of Technology.pdf
 

Majumdar, Dave - 2015, 141-2-1344

  • 1. 42 • February 2015 www.usni.org Trouble I f the United States is serious about its “pivot” to the Pacific, the Department of Defense should invest in a long-range stealthy autonomous combat air vehicle that can operate deep behind enemy lines and engage adversaries without human intervention. Autonomy might be the only effective way for a long-range unmanned air- craft to operate in a theater where the United States faces By Dave Majumdar The legal and moral questions that come with unleashing weapons that can seemingly think for themselves pose a dilemma for the U.S. military. High-Tech The with off against a near-peer potential adversary—but the prospect of such a system raises legal and moral questions. In the Western Pacific, China is building up its anti- access/area-denial (A2/AD) capabilities—including com- munications jamming, cyber-warfare, and antisatellite weapons. In the event of a conflict, Chinese forces are likely to attack those vital communication links that enable
  • 2. www.usni.org PROCEEDINGS • 43 U.S. NAVY (KELLY SCHINDLER) An MQ-4C Triton unmanned aircraft system prepares to land at Naval Air Station Patuxent River, Maryland, after completing a flight from Northrop Grumman’s California facility. The Triton and the RQ-4B Global Hawk “have far more autonomy but are not armed,” the author notes. “Nonetheless, a number of autonomous weapon systems are already either in service or have reached the prototype stage.” stealth aircraft may not have the range or the survivability to operate in those theaters.2 ‘The Best Option’ In that case, the best option for the Navy might be to develop a long-range unmanned strike aircraft with wide- band all-aspect stealth technology that would be able to persist inside even the densest of enemy air defenses. By necessity, given that such an advanced adversary would be able to deny or degrade communications significantly, such an aircraft would have to be fully autonomous. In other words, the unmanned aircraft would have to be able to operate independently of prolonged communications with its human masters, and it would also need to be able to make the decision to release weapons without phoning home for a human operator’s consent. Moreover, the U.S. Air Force also faces basing chal- lenges in the Western Pacific, as existing air bases such as Kadena and Misawa in Japan and Andersen Air Force Base in Guam are vulnerable to concerted air and mis- sile attacks.3 A very stealthy long-range autonomous un- manned strike aircraft could be used to complement the service’s prospective long-range strike bomber—going into places that are far too dangerous for a manned aircraft or to perform missions like stand-in jamming from inside hostile territory. While the initial cost of developing such an autono- mous unmanned aircraft might be high, there could be significant savings longer-term. An autonomous unmanned aircraft only needs to be flown occasionally during peace- time to keep up the proficiency of maintainers. Further, an autonomous aircraft has no need to fly training sorties or to practice—a computer can simply be programmed to do what needs to be done. Such an autonomous unmanned aircraft also would not need downtime between deployments—just the occasional depot-level maintenance overhaul. That means that the Navy—or the Air Force if it bought some—would need only as many aircraft as required to fill the number of deployed carriers and account for attrition reserves and planes laid up in depot maintenance. There could also be significant personnel cost savings because a fully autono- mous aircraft would not require pilots, and the smaller fleet would require fewer maintainers. We Have the Technology The technology to develop and build such an aircraft mostly already exists. Most current unmanned aircraft like the General Atomics Aeronautical Systems MQ-1 Predator and MQ-9 Reaper are remotely controlled by a U.S. forces to operate cohesively. In those communica- tions-degraded/-denied environments, unless a system is manned, autonomy might be the only way to go. For the U.S. Navy, there is an added dimension—as was postulated by Jan van Tol, Mark Gunzinger, Andrew Krepinevich, and Jim Thomas at the Center for Strategic and Budgetary Assessments. The service’s aircraft carriers no longer have a safe haven in coastal waters 200 nautical miles off shore.1 With the rising threats to the carrier in the form of antiship cruise and ballistic missiles, those ships may be forced to stand off a significant distance—more than 1,000 nautical miles—from the enemy shoreline. In addition, with the proliferation of advanced inte- grated air-defense networks and low-frequency radars that can detect and track low-observable targets, existing
  • 3. 44 • February 2015 www.usni.org human operator. Others—such as the Northrop Grumman MQ-4C Triton or RQ-4B Global Hawk—have far more autonomy but are not armed. Nonetheless, a number of autonomous weapon systems are already either in service or have reached the prototype stage that can engage hostile targets without human intervention. Perhaps the two most obvious examples are cruise mis- siles and intercontinental ballistic missiles. Once those weapons are launched, they proceed autonomously to their preprogrammed targets—without human intervention. If one were to imagine a U.S. Navy destroyer launching a Tomahawk cruise missile at a fixed target somewhere in the Western Pacific, a sequence of events would be followed. The crew of the destroyer would receive orders to attack a particular target. The crew would then pro- gram that information into the missile. Once launched, the Tomahawk navigates its way to the target in a man- ner similar to a manned aircraft—but completely without human intervention. Against a fixed target—a bunker or factory, for exam- ple—a fully autonomous unmanned aerial vehicle (UAV) would be very similar to a cruise missile. Like a Toma- hawk, the UAV would receive a particular target location and instructions for how to engage that target with the cor- rect weapons. Also like the Tomahawk, the UAV would be able to navigate to that target completely autonomously. If the UAV were then to engage that fixed target with a joint direct-attack munition (JDAM) or some other weapon, there is no real difference in practical terms between an unmanned aircraft and a cruise missile. The effect would be identical. The only change would be that the UAV could make a second pass, fly on to another target, or fly home to be rearmed. And it could be argued that with its jet engine and wings, a Tomahawk is really just a small UAV on a one-way trip. The more challenging scenario comes when there is an unexpected “pop-up” threat such as an S-400 surface-to- air missile battery that might be encountered by an au- tonomous unmanned combat air vehicle (UCAV) during a wartime sortie. Human pilots are assumed to inherently have the judgment to decide whether or not to engage such a threat. But those human pilots are making their decisions based on sensor information being processed by the aircraft’s computer. In fact, the pilot is often entirely dependent on the aircraft’s sensors and the avionics to perform a combat identification of a contact. The Lockheed Martin F-22 Raptor and F-35 Joint Strike Fighter epitomize this concept—especially in the realm of beyond-visual-range air-to-air combat. Both the F-22 and the F-35 fuse data correlated from the aircraft’s radar, electronic support measures, and other sensors into a track file that the computer identifies as hostile, friendly, or unknown. The pilot is entirely reliant on the computer to determine a proper combat identification. It would be a very small technological step for the sys- tem to engage targets autonomously—without human intervention. The air-to-ground arena is somewhat more challeng- ing because of target-location errors that are inherent in sensors and navigation systems (and also environmental effects and enemy camouflage). But with a combination of electro-optical/infrared cameras, synthetic aperture radar (SAR), ground moving-target indication (GMTI) radar or even hyperspectral sensors, a computer can ascertain a positive combat identification of ground targets—assum- ing that the data being gathered are geo-registered. Once the computer can determine a positive identification, either a manned or unmanned aircraft can engage a target. But ultimately, the computer is still making the determination that a contact is hostile. ‘Zero in for the Kill’ In fact, autonomous systems capable of identifying and attacking targets at their own discretion have existed in the Civil engineers fold a fiberglass mat at Kadena Air Base, Japan. “The U.S. Air Force faces basing challenges in the Western Pacific, as existing air bases such as Kadena and Misawa in Japan and Andersen Air Force Base in Guam are vulnerable to concerted air and missile attacks.” One way to rectify the basing shortfall could be “a very stealthy long-range autonomous unmanned strike aircraft.” U.S.AIRFORCE(JARVIEZ.WALLACE)
  • 4. www.usni.org PROCEEDINGS • 45 past. One example is the Northrop AGM-136 Tacit Rain- bow antiradiation cruise missile that was canceled in 1991. The weapon was designed to be preprogrammed for a designated target area over which it would loiter. It would remain in that designated box until it detected emissions from a hostile radar. Once the Tacit Rainbow detected and identified an enemy emitter, the missile would zero in for the kill—all without human intervention. A later example is the Lockheed Martin low-cost au- tonomous attack system (LOCASS). The now-defunct miniature loitering cruise missile demonstrator was guided by GPS/INS to a target box. It would then use a laser radar to illuminate targets and match them with preloaded signatures. The weapon would then go after the highest- priority target while at the same time selecting the ap- propriate mode for the warhead to best engage the target autonomously—without human intervention. Other prominent examples include the Aegis combat system, which in its full automatic mode can engage mul- tiple aircraft or missiles simultaneously—without human intervention. Similarly, the shipboard close-in weapon sys- tem (CIWS, or Phalanx) has an autonomous-engagement capability. What all of that means is that fully autonomous combat identification and engagement are technically feasible for unmanned aircraft—given sophisticated avionics and smart precision-guided weapons. But while feasible, what of the moral and legal implications? Moral and Legal The Pentagon already preemptively issued policy guid- ance on the development and operational use of autono- mous and semi-autonomous weapons in November 2012. DOD directive 3000.09 states: “Autonomous and semi- autonomous weapon systems shall be designed to allow commanders and operators to exercise appropriate lev- els of human judgment over the use of force.”4 But the policy does not expressly forbid the development of fully autonomous lethal weapon systems, it merely states that senior DOD leadership would closely supervise any such de- velopment.5 To prevent what the DOD calls an “unintended engage- ment,” those who authorize or direct the operation of autono- mous and semi-autonomous weapon systems are required to use “appropriate care and in accordance with the law of war, applicable treaties, weapon sys- tem safety rules, and applicable rules of engagement (ROE).”6 Thus it would seem that the U.S. government views the use of autonomous weapon systems as legal under the laws of war— provided certain conditions are met. A number of lawyers spe- cializing in national-security law have suggested that fully autonomous weapons are law- ful. The responsibility for the use of such a weapon would ultimately fall to the person who authorized its employ- ment—which is similar to any other manned weapon. But there are those who are adamantly opposed to any fully autonomous weapon systems—organizations such as Human Rights Watch (HRW). HRW has called for an international treaty that would preemptively ban all autonomous weapons in a November 2012 report titled “Losing Humanity: The Case against Killer Robots.”7 In fact, it seems likely that the DOD policy guidance on the development of autonomous weapons stems from the conclusions of the HRW report. The report makes three recommendations.8 The first is to “Prohibit the development, production, and use of fully autonomous weapons through an international legally bind- ing instrument.” The second is to “Adopt national laws and policies to prohibit the development, production, and use of The guided-missile destroyer USS Arleigh Burke (DDG-51) launches Tomahawk cruise missiles from the Red Sea in strikes against the Islamic State of Iraq and the Levant, or ISIL, on 23 September 2014. “Against a fixed target, a fully autonomous unmanned aerial vehicle (UAV) would be very similar to a cruise missile. . . . With its jet engine and wings, a Tomahawk is really just a small UAV on a one-way trip.” U.S.NAVY(CARLOSM.VAZQUEZ)
  • 5. 46 • February 2015 www.usni.org fully autonomous weapons.” And the third is to “Commence reviews of technologies and components that could lead to fully autonomous weapons. These reviews should take place at the very beginning of the development process and continue throughout the development and testing phases.”9 HRW asserts that autonomous systems are unable to meet the standards set forth under international humani- tarian law. “The rules of distinction, proportionality, and military necessity are especially important tools for pro- tecting civilians from the effects of war, and fully autono- mous weapons would not be able to abide by those rules,” the report states.10 Machines Are Better than Humans But critics such as legal scholar Benjamin Wittes at the Brookings Institution have challenged such statements. Wittes has written that there are situations where machines can “distinguish military targets far better and more ac- curately than humans can.” Those familiar with unmanned technology, sensor hardware, and software can attest that is indeed the case.11 If a computer is given a certain set of parameters—for example, a series of rules of engagement—it will follow those instructions precisely. If the autonomous weapon is designed and built to operate within the laws of war, then there should be no objection to their use.12 Under Article 36 of the 1977 Additional Protocol to the Geneva Conventions, weapons cannot be inherently indiscriminate and are prohibited from causing unnecessary suffering or superfluous injury. “The fact that an autonomous weapon system selects the target or undertakes the attack does not violate the rule,” write Hoover Institution legal scholars Kenneth Anderson and Matthew Waxman in a paper titled “Law and Ethics For Autonomous Weapon Systems.”13 Technology is continuously moving forward, and while autonomous systems may not be able to operate under all circumstances, it may only be a matter of time before engineers find a technical solution. Under many circum- stances—with the right sensors and algorithms—autono- mous systems would have the ability to distinguish lawful targets from unlawful ones, but that is not currently the case under all circumstances. Thus there are some limitations inherent to autonomous weapon systems for the time being. A Matter of Proportion However, those will not always be there, as technol- ogy continues to advance and engineers make progress. As Wittes points out correctly, “To call for a per se ban on autonomous weapons is to insist as a matter of IHL [international humanitarian law] on preserving a minimum level of human error in targeting.”14 Machines are gener- ally far more precise than human beings. Along with being able to distinguish between targets, the law requires that combatants weigh the proportionality of their actions.15 “Any use of a weapon must also involve evaluation that sets the anticipated military advantage to be gained against the anticipated civilian harm,” Anderson and Waxman write. “The harm to civilians must not be excessive relative to the expected military gain.” Though technically challenging, a completely autono- mous weapon system would have to be required to address proportionality as well as distinction. But the difficulty is entirely dependent on the specific operational scenario. For example, while an unmanned aircraft could identify and The technology is here. Air Force F-22 Raptors line up on the runway at Hill Air Force Base in Utah. The Raptor and the F-35 Joint Strike Fighter are specially equipped for “beyond-visual-range air-to-air combat. Both the F-22 and the F-35 fuse data correlated from the aircraft’s radar, elec- tonic support measures, and other sensors into a track file that the computer identifies as hostile, friendly, or unknown.” U.S.AIRFORCE
  • 6. www.usni.org PROCEEDINGS • 47 attack a hostile surface-to-air missile system deep behind enemy lines or an enemy warship at sea—where there is little chance of encountering civilians—targets inside a highly populated area are more difficult to prosecute. Some of the most problematic scenarios—which would not necessarily be factors in a high-end campaign against an A2/AD threat—would be challenging for a human pilot, let alone a machine. For ex- ample, during a counter-insurgency campaign, if two school buses were driving side-by-side in a built-up area, but one of the vehicles was carrying nuns and the other carrying heavily armed terrorists, it would be very dif- ficult for a human pilot to determine which bus is the proper target until one of them commits a hostile act. The same would be true for an autonomous system, but in the near term it could be a technological challenge. The human pilot would also have to determine what kind of weapon to use—judging the proportionality. Does he or she select a 2,000-pound JDAM, a smaller 250- pound small-diameter bomb, a 20-mm cannon, or do nothing since the risk of civilian casualties is too high? Likewise, an autonomous weapon system would need to be programmed to select an appropriate low-collateral- damage munition or to disengage if the danger of civilian casualties is too great once the target has been positively identified. But it would take time and investment before such an autonomous system could become a reality. Playing Fair? Thus, for the near future, autonomous weapons would have to be developed incrementally, starting with systems that could engage fixed targets and “obviously” military targets like surface-to-air missile sites or tank columns on the open battlefield during a conventional war. Likewise, in the maritime environment, where there are few civil- ians to speak of, autonomous systems could offer huge advantages with little in the way of drawbacks. For the time being, autonomous weapons should not be used in complex scenarios such as counterinsurgency, where there is significant possibility that it could cause inadvertent civilian casualties or unintended collateral damage. It may also be unwise to use a fully autono- mous UCAV for missions like close-air support—particu- larly during “danger close”–type situations where friendly troops are in close contact with the enemy—until the technology has been proven operationally in other roles. Human pilots have a hard enough time with those types of missions. Oftentimes, even human pilots cannot determine friend from foe—depending on the scenario—as the U.S. Central Command is discovering over Iraq and Syria with the ongoing fight against the Islamic State terrorist group. While at present some technological limitations do exist, those are not likely to remain roadblocks forever. Autono- mous technology is advancing rapidly and could one day be precise and reliable enough to not only distinguish cor- rect targets but also make proportionality judgments in complex scenarios based on parameters programmed into the machine. Those parameters would not be unlike rules of engagement given to human operators. Already, cam- eras and software exist that can identify individual human faces. Once a target is precisely identified, it would not be a huge leap then for an autonomous system to use a low-collateral-damage weapon to eliminate hostile targets while minimizing harm to civilians. Much of the objection to fully au- tonomous weapons seems to stem from a sense of “fair play” rather than any real legal issues—most of which are likely to be overcome. But any time new weapon technology emerges, there is opposition from those who believe that the technology fundamentally unbalances war.16 Objec- tions have been raised throughout history to new technolo- gies ranging from crossbows and longbows to machine guns and submarines because the use of such weapons was considered to be “unfair” or “unsporting.” But ultimately, the use of such weapons became a fact of life. War is not a game, and as Anderson and Waxman write: “The law, to be sure, makes no requirement that sides limit themselves to the weapons available to the other side; weapons supe- riority is perfectly lawful and indeed assumed as part of military necessity.”17 Indeed, there is no legal requirement for war to be fair— in fact throughout history war has been anything but. 1. Jan van Tol, Mark Gunzinger, Andrew Krepinevich, and Jim Thomas, “AirSea Battle: A Point-of-Departure Operational Concept,” Center for Strategic and Budgetary As- sessments, www.csbaonline.org/publications/2010/05/airsea-battle-concept/. 2. Ibid. 3. Ibid. 4. DOD directive 3000.09, www.dtic.mil/whs/directives/corres/pdf/300009p.pdf. 5. Ibid. 6. Ibid. 7. Human Rights Watch, “Losing Humanity: The Case against Killer Robots,” www. hrw.org/reports/2012/11/19/losing-humanity-0. 8. Ibid. 9. Ibid. 10. Ibid. 11. Benjamin Wittes, Brookings Institution, www.lawfareblog.com/2012/12/does- human-rights-watch-prefer-disproportionate-and-indiscriminate-humans-to-discrimi- nating-and-proportionate-robots/#.UunS7GRdXB8 12. Kenneth Anderson and Matthew Waxman, “Law and Ethics For Autono- mous Weapon Systems,” Hoover Institution http://papers.ssrn.com/sol3/papers. cfm?abstract_id=2250126. 13. Ibid. 14. Wittes, Brookings Institution. 15. Anderson and Waxman, Hoover Institution. 16. COL Lawrence Spinetta, USAF, commander of the 69th Reconnaissance Group, interview with author. 17. Anderson and Waxman, Hoover Institution. Mr. Majumdar has been covering defense since 2004. In addition to Proceedings, he writes for USNI News, Aviation Week & Space Technology, and The Daily Beast, among others. He previously covered national-security issues at Flight International, Defense News, and C4ISR Journal. Human Rights Watch has called for an international treaty that would ban all autonomous weapons.